Hacker News new | past | comments | ask | show | jobs | submit login

I'm far from the first to make this observation but LLMs are like anti-compression algorithms when used like that, a simple idea gets expanded into a bloated mess by an LLM, then sent to someone else who runs it through another LLM to summarize it back to something approximating the original prompt. Nobody benefits aside from Sam Altman and co, who get to pocket a cool $0.000000001 for enabling this pointless exercise.





> LLMs are like anti-compression algorithms when used like that, a simple idea gets expanded into a bloated mess by an LLM,

I think that's the answer:

LLMs are primarily useful for data and text translation and reduction, not for expansion.

An exception is repetitive or boilerplate text or code where a verbose format is required to express a small amount of information.


There is one other very useful form of "expansion" that LLMs do.

If you aren't aware: (high-parameter-count) LLMs can be used pretty reliably to teach yourself things.

LLM base models "know things" to about the same degree that the Internet itself "knows" those things. For well-understood topics — i.e. subjects where the Internet contains all sorts of open-source textbooks and treatments of the subject — LLMs really do "know their shit": they won't hallucinate, they will correct you when you're misunderstanding the subject, they will calibrate to your own degree of expertise on the subject, they will make valid analogies between domains, etc.

Because of this, you can use an LLM as an infinitely-patient tutor, to learn-through-conversation any (again, well-understood) topic you want — and especially, to shore up any holes in your understanding.

(I wouldn't recommend relying solely on the LLM — but I've found "ChatGPT in one tab, Wikipedia open in another, switching back and forth" to be a very useful learning mode.)

See this much-longer rambling https://news.ycombinator.com/item?id=43797121 for details on why exactly this can be better (sometimes) than just reading one of those open-source textbooks.


I use LLMs a huge amount in my work as a senior software engineer to flesh out the background information required to make my actual contributions understandable to those without the same background as me. eg, if I want to write a proposal on using SLO's and error budgets to make data driven decisions about which errors need addressing and which don't, inside a hybrid kubernetes and serverless environment, I could do a few things:

* Not provide background information and let people figure it out for themselves. This will not help me achieve my goals.

* Link them to Google's SRE book and hope they read it. Still not achieving my goals, because they won't.

* Spend 3 hours writing the relevant background information out for them to read as part of my proposal. This will achieve my goals, but take an extra 3 hours.

* Tell the LLM what I'm looking for and why, then let it write it for me in 2 minutes, instead of 3 hours. I can check it over, make sure it's got everything, refine it a little, and I've still saved 2.5 hours.

So for me, I think the author has missed a primary reason people use LLMs. It saves a bunch of time.


For a teacher, they want to see you spend those 3h to see what you come up with, and if there's something they should direct your attention to, or something they should change in their instruction.

But ultimately, getting the concise summary for a complex topic (like SLIs and SLOs are) is brilliant, but would be even better if it was full of back-links to deeper dives around the Internet and the SRE book.


This usage had never occurred to me but including adequate background information is definitely something I struggle with - I'll definitely try this!

Yeah you're totally right with this use case.

It feels like the information is there strewn across the internet, in forums, Reddit posts, stack overflow, specs, books. But to trawl though it all was so time consuming. With an LLM you can quickly distill it down to just the information you need.

Saying that, I do feel like reading the full spec for something is a valuable exercise. There may be unknown unknowns that you can't even ask the LLM about. I was able to become a subject expert in different fields just but sitting down and reading through the specs / RFCs, while other colleagues continued to struggle and guess.


To me:

If an LLM can help you understand an RFC, it's great. You're now relying on the RFC.

If an LLM can help you not rely on the RFC, you're doing it wrong.


This is how I currently am relearning upper high school math. It’s tremendously helpful as I am a why guy.

Why is the angle called m? Why is a combination nPr * (1/r)? What is 1/r doing there?

I use mathacademy.com as my source of practice. Usually that’s enough but I tend to fall over if small details aren’t explained and I can’t figure out why those details are there.

In high school this was punished. With state of the art LLMs, I have a good tutor.

Also it’s satisfying to just upload a page in my own handwriting and it understands what I did, and is able to correct me there.


The fact that we see lower grades as "punishment" is the root of the problem: grades are an assessment of our understanding level and competence on a topic.

Now, I know psychologically it's not as simple, and both society and ourselves equate academic (and professional, later on) success with personal worth, but that's a deeper, harder topic.


Yeah, this is the usual interaction with LLMs when coding: Ask GPT/Claude to write me a simple function. It writes 100 lines of code, trying to infer the rest of your codebase. Tell it to only solve my issue and do it in 5 lines of code. Get something close to working.

They are also useful for association. Imagine an LLM trained on documentation. Then you can retrieve info associated with your question.

This can go beyond just specific documentation but also include things like "common knowledge" which is what the other poster meant when they talked about "teaching you things".


In my experience, if the information I need is in the documentation, then I don't need the LLM. If it is not in the documentation, then the LLM will invent stuff that could be there but that isn't, and it's actually a loss of time.

> In my experience, if the information I need is in the documentation, then I don't need the LLM

need is a strong word. Do you need to be able to do ctrl+F? Not really, you can just read it all. But maybe it's easier to do ctrl+F. Same with LLM. Just imagine it as a fuzzy ctrl+F. Can be useful.


Fuzzy isn't what I want to be when referring documentation. So much documentation is incomplete to begin with.

That's the big issue with LLMs as of now; They reflect their American creators and never want to admit when they just can't answer a question. CTRL+F will in fact give me 0 results, which is more useful than a wrong result.


I agree.

However, as a non-native speaker, I really like the fact that I can give the LLM some description of a word that I don't know and it'll figure it out. "those fruits that are kind of bitter but also make my tongue feel like it wants to curl" (= "adstringend" [sic] or so, but who remembers that word or can spell it correctly?)

LLMs are basically like all those type-correction algorithms on steroids. Very helpful sometimes, even if it means I have to doublecheck their output.


The thing is, being able to read documentation is a skill.

Being really good at ctrl+F / LLM is not the same. I learn a lot just browsing through documentation, without searching anything in particular.


IMO the main uses of LLM are guided problem solving and a better Google search that you can ask questions in a natural way.

Producing text as output is not the way


Actually, pretty sure Sam Altman gets to spend that money on power and compute, not sure they’ve figured out how to turn it into an income stream yet.

I think they actually lose the $0.000001

Yep. They're very closely linked.

http://prize.hutter1.net/

Note the preamble, FAQs, and that all of the winning entries are now neural networks.


I blame humans. I never understand why unnecessarily long writing is required in a lot of places.

Rituals are significant because they are long. A ritual that consisted of the words “Rain please” wouldn’t convince the gods, much less their human followers.

"I have made this longer than usual because I have not had time to make it shorter." — attributed to many[1]

The funny thing is that people use LLMs to do the opposite instead of what is implied to be a smarter thing to do with the above quote.

[1] https://quoteinvestigator.com/2012/04/28/shorter-letter/


Depends on what you are looking for. I’ve turned half baked ideas into white papers for plenty of praise. I’ve used them to make my Jira tickets seem complicated and complete. I’ve used them to get praised for writing comprehensive documentation.

Part of my performance review is indirectly using bloat to seem sophisticated and thorough.


> comprehensive documentation

Documentation is an interesting use case. There are various kinds of documentation (reference, tutorial, architecture, etc.) and LLMs might be useful for things like

- repetitive formatting and summarization of APIs for reference

- tutorials which repeat the same information verbosely in an additive, logical sequence (though probably a human would be better)

- sample code (though human-written would probably be better)

The tasks that I expect might work well involve repetitive reformatting, repetitive expansion, and reduction.

I think they also might be useful for systems analysis, boiling down a large code base into various kinds of summaries and diagrams to describe data flow, computational structure, signaling, etc.

Still, there is probably no substitute for a Caroline Rose[1] type tech writer who carefully thinks about each API call and uses that understanding to identify design flaws.

[1] https://folklore.org/Inside_Macintosh.html?sort=date


Yes, but none of the current LLMs are even remotely useful doing that kind of work for even something moderately complex. I have a 2k LOC project that no LLM even "understands" *. They can't grasp what it is: It's a mostly react-compatible implementation of "hooks" to be used for a non-DOM application. Every code assistant thinks it's a project using React.

Any documentation they write at best re-states what is immediately obvious from the surrounding code (Useless: I need to explain why), or is some hallucination trying to pretend it's a React app.

To their credit they've slowly gotten better now that a lot of documentation already exists, but that was me doing the work for them. What I needed them to do was understand the project from existing code, then write documentation for me.

Though I guess once we're at the point AI is that good, we don't need to write any documentation anymore, since every dev can just generate it for themselves with their favorite AI and in the way they prefer to consume it.

* They'll pretend they understand by re-stating what is written in the README, then proceed to produce nonsense.


I've found "Claude 3.7 Sonnet (Thinking)" to be pretty good at moderately complex code bases, after going through the effort to get it to be thorough.

Without that effort it's a useless sycophant and is functionally extremely lazy (ie takes short cuts all the time).

Don't suppose you've tried that particular model, after getting it to be thorough?


Delivering a library with an llm to explain the api and idiomatic usage seems like an interesting use case.

I’d rather be homeless in Philadelphia than work where you work

This kind of "perf review" hacking works ~everywhere; how well it works correlates with how entrenched the organization is (i.e., how hard it is for new players to disrupt).

You don't have to play the game the same way to work there. But it helps to accept that others will play it, and manage your own expectations accordingly.


> This kind of "perf review" hacking works ~everywhere

I don't have tons of examples, but in my experience:

* This worked in toxic environments. They deserve it.

* This doesn't work in a functional environment, because they don't have those bullshit metrics.

If you have to rely on those tricks, it's time to look for another job.


Which big, well-paying companies do not have "those bullshit metrics"? I know for a fact that Meta, Google, Stripe, Airbnb, and Oracle all lean heavily on performance review cycles based entirely on ridiculous metrics. Getting ahead there requires you to play the stupid games GP is suggesting.

The original post doesn't mention anything quantitative ("metrics"). Did this get sidetracked?

> Depends on what you are looking for. I’ve turned half baked ideas into white papers for plenty of praise. I’ve used them to make my Jira tickets seem complicated and complete. I’ve used them to get praised for writing comprehensive documentation.

This is about giving people a good impression of you so they'll write strong peer feedback.


And those are all examples of companies nobody should work for

If you work there, as OP said, it's time to look for another job.

Totally, if you're willing to trade in your half-million or more annual compensation for $150k or less. My point is that it's an unfortunate game you have to play if you're working for places like that.

>Totally, if you're willing to trade in your half-million or more annual compensation for $150k or less.

I will take 60k at this point. I've been living ono half of that for almost 2 years now.

I have no idea how anyone is navigating this job market. Maybe it's just 10x worse and most people here are in the Bay area that's a tiny bit more shielded from this.

>Their number is absolutely miniscule compared to the number of big-tech jobs.

give it another year of layoffs. We'll get there.


the coolest thing "the big tech" was able to do is convince a whole bunch ridiculously smart people that they only place they can make $500k is with them. I personally (I am just one person) know more than 10 people not working for big tech making (some significantly) north of $500k as SWEs and doing awesome sh*t (unlike I would un-intelligently guess most of big tech employees)

Sure, there are some jobs that match your description, likely for people that can brand themselves "AI engineers" for a rocketship company like OpenAI. Their number is absolutely miniscule compared to the number of big-tech jobs.

Yeah, I think another problem is that TooBigTech is able to pay insanely for... what do they bring to society again? Was it social media and ads? Anyway, all that desirable stuff /s.

But it helps to accept that others will play it

Feel for you or anyone surrounded by such others but it is most definitely not everywhere - that is used to justify your presence in a place of work you should not be


Would be nice to fix the performance reviews so we don't end up in a arms race of creating bloat until it becomes so unproductive it kills the host.

Over-fitting proxy measures is one of the scourges of modernity.

The only silver lining is if it becomes so wide spread and easy it loses the value of seeming sophisticated and thorough.


> creating bloat until it becomes so unproductive it kills the host

Maybe we should let/encourage this to happen. Maybe letting bloated zombie-like organisations bloat themselves to death would thin the herd somewhat, to make space for organisations that are less “broken”.


"But at what price?" is probably the right question here, and that'd be a case by case basis thing. ;)

I am on the mind that every organization should eventually die before it becomes a monster. I am also not a huge fan of inheritance for the same reason.

All FAANG/MAGMA dying is a bonus. The cherry on top. Net positive for humanity. A best case scenario.

I fully believe you and I am saddened by the reality of your situation.

At the same time, I strive really hard to influence the environment I am in so it does not value content bloat as a unit of productivity, so hopefully there are at least some places where people can have their sanity back!


If your organisation is such that you have to do this even though you are competent for your job, then they deserve it. They lose money because they do it wrong.

If your organisation is functional and you are abusing it by doing that, then you deserve to get fired.


...thinking about it, there are probably situations where making something more verbose makes it take less effort to read. I can see how an LLM might be useful in that situation.



Join us for AI Startup School this June 16-17 in San Francisco!

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: