Hacker News new | past | comments | ask | show | jobs | submit login

It's more about UX, to reduce the perceived delay. LLMs inherently stream their responses, but if you wait until the LLM has finished inference, the user is sitting around twiddling their thumbs.



Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: