Each planet has its own gimmick which throws a spanner into standard builds in its own unique way - one planet is essentially a farm where your factory is growing and processing fruits, which will rot and spoil if they aren't processed immediately -- so you need to design a factory which processes small packets at high speed without any buffering.
A well done PWA will absolutely beat SSR on a shitty connection if it's actually an app.
Cache-control immutable the code and assets of the app and it will only be reloaded on changes. Offline-first and/or stale-while-revalidate approaches (as in the React swr library) can hugely help with interactivity while (as quickly as possible) updating in the background things that have changed and can be synced. (A service worker can even update the app in the background so it's usable while being updated.) HTTP3/QUIC solves the "many small requests" and especially the "head of line blocking" problems of earlier protocols (though only good app/API design can prevent waterfalls). The client can automatically redo bad connections/requests as needed. Once the app is loaded (you can still use code splitting), the API requests will be much smaller than redownloading the page over and over again
Of course this requires a lot of effort in non-trivial cases, and most don't even know how to do it/that it is possible to do.
> HTTP3/QUIC solves the "many small requests" and especially the "head of line blocking" problems of earlier protocols (though only good app/API design can prevent waterfalls).
I love the autocomplete, honestly use it more than any other AI feature.
But I'm forced to write in Go which has a lot of boilerplate (and no, some kind of code library or whatever would not help... it's just easier to type at that point).
It's great because it helps with stuff that's too much of a hassle to talk to the AI for (just quicker to type).
I also read very fast so one line suggestions are just instant anyway (like non AI autocomplete), and longer ones I can see if it's close enough to what I was going to type anyway. And eventually it gets to the point where you just kinda know what it's going to do.
Not an amazing boost, but it does let me be lazy writing log messages and for loops and such. I think you do need to read it much faster than you can write it to be helpful though.
(This is a big advantage of open weight models; even if they're too big to host yourself, if it's worth anything there's a lot of competition for inference)