> Your first leadership principle is customer obsession: “Leaders start with the customer and work backwards”.
> I’m your customer, and I’m begging you: please let me be a cloud engineer again.
Only AWS knows how many H100 GPUs they have, how busy they are. How many people are paying for them, how many people want them and can't get them, and how many people just don't care at all.
It's possible that the focus on GenAI for Re:Invent 2023 wasn't based on any hard data like that, and is really just up to the whims of Adam Selipsky since Jassy moved over, but maybe someone who better knows their planning process can comment.
I’d wager AWS makes more money from GenAI than any other domain. So it makes financial sense for them to sell that part of the business hard at the moment.
This opinion is based on admittedly anecdotal experience, but I’ve worked in a large range of domains on AWS over the years and by far the biggest AWS bills were for startups specialising in GenAI.
Interesting. I figured all their AI efforts were motivated by FOMO rather than actual returns. Why is AI stuff making so much money? Wouldn’t a new area like AI be a loss leader as they try to get market share?
The only way cloud providers are making money on "AI" is via obscene mark-ups on access to GPUs used by companies who are deluded into thinking that (re-)training their own LLM is what their own shareholders want to see.
The problem is that legal world are still undecided about the safety of public models.
Plus often businesses need GenAI to be trained on their own IP (ie stuff sensitive to their own business that they don’t want in the public domain).
—-
Point 1 will be decided over the next few years as creators take companies to court (or “ethical AI” starts to displace the current models trained on unlicensed content)
Point 2 cannot be resolved without training your own models.
——
Let’s also not forget that LLMs are just one part of the GenAI movement. There’s audio and image generation too (Plus video, but that’s more an extension of image). In fact it was the latter that I worked on.
And then you have other areas of AI outside of the generative space too. From hundreds of different applications of image recognition to sound processing to searching for other kinds of bespoke patterns. These are all areas I’ve worked in too.
Often a GenAI product will require multiple different “AIs” to function, as part of a larger pipeline that appears like a single opaque box to the customer. And most of those models in the pipeline likely aren’t generative, let alone LLMs.
You make it out as if all the executives at AWS have some master plan surrounding the probably absurd number of GPUs they bought, but the likely answer is its just a bunch of fallible people bandwagoning on the latest trend.
> I’m your customer, and I’m begging you: please let me be a cloud engineer again.
Only AWS knows how many H100 GPUs they have, how busy they are. How many people are paying for them, how many people want them and can't get them, and how many people just don't care at all.
It's possible that the focus on GenAI for Re:Invent 2023 wasn't based on any hard data like that, and is really just up to the whims of Adam Selipsky since Jassy moved over, but maybe someone who better knows their planning process can comment.