That's not a typical use case during development though. People need fast feedback loops: they'd rather rent multiple GPUs and have the result the next morning, than waiting for days with no guarantee of success.
So unless you have stable tasks that need to run continuously, or have enough users to keeps your GPU clusters busy, your GPUs usage would be quite bursty: some period of high activity then a lot of time idling.
I'd argue the opposite. Having to spin up and down instances for development is a huge PITA, the tooling sucks, and the instance might not even be available the next time you need them. It also stresses me out personally, because I'm worrying about getting productive use out of every minute. Whereas my little GPU cluster (Despite a big upfront cost) costs nothing but electricity and runs 24/7.
So unless you have stable tasks that need to run continuously, or have enough users to keeps your GPU clusters busy, your GPUs usage would be quite bursty: some period of high activity then a lot of time idling.