Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

Are consumer grade RTX 4090 cards going to be suitable for running full tilt 24/7 for a year? Those things are fine to stress on the latest game for a few hours at a time, but would probably cause some defects from significant heat stress after just a few days at 100%.

This is inconsequential when you're playing Overwatch for a few hours a night and a frame drops now and again. If you're training an iteratively developed LLM though, physical defects could propagate into huge deficiencies in the final model.



Yep absolutely, crypto miners have been doing it for years.

I still think it would be impractical at scale because they are so much more hot and power hungry than the datacenter cards, and you would be lucky to score one or two if you’re on a wait list.


Except you can absolutely obtain 4090s today, while enterprise hardware is (was? haven't looked at the data) recently, which is the exact opposite scenario you mentioned.

I'm actually really surprised that you can still buy 4090s for under $2,000 (cheapest available I saw was $1,800 new and I only took 30 seconds to look), but you can usually sell certain models for quite a bit more. For example, my used 4090 FE is currently worth more than I paid for it.

I've played with AI, and while admittedly I've not done anything super serious, I can tell you that both the 3090 and 4090 are more than capable of performing. Tie them with a power efficient AMD CPU and you have something that can be competitive with enterprise (somewhat).

I've seen the pricing of "cloud" offerings and I've toyed with the idea of creating an "AI Cloud" because I have access to really fast internet and super cheap electricity, but I haven't executed because I'm most certainly not a salesperson. I do, however, know enough about marketing that one should not target price, so there is that...


You could under-volt or watt-limit a bit and lose just a fraction of FLOPS for much less heat/power though, depending on the workload


I don't think they'd become a fire hazard, but it is true that one would likely pick something else for this application.

Having said that, switching to something like the Tesla V100-SXM2-16GB wouldn't cost that much more.

TBH, I'm shocked at how many people treat Amazon as the first choice for this stuff. Much of it isn't even what most would consider a "production" workload. You are paying for a lot of enterprise-readiness that you don't need for training.


If you wanted to finetune a Mixtral 8x7B, what would you use?


Given the relative availability, I'd probably try to do it with a couple of rtx4090s on tensordock.


> TBH, I'm shocked at how many people treat Amazon as the first choice for this stuff

You can thank Amazon's legions of salespeople for that, particularly the end of year junket in Las Vegas where attendees are so pampered that about the only thing they won't do is suck your dick

Oh, yeah, they'll also yell at you on stage if you complain about their UI




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: