Comment by abraae

Comment by abraae 17 hours ago

33 replies

It's just the same dynamic as old servers. They still work fine but power costs make them uneconomical compared to latest tech.

acdha 17 hours ago

It’s far more extreme: old servers are still okay on I/O, and memory latency, etc. won’t change that dramatically so you can still find productive uses for them. AI workloads are hyper-focused on a single type of work and, unlike most regular servers, a limiting factor in direct competition with other companies.

  • matt-p 10 hours ago

    I mean you could use training GPUs for inference right? That would be use case number 1 for a 8 * a100 box in a couple of years. It can also be used for non IO limited things like folding proteins or other 'scientific' use cases. Push comes to shove im sure an old A100 will run crysis.

    • physicsguy 6 hours ago

      > Push comes to shove im sure an old A100 will run crysis.

      They don’t have video out ports!

      • [removed] 3 hours ago
        [deleted]
    • oblio 8 hours ago

      All those use cases would probably use up 1% of the current AI infrastructure, let alone ahat they're planning to build.

      Yeah, just like gas, possible uses will expand if AI crashes out, but:

      * will these uses cover, say, 60% of all this infra?

      * will these uses scale up to use that 60% within the next 5-7 years, while that hardware is still relevant and fully functional?

      Also, we still have railroad tracks from the 1800s rail mania that were never truly used to capacity and dot com boom dark fiber that's also never been used fully, even with the internet growing 100x since. And tracks and fiber don't degrade as quickly as server hardware and especially GPUs.

m00x 10 hours ago

LambdaLabs is still making money off their Tesla V100s, A100s, and A6000s. The older ones are cheap enough to run some models and very cheap, so if that's all you need, that's what you'll pick.

The V100 was released in 2017, A6000 in 2020, A100 in 2021.

Havoc 14 hours ago

That could change with a power generation breakthrough. If power is very cheap then running ancient gear till it falls apart starts making more sense

  • rgmerk 5 hours ago

    Hugely unlikely.

    Even if the power is free you still need a grid connection to move it to where you need it, and, guess what, the US grid is bursting at the seams. This is not just due to data center demand; it was struggling to cope with the transition away from coal well before that point.

    You also can’t buy a gas turbine for love nor money at the moment, and they’re not ever going to be free.

    If you plonked massive amounts of solar panels and batteries in the Nevada desert, that’s becoming cheap but it ain’t free, particularly as you’ll still need gas backup for a string of cloudy days.

    If you think SMRs are going to be cheap I have a bridge to sell you, you’re also not going to build them right next to your data centre because the NRC won’t let you.

    So that leaves fusion or geothermal. Geothermal is not presently “very cheap” and fusion power has not been demonstrated to work at any price.

  • overfeed 9 hours ago

    Power consumption is only part of the equation. More efficient chips => less heat => lower cooling costs and/or higher compute density in the same space.

    • nish__ 9 hours ago

      Solution: run them in the north. Put a server in the basement of every home in Edmonton and use the excess heat to warm the house.

zppln 16 hours ago

I'm a little bit curious about this. Where do all the hardware from the big tech giants usually go once they've upgraded?

  • q3k 11 hours ago

    In-house hyperscaler stuff gets shredded, after every single piece of flash storage gets first drilled through and every hard drive gets bent by a hydraulic press. Then it goes into the usual e-waste recycling stream (ie. gets sent to poor countries where precious metals get extracted by people with a halved life expectancy).

    Off-the-shelf enterprise gear has a chance to get a second life through remarketing channels, but much of it also gets shredded due to dumb corporate policies. There are stories of some companies refusing to offload a massive decom onto the second hand market as it would actually cause a crash. :)

    It's a very efficient system, you see.

    • oblio 7 hours ago

      Similar to corporate laptops where due to stupid policies, for most BigCos you can't really buy or otherwise get a used laptop, even as the former corporate used of said laptop.

      Super environmentally friendly.

  • trollbridge 16 hours ago

    I used (relatively) ancient servers (5-10 years in age) because their performance is completely adequate; they just use slightly more power. As a plus it's easy to buy spare parts, and they run on DDR3, so I'm not paying the current "RAM tax". I generally get such a server, max out its RAM, max out its CPUs and put it to work.

    • taneq 11 hours ago

      Same, the bang for buck on a 5yo server is insane. I got an old Dell a year ago (to replace our 15yo one that finally died) and it was $1200 AUD for a maxed out recently-retired server with 72TB of hard drives and something like 292GB of RAM.

      • PunchyHamster 11 hours ago

        Just not too old. Easy to get into "power usage makes it not worth it" for any use case when it runs 24/7

  • wmf 16 hours ago

    Some is sold on the used market; some is destroyed. There are plenty of used V100 and A100 available now for example.

dogman144 16 hours ago

Manipulating this for creative accounting seems to be the root of Michael Burry’s argument, although I’m not fluent enough in his figures to map here. But, commenting that it interesting to see IBM argue a similar case (somewhat), or comments ITT hitting the same known facts, in light of Nvidia’s counterpoints to him.

PunchyHamster 11 hours ago

Eh, not exactly. If you don't run CPU at 70%+ the rest of the machine isn't that much more inefficient that model generation or two behind.

It used to be that new server could use half power of the old one at idle but vendors figured out that servers also need proper power management a while ago and it is much better.

Last few gens increase could be summed up to "low % increase in efficiency, with TDP, memory channels and core count increase".

So for loads not CPU bound the savings on newer gen aren't nearly worth it to replace it, and for bulk storage the CPU power usage is even smaller part

  • matt-p 10 hours ago

    Definitely single thread performance and storage are the main reasons not to use an old server. A 6 year old server didn't have nvme drives, so SATA SSD at best. That's a major slow down if disk is important.

    Aside from that there's no reason to not use a dual socket server from 5 years ago instead of a single socket server of today. Power and reliability maybe not as good.

    • zozbot234 6 hours ago

      NVMe is just a different form factor for what's essentially a PCIe connection, and adapters are widely available to bridge these formats. Surely old servers will still support PCIe?

knowitnone3 11 hours ago

that was then. now, high-end chips are reaching 4,3,2 nm. power savings aren't that high anymore. what's the power saving going from 4 to 2nm?

  • monster_truck 11 hours ago

    +5-20% clockspeed at 5-25% lower voltages (which has been and continues to be the trend) add up quick from gen to gen, nevermind density or ipc gains.

    • baq 3 hours ago

      We can’t really go lower on voltage anymore without a very significant change in the materials used. Silicon band gap yadda yadda.

  • [removed] 11 hours ago
    [deleted]