What happens to older datacenter GPUs? Do they have a second life somewhere outside of datacenters?<p>I could see Nvidia adding terms of sale requiring disposal rather than resale.
Plenty of enterprise server hardware (racks, servers, RAM, disks) does have an active secondhand market after 3-5 years of use, but I think GPUs are too specialized for it to be viable. I doubt anyone has the setup to run a H200 in their home rig.<p>I also don't think companies are going to have mandatory replacement cycles for GPU hardware the same way they do for everything else, because:<p>1. It is an order or magnitude (or more) more expensive.<p>2. It isn't clear whether Moore's law will apply to the AI GPU space the same way it has for everything else.<p>Unless Nvidia can launch a new chip every 2-3 years with massively improved performance-per-watt at a <i>lower</i> price no one is going to rush to recycle the old one.
> Unless Nvidia can launch a new chip every 2-3 years with massively improved performance-per-watt at a lower price no one is going to rush to recycle the old one.<p>That's exactly the point.<p>Performance/watt is increasing so much gen-to-gen that it makes no longer sense to run older hardware.<p>Not my words, Jensen's.
you can absolutely run e.g. datacenter-level A100 at home, there are adapters from the SXM to the PCIe socket. Haven't seen people running SXM versions of H100s this way but this could be due to the price factor only
Well technically true, I would wager that the home lab is going to require increasingly distinct and unusual adaptations to retrofit the hardware to home use.<p>New stuff is all liquid cooled by default and that's a paradigm shift for your average home lab.<p>I'm less aware of exactly what's happening on the power side of things but I think some of the architectures are now moving to relatively high voltage DC throughout and then down converting it to low voltage right before it's used. So not exactly just plug-and-play with your average nema15 outlet.
> I doubt anyone has the setup to run a H200 in their home rig.<p>There are PCIe versions of these right? And another comment is saying there are PCI adapters too. It "only" requires 600 to 700W. It's not out of reach for everybody.<p>If the used regular server market is any indication, you can find, after a few years, a <i>lot</i> of enterprise gear at totally discounted prices. CPU costing $4K brand new for $100 after a few years: stuff like that.<p>A friend has got a 42U rack and so do some homelab'ers. People have been running GPU farms mining cryptocurrencies or doing "transcoding" (for money).<p>It's not just CPUs at 1/40th of their brand new price: network gear too. And ECC RAM (before the recent RAM craze).<p>I'm pretty sure that <i>if</i> H200 begin to flood the used market, people shall quickly adapt.<p>> Unless Nvidia can launch a new chip every 2-3 years with massively improved performance-per-watt at a lower price no one is going to rush to recycle the old one.<p>I agree with that. But if they resell old H200s, people are resourceful and shall find a way to run these.
Last I checked AWS is still offering g4dn instances that run on NVIDIA T4 GPUs, which were first released in 2018. I think most people underestimate how long superscalers can keep these things running profitably after they depreciate, and you probably don’t want anything they throw away.<p>My last employer is still running a bunch of otherwise discontinued g3 instances with 2015 era GPUs.
Depending on the elemental composition, it could definitely be worthwhile to recycle wherever scale is practical. For giant datacenters and companies using hundreds of thousands or millions of gpus, that adds up to a lot of gold and other valuable elements.<p>In order to take advantage of that, someone needs to be positioned to process all that material economically, and to make the logistics achievable by the big players. If it costs Facebook $10million to store and transport phased out gpus vs just sending them to a landfil, they're not going to do it. If they get $100k for recycling - probably not going to do it. If they pocket $5 million, they will definitely contract that out, especially if it costs $50 million to build out the infrastructure to handle it.<p>Probably a good company idea - transport, disposal, refurbishment of out of cycle GPUs and datacenter assets, creating a massive recycling pipeline for recapturing all the valuable elements is a pretty good niche.
It's likely the GPU boards are designed for water cooled data center racks and might not fit in a regular PC case. It's also possible the PCB the GPU's are mounted to might not be standard PCIe cards that fit into an ATX case.<p>I bought a used NEC SX Aurora TSUBASA (PCIe x16 board that looks like a GPU board) and realized it has no fans. The server case it is designed to fit into is pressurized by fans forcing air through eight cards on a special 4 + 4 slot motherboard. I have to stack and mount three 40mm fans on the back.
They are build to physically last 5-7 years in 24/7 datacenter use, but they have effective lifetime just 3-4 years, then their value has deprecated and electricity and infrastructure cost dominates. Meta did a benchmark where 9% of the chips failed every year, 'infant mortality' is much higher in the first 3 months of use.
I've written about this elsewhere but I predict there will be a significant secondary market for repurposing parts of datacenter GPUs (for example, RAM chips) by desoldering them and soldering them onto new PCBs that fit PC/consumer use cases.
It seems like GPUs with a high utilization rate (60%+) degrade after 1-3 years: <a href="https://www.tomshardware.com/pc-components/gpus/datacenter-gpu-service-life-can-be-surprisingly-short-only-one-to-three-years-is-expected-according-to-unnamed-google-architect" rel="nofollow">https://www.tomshardware.com/pc-components/gpus/datacenter-g...</a><p>Would be interested to know if others have takes on this.
I previously ran 150,000 AMD gpus in all conditions at 100% utilization for years. I currently have a multi-million $ cluster of enterprise AMD GPUs.<p>A couple real world points:<p>1. They generally don't just fail. More likely a repairable component on a board fails and you can send it out to be repaired.<p>2. For my current stuff, I have a 3 year pro support contract that can be extended. Anything happens, Dell goes and fixes it. We also haven't had someone in our cage at the DC in over 6 months now.
You send them back to Nvidia or a third party e-waste recycler at end of life. Sometimes they're resold and reused, but my understanding is that most are eventually processed for materials.
Bin!!<p>Why would them sell it cheaper to the 2nd market??<p>It will hurt the sales of new ones. This is the way even with food, let alone technology. Don't expect to buy cheaper 2nd GPU any century soon.
This article misses the point that people are still actively running older compute.