AI, power availability and Intel’s future

6) Liquid cooling takes off

The heat generated by AI systems has already driven the need for liquid cooling because air cooling is simply not enough. But it’s not just GPUs that run hot – new CPUs from Intel and AMD are pretty toasty as well. I expect to see an increase in self-contained liquid cooling that will go into existing data centers and not require a hefty retrofit. I also think that HPE and Dell will finally do their own liquid cooling, similar to Lenovo’s Project Neptune. Up until now, HPE and Dell have been content to leave liquid cooling to third parties, but they may have to finally do their own thing.

7) Intel splits

There’s just no way around it. Intel must spin off its fabrication business just like AMD did in 2008. It was expensive, painful and necessary for long-term success. Gelsinger simply didn’t have the bandwidth to manage Intel foundries and Intel products, and all three suffered for it: The fab business was slow to get started, the chip business fell behind AMD, and Gelsinger’s tenure was cut short. It’s time to cut the fabs loose, Intel.

8) Continued predictions of on-prem exodus

This completely inaccurate prognostication will continue, and it will continue to be wrong. There are too many reasons to keep an on-premises data center, starting with data privacy and integrity. Repatriation of data from the cloud to on-premises infrastructure goes on every year. On-premises data centers will die when mainframes do.

9) GPU utilization becomes paramount

Nvidia is showing no signs of cutting power draw, so it’s up to others to make these things work as efficiently as possible. That means maximum utilization and scaling of hardware. As a result, maximizing GPU utilization will become the primary design goal for modern data centers. This will drive innovations in hardware and software to sustain the infrastructure necessary for training and minimize latency, lagging, and other issues that cause pauses in training. Success will be measured by how effectively data centers can keep their GPU resources utilized.

10) Power constraints impact data center locations

With almost 500 data centers in the Virginia area, it’s safe to say that region is reaching its limit. The same applies for Texas and Santa Clara. The demand for large-scale processing of data for AI, data analytics, and quantum computing will shift where new data centers are built. The good news is that these applications, especially AI training, are not latency sensitive, so they can afford to be placed in a remote location that has a lot of land and a lot of cheap power. This will largely involve data centers dedicated to massive computational processes, so there’s no need to worry that your colocation provider will open up shop in the Sierra Mountains.



Source link

Leave a Comment