The data centre has become the cornerstone that links our digitally interconnected world. At the same time, the rapid growth and application of artificial intelligence (AI) and machine learning (ML) is shaping the design and operation of data centres.
By Ben Selier, vice-president: secure power for Anglophone Africa at Schneider Electric
The training requirements associated with AI are driving new chip and server technologies and the need for extreme rack power densities.
The distinction between training and inference is critical when designing AI systems. Training workloads are used to train AI models like large language models (LLMs). These workloads require massive amounts of data fed to specialised servers with processors known as accelerators.
Inference models might be deployed on edge devices or cloud servers, depending on the application’s needs, and take the previously trained AI model into production to predict the output of new queries (inputs).
This demand for high-performance computing has led to increased requirements for powerful servers, GPUs (graphics processing units), and other specialised hardware within data centres to support workloads.
At the same time, the rise of edge computing, where computing resources are brought closer to the location where data is generated, is driven in part by AI applications.
Edge data centres are deployed to reduce latency and enhance the performance of AI applications in scenarios where real-time processing is critical like monitoring the movements of hospital patients to keep them safe.
The servers supporting these AI applications use advanced AI chipsets, more commonly known as AI accelerators. These chipsets play a crucial role in enhancing the performance of AI applications across various domains.
Energy consumption and demand for power
Modern data centres use high-density servers and equipment that demand more power for processing. This leads to concentrated energy usage in smaller spaces, increasing the overall energy footprint, which can raise concerns about the amount of energy required to process AI operations.
For example, if your AI model is operating emergency vehicle routing, it will need to process high-definition videos and traffic patterns to perform real-time operations to clear the traffic.
This would probably be an IT network of high speed, networked edge AI data centres that could process a larger amount of data and deliver real time decisions and predictions.
As the dependence on AI continues, new technologies and techniques will be deployed to make AI faster and more accurate and efficient.
One of these techniques is moving compressed models to the edge in the form of edge AI data centres. This will enable businesses to match the application to the model and optimise performance and the energy use.
The rapid growth in data traffic
The emergence of autonomous AI agents and decision-making programs holds the potential to revolutionise various aspects of business operations. These intelligent programs perform tasks independently, adapting and learning from its environment.
With increasing AI-driven automation, data centre employees can then automate routine tasks, reduce manual workloads and enhance overall efficiency as responsibilities such as server system maintenance or system monitoring can be handled by these intelligence programs.
AI and data centre evolution
As AI technology advances, it will continue to influence the design and operation of data centres. While these advancements bring efficiency and innovation, it also poses challenges related to energy consumption, and power and cooling systems.
This relentless advancement of AI is only going to continue, and to meet these evolving needs, the data centre industry needs to adapt.