What are data centers? How they work and how they are changing in size and scope
A data center is a physical facility that enterprises use to house their business-critical applications and information. As they evolve from centralized on-premises facilities to edge deployments to public cloud services, it’s important to think long-term about how to maintain their reliability and security.
What is a data center?
Data centers are often referred to as a singular thing, but in actuality they are composed of a number of technical elements. These can be broken down into three categories:
- Compute: The memory and processing power to run the applications, generally provided by high-end servers
- Storage: Important enterprise data is generally housed in a data center, on media ranging from tape to solid-state drives, with multiple backups
- Networking: Interconnections between data center components and to the outside world, including routers, switches, application-delivery controllers, and more
These are the components that IT needs to store and manage the most critical resources that are vital to the continuous operations of an organization. Because of this, the reliability, efficiency, security and constant evolution of data centers are typically a top priority. Both software and hardware security measures are a must.
In addition to technical equipment, a data center also requires a significant amount of facilities infrastructure to keep the hardware and software up and running. This includes power subsystems, uninterruptable power supplies (UPS), ventilation and cooling systems, backup generators and cabling to connect to external network operators.
Data center architecture
Any company of significant size will likely have multiple data centers, possibly in multiple regions. This gives the organization flexibility in how it backs up its information and protects against natural and man-made disasters such as floods, storms and terrorist threats. How the data center is architected can require some of the difficult decisions because there are almost unlimited options. Some of the key considerations are:
- Does the business require mirrored data centers?
- How much geographic diversity is required?
- What is the necessary time to recover in the case of an outage?
- How much room is required for expansion?
- Should you lease a private data center or use a co-location/managed service?
- What are the bandwidth and power requirements?
- Is there a preferred carrier?
- What kind of physical security is required?
Answers to these questions can help determine how many data centers to build and where. For example, a financial services firm in Manhattan likely requires continuous operations as any outage could cost millions. The company would likely decide to build two data centers within close proximity, such as New Jersey and Connecticut, that are mirror sites of one another. An entire data center could then be shut down with no loss of operations because the company could run off just one of them.
However, a small professional-services firm may not need instant access to information and can have a primary data center in their offices and back the information up to an alternate site across the country on a nightly basis. In the event of an outage, it would start a process to recover the information but would not have the same urgency as a business that relies on real-time data for competitive advantage.
While data centers are often associated with enterprises and web-scale cloud providers, actually any size company can have a data center. For some SMBs, the data center could be a room located in their office space.
Industry standards
To help IT leaders understand what type of infrastructure to deploy, in 2005, the American National Standards Institute (ANSI) and Telecommunications Industry Association (TIA) published standards for data centers, which defined four discrete tiers with design and implementation guidelines. A Tier 1 data center is basically a modified server room, while a Tier Four data center has the highest levels of system reliability and security.
As is the case with all things technology, data centers are currently undergoing a significant transition, and the data center of tomorrow will look significantly different from the one most organizations are familiar with today.
Businesses are becoming increasingly dynamic and distributed, which means the technology that powers data centers needs to be agile and scalable. As server virtualization has increased in popularity, the amount of traffic moving laterally across the data center (East-West) has dwarfed traditional client-server traffic, which moves in and out (North-South). This is playing havoc with data-center managers as they attempt to meet the demands of this era of IT.
Here are key technologies that will evolve data centers from being static and rigid environments that are holding companies back to fluid, agile facilities capable of meeting the demands of a digital enterprise.
Edge computing and micro data centers
Edge computing is an increasingly popular paradigm in which much of the computational work that would traditionally have happened in a centralized data center happens closer to the edge of the network, where data is gathered. That means less delay for applications that need near-real-time action, and a reduction in the amount of data bandwidth needed. Micro data centers are compact units that can gather, process, analyze and store data physically close to the devices that collect it, and placing them on-site makes edge computing possible. Micro data centers are deployed in support of a number of uses, including 5G networks, Internet of Things rollouts, and content delivery networks.
There are a number of vendors in the micro data center space, some with background in other adjacent areas like IaaS or colocation services. Micro data centers are often (but not always) sold as pre-assembled appliances, and “micro” covers a fairly wide range of sizes—they can range from a single 19-inch rack to a 40-foot shipping container—and administration may be taken care of by the vendor or outsourced to a managed service provider.
The role of cloud
Historically, businesses had a choice of building their own data centers or using a hosting vendor or a managed service partner. Going the latter routes shifted ownership and the economics of running a data center, but the long lead times required to deploy and manage the technology still remained. The rise of Infrastructure as a Service (IaaS) from cloud providers like Amazon Web Services and Microsoft Azure has given businesses an option where they can provision a virtual data center in the cloud with just a few mouse clicks. In 2019, for the first time enterprises spent more annually on cloud infrastructure services than they did on physical data center hardware, and more than half of servers sold went into cloud providers’ data centers.
Nevertheless, the local on-premises data center isn’t going away any time soon. In a 2020 survey from the Uptime Institute, 58% of respondents said that most of their workloads remained in corporate data centers, and they cited a lack of visibility into public clouds and responsibility for uptime as a reason to resist the switch.
Many organizations are getting the best of both worlds by using a hybrid-cloud approach, in which some workloads are offloaded to a public cloud while others that require more hands-on control or security still run in the local data center.
Software-defined networking (SDN)
A digital business can only be as agile as its least agile component. and that’s often the network. SDN can bring a level of dynamism never experienced before.
Hyperconverged infrastructure (HCI)
One of the operational challenges of data centers is having to cobble together the right mixture of servers, storage, and networks to support demanding applications. Then, once the infrastructure is deployed, IT operations needs to figure out how to scale up quickly without disrupting the application. HCI simplifies that by providing an easy-to-deploy appliance, based on commodity hardware, that can scale out by adding more nodes into the deployment. There are HCI offerings from a number of high-profile vendors.
HCI can deliver a number of advantages over traditional data centers, including scalability, cloud integration, and easier configuration and administration. Early use cases for HCI revolved around desktop virtualization, but the technology now occupies a number of niches, including remote office/branch office deployments, test and development environments, backup and data recovery, and logging and analytics.
Containers, microservices, service meshes
Application development is often slowed down by the length of time it takes to provision the infrastructure it runs on. This can significantly hamper an organization’s ability to move to a DevOps model. Containers are a method of virtualizing an entire runtime environment that allows developers to run applications and their dependencies in a self-contained system. Containers are very lightweight and can be created and destroyed quickly so they are ideal to test how applications run under certain conditions.
Containerized applications are often broken into individual microservices, each encapsulating a small, discreet chunk of functionality, which interact with one another to form a complete application. The job of coordinating those individual containers falls to an architectural form known as a service mesh, and while the service mesh does a lot of work to abstract complexity away from developers, it needs its own care and maintenance. Service-mesh automation and management information should be integrated into your comprehensive data-center networking-management system—especially as container deployments become more numerous, complex and strategic.
Microsegmentation
Traditional data centers have all the security technology at the core, so as traffic moves in a North-South direction, it passes through the security tools and protects the business. The rise of East-West traffic means the traffic bypasses firewalls, intrusion prevention systems and other security systems and enabling malware to spread very quickly. Microsegmentation is a method of creating secure zones within a data center where resources can be isolated from one another so if a breach happens, the damage is minimized. Microsegmentation is typically done in software, making it very agile.
Non-volatile memory express (NVMe)
Everything is faster in a world that is becoming increasingly digitized, and that means data needs to move faster. Traditional storage protocols such as the small computer system interface (SCSI) and Advanced Technology Attachment (ATA) have been around for decades and are reaching their limit. NVMe is a storage protocol designed to accelerate the transfer of information between systems and solid state drives greatly improving data transfer rates.
And NVMe isn’t just limited to connecting to solid-state memory chips: NVMe over Fabrics (NVMe-oF) allows the creation of super-fast storage networks with latencies that rival direct attached storage.
GPU computing
Central processing units (CPUs) have powered data-center infrastructure for decades, but Moore’s Law is running up against physical limitations. Also, new workloads such as analytics, machine learning and IoT are driving the need for a new type of compute model that exceeds what CPUs can do. Graphics processing units (GPUs), once only used for games, operate fundamentally different as they are able to process many threads in parallel.
As a result, GPUs are finding a place in the modern data center, which is increasingly tasked with taking on AI and neural networking tasks. This will result in a number of shifts in how data centers are architected, from how they’re connected to the network to how they’re cooled.
Data centers have always been critical to the success of businesses of almost all sizes, and that won’t change. However, the number of ways to deploy a data center and the enabling technologies are undergoing a radical shift. To help build a roadmap to the future data center, recall that the world is becoming increasingly dynamic and distributed. Technologies that accelerate that shift are the ones that will be needed in the future. Those that don’t will likely stick around for a while but will be increasingly less important.
Copyright © 2020 IDG Communications, Inc.