Ecosystems at The Edge: Cloud vs. Colocation

Sept. 28, 2021
Cloud computing providers are extending their existing platforms to core and edge data centers so that customers have identical functionality regardless of location. A new special report courtesy of Iron Mountain looks at how the edge can help small data centers transform into full-fledged service providers and fuel digital transformation.

Last week in our special report series, we looked at why interconnection is key for edge data centers. We also explored some of the factors that should be considered when selecting a colocation provider. In our final article in the series, we’ll compare the cloud versus colocation.

Get the full report.

Cloud vs. Colocation

The major cloud computing providers are all aggressively deploying edge infrastructure in similar ways. Their objective is to extend their existing platforms to core and edge data centers so that customers have identical functionality regardless of location. To do this, the big providers are pursuing a two-pronged strategy of installing their infrastructure inside customers’ data centers and also in local telecommunications facilities. There are several advantages to this approach, chief among them being the ability for cloud providers to quickly provision infrastructure that fully compatible with the cloud regions.

However, limitations on size, bandwidth and compute capacity may challenge some to provide their customers with the full range of capabilities needed to build scalable edge architectures. Regional service breakdowns show that service availability can vary significantly by location.

Most of the major public cloud providers don’t operate a network of wholly-owned edge colocation facilities, so they must rely on partnerships and workforces they don’t control. These relationships will take some time to develop and can be fragile.

Edge infrastructure that isn’t designed to be managed remotely or that must be administered by contract personnel can be a risky proposition in far-flung locations. And self-contained miniature data centers are prone to theft, weather damage and unplanned outages. Most companies building out edge networks will want infrastructure that can’t be taken offline by a traffic accident.

Data Center Frontier’s Editor in Chief Rich Miller and executives from Iron Mountain discuss Ecosystems and the Edge. Watch on demand today!

Over time, cloud providers will no doubt address these structural limitations, but some customers may not want to wait. The fastest, safest and most flexible option for them is to locate edge infrastructure in established colocation regions that are fully equipped and staffed by trained personnel and that already support local ecosystems.

  • ROI is faster because regional colocation providers already have the facilities and relationships to support edge buildouts.
  • Regional colocation centers can scale to meet capacity demands and, in most cases, are fully compliant with relevant regulations.
  • The staff is trained and familiar with the equipment they work with every day.
  • Facilities are secure and reliable with power, environmental and seismic controls that already meet local requirements.
  • The staff speaks the local language and understands the culture and expectations of the ultimate end-users of edge services.
  • Connections to nearby services can be quickly facilitated through peering and interconnection.

Most colocation providers also have existing ecosystems of customers and partners to whom new customers can connect. For example, Iron Mountain Data Centers’ rapidly growing ecosystem with hundreds of third-party partners provides specialized network connectivity, access to fiber networks, software-defined networks, streaming services, specialized peering solutions, vertical market expertise and access to data centers in specific geographic and remote locations.

Many established co-location providers have also adopted public application program interfaces for rapid onboarding of customers and partners and have existing backend integration with a wide variety of other providers, including telcos, cloud platforms and colocation partners.

Download the full report, “The State of Data Center Cooling: A Key Point in Industry Evolution and Liquid Cooling” courtesy of TMGcore for an exclusive interview with Paul Gillin, Rich Miller, and Mark Lewis and Mark Kidd of Iron Mountain.  

About the Author

Paul Gillin

Paul Gillin is a speaker, writer and technology journalist who has written five books and more than 400 articles on the topic of social media and digital marketing. A technology journalist for 25 years, he has served as Enterprise Editor of the tech news site SiliconAngle since 2014. He was founding editor-in-chief of B2B technology publisher TechTarget, and served as editor-in-chief and executive editor of the technology weekly Computerworld.

Sponsored Recommendations

In this executive brief, we discuss the growing need for liquid cooling in data centers due to the increasing power demands of AI and high-performance computing. Discover how ...
AI hype has put data centers in the spotlight, sparking concerns over energy use—but they’re also key to a greener future. With renewable power and cutting-edge cooling, data ...
After a decade of stability, data center energy consumption is now set to surge—but can we change the trajectory? Discover how small efficiency gains could cut energy growth by...
Traditional data center power, cooling, and racks aren’t sufficient for GPU-based servers arranged in high-density AI clusters...

Adobe Stock, courtesy of Radix IoT
Source: Adobe Stock, courtesy of Radix IoT
Michael Skurla, co-founder of Radix IoT, explains how active, intelligent monitoring can extend data center lifespans while unlocking hidden capacity for AI innovation.

White Papers

Get the full report.
July 8, 2022
DartPoints outlines five new realities at the edge from decentralization to cloud-native applications and composable infrastructure.