Groundbreaking to commissioned data center in less than 10 months demonstrates leading speed-to-market development capabilities.
QTS Realty Trust (NYSE: QTS), a leading provider of hybrid colocation and mega scale data center solutions, today announced the official opening of a new 495,000 gross square foot mega data center on its Atlanta Metro campus.
Constructed and commissioned in less than 10 months, QTS Atlanta-Metro Data Center 2 (DC2) features 240,000 square feet of data hall space and 72 megawatts of power capacity designed for large-scale enterprise colocation and hyperscale deployments. The facility is purpose-built utilizing QTS’ innovative standardized Freedom Building design and specifications.
Consistent with QTS’ de-risked approach to development and capital allocation, the Company previously announced the signing of anchor tenant leases with two existing strategic hyperscale customers totaling more than 16 megawatts at the new Atlanta-Metro DC2 site. These customers chose to expand with QTS based on QTS’ operational maturity, speed to market and commitment to a premium customer experience.
QTS Atlanta-Metro DC2 sits adjacent to the Company’s flagship Atlanta-Metro Data Center (DC1) on a 95+ acre site that now encompasses 200+ megawatts of utility capacity fed from two of the greatest pre-positioned, data center owned substations in the country. These substations enable QTS to deliver, it believes, the lowest cost of power to its customers in the southeast data center market at less than 4 cents per kilowatt hour. Upon full development, the Atlanta-Metro campus is expected to support more than 275 megawatts of power capacity.
QTS’ Atlanta-Metro data center campus represents one of the most strategic data center properties in the Southeast with approximately 250 embedded customers. The campus features abundant networking including 2,000+ cross connects supported by diverse connectivity for cloud and hybrid colocation, direct fiber access to a multitude of carriers, access to multiple fiber routes and IP providers, multiple dark and lit fiber providers, redundant transport paths and access to the world’s largest cloud providers.
“We are opening our new mega data center on our Atlanta Metro campus to expand our growth opportunity with hyperscale, large enterprise and government organizations in the Southeast,” said Chad Williams, Chief Executive Officer, QTS. “Atlanta Metro DC2 represents QTS’ third mega data center in the greater Atlanta area, further solidifying our presence as Atlanta’s market leader.”
Hivelocity, a provider of edge computing and IaaS services in 26 global markets, announced today the addition of Telefonica to its already diverse network blend. The addition of Telefonica will specifically enhance the performance of Hivelocity’s network throughout Latin America where Hivelocity has a large customer base.
With data centers in Tampa, Miami and Dallas, Hivelocity has long been a favorite IaaS provider among businesses who cater to users in Central and South America. “Brazil is actually our second largest customer market outside of the United States. We, of course, are always eager to find ways to build upon our traction there,” said Hivelocity COO, Steve Eschweiler.
“The addition of Telefonica should further improve upon what is already great network performance to that part of the world. As an edge computing provider, we are always working to minimize latency. Having Telefonica added to our network blend will certainly achieve that for many of our customers,” he added.
Hivelocity provides bare metal edge computing services to customers in over 130 countries from its data centers across the USA, Europe, Asia and Australia. “As we continue to expand the geographic footprint of our edge computing platform, we continue to seek ways to refine our service and performance. Having data centers strategically spread across the globe is only one piece of the edge puzzle. The proper blend of network providers at each location is just as critical to the performance of our products and enabling our users to maximize the potential of our platform.”
Currently Hivelocity offers its bare-metal edge computing cloud platform in Amsterdam, Atlanta, Chicago, Dallas, Frankfurt, Hong Kong, London, Los Angeles, Madrid, Miami, Milan, Newark, New York City, Paris, Reston, Seattle, Stockholm Sunnyvale, Seoul, Singapore, Sydney, Tampa, Tokyo, Toronto and Vancouver.
Through Hivelocity’s platform and API, customers are able to instantly deploy, manage and automate Windows and Linux bare metal servers at each of these locations. You can find out more about Hivelocity and its services at https://www.hivelocity.net/.
Historic Wells Building Becomes Latest Portfolio Addition
1547 Data Center Real Estate Fund II, LP (“The Fund”), an affiliate of fifteenfortyseven Critical Systems Realty, LLC (“1547”), a leading owner, operator and developer of data center space, in partnership with Harrison Street, one of the leading investment management firms focused on alternative real assets, today announced their expansion into Milwaukee, Wisconsin with the acquisition of the city’s historic Wells Building.
The 15-story, 165,000-square-foot building located at 324 East Wisconsin Avenue was purchased from Ascendant Holdings LLC, a Wisconsin-focused commercial real estate development and investment company.
Strategically located in downtown Milwaukee, the building has a long history as a communications center. Once the home of Western Union Telegraph Company’s Milwaukee headquarters, the building continues to be a communications and technology hub as a carrier hotel and network-dense data center.
“We are thrilled about this acquisition, our first with Harrison Street,” says J. Todd Raymond, CEO of 1547. ”The Wells Building is a strategic fiber convergence point in the Upper Midwest with significant upside value as edge data center requirements continue to grow exponentially. Our plan is to build on the great work of the Ascendant team and make significant investments in the power and cooling infrastructure. We are excited to resurrect this century old, purpose-built, communications property and redevelop it into the premier network and content distribution data center facility in the Upper Midwest.”
Michael Hochanadel, Managing Director and Head of Digital Real Estate at Harrison Street, said, “The Wells Building investment represents an important step in growing our digital platform. Strategic connectivity assets like the Wells Building are irreplaceable. We look forward to expanding our relationship with 1547 as we continue to pursue opportunities in digital infrastructure.”
The growth in the need for online accessibility to services such as schooling, virtual healthcare, shopping, and entertainment makes the availability of data storage and computer infrastructure, bandwidth, and speed more critical than ever.
The Milwaukee metropolitan area has over 2,000,000 people whose reliance and dependence on technology continues to grow exponentially with few options for hosting their data center needs. With over 25 network providers, the dense telecommunication infrastructure and multiple diverse entry points into the Wells Building is already appealing to data center customers seeking a location for low latency content distribution in the Upper Midwest.
With the planned robust infrastructure upgrades, the Wells Building will be the most network-dense, fully redundant, high-security data center facility in the region.
Houliahn Lokey, formerly MVP Capital, acted as the seller’s financial advisor on the transaction.
Are you moving your files to the cloud? Shifting from an in-house solution to an external cloud environment might feel daunting, but it’s an important move for most businesses as they grow and as their technological needs change and expand.
Let’s take a look at some common problems involving cloud migration – and how you can solve them.
Problem #1: Rushing the migration without taking enough time to plan
Moving to a cloud based file server isn’t something you want to rush. Unfortunately, many organizations make the mistake of hurrying their migration without taking the time to plan and create a proper strategy.
The last thing you want is to end up with unexpected downtime, or worse, the loss of important files or data.
Solution: Analyze your current infrastructure … and plan accordingly
As Daniel Hein points out in an article for SolutionsReview, migrating to the cloud can potentially take several months, depending on the size of your organization and the amount of data you need to move.
You’ll need to pay particular attention to assets that may need adjusting or even rebuilding completely once you migrate to the cloud.
Be realistic about the costs and timescales that your organization will face, too. It’s much better to be clear about these upfront than to find them spiraling out of control partway through the process.
Problem #2: Not training your employees adequately
When you deal with IT a lot, it’s all too easy to assume that others will be just as quick as you at picking up new technologies.
Some employees, though, may not find your new cloud-based systems at all intuitive to use. If you don’t provide enough training, you’re likely to face a drop in productivity (and even in job satisfaction). Plus, your IT team may be overwhelmed by support requests.
Solution: Allow time and resources for employee training
Make sure that your cloud migration strategic plan also includes the time and resources to train employees on the new systems. That could involve a mix of:
Hands-on live training where employees are shown what to do and have a chance to try it out in real-time, so they can easily ask questions if they’re confused.
Pre-recorded video demos or written documentation on how to use the new cloud-based system.
One-to-one training in a small company where specific employees will be using the new system a lot.
Problem #3: Not accounting for ongoing costs
When moving to a cloud-based solution, it’s not just about the upfront cost of servers or even the ongoing costs of bandwidth and your IT team’s time. You also want to take into account the other ongoing costs that you’re likely to face. As Sulakshana Iyer explains:
“Cloud server management includes ongoing operations such as industry compliance, security certificates, monitoring application performance, up-scaling servers, and more.”
Solution: Be clear about ongoing, not just upfront, costs of cloud migration
While your cloud-based systems may well be cheaper than your previous ones, you still need to be clear about the ongoing costs that will be faced – both in terms of direct money paid out and in employee time.
Make sure the company you’re using for your cloud-based server clearly lays out the costs for you and be sure to factor in the indirect costs as well. You may want to err on the side of overestimating how much staff time the migration will take: that way, you’re covered even if something doesn’t go as smoothly as you hoped.
Cloud-based servers are a great option for both big and small businesses. By ensuring that you consider and face up to potential problems upfront, you’ll pave the way for a smooth and easy migration.
Research conducted by Omdia reveals that the global market for prefabricated modular data centres, from the edge to the core, increased by more than 65%
Vertiv (NYSE: VRT) (www.Vertiv.com/en-emea), a global provider of critical digital infrastructure and continuity solutions, has been ranked by technology analyst firm Omdia as one of the leading suppliers in the prefabricated modular (PFM) data centre market with the second highest market share worldwide. The newly released research has highlighted that benefits such as the ability to scale with confidence are driving significant growth in the adoption of PFM solutions in all geographies.
The Omdia report, Prefabricated Modular Data Centres, published in early 2020 and based on 2018 and 2019 data, valued PFM shipments at more than $1.2bn USD in 2018 with growth in deployments set to increase by more than 65% for 2019. The analyst group attributed this strong growth to a number of factors including scalability, the benefits of offsite manufacturing and integration, and speed of deployment.
Lucas Beran, principal analyst for Omdia’s cloud and data centre research practice and the report’s author, identified speed of deployment as the primary driver for many owners and operators. “The rapid growth of data and insatiable demand for compute is driving the rapid growth of data centres. Given that a traditional data centre takes 18 to 24 months to deploy, a quicker solution is often needed,” he said. “On average, suppliers of prefabricated modular data centres can deliver a solution in four to six months.”
“We have seen strong growth in demand for PFM data centres as owners and operators realise the benefits of scalability, cost-efficiency and speed of deployment that they provide. Our customers also value our design flexibility and the customizations that we offer,” said Viktor Petik, vice president of Integrated Modular Solutions, Vertiv. “To make the offerings more resilient and quickly available, we gained PFM TIER-Ready certification in EMEA through a recent agreement with Uptime Institute, increased operations in North America and are working on plans to expand our PFM facility in Croatia to increase capacity and reinforce our capabilities in this key market sector.”
The Omdia report defined a number of different forms of PFM data centres with varying use cases, including IT and facility-specific designs and so-called all-in-one modules (with integrated IT, power, and cooling infrastructure) which are commonly used in education, industrial, and healthcare applications as well as remote and harsh environments.
According to Omdia, demand for edge computing is driving uptake of all-in-one modules for edge locations that need a small data centre presence close to end users. The ‘plug and play’ approach has the benefit of not only cutting the time for start-up and commissioning to just a few days instead of weeks, or months, but also reducing the potential for quality issues, as components are pre-integrated and pre-tested off-site.
As described in the report, Vertiv’s PFM designs cover a range of different offerings to meet specific customer needs, such as the Vertiv™ SmartMod™ – a fully self-contained, easily-configurable, and ready-to-order PFM product range that enables new data centre whitespace to be rapidly deployed.
SmartMod designs have recently been awarded Uptime Institute’s TIER-Ready designation of performance resiliency in Europe, Middle East and Africa (EMEA), also allowing Vertiv to sell Uptime Institute Tier Certification of Constructed Facilities (TCCF) services with TIER-Ready modular units, thus providing the benefits of full tier certification directly to Vertiv customers. Vertiv also provides completely custom PFM solutions, which include design, project management and system integration.
Bandwidth IG serves mission critical data centers, hyperscale customers and enterprises with high capacity dark fiber services. Bandwidth IG recently announced that it is deploying a metro dark fiber network in Atlanta intentionally placed to ensure minimal overlap with other networks. The high-density cables allow Bandwidth IG to deliver a diverse, low-latency, high-count fiber network at a competitive value.
Bandwidth IG’s greater Atlanta network has more than 40 newly built route miles and 75,000 fiber miles. The network reaches 16 data centers, representing nearly 400 megawatts of IT load. As part of this initiative, Bandwidth IG is deploying eight separate fiber rings that all interconnect in the QTS Atlanta Metro data center campus.
“We are pleased to establish our core presence in QTS, one of the most important data center operators in Atlanta,” said Jim Nolte, Chief Executive Officer, Bandwidth IG. “Bandwidth IG’s metro dark fiber rings interconnected with QTS’ flexible connectivity ecosystem offers unlimited solutions for solving high capacity broadband needs with diversity, speed and scalability.”
The QTS connectivity ecosystem features diverse connectivity for cloud and hybrid colocation including carrier-neutral cloud interconnection, multiple fiber routes and IP providers. This includes the world’s largest IP networks, multiple dark and lit fiber providers, redundant transport paths and access to the world’s largest cloud providers.
“The introduction of Bandwidth IG’s high count, diverse fiber rings into our Atlanta Metro facility adds additional value to the dense ecosystem already in place and gives our customers yet another compelling connectivity option,” said Sean Baillie, Executive Vice President, Connectivity Strategy, QTS. “Bandwidth IG’s dark fiber rings all converge and interconnect at our Atlanta campus making QTS the epicenter of next generation fiber connectivity in Atlanta.”
In addition to the network development in Atlanta, Bandwidth IG is 90 days from completing a high-capacity ring in Santa Clara, California to connect another strategic QTS data center campus.
“QTS serves as host to many large technology companies in Silicon Valley who are looking for diverse network options and access to a variety of end points throughout the South Bay – which we can deliver,” said Nolte.
PCCW Solutions, the IT services flagship of PCCW Limited, unveils its SLC Data Center in Fo Tan, Hong Kong, which is designed to cater for high density and capacity requirements to support growing digital and cloud business needs in the Asia Pacific region.
Riding on its proven track record in delivering world-class data center services, PCCW Solutions has secured substantial client pre-commitments prior to construction completion. The first stage of the new data center has been fully occupied by global financial institutions, local enterprises and cloud providers. The current pandemic has spurred further demand for resilient and scalable infrastructure to support business continuity and accelerate the uptake of digital services.
“The successful launch of this new facility underscores our market leadership in attracting international hyper-cloud service providers to base their data centers in Hong Kong, supporting Hong Kong to become a strategic data center hub in Asia,” said Mr. Ramez Younan, Managing Director of PCCW Solutions.
“PCCW Solutions will look at further expansion in data center capacity in Hong Kong and Southeast Asia to address our rapidly growing customers demand across the region,” added Mr. Younan.
Adopting the modular design principle, the first stage of this data center has been completed within 12 months from design to build, and awarded with Uptime Institute Tier III Certification. This scalable and flexible design aligns capacity to business needs while accelerating time-to-market and achieving energy efficiency. The 13-story SLC Data Center is on target to be fully completed by the end of 2020.
In the picture: Mr. Ramez Younan, Managing Director of PCCW Solutions (2nd from right), Ms. Hester Shum, Group Chief Human Resources Officer of PCCW Group, Mr. Brian Groen, Senior Vice President of Data Center of PCCW Solutions (2nd from left) and Mr. Sinko Choy, Senior Vice President and Head of Client and Market Development of PCCW Solutions, officiate at the SLC Data Center opening ceremony.
In recent years, the many technological advancements and changing customer demands have drastically changed the way organizations operate. Organizations are now becoming more lean and agile. And to become agile, most of them are transforming by utilizing a combination of automation and DevOps approach.
So, what is DevOps? The term DevOps is strongly aligned with lean and agile concepts. This involves practices and tools that enhance the ability of an organization to deliver its services quickly and efficiently. It breaks the barriers between the IT department in an organization and enables more collaboration making the whole development and delivery process more effective and efficient.
Some other benefits of DevOps include:
Reduction in environment downtime by more than 70 percent
Reduction in time spent or coding by almost 70-80 percent
Reduction in development effort by 30 percent
With every passing day, more and more organizations see DevOps as their preferred software development model. According to Grand View Research, the coming years are going to be big for DevOps: Globally, the DevOps market size is expected to be worth $12.85 billion by 2025.
Cloud computing forms the basis of DevOps tools available today and makes DevOps implementation part easier. It helps organizations by making every step of DevOps operation successful. Through cloud technology, apps can be easily developed and tested in multiple environments. This saves time, reduces cost, increases efficiency, and is more secure too. The best part is the fact that you pay for only what you use.
DevOps services of leading vendors
If you are someone looking for information about DevOps service on the cloud, then this article can help you. We have researched and shared information about the five best cloud service providers from the DevOps perspective: Alibaba Cloud vs AWS vs Google Cloud vs IBM vs Microsoft.
#1 Alibaba Cloud
Alibaba Cloud offers powerful DevOps services which can cover all life cycle of DevOps. Most of its tools and services support multi-cloud environment.
It provides some open source tools like Terraform, Ansible, Chef, and Packer to allow users to easily build their environment. With the support of Terraform and Packer, users can swiftly deploy their applications and infrastructure on Alibaba Cloud. Subsequently, Alibaba Cloud users can save time and focus on everyday business-critical needs.
Furthermore, Alibaba Cloud provides Cloud Container Service that supports the use of Kubernetes clusters to manage container applications in it. It provides some other features like service orchestration, service discovery, auto-scaling, and failure schedule.
There are other products such as Function Compute and ROS by Alibaba Cloud to support the DevOps lifecycle.
Main benefits of Alibaba Cloud DevOps services:
They cover all life cycles of DevOps and support multi-cloud.
There are wide varieties of open source tools and services that Alibaba Cloud provides. Hence, giving you more power to choose your own products as per your business requirements.
The automated O&M tools help users manage complex cloud infrastructure, quickly create images, and deploy applications on multi-cloud infrastructure.
Alibaba Cloud provides data protection from DDoS attacks for free.
AWS (Amazon Web Services) offers a variety of tools aimed at supporting DevOps services practices in the cloud. Some of the core services provided by AWS for DevOps are CodeBuild, CodePipeline, and CodeDeploy. CodeBuild and CodeDeploy help users to automate the container’s delivery and deployment process.
AWS CodeBuild helps to automate the process of building code where it gets compiled, tested, and deployed. It makes sure that your copies of produced artifacts are maintained and stay secure. Then there is AWS Code Deploy that makes it simple for you to manage the builds and helps in instant patch and upgrade at once.
CodePipeline is a continuous delivery service that lets AWS users create a channel or pipeline where all the processes can be automated, required to release the software.
AWS also supports open-source tools like Chef, Puppet, Jenkins, and more.
Main benefits of AWS DevOps services:
It simplifies provisioning and infrastructure management and helps you deliver the services at high velocity.
It increases the efficiency of the overall software development lifecycle processes.
It gives you more control over who can access your resources through AWS Identity and Access Management (IAM).
You only pay for services as and when you use them.
The search giant also provides DevOps related tools and services. Google Cloud Platform manages the software development lifecycle by incorporating configuration management tools – Stackdriver Monitoring, Stackdriver debugger, Stackdriver Logging, and security scanner service (App Engine).
It also provides full cloud development solutions for platforms like Android Studio, Visual Studio, Eclipse, PowerShell, etc. Google Cloud offers management tools like Cloud Deployment Manager and Cloud Console.
Cloud Deployment Manager is an infrastructure deployment service that lets you completely focus on the development part as it automates the creation and management of Google Cloud Platform resources. DevOps team can describe the Deployment Manager about what their deployment should look like and they will apply the relevant tools and processes for you. Once the final procedure is created, it is further repeatable and scalable as required.
Cloud Console is a web-based interface that lets the DevOps team manage resources, monitor issues, reduce costs, and more.
Google Cloud also integrates with Jenkins on Container Engine which lets you manage containers and deploy resources when needed, making your pipeline more organized.
Main benefits of Google Cloud DevOps services:
It automates builds, test, and deployment operations.
It provides monitoring feature that accelerates the troubleshooting, and resumes the apps and services back to their correct state.
IBM introduced DevOps Insights as a set of DevOps services in the cloud. DevOps insights provides comprehensive insights of data across your DevOps tools to boost speed with quality and control. The DevOps Insights integrates with Jenkins, urban code, Travis CI, and many other open-source DevOps tools. It automatically monitors your deployment risks on the basis of test data that you publish.
IBM DevOps Insights can accept data from IBM, third party, or open-source tools. It also supports multi-cloud deployment. With Insights, you can access your apps and services’ status across all DevOps tools in your pipeline from a single dashboard. You can track how they are running, analyze, and yield valuable information as to which jobs are having problems while it runs, and which one is taking the longest time.
Further, you can troubleshoot the issue with your pipelines to increase your deployment velocity and get more control over the process.
Main benefits of IBM DevOps services:
It provides high speed with quality and control.
It has a better security model. It keeps your data encrypted when at rest or in transit and limits the data access to only those users who require it.
It automates the set-up process of the deployment pipeline.
Azure DevOps services involve continuous integration, continuous delivery, infrastructure as a code, monitoring tools, easy and automated provisioning, and configurations. It integrates with most leading DevOps tools in the market like Terraform, Ansible, Chef, and Puppet.
Azure DevOps comprises a wide range of services that cover the complete software development lifecycle. These are:
Azure Boards: Azure Boards is a web service that DevOps teams can use to manage their software projects. It offers features like customizable dashboards, native support for Scrum and Kanban, and integrated reporting. With this, you can also track your work quickly and easily.
Azure Pipelines: This is a cloud, CI/CD platform that supports containers or Kubernetes enabling you to build and test your code project automatically, quickly, and consistently.
Azure Repos: Use this set of tools to manage and track your code performance.
Azure Artifacts: It offers support for npm, Maven, Python, and NuGet package feeds from public and private sources.
Azure Test Plans: It offers rich and powerful solutions for planned manual testing, user acceptance testing, and exploratory testing.
Main benefits of Microsoft DevOps services:
They are globally available, reliable, and scalable. They also provide an SLA of 99.9% uptime.
You can use this to build, test, and deploy in any language to any cloud or on-premises.
Any size team can work on the same solution seamlessly.
In this article, we discussed the concept of DevOps, its benefits, and how it fits in an organization’s software development lifecycle. We also have gone through and added information about major cloud service providers from a DevOps perspective.
Please help us improve this article by adding your suggestions in the comments section.
Disclaimer: The information contained in this article is for general information purpose only. Price, product and feature information are subject to change. This information has been sourced from the websites and relevant resources available in the public domain of the named vendors on 4 August 2020. DHN makes best endeavors to ensure that the information is accurate and up to date, however, it does not warrant or guarantee that anything written here is 100% accurate, timely, or relevant to the website visitors.
StormWall, an international cybersecurity company, is launching an infosec grant program for African ISPs. Internet service providers will get an opportunity to receive a $3,960 grant, allowing them to create a robust cyber defense system for their online resources. The grant will equip companies with a 6-month BGP network protection against DDoS attacks with a legitimate bandwidth of up to 100 megabits per second.
The African internet service market has recently experienced a period of rapid expansion, and the coronavirus epidemic has increased the growth dynamics in the industry even further. The emerging trend may become a turning point in development for many African countries. However, due to the boom in internet consumption, the number of cyberattacks may also rise.
Cyberthreats can shut down internet providers indefinitely, leaving customers without communication. Unfortunately, most African companies lack the financial capacity to shield their online resources against cyber threats and become easy prey for hackers. StormWall has allocated cybersecurity grants to help African providers secure their networks, giving them the opening for further development and growth.
StormWall has outlined several conditions for receiving grants. To qualify for a grant, a company must have valid registration. It must run a business providing internet access in Africa and must be working for at least three years. Businesses can submit a claim on the StormWall’s website by completing the protection request form. StormWall will contact eligible companies within 5 working days. Each application is considered individually. StormWall provides grants exclusively to help applicants protect their networks. The right to resell services is not provided. The collection of applications will be carried out until September 1st, 2020.
StormWall has extensive working experience in African countries. The company has already implemented several large-scale projects, that defend the leading African providers from cyberattacks. StormWall’s practice allowed experts to conclude that the requirements of African companies fully match with the needs of organizations in other nations.
“Over the last 15 years, Africa has made a significant technological leap. Today, African technology development dynamics is ahead of that in Europe and the United States. The most promising sectors are e-commerce and fintech. I am sure that the African continent has exceptional potential in the field of information technology. Of course, historically developed challenges of the continent — poverty, and crime — have not disappeared, and they complicate the evolution of the information society. Our mission is to help African countries achieve a new level of technological advancement. For this purpose, we launched a cybersecurity grant program for African Internet providers. We hope other companies in the IT and information security market will follow our example,” states Ramil Khantimirov, co-founder and CEO of StormWall.
Hivelocity, a provider of global IaaS and Edge Compute services, has announced the opening of 17 new data center locations as it further expands its edge computing platform. The addition of over a dozen new data centers across North America, Europe, Asia and Australia provide Hivelocity customers with a total of 32 available data centers in 26 markets across the globe in which to deploy edge compute services. Further, Hivelocity announced the immediate availability of its public API which provides users with advanced functions and data points that help automate, simplify and maximize their edge computing solution.
With a total of 32 data centers in 26 cities across 4 continents, Hivelocity has created one of the most geographically diverse and comprehensive edge computing platforms and infrastructures in the world. Hivelocity enables its customers to instantly deploy bare-metal servers across any of its data centers with ease. By leveraging the Hivelocity platform users can easily manage and scale their edge compute solutions when desired.
With the release of its robust API users can also now control their infrastructure with code and integrate existing tools with their Hivelocity compute solutions. Any action performed within the platform; deploying servers, accessing servers, managing bandwidth, networking etc., can be accomplished through the API. With integrations to Terraform, Ansible and other favorite Ops tools, repeatable single tenant bare-metal is just a config file away.
“Our goal is to make our bare-metal at the edge as simple to consume, manage and scale as possible,” says Steve Eschweiler, Hivelocity’s COO. “The more powerful and simplified we make our platform, the better armed our customers are to dominate in their sector. With our platform exclusively deploying and managing bare-metal dedicated servers, we provide our customers with maximum resources, control and predictable costs.”
Hivelocity’s new edge-compute markets are Ashburn, Chicago, Hong Kong, London, Madrid, Milan, Newark, Paris, Reston, Seoul, Singapore, Stockholm, Sunnyvale, Sydney, Tokyo, Toronto and Vancouver. These new locations join Hivelocity’s previously established data centers in Amsterdam, Atlanta, Dallas, Frankfurt, Los Angeles, Miami, New York City, Seattle and Tampa.
Hivelocity was founded in 2002 and currently serves thousands of businesses from over 130 countries worldwide. To learn more about Hivelocity please visit https://www.hivelocity.net/.