The adoption of server virtualization technology follows several trajectories. We could consider the breadth of its penetration in terms of the number of organisations using it today, or we could consider the depth of its use in individual organisations, in terms of what they actually do with it.
The latter is of more relevance to IT departments that have already taken first steps down the server virtualization path. For other organizations, if we think beyond the workloads already being run in a virtualized environment, is there a ‘next’. And if there is, what is it?
Perhaps this is simplistic. When you think about virtualization do you think in terms of a proportion of the x86 server estate in your datacenter that ‘could’ be virtualized or do you think about the different types of workloads that need executing? Vendors with a vested interest in shifting virtualization technology tend to presume that something that could be a candidate for virtualization, automatically will be.
However, we know from research that decision-making is typically focused on a simpler criterion: can it save money right now? As a result, virtualization tends to be employed for the more straightforward workloads that can be easily consolidated.
Admittedly, the notion of ‘straight forward’ is relative, although there are some commonly accepted candidates such as print servers, web servers and the like. Whether these are chosen because they are seen as cost-saving, low risk, ‘non-core’ or ‘non-critical’ areas, it’s where most organizations cut their teeth. So where do we go from here? The answer has to be into areas of higher potential risk, and less evident cost-benefit. So then: what is the rationale for making decisions?
Work up the list
To reiterate, the factors at play are: cost savings; virtualization benefit; business importance; and migration of risk. Does IT simply ‘work up the list’ from least risk / importance? Or are those with prior experience now applying virtualization to areas which would benefit specifically from it, regardless of their importance to the business?
Factors around migration-risk bring into question enough experience and confidence exists in the technology itself and on the periphery (availability, resilience and back-up and recovery systems), as well as the skills of the IT department itself to be able to consider higher-risk workloads as virtualization candidates.
One must also take into consideration the socio-political aspects of IT ownership. A line of business leader might have concerns about ‘his’ application running in a virtualized environment, even if he’s perfectly happy with the service he gets from ‘lower value’ services. But if the technology is proven elsewhere, what’s the fuss?
Part of the answer could lie in how big the first step down the virtualization route was. Did the IT department have to fight to make it to happen, or did someone in the business make a request for it directly or indirectly – e.g., a demand that could only be fulfilled by employing technology in this way?
One argument suggests that had it not been for the economic crisis in 2008, many organizations would not have felt it necessary to virtualise any server infrastructure.
Have You Read This?
Generative AI Checkpoint
From Barcode Scanning to Smart Data Capture
Beyond the Barcode: Smart Data Capture
The Evolving Role of Converged Infrastructure in Modern IT
Evaluating the Potential of Hyper-Converged Storage
Kubernetes as an enterprise multi-cloud enabler
A CX perspective on the Contact Centre
Automation of SAP Master Data Management