CIO - Not so long ago, if Underwriters Laboratories Inc. needed to add three additional servers worth of computing power in three weeks, the company just bought three new units, says Kent Walker, manager of computer operations at UL. With more time for capacity analysis -- which is both labor- and time-intensive -- Walker might be able to shift resources around and stave off such purchases. But when the need is immediate, there's no time for that.
That's why CiRBA Inc.'s data center intelligence tool appealed to Walker. The product is designed to automatically analyze a server installation and recommend ways to consolidate hardware using virtual servers instead of physical ones. Walker asked CiRBA for a 100-server trial evaluation, to see if the vendor could help rein in his sprawl of underutilized servers.
"They dazzled us," Walker says.
Other tools can tackle pieces of the analysis, such as examining server utilization levels to plan capacity. But CiRBA's tool is comprehensive, letting users factor in technical, business and workload constraints specific to their organizations, analysts say. A financial institution, for instance, might need to stipulate that its trading and research groups not share the same virtual servers because of regulatory requirements.
CiRBA gathers information about the data center in a variety of ways, including agent and agentless discovery mechanisms, and stores the details in a central repository. The tool then applies rules and constraints, builds a multidimensional model of the data center and spits out answers on the best way to optimize it, whether it's a physical server farm, a virtual environment or a mixture of both.
"Very few people have the majority of their servers virtualized, so our customers tend to put it on everything. They get a lot of visibility into both" their physical and virtual environments, says Andrew Hillier, CiRBA's co-founder and chief technology officer.
CiRBA's Version 4.6, released last month, added advanced benchmarking to enable users to analyze how specific server workloads would perform on any virtualization platform and fit with other resources. It also built in probability analysis and workload-scoring strategies that factor in the risk and service levels an organization is willing to accept, as well as advanced network and storage analysis.
With those new features, the tool for the first time can be used to model how applications will perform on IBM's System z mainframes and determine which workloads running in Unix- or x86-based environments might be better suited to mainframes.
The upcoming 5.0 version, due in the summer, will build dynamic models of the data center, tracking daily changes, according to Hillier. A subsequent fall release is expected to provide tighter integration with major system management frameworks, as well as enhanced support for storage analysis, he adds.
- Considerations For Effective Software License Management For many reasons, software license management has become a critical issue for many IT organizations and enterprise's alike. With many licensing options, hurdles...
- Vblock™ Specialized System for High Performance Databases Learn how Vblock™ Specialized Systems have been designed to deliver high performance in the millions of IOPS - with millisecond response times.
- Innovating Cloud Backup with Veeam Veeam leverages virtualization and cloud technologies to innovate and protect the modern data center.
- Virtualization Best Practices Guide Whether you're running applications on physical or virtual machines, you still need to stay vigilant to guard against the constant and growing hazard...
- Building Tomorrow's Data Center with Converged Technologies A number of forces are converging: the cloud, converged infrastructure, big data and fabric architectures to name a few.
- Virtual Desktop Must-haves: Cost-effectiveness, Scalability and Security Employees today expect to take their work anywhere and on any device. This BYOD mindset poses a dilemma for organizations that want to... All Virtualization White Papers | Webcasts