Cassandra 1.2 geared to 'fat servers'
Updated Apache database offers approach to managing data on nodes
IDG News Service - Adjusting to changes in corporate hardware buying habits, the Apache Software Foundation's Cassandra NoSQL distributed database has been updated to better use larger servers through the introduction of virtual nodes and configurable policies for disk failure.
The newly released Cassandra 1.2 also features the ability to perform atomic batch operations, and comes with a new version of the Cassandra Query Language, CQL3.
Traditionally, "Cassandra's sweet spot has been in scaling out across a lot of relatively lightweight machines," said Jonathan Ellis, Apache project chair of Cassandra and a co-founder of DataStax, which offers commercial support for the software.
Recently however, more organizations have been buying "denser" servers with more memory and hard drive space because such servers now offer the best price-performance value, Ellis said. So much of the work to update Cassandra has been around better supporting such machines.
First developed internally at Facebook and released to the public in 2008, Cassandra was designed to store massive data sets across multiple servers. Adobe, Cisco, Disney, eBay, IBM, Netflix, Reddit, Spotify, Twitter and Williams-Sonoma all use the technology.
The new version of the software features the ability to create virtual nodes (vnodes), designed to streamline the recovery process should an individual server in a Cassandra cluster fail. Vnodes should also improve performance in general.
Vnodes was one of the chief features of Amazon's Dynamo distributed data store, which the developers of Cassandra used as the model for Cassandra. The developers initially opted for a simpler architecture of assigning one node to each server.
The new virtual node technology should simplify the process of managing clusters, particularly when adding and rebuilding individual nodes. With vnodes, each server can hold multiple nodes.
Because individual vnodes don't have to occupy the entire space of a server, multiple nodes can be created on each server, which then can be re-created more quickly from replicas on other servers should the drive fail, compared to the time it would take to copy the replicated data from a single server.
Smaller nodes spread across a greater number of servers also better balances the workload among all the machines in a cluster.
"Each virtual node is managed by one Java process per machine, so we're not adding a lot of operating system processes. We're just virtualizing out that storage responsibility," Ellis said.
Another new feature, Atomic batching, should help organizations that require transactional integrity across business processes, such as an online merchant that needs to make sure orders are captured even when a component such as a hard drive fails right in the middle of a transaction. Previously, developers would have to build processes, such as retry mechanisms, into their code to guarantee transactional integrity.
- 15 Non-Certified IT Skills Growing in Demand
- How 19 Tech Titans Target Healthcare
- Twitter Suffering From Growing Pains (and Facebook Comparisons)
- Agile Comes to Data Integration
- Slideshow: 7 security mistakes people make with their mobile device
- iOS vs. Android: Which is more secure?
- 11 sure signs you've been hacked
- Aberdeen Group: Marketing Analytics for Manufacturing: Forging Customer Insights There are no recalls for poor marketing. Manufacturers need to get their customer intelligence and messaging right the first time. Learn how.
- SIEM: Keeping Pace with Big Security Data Learn how SIEM can have the right database back-end and offer security intelligence that leverages contextual data to achieve a strong security posture...
- Pay-as-you-Grow Data Protection: IBM Tivoli's Full-featured Data Protection Suite for Small to Medium Businesses IBM Tivoli Storage Manager Suite for Unified Recovery gives small and medium businesses the opportunity to start out with only the individual solutions...
- Streamline Data Protection with IBM Tivoli Storage Manager Operations Center IBM Tivoli Storage Manager (TSM) has been an industry-standard data protection solution for two decades. But, where most competitors focus exclusively on Backup...
- Meg Whitman presents Unlocking IT with Big Data During this Web Event you will hear Meg Whitman, President and CEO, HP discuss HAVEn - the #1 Big Data platform, as well...
- The New Way to Work Knowledge Vault This Knowledge Vault focuses on how, in today's increasingly virtual world, it's more important than ever to engage deeply with employees, suppliers, partners,... All Databases White Papers | Webcasts