Skip the navigation

Hortonworks releases its Hadoop version

For its first Hadoop release, Hortonworks focused on making the data analysis software easy to deploy and monitor

By Joab Jackson
June 12, 2012 08:00 AM ET

IDG News Service - For the first production release of what will be its flagship Apache Hadoop distribution, Hortonworks has focused on providing a set of tools to help deploy, manage and extend the data analysis platform.

"Hortonworks' goal is to make Hadoop easy to use and consume," said John Kreisa, Hortonworks vice president of marketing.

Version 1 of the Hortonworks Data Platform (HDP), to be released June 15, will be Hortonworks' first production-ready product release. Hortonworks was set up a year ago by Yahoo, along with Benchmark Capital, to provide enterprise support for Hadoop, the large-scale data analysis platform. Yahoo played a pivotal role in the early development of Hadoop.

Hortonworks now competes with a number of other companies also offering support packages, including Cloudera, MapR and IBM. Microsoft has chosen Hortonworks' Hadoop distribution for use on its Azure cloud service, though that service, promised by the end of 2011, has not debuted yet.

Like other commercial Hadoop packages, HDP packages a number of different open-source Hadoop components, including the latest versions of the Pig scripting engine, the Hive data warehousing software and the HBase database.

In addition to these basic components, Hortonworks added a number of additional management and interoperability tools to the package, all of them based on open-source projects as well.

To aid in management, the package includes a customized version of Apache Ambari, a Hadoop monitoring and lifecycle management program. With this software, an administrator can set up a single Hadoop instance across a number of servers. Once Hadoop is installed, the software then monitors performance of the servers as well as the Hadoop jobs themselves, presenting the data on a dashboard.

"The dashboards are customizable and the APIs [application programming interfaces] allow the management and monitoring functionality to be tied into third-party dashboards like Hewlett-Packard's OpenView or Teradata's Viewpoint," Kreisa said.

With this release, the management tools will only be able to manage a single cluster, though future versions may be able to manage multiple clusters, said Ari Zilka, Hortonworks chief products officer. Specific metrics that are being captured include network utilization, throughput and latency, and usage of CPUs, memory and disks. Jobs in Hadoop are also measured, including the time it takes for a task to start, how many tasks there are on backlog, how many data blocks a task uses and where these data blocks are located.

For data interoperability, the package includes a metadata catalogue that should make it easier for business intelligence and other data analysis products to query Hadoop datasets. Based on Apache HCatalog, this metadata repository provides pointers to Hadoop data in a set of tables that can be easily queried by tools commonly used for relational databases, enterprise data warehouses and other structured data systems.

Reprinted with permission from Story copyright 2014 International Data Group. All rights reserved.
Our Commenting Policies
Internet of Things: Get the latest!
Internet of Things

Our new bimonthly Internet of Things newsletter helps you keep pace with the rapidly evolving technologies, trends and developments related to the IoT. Subscribe now and stay up to date!