Amazon launches workflow orchestration service
The AWS Data Pipeline can stream big data analysis jobs
IDG News Service - Users of Amazon Web Services will soon be able to orchestrate workflows across different AWS services and their own internal resources, using a new orchestration engine called the AWS Data Pipeline.
Amazon Chief Technology Officer Werner Vogels introduced the technology at the company's Re:Invent conference, held this week in Las Vegas. The service is now available in limited beta preview, though Vogels did not say when it would be commercially available, nor what the price would be.
The service can "automate the movement and processing of any amount of data using data-driven workflows and built-in dependency checking," according to a blog post AWS issued that further explained the technology.
Amazon designed the service to automate the process of parsing large sets of data. For example, one pipeline can move log data from an AWS EC2 (Elastic Cloud Compute) instance to the AWS S3 (Simple Storage Service) once a day, and then, once a week, evoke an analysis job on the data on an AWS Elastic MapReduce cluster.
To set up a workflow pipeline, the user identifies some data sources and describes the steps that AWS should take to process the data. The user would also identify the destination for the processed data as well as a schedule for when the pipeline should be executed. Preconditions can also be established that the service will check before executing a job, such as checking if a file that is needed for the operation exists.
Pipelines can run across EC2, Elastic MapReduce clusters, and the user's own hardware. Pipelines can be set up in the AWS Management Console or by writing a script.
This is not the first workflow engine on AWS. The company also launched the Amazon Simple Workflow in February. However, AWS Data Pipeline is more focused on executing data-driven jobs.
The AWS Data Pipeline is one of a number of announcements Amazon made at the conference. The company also unveiled a data warehouse service and an auto-discovery service to ease the management of its ElastiCache. It has also cut the prices of some of its storage services and created two new EC2 instance types, for high-memory usage and large data usage.
- Improving IT Efficiencies: Four Advantages of Multi-Tenant Data Centers Increasing demands on IT are forcing organizations to rethink their data center options. For many organizations, that means turning to the flexibility afforded...
- Accelerating Cloud Deployment and Operations with Managed Services Companies that do not have sufficient in-house expertise to either deploy or maintain an IaaS cloud should turn to Managed Service Providers .
- Rethinking IT Operations in the Cloud This paper breaks down the challenges that often prevent the cloud from delivering the fast, flexible and affordable infrastructure companies seek - and...
- Gartner Magic Quadrant for Cloud-Enabled Managed Hosting, North America Cloud-enabled managed hosting brings cloudlike consumption and provisioning attributes to the traditional managed hosting market
- The Key to Happiness: Throw out Your Data Warehouse In this webinar, Kerry Reitnauer, Director, Solution Architect at FairPoint Communications will discuss the challenges the data warehouse brought, how they migrated to...
- Building Tomorrow's Data Center with Converged Technologies A number of forces are converging: the cloud, converged infrastructure, big data and fabric architectures to name a few. All Data Center White Papers | Webcasts
Our new bimonthly Internet of Things newsletter helps you keep pace with the rapidly evolving technologies, trends and developments related to the IoT. Subscribe now and stay up to date!