By powering parallel processing across large clusters of commodity server hardware, Hadoop provides a means of cost-effectively analyzing "big data" as well as a robust ecosystem of open-source analytics software options. The business case for deploying a data lake – is so compelling that for most enterprises the question now is not whether to do it but how to do it. Key implementation questions including which of the several Hadoop software distributions to use, whether to use Hadoop on-premises or in the cloud, and what Hadoop data ingestion tool to use to load data from your source systems into Hadoop. Increasingly, businesses are having an easy time selecting a Hadoop data ingestion tool as they've found that Attunity Replicate provides features and benefits that set it apart from competing tools.
Attunity's innovative solutions for data migration and integration are used by thousands of businesses, from high tech start-ups to Fortune 500 enterprises. While businesses count on Attunity technology to solve a wide range of modern data management challenges, an increasingly popular use case is to employ Attunity Replicate as a Hadoop data ingestion tool.
For organizations wanting to move quickly to capture the benefits of Hadoop-powered data lake analytics, Attunity Replicate offers key advantages as a Hadoop data ingestion tool. Attunity Replicate:
Unlike a typical narrowly restrictive Hadoop data ingestion tool, Attunity Replicate's business value extends well beyond loading data into your Hadoop cluster. For example, a common Hadoop workflow entails moving processed data --- the output of Hadoop map-reduce jobs – out of the data lake and into some other system where it can be preserved and/or accessed by applications and users. Attunity supports all such workflows, allowing for easily created and executed migrations of processed data from Hadoop to destination systems such as databases, conventional data warehouses and data marts, or cloud-based targets such as Amazon S3.
More broadly, while excelling as a Hadoop data ingestion tool, Attunity Replicate can serves all your organization's needs for moving and integrating big data. With a single unified data integration platform, you can not only meet your data warehouse and Hadoop operational needs, but also manage tasks such as database migrations, replicating data across multiple data centers, or replicating data from on-premises systems to the cloud.
Once data is ingested and landed in Hadoop, IT often still struggles to create usable analytics data stores. Traditional methods require Hadoop-savvy ETL programmers to manually code the various steps – including data transformation, the creation of Hive SQL structures, and reconciliation of data insertions, updates and deletions to avoid locking and disrupting users. The administrative burden of ensuring data is accurate and consistent can delay and even kill analytics projects.
Attunity Compose for Hive automates the creation, loading and transformation of enterprise data into Hadoop Hive structures. Our solution fully automates the pipeline of BI ready data into Hive, enabling you to automatically create both Operational Data Stores (ODS) and Historical Data Stores (HDS). And we leverage the latest innovations in Hadoop such as the new ACID Merge SQL capabilities, available today in Apache Hive (part of the Hortonworks 2.6 distribution), to automatically and efficiently process data insertions, updates and deletions.
Attunity Replicate integrates with Attunity Compose for Hive to simplify and accelerate data ingestion, data landing, SQL schema creation, data transformation and ODS and HDS creation/updates.