New Hive and performance features in WANdisco Fusion 2.9

WANdisco Fusion 2.9 delivers new levels of performance and scalability, with a number of enhancements that make it easier to support hybrid big data deployments with continuous and consistent access across any mix of on-premise and cloud environments.

One of the most significant new features is the application of WANdisco’s patented active transactional replication to the Hive metastore, used to support familiar SQL-like access to Hadoop. Changes made to table definitions and other objects in the Hive metastore are replicated to Hive metastore instances deployed with other Hadoop clusters. Data added to the underlying Hadoop cluster referencing the new object definitions is replicated to other Hadoop clusters in a subsequent step, to guarantee consistent access across clusters and locations.

Hadoop clusters are often built around commodity storage, but architects now have a wide selection of Big Data storage choices, including solid-state or spinning disk for clusters and enterprise storage for compatibility layers and connectors.

In this webinar, our experts will review the storage options available to Hadoop architects and provide recommendations for each use case, including an active-active replication option that makes data available across multiple storage systems.

WANdisco Fusion 2.9 delivers new levels of performance and scalability, with a number of enhancements that make it easier to support hybrid big data deployments with continuous and consistent access across any mix of on-premise and cloud environments.

One of the most significant new features is the application of WANdisco’s patented active transactional replication to the Hive metastore, used to support familiar SQL-like access to Hadoop. Changes made to table definitions and other objects in the Hive metastore are replicated to Hive metastore instances deployed with other Hadoop clusters. Data added to the underlying Hadoop cluster referencing the new object definitions is replicated to other Hadoop clusters in a subsequent step, to guarantee consistent access across clusters and locations.

WANdisco Fusion 2.9 delivers new levels of performance and scalability, with a number of enhancements that make it easier to support hybrid big data deployments with continuous and consistent access across any mix of on-premise and cloud environments.

One of the most significant new features is the application of WANdisco’s patented active transactional replication to the Hive metastore, used to support familiar SQL-like access to Hadoop. Changes made to table definitions and other objects in the Hive metastore are replicated to Hive metastore instances deployed with other Hadoop clusters. Data added to the underlying Hadoop cluster referencing the new object definitions is replicated to other Hadoop clusters in a subsequent step, to guarantee consistent access across clusters and locations.

GigaOM analyst William McKnight will be joined by experts from Qubole and WANdisco, who will explain the benefits of moving to the cloud and review the tools available for cloud migration and hybrid cloud deployments.

Learn what's required to avoid the downtime and business disruption that often accompany cloud migration projects.

Global enterprises have quietly funneled enormous amounts of data into Hadoop over the last several years. Hadoop has transformed the way organizations deal with big data. By making vast quantities of rich unstructured and semi-structured data quickly and cheaply accessible, Hadoop has opened up a host of analytic capabilities that were never possible before, to drive business value.

The challenges have revolved around operationalizing Hadoop to enterprise standards, and leveraging cloud-based Hadoop as a service (HaaS) options offering a vast array of analytics applications and processing capacity that would be impossible to deploy and maintain in-house.

This webcast will explain how solutions from IBM and WANdisco address these challenges by supporting:

- Continuous availability with guaranteed data consistency across Hadoop clusters any distance apart, both on-premises and in the cloud.
- Migration to cloud without downtime and hybrid cloud for burst-out processing and offsite disaster recovery.
- Flexibility to eliminate Hadoop distribution vendor lock-in and support migration to cloud without downtime or disruption.
- IBM's BigInsights in the cloud, and BigSQL, which allows you to run standard ANSI compliant SQL against your Hadoop data.

Data is coming from everywhere. The first challenge is just being able to get hold of it, curate and convey it in a secure and transparent manner. Hortonworks Data Flow is the tool that collects data at the edge, processes and secures data in motion and brings data into your data-at-rest platform (HDP).

Once you have your data collected in a valuable data lake, you need resilience, control over its location, and safety against failure. That’s where Wandisco Fusion & Hortonworks HDP come in. With Wandisco Fusion on HDP, an enterprise can now build an effective, fast and secure data engine out of multiple Hadoop clusters, getting the most business value out of its HDP deployment with a reliable and high-performing Big Data service.

Join Hortonworks & WANdisco on this webinar to learn how to make this into reality.

Global banks have the most rigorous availability, performance and data security standards. Join 451 Research and WANdisco as we explore the cutting-edge techniques leading financial services firms are using to fully operationalize Hadoop to meet these standards and leap ahead of their competition. Register for this webinar and get the free white paper entitled "Bringing Hadoop into the Banking Mainstream"

Solutions for seamlessly moving data between on-premise and cloud environments are virtually non-existent. This webinar explains how to achieve a true hybrid cloud deployment that supports on-demand burst-out processing in which data moves in and out of the cloud as it changes, and enable the cloud to be used for offsite disaster recovery without downtime or data loss.

In the traditional world of EDW, ETL pipelines are a troublesome bottleneck when preparing data for use in the data warehouse. ETL pipelines are notoriously expensive and brittle, so as companies move to Hadoop they look forward to getting rid of the ETL infrastructure.

But is it that simple? Some companies are finding that in order to move data between clusters for backup or aggregation purposes, they are building systems that look an awful lot like ETL.

Disaster recovery sites are typically underutilized with idle hardware and software that are only used in an emergency. Why let your valuable resources remain idle?

In this webinar, you’ll learn how you can take full advantage of the resources in what would ordinarily be your DR site by using active-active replication to provide full utilization as well as complete failover with lower RPO and RTO.

Migrating your Hadoop cluster between versions or distributions is difficult. It is a critical process that if done incorrectly can lead to the loss of data, system downtime, and disruption of business activities.

In this webinar, learn how you can mitigate the risk in a migration through the development of a comprehensive migration strategy and leveraging tools like those from WANdisco to simplify and automate your migration.

Migrating your Hadoop cluster between versions or distributions is difficult. It is a critical process that if done incorrectly can lead to the loss of data, system downtime, and disruption of business activities.

In this webinar, learn how you can mitigate the risk in a migration through the development of a comprehensive migration strategy and leveraging tools like those from WANdisco to simplify and automate your migration.

To make your Hadoop deployment enterprise-class, robust data replication is required to support business-critical functions that depend on Hadoop. Firms can no longer rely on the status quo of traditional, slow data backups into underutilized hardware.

Paul Scott-Murphy, Vice President of Field Technical Services for WANdisco will introduce WANdisco Fusion, the only active-active solution for total data protection and availability across Hadoop distributions and storage and the advantages of implementing this in your business.

A growing number of Hadoop adopters are making use of both on-premise and cloud clusters. Production workloads run on-premise to provide the best performance and security, while cloud clusters are reserved for testing, development, and burst-out processing power. In this webinar we'll review best practices for managing a hybrid environment including security, data transfer, and performance.

Hadoop clusters are often built around commodity storage, but architects now have a wide selection of Big Data storage choices. Hadoop clusters can use a mix of solid-state and spinning disk storage, while Hadoop compatibility layers and connectors can use enterprise storage systems or share storage between Hadoop and legacy applications.

In this webinar, 451 Research Director Matt Aslett will review the storage options available to Hadoop architects and provide recommendations for each use case. WANdisco's Randy DeFauw will then present an active-active replication option that makes data available across multiple storage systems.

Once believed to be impossible, WANdisco's patented technology allows Big Data to be stored and queried with absolute reliability and security, unleashing limitless possibilities for innovation. That's Hadoop without limits. We cover topics such as hardening Hadoop for the enterprise, simplifying audit and compliance, and getting the most out of your multi-data center Hadoop investment. These interactive presentations are targeted at enterprise architects and IT infrastructure staff who are designing and implementing big data environments with Hadoop, HBase and related technologies.