Hi [[ session.user.profile.firstName ]]

WANdisco | Big Data & Cloud

  • Date
  • Rating
  • Views
  • Three ways to get a handle on data governance in hybrid cloud environments Three ways to get a handle on data governance in hybrid cloud environments Experts from New Context and WANdisco Recorded: Nov 3 2016 37 mins
    Join experts from New Context and WANdisco as they explain the tools and techniques for securely managing big data in hybrid cloud environments.

    Attendees will learn what's required for effective data governance with the highest levels of availability and performance on-premises and in the cloud.
  • Big data storage: options & recommendations Big data storage: options & recommendations Jagane Sundar, WANdisco CTO Recorded: Oct 20 2016 51 mins
    Hadoop clusters are often built around commodity storage, but architects now have a wide selection of Big Data storage choices, including solid-state or spinning disk for clusters and enterprise storage for compatibility layers and connectors.

    In this webinar, our experts will review the storage options available to Hadoop architects and provide recommendations for each use case, including an active-active replication option that makes data available across multiple storage systems.
  • New Hive and performance features in WANdisco Fusion 2.9 New Hive and performance features in WANdisco Fusion 2.9 WANdisco Recorded: Oct 6 2016 47 mins
    WANdisco Fusion 2.9 delivers new levels of performance and scalability, with a number of enhancements that make it easier to support hybrid big data deployments with continuous and consistent access across any mix of on-premise and cloud environments.

    One of the most significant new features is the application of WANdisco’s patented active transactional replication to the Hive metastore, used to support familiar SQL-like access to Hadoop. Changes made to table definitions and other objects in the Hive metastore are replicated to Hive metastore instances deployed with other Hadoop clusters. Data added to the underlying Hadoop cluster referencing the new object definitions is replicated to other Hadoop clusters in a subsequent step, to guarantee consistent access across clusters and locations.

    WANdisco Fusion 2.9’s patented technology also:

    • Enables throughput in excess of 100,000 transactions per minute - well beyond the largest enterprise requirements
    • Ensures operations are not affected by transient network outages or “flapping WAN” links
    • Supports scenarios in hybrid cloud deployments where on-premise environments cannot accept inbound network connection requests

    Join us on October 6th at 10AM Pacific, 1PM Eastern as we present and demo the latest release of WANdisco Fusion.
  • New Hive and performance features in WANdisco Fusion 2.9 (EMEA) New Hive and performance features in WANdisco Fusion 2.9 (EMEA) Paul Scott-Murphy, VP Product Management Recorded: Oct 5 2016 50 mins
    WANdisco Fusion 2.9 delivers new levels of performance and scalability, with a number of enhancements that make it easier to support hybrid big data deployments with continuous and consistent access across any mix of on-premise and cloud environments.

    One of the most significant new features is the application of WANdisco’s patented active transactional replication to the Hive metastore, used to support familiar SQL-like access to Hadoop. Changes made to table definitions and other objects in the Hive metastore are replicated to Hive metastore instances deployed with other Hadoop clusters. Data added to the underlying Hadoop cluster referencing the new object definitions is replicated to other Hadoop clusters in a subsequent step, to guarantee consistent access across clusters and locations.

    WANdisco Fusion 2.9’s patented technology also:

    •Enables throughput in excess of 100,000 transactions per minute - well beyond the largest enterprise requirements
    •Ensures operations are not affected by transient network outages or “flapping WAN” links
    •Supports scenarios in hybrid cloud deployments where on-premise environments cannot accept inbound network connection requests

    Join us on October 5th as we present and demo the latest release of WANdisco Fusion.
  • Migrating your big data infrastructure to cloud Migrating your big data infrastructure to cloud GigaOM analyst William McKnight with experts from Qubole and WANdisco Recorded: Sep 8 2016 59 mins
    GigaOM analyst William McKnight will be joined by experts from Qubole and WANdisco, who will explain the benefits of moving to the cloud and review the tools available for cloud migration and hybrid cloud deployments.

    Learn what's required to avoid the downtime and business disruption that often accompany cloud migration projects.

    Limited Time Offer - View Qubole and WANdisco's Special Quick Start Package: http://bit.ly/2cGupC6
  • Deploying mission critical applications on Hadoop, on-premises and in the cloud Deploying mission critical applications on Hadoop, on-premises and in the cloud Jim Wankowski, IBM Worldwide Cloud Data Services and James Campigli, Co-Founder, Chief Product Officer of WANd Recorded: Jul 21 2016 63 mins
    Global enterprises have quietly funneled enormous amounts of data into Hadoop over the last several years. Hadoop has transformed the way organizations deal with big data. By making vast quantities of rich unstructured and semi-structured data quickly and cheaply accessible, Hadoop has opened up a host of analytic capabilities that were never possible before, to drive business value.

    The challenges have revolved around operationalizing Hadoop to enterprise standards, and leveraging cloud-based Hadoop as a service (HaaS) options offering a vast array of analytics applications and processing capacity that would be impossible to deploy and maintain in-house.

    This webcast will explain how solutions from IBM and WANdisco address these challenges by supporting:

    - Continuous availability with guaranteed data consistency across Hadoop clusters any distance apart, both on-premises and in the cloud.
    - Migration to cloud without downtime and hybrid cloud for burst-out processing and offsite disaster recovery.
    - Flexibility to eliminate Hadoop distribution vendor lock-in and support migration to cloud without downtime or disruption.
    - IBM's BigInsights in the cloud, and BigSQL, which allows you to run standard ANSI compliant SQL against your Hadoop data.
  • Build An Effective, Fast And Secure Data Engine With Hortonworks & WANdisco Build An Effective, Fast And Secure Data Engine With Hortonworks & WANdisco Dave Russell, Hortonworks and Mark Lewis, WANdisco Recorded: Jun 23 2016 45 mins
    Data is coming from everywhere. The first challenge is just being able to get hold of it, curate and convey it in a secure and transparent manner. Hortonworks Data Flow is the tool that collects data at the edge, processes and secures data in motion and brings data into your data-at-rest platform (HDP).

    Once you have your data collected in a valuable data lake, you need resilience, control over its location, and safety against failure. That’s where Wandisco Fusion & Hortonworks HDP come in. With Wandisco Fusion on HDP, an enterprise can now build an effective, fast and secure data engine out of multiple Hadoop clusters, getting the most business value out of its HDP deployment with a reliable and high-performing Big Data service.

    Join Hortonworks & WANdisco on this webinar to learn how to make this into reality.
  • Bringing Hadoop into the Banking Mainstream Bringing Hadoop into the Banking Mainstream James Curtis (451 Research) & Jim Campigli (WANdisco) Recorded: Jun 9 2016 60 mins
    Global banks have the most rigorous availability, performance and data security standards. Join 451 Research and WANdisco as we explore the cutting-edge techniques leading financial services firms are using to fully operationalize Hadoop to meet these standards and leap ahead of their competition. Register for this webinar and get the free white paper entitled "Bringing Hadoop into the Banking Mainstream"
  • Making Hybrid Cloud a Reality Making Hybrid Cloud a Reality Jim Campigli and Jagane Sundar Recorded: Apr 21 2016 39 mins
    Solutions for seamlessly moving data between on-premise and cloud environments are virtually non-existent. This webinar explains how to achieve a true hybrid cloud deployment that supports on-demand burst-out processing in which data moves in and out of the cloud as it changes, and enable the cloud to be used for offsite disaster recovery without downtime or data loss.
  • ETL and Big Data: Building Simpler Data Pipelines ETL and Big Data: Building Simpler Data Pipelines Chris Almond, Solutions Architect Recorded: Feb 11 2016 53 mins
    In the traditional world of EDW, ETL pipelines are a troublesome bottleneck when preparing data for use in the data warehouse. ETL pipelines are notoriously expensive and brittle, so as companies move to Hadoop they look forward to getting rid of the ETL infrastructure.

    But is it that simple? Some companies are finding that in order to move data between clusters for backup or aggregation purposes, they are building systems that look an awful lot like ETL.