Hi [[ session.user.profile.firstName ]]

Big Data Solutions with Aptus Data Labs and WANdisco

Ravindra Swamy, Chief Technologist and Ankur Gupta, VP –Client Services will present Aptus Data Labs’ data to decision (D2D) framework, Big Data / Hadoop Services & best practices.

Paul Scott-Murphy, Vice President of Field Technical Services for WANdisco will introduce WANdisco Fusion, the only active-active solution for total data protection and availability across Hadoop distributions and storage and the advantages of implementing this in your business.
Recorded Aug 10 2015 59 mins
Your place is confirmed,
we'll send you email reminders
Presented by
Ravindra Swamy, Ankur Gupta, Paul Scott-Murphy
Presentation preview: Big Data Solutions with Aptus Data Labs and WANdisco
  • Channel
  • Channel profile
  • Big data replication to Amazon S3 Dec 14 2016 6:00 pm UTC 60 mins
    Paul Scott-Murphy, VP Product Management
    Paul Scott-Murphy, WANdisco VP of Product Management will explain the benefits of moving to the cloud and review the AWS tools available for cloud migration and hybrid cloud deployments.
  • Three ways to get a handle on data governance in hybrid cloud environments Recorded: Nov 3 2016 37 mins
    Experts from New Context and WANdisco
    Join experts from New Context and WANdisco as they explain the tools and techniques for securely managing big data in hybrid cloud environments.

    Attendees will learn what's required for effective data governance with the highest levels of availability and performance on-premises and in the cloud.
  • Big data storage: options & recommendations Recorded: Oct 20 2016 51 mins
    Jagane Sundar, WANdisco CTO
    Hadoop clusters are often built around commodity storage, but architects now have a wide selection of Big Data storage choices, including solid-state or spinning disk for clusters and enterprise storage for compatibility layers and connectors.

    In this webinar, our experts will review the storage options available to Hadoop architects and provide recommendations for each use case, including an active-active replication option that makes data available across multiple storage systems.
  • New Hive and performance features in WANdisco Fusion 2.9 Recorded: Oct 6 2016 47 mins
    WANdisco
    WANdisco Fusion 2.9 delivers new levels of performance and scalability, with a number of enhancements that make it easier to support hybrid big data deployments with continuous and consistent access across any mix of on-premise and cloud environments.

    One of the most significant new features is the application of WANdisco’s patented active transactional replication to the Hive metastore, used to support familiar SQL-like access to Hadoop. Changes made to table definitions and other objects in the Hive metastore are replicated to Hive metastore instances deployed with other Hadoop clusters. Data added to the underlying Hadoop cluster referencing the new object definitions is replicated to other Hadoop clusters in a subsequent step, to guarantee consistent access across clusters and locations.

    WANdisco Fusion 2.9’s patented technology also:

    • Enables throughput in excess of 100,000 transactions per minute - well beyond the largest enterprise requirements
    • Ensures operations are not affected by transient network outages or “flapping WAN” links
    • Supports scenarios in hybrid cloud deployments where on-premise environments cannot accept inbound network connection requests

    Join us on October 6th at 10AM Pacific, 1PM Eastern as we present and demo the latest release of WANdisco Fusion.
  • New Hive and performance features in WANdisco Fusion 2.9 (EMEA) Recorded: Oct 5 2016 50 mins
    Paul Scott-Murphy, VP Product Management
    WANdisco Fusion 2.9 delivers new levels of performance and scalability, with a number of enhancements that make it easier to support hybrid big data deployments with continuous and consistent access across any mix of on-premise and cloud environments.

    One of the most significant new features is the application of WANdisco’s patented active transactional replication to the Hive metastore, used to support familiar SQL-like access to Hadoop. Changes made to table definitions and other objects in the Hive metastore are replicated to Hive metastore instances deployed with other Hadoop clusters. Data added to the underlying Hadoop cluster referencing the new object definitions is replicated to other Hadoop clusters in a subsequent step, to guarantee consistent access across clusters and locations.

    WANdisco Fusion 2.9’s patented technology also:

    •Enables throughput in excess of 100,000 transactions per minute - well beyond the largest enterprise requirements
    •Ensures operations are not affected by transient network outages or “flapping WAN” links
    •Supports scenarios in hybrid cloud deployments where on-premise environments cannot accept inbound network connection requests

    Join us on October 5th as we present and demo the latest release of WANdisco Fusion.
  • Migrating your big data infrastructure to cloud Recorded: Sep 8 2016 59 mins
    GigaOM analyst William McKnight with experts from Qubole and WANdisco
    GigaOM analyst William McKnight will be joined by experts from Qubole and WANdisco, who will explain the benefits of moving to the cloud and review the tools available for cloud migration and hybrid cloud deployments.

    Learn what's required to avoid the downtime and business disruption that often accompany cloud migration projects.

    Limited Time Offer - View Qubole and WANdisco's Special Quick Start Package: http://bit.ly/2cGupC6
  • Deploying mission critical applications on Hadoop, on-premises and in the cloud Recorded: Jul 21 2016 63 mins
    Jim Wankowski, IBM Worldwide Cloud Data Services and James Campigli, Co-Founder, Chief Product Officer of WANd
    Global enterprises have quietly funneled enormous amounts of data into Hadoop over the last several years. Hadoop has transformed the way organizations deal with big data. By making vast quantities of rich unstructured and semi-structured data quickly and cheaply accessible, Hadoop has opened up a host of analytic capabilities that were never possible before, to drive business value.

    The challenges have revolved around operationalizing Hadoop to enterprise standards, and leveraging cloud-based Hadoop as a service (HaaS) options offering a vast array of analytics applications and processing capacity that would be impossible to deploy and maintain in-house.

    This webcast will explain how solutions from IBM and WANdisco address these challenges by supporting:

    - Continuous availability with guaranteed data consistency across Hadoop clusters any distance apart, both on-premises and in the cloud.
    - Migration to cloud without downtime and hybrid cloud for burst-out processing and offsite disaster recovery.
    - Flexibility to eliminate Hadoop distribution vendor lock-in and support migration to cloud without downtime or disruption.
    - IBM's BigInsights in the cloud, and BigSQL, which allows you to run standard ANSI compliant SQL against your Hadoop data.
  • Build An Effective, Fast And Secure Data Engine With Hortonworks & WANdisco Recorded: Jun 23 2016 45 mins
    Dave Russell, Hortonworks and Mark Lewis, WANdisco
    Data is coming from everywhere. The first challenge is just being able to get hold of it, curate and convey it in a secure and transparent manner. Hortonworks Data Flow is the tool that collects data at the edge, processes and secures data in motion and brings data into your data-at-rest platform (HDP).

    Once you have your data collected in a valuable data lake, you need resilience, control over its location, and safety against failure. That’s where Wandisco Fusion & Hortonworks HDP come in. With Wandisco Fusion on HDP, an enterprise can now build an effective, fast and secure data engine out of multiple Hadoop clusters, getting the most business value out of its HDP deployment with a reliable and high-performing Big Data service.

    Join Hortonworks & WANdisco on this webinar to learn how to make this into reality.
  • Bringing Hadoop into the Banking Mainstream Recorded: Jun 9 2016 60 mins
    James Curtis (451 Research) & Jim Campigli (WANdisco)
    Global banks have the most rigorous availability, performance and data security standards. Join 451 Research and WANdisco as we explore the cutting-edge techniques leading financial services firms are using to fully operationalize Hadoop to meet these standards and leap ahead of their competition. Register for this webinar and get the free white paper entitled "Bringing Hadoop into the Banking Mainstream"
  • Making Hybrid Cloud a Reality Recorded: Apr 21 2016 39 mins
    Jim Campigli and Jagane Sundar
    Solutions for seamlessly moving data between on-premise and cloud environments are virtually non-existent. This webinar explains how to achieve a true hybrid cloud deployment that supports on-demand burst-out processing in which data moves in and out of the cloud as it changes, and enable the cloud to be used for offsite disaster recovery without downtime or data loss.
  • ETL and Big Data: Building Simpler Data Pipelines Recorded: Feb 11 2016 53 mins
    Chris Almond, Solutions Architect
    In the traditional world of EDW, ETL pipelines are a troublesome bottleneck when preparing data for use in the data warehouse. ETL pipelines are notoriously expensive and brittle, so as companies move to Hadoop they look forward to getting rid of the ETL infrastructure.

    But is it that simple? Some companies are finding that in order to move data between clusters for backup or aggregation purposes, they are building systems that look an awful lot like ETL.
  • No More DR Sites Recorded: Oct 22 2015 42 mins
    Brett Rudenstein
    Disaster recovery sites are typically underutilized with idle hardware and software that are only used in an emergency. Why let your valuable resources remain idle?

    In this webinar, you’ll learn how you can take full advantage of the resources in what would ordinarily be your DR site by using active-active replication to provide full utilization as well as complete failover with lower RPO and RTO.
  • EMEA/APAC - Hadoop Migration and Upgrade without Downtime or Data Loss Recorded: Oct 8 2015 41 mins
    Paul Scott Murphy
    Migrating your Hadoop cluster between versions or distributions is difficult. It is a critical process that if done incorrectly can lead to the loss of data, system downtime, and disruption of business activities.

    In this webinar, learn how you can mitigate the risk in a migration through the development of a comprehensive migration strategy and leveraging tools like those from WANdisco to simplify and automate your migration.
  • Americas - Hadoop Migration and Upgrade without Downtime or Data Loss Recorded: Oct 6 2015 37 mins
    Paul Scott Murphy
    Migrating your Hadoop cluster between versions or distributions is difficult. It is a critical process that if done incorrectly can lead to the loss of data, system downtime, and disruption of business activities.

    In this webinar, learn how you can mitigate the risk in a migration through the development of a comprehensive migration strategy and leveraging tools like those from WANdisco to simplify and automate your migration.
  • Running Globally Distributed Hadoop Clusters Recorded: Sep 24 2015 44 mins
    Brett Rudenstein
    Join us for this webinar to learn how active-active replication available only through WANdisco Fusion allows you to run a single Hadoop namespace across multiple clusters located around the world.

    In this webinar, you’ll see:

    * How slower, lower bandwidth WAN connections can impact distributed Hadoop clusters and how to mitigate that impact

    * How to use 100% of your data center resources for robust disaster recovery without leaving hardware idle

    * How to keep your Hadoop clusters running during necessary maintenance and upgrades
  • Enterprise Class Replication for Hadoop and Why You Need It Recorded: Sep 2 2015 63 mins
    Jim Campigli and Brett Rudenstein
    To make your Hadoop deployment enterprise-class, robust data replication is required to support business-critical functions that depend on Hadoop. Firms can no longer rely on the status quo of traditional, slow data backups into underutilized hardware.

    Enterprise class active-active replication delivers several benefits, including:
    - Maximum infrastructure resource utilization
    - Better performance across geographies
    - Faster disaster recovery.

    In this webinar, learn how WANdisco Fusion enables true enterprise class replication for Hadoop.
  • Big Data Solutions with Aptus Data Labs and WANdisco Recorded: Aug 10 2015 59 mins
    Ravindra Swamy, Ankur Gupta, Paul Scott-Murphy
    Ravindra Swamy, Chief Technologist and Ankur Gupta, VP –Client Services will present Aptus Data Labs’ data to decision (D2D) framework, Big Data / Hadoop Services & best practices.

    Paul Scott-Murphy, Vice President of Field Technical Services for WANdisco will introduce WANdisco Fusion, the only active-active solution for total data protection and availability across Hadoop distributions and storage and the advantages of implementing this in your business.
  • Global Hadoop: Storage and Compute Challenges Recorded: Jul 23 2015 45 mins
    Brett Rudenstein
    Enterprise Hadoop applications require continuous operation in the face of complete data center failure. To address this, businesses have taken a multi-data center, global approach to Hadoop.

    In addition, organizations have begun to utilize multi-directional data sharing between clusters to get more out of their existing infrastructure.

    In this webinar, we'll examine storage and compute challenges in operating Hadoop over a WAN, and lay out the blueprint for an ideal solution.
  • Managing hybrid on-premise/cloud Hadoop deployments Recorded: Jun 18 2015 54 mins
    George Gilbert of Wikibon, WANdisco's Randy DeFauw
    A growing number of Hadoop adopters are making use of both on-premise and cloud clusters. Production workloads run on-premise to provide the best performance and security, while cloud clusters are reserved for testing, development, and burst-out processing power. In this webinar we'll review best practices for managing a hybrid environment including security, data transfer, and performance.
  • Big Data Storage: Options & Recommendations Recorded: May 14 2015 50 mins
    Matt Aslett of 451 Research, Randy DeFauw of WANdisco
    Hadoop clusters are often built around commodity storage, but architects now have a wide selection of Big Data storage choices. Hadoop clusters can use a mix of solid-state and spinning disk storage, while Hadoop compatibility layers and connectors can use enterprise storage systems or share storage between Hadoop and legacy applications.

    In this webinar, 451 Research Director Matt Aslett will review the storage options available to Hadoop architects and provide recommendations for each use case. WANdisco's Randy DeFauw will then present an active-active replication option that makes data available across multiple storage systems.
The World Leaders in Active Transactional Data Replication
Once believed to be impossible, WANdisco's patented technology allows Big Data to be stored and queried with absolute reliability and security, unleashing limitless possibilities for innovation. That's Hadoop without limits. We cover topics such as hardening Hadoop for the enterprise, simplifying audit and compliance, and getting the most out of your multi-data center Hadoop investment. These interactive presentations are targeted at enterprise architects and IT infrastructure staff who are designing and implementing big data environments with Hadoop, HBase and related technologies.

Embed in website or blog

Successfully added emails: 0
Remove all
  • Title: Big Data Solutions with Aptus Data Labs and WANdisco
  • Live at: Aug 10 2015 5:00 am
  • Presented by: Ravindra Swamy, Ankur Gupta, Paul Scott-Murphy
  • From:
Your email has been sent.
or close