Hi [[ session.user.profile.firstName ]]

Case Study in Big Data and Data Science: University of Georgia

Watch this on-demand webinar to learn how the University of Georgia (UGA) uses Apache Spark and other tools for Big Data analytics and data science research.

UGA needs to give its students and faculty the ability to do hands-on data analysis, with instant access to their own Spark clusters and other Big Data applications.

So how do they provide on-demand Big Data infrastructure and applications for a wide range of data science use cases? How do they give their users the flexibility to try different tools without excessive overhead or cost?

In this webinar, you’ll learn how to:

-Spin up new Spark and Hadoop clusters within minutes, and quickly upgrade to new versions

-Make it easy for users to build and tinker with their own end-to-end data science environments

-Deploy cost-effective, on-premises elastic infrastructure for Big Data analytics and research
Recorded May 11 2016 61 mins
Your place is confirmed,
we'll send you email reminders
Presented by
Shannon Quinn, Assistant Professor at University of Georgia; and Nanda Vijaydev, Director of Solutions Management at BlueData
Presentation preview: Case Study in Big Data and Data Science: University of Georgia

Network with like-minded attendees

  • [[ session.user.profile.displayName ]]
    Add a photo
    • [[ session.user.profile.displayName ]]
    • [[ session.user.profile.jobTitle ]]
    • [[ session.user.profile.companyName ]]
    • [[ userProfileTemplateHelper.getLocation(session.user.profile) ]]
  • [[ card.displayName ]]
    • [[ card.displayName ]]
    • [[ card.jobTitle ]]
    • [[ card.companyName ]]
    • [[ userProfileTemplateHelper.getLocation(card) ]]
  • Channel
  • Channel profile
  • AI and Machine Learning: Enterprise Use Cases and Challenges Jun 28 2018 5:00 pm UTC 60 mins
    Radhika Rangarajan Director, Big Data, Intel; Nanda Vijaydev Director, Solutions, BlueData
    Join this webinar to learn how you can accelerate your AI initiative to deliver faster time-to-value with machine learning.

    AI has moved into the mainstream. Innovators in every industry are adopting machine learning for AI and digital transformation, with a wide range of different use cases. But these technologies are difficult to implement for large-scale distributed environments with enterprise requirements.

    This webinar will discuss:

    -The game-changing business impact of AI and machine learning (ML) in the enterprise
    -Example use cases: from fraud detection to medical diagnosis to autonomous driving
    -The challenges of building and deploying distributed ML pipelines and how to overcome them
    -A new turnkey solution to accelerate enterprise AI initiatives and large-scale ML deployments

    Find out how to get up and running quickly with a multi-node sandbox environment for TensorFlow and other popular ML tools.
  • Deep Learning with TensorFlow and Spark: Using GPUs & Docker Containers Recorded: May 3 2018 62 mins
    Tom Phelan, Chief Architect, BlueData; Nanda Vijaydev, Director - Solutions, BlueData
    Watch this on-demand webinar to learn about deploying deep learning applications with GPUs in a containerized multi-tenant environment.

    Keeping pace with new technologies for data science and machine learning can be overwhelming. There are a plethora of open source options, and it's a challenge to get these tools up and running easily and consistently in a large-scale distributed environment.

    This webinar will discuss how to deploy TensorFlow and Spark clusters running on Docker containers, with a shared pool of GPU resources. Learn about:

    *Quota management of GPU resources for greater efficiency
    *Isolating GPUs to specific clusters to avoid resource conflict
    *Attaching and detaching GPU resources from clusters
    *Transient use of GPUs for the duration of the job

    Find out how you can spin up (and tear down) GPU-enabled TensorFlow and Spark clusters on-demand, with just a few mouse clicks.
  • Deployment Use Cases for Big-Data-as-a-Service (BDaaS) Recorded: Mar 15 2018 55 mins
    Nick Chang, Head of Customer Success, BlueData; Yaser Najafi, Big Data Solutions Engineer, BlueData
    Watch this on-demand webinar to learn about use cases for Big-Data-as-a-Service (BDaaS) – to jumpstart your journey with Hadoop, Spark, and other Big Data tools.

    Enterprises in all industries are embracing digital transformation and data-driven insights for competitive advantage. But embarking on this Big Data journey is a complex undertaking and deployments tend to happen in fits and spurts. BDaaS can help simplify Big Data deployments and ensure faster time-to-value.

    In this webinar, you'll hear about a range of different BDaaS deployment use cases:

    -Sandbox: Provide data science teams with a sandbox for experimentation and prototyping, including on-demand clusters and easy access to existing data.

    -Staging: Accelerate Hadoop / Spark deployments, de-risk upgrades to new versions, and quickly set up testing and staging environments prior to rollout.

    -Multi-cluster: Run multiple clusters on shared infrastructure. Set quotas and resource guarantees, with logical separation and secure multi-tenancy.

    -Multi-cloud: Leverage the portability of Docker containers to deploy workloads on-premises, in the public cloud, or in hybrid and multi-cloud architectures.
  • Decoupling Compute and Storage for Big Data Recorded: Jan 31 2018 64 mins
    Tom Phelan, Chief Architect, BlueData; Anant Chintamaneni, Vice President, Products, BlueData
    Watch this on-demand webinar to learn how separating compute from storage for Big Data delivers greater efficiency and cost savings.

    Historically, Big Data deployments dictated the co-location of compute and storage on the same physical server. Data locality (i.e. moving computation to the data) was one of the fundamental architectural concepts of Hadoop.

    But this assumption has changed – due to the evolution of modern infrastructure, new Big Data processing frameworks, and cloud computing. By decoupling compute from storage, you can improve agility and reduce costs for your Big Data deployment.

    In this webinar, we discussed how:

    - Changes introduced in Hadoop 3.0 demonstrate that the traditional Hadoop deployment model is changing
    - New projects by the open source community and Hadoop distribution vendors give further evidence to this trend
    - By separating analytical processing from data storage, you can eliminate the cost and risks of data duplication
    - Scaling compute and storage independently can lead to higher utilization and cost efficiency for Big Data workloads

    Learn how the traditional Big Data architecture is changing, and what this means for your organization.
  • Big-Data-as-a-Service for Hybrid and Multi-Cloud Deployments Recorded: Dec 14 2017 64 mins
    Anant Chintamaneni, Vice President, Products, BlueData; Saravana Krishnamurthy, Senior Director, Product Management, BlueData
    Watch this on-demand webinar to see how BlueData's EPIC software platform makes it easier, faster, and more cost-effective to deploy Big Data infrastructure and applications.

    Find out how to provide self-service, elastic, and secure Big Data environments for your data science and analyst teams – either on-premises; on AWS, Azure, or GCP; or in a hybrid architecture.

    In this webinar, learn how you can:

    *Simplify Big Data deployments with a turnkey Big-Data-as-a-Service solution, powered by Docker containers

    *Increase business agility with the ability to create on-demand Hadoop and Spark clusters, in just a few mouse clicks

    *Deliver faster time-to-insights with pre-integrated images for common data science, analytics, visualization, and machine learning tools

    *Separate compute and storage, and while ensuring security and control in a multi-tenant environment

    See an EPIC demo – including our latest innovations – and discover the flexibility and power of Big-Data-as-a-Service with BlueData. It's BDaaS!
  • Panera Case Study in Big Data Analytics and Data Science Recorded: Nov 15 2017 64 mins
    Darren Darnell, Jim Foppe, and Mike Steimel (Panera Bread); Nanda Vijaydev (BlueData)
    Watch this on-demand webinar to learn how Panera Bread uses Big Data analytics to drive their business, with #1 ranked customer loyalty.

    Panera Bread – with over 2,000 locations and 25 million customers in its loyalty program – relies on analytics to fine-tune its menu, operations, marketing, and more. Find out how they solve key business challenges using Hadoop and next generation Big Data technologies, including real-time data to analyze consumer behavior.

    In this webinar, Panera Bread discussed how they:

    -Use a data-driven approach to improve customer acquisition, customer retention, and operational efficiency

    -Spin up instant clusters for rapid prototyping and exploratory analytics, with real-time streaming platforms like Kafka

    -Operationalize their data science and data pipelines in a hybrid deployment model, both on-premises and in the cloud

    Don’t miss watching this case study webinar. Discover your own recipe for success with Big Data analytics and data science!
  • Big Data Customer Case Study: The Advisory Board Company Recorded: Sep 14 2017 64 mins
    Ramesh Thyagarajan, Advisory Board; Roni Fontaine, Hortonworks; Anant Chintamaneni, BlueData
    Watch this on-demand webinar and learn how a leading healthcare company is yielding big dividends from Big Data.

    Advisory Board, a healthcare firm serving 90% of U.S. hospitals, has multiple different business units and data science teams within their organization. In this webinar, they'll share how they use technologies like Hadoop and Spark to address the diverse use cases for these different teams – with a highly flexible and elastic platform leveraging Docker containers.

    In this webinar, Advisory Board discussed how they:

    -Migrated their analytics from spreadsheets and RDBMS to a modern architecture using tools such as Hadoop, Spark, H2O, Jupyter, RStudio, and Zeppelin.

    -Provide the ability to spin up instant clusters for greater agility, with shared and secure access to a treasure trove of data in their HDFS data lake.

    -Shortened time-to-insights from days to minutes, slashed infrastructure costs by more than 80 percent, and freed up staff to innovate and build new capabilities.

    Don’t miss watching this case study webinar. Find out how you can improve agility, flexibility, and ROI for your Big Data journey.
  • Hadoop and Spark on Docker: Container Orchestration for Big Data Recorded: Jul 27 2017 63 mins
    Anant Chintamaneni, Vice President, Products, BlueData; Tom Phelan, Chief Architect, BlueData
    Watch this on-demand webinar to learn the key considerations and options for container orchestration with Big Data workloads.

    Container orchestration tools such as Kubernetes, Marathon, and Swarm were designed for a microservice architecture with a single, stateless service running in each container. But this design is not well suited for Big Data clusters constructed from a collection of interdependent, stateful services. So what are your options?

    In this webinar, we discussed:

    - Requirements for deploying Hadoop and Spark clusters using Docker containers

    - Container orchestration options and considerations for Big Data environments

    - Key issues such as management, security, networking, and petabyte-scale storage

    - Best practices for a scalable, secure, and multi-tenant Big Data architecture

    Don’t miss watching this webinar on container orchestration for Hadoop, Spark, and other Big Data workloads.
  • Nasdaq Runs Big Data Analytics on BlueData Recorded: Jul 25 2017 5 mins
    Nasdaq, Intel, BlueData
    Watch this video to find out how Nasdaq improves agility and reduces costs for their Big Data infrastructure, while ensuring performance and security. To learn more about the BlueData software platform, visit www.bluedata.com
  • BlueData EPIC on AWS Demo Recorded: Jul 25 2017 4 mins
    BlueData
    The BlueData EPIC software platform makes deployment of Big Data infrastructure and applications easier, faster, and more cost-effective – whether on-premises or on the public cloud.

    With BlueData EPIC on AWS, you can quickly and easily deploy your preferred Big Data applications, distributions and tools; leverage enterprise-class security and cost controls for multi-tenant deployments on the Amazon cloud; and tap into both Amazon S3 and on-premises storage for your Big Data analytics.

    Sign up for a free two-week trial at www.bluedata.com/aws
  • Simplifying Big Data Deployment Recorded: Jul 25 2017 4 mins
    BlueData
    The BlueData software platform is a game-changer for Big Data analytics. Watch this video to see how BlueData makes it easier, faster, and more cost-effective to deploy Big Data infrastructure and applications on-premises.

    With BlueData, you can spin up Hadoop or Spark clusters in minutes rather than months – at a fraction of the cost and with far fewer resources. Leveraging Docker containers and optimized to run on Intel architecture, BlueData’s software delivers agility and high performance for your Big Data analytics.

    Learn more at www.bluedata.com
  • Top 5 Worst Practices for Big Data Deployments and How to Avoid Them Recorded: Jun 28 2017 63 mins
    Matt Maccaux, Global Big Data Lead, Dell EMC; Anant Chintamaneni, Vice President, Products, BlueData
    Watch this on-demand webinar to learn how to deploy a scalable and elastic architecture for Big Data analytics.

    Hadoop and related technologies for Big Data analytics can deliver tremendous business value, and at a lower cost than traditional data management approaches. But early adopters have encountered challenges and learned lessons over the past few years.

    In this webinar, we discussed:

    -The five worst practices in early Hadoop deployments and how to avoid them

    -Best practices for the right architecture to meet the needs of the business

    -The case study and Big Data journey for a large global financial services organization

    -How to ensure highly scalable and elastic Big Data infrastructure

    Discover the most common mistakes for Hadoop deployments – and learn how to deliver an elastic Big Data solution.
  • Scalable Data Science with Spark, R, RStudio, & sparklyr Recorded: May 25 2017 62 mins
    Nanda Vijaydev, Director of Solutions Management, BlueData; and Anant Chintamaneni, Vice President, Products, BlueData
    Watch this on-demand webinar to learn how to get started with large-scale distributed data science.

    Do your data science teams want to use R with Spark to analyze large data sets? How do you provide the flexibility, scalability, and elasticity that they need – from prototyping to production?

    In this webinar, we discussed how to:

    -Evaluate compute choices for running R with Spark (e.g., SparkR or RStudio Server with sparklyr)

    -Provide access to data from different sources (e.g., Amazon S3, HDFS) to run with R and Spark

    -Create on-demand environments using Docker containers, either on-premises or in the cloud

    -Improve agility and flexibility while ensuring enterprise-grade security, monitoring, and scalability

    Find out how to deliver a scalable and elastic platform for data science with Spark and R.
  • Hybrid Architecture for Big Data: On-Premises and Public Cloud Recorded: Apr 13 2017 62 mins
    Anant Chintamaneni, Vice President, Products, BlueData; Jason Schroedl, Vice President, Marketing, BlueData
    Watch this on-demand webinar to learn how to deploy Hadoop, Spark, and other Big Data tools in a hybrid cloud architecture.

    More and more organizations are using AWS and other public clouds for Big Data analytics and data science. But most enterprises have a mix of Big Data workloads and use cases: some on-premises, some in the public cloud, or a combination of the two. How do you support the needs of your data science and analyst teams to meet this new reality?

    In this webinar, we discussed how to:

    -Spin up instant Spark, Hadoop, Kafka, and Cassandra clusters – with Jupyter, RStudio, or Zeppelin notebooks

    -Create environments once and run them on any infrastructure, using Docker containers

    -Manage workloads in the cloud or on-prem from a common self-service user interface and admin console

    -Ensure enterprise-grade authentication, security, access controls, and multi-tenancy

    Don’t miss watching this webinar on how to provide on-demand, elastic, and secure environments for Big Data analytics – in a hybrid architecture.
  • Data Science Operations and Engineering: Roles, Tools, Tips, & Best Practices Recorded: Feb 2 2017 64 mins
    Nanda Vijaydev, Director of Solutions Management, BlueData and Anant Chintamaneni Vice President, Products, BlueData
    Watch this on-demand webinar to learn how to bring DevOps agility to data science and big data analytics.

    It’s no longer just about building a prototype, or provisioning Hadoop and Spark clusters. How do you operationalize the data science lifecycle? How can you address the needs of all your data science users, with various skillsets? How do you ensure security, sharing, flexibility, and repeatability?

    In this webinar, we discussed best practices to:

    -Increase productivity and accelerate time-to-value for data science operations and engineering teams.

    -Quickly deploy environments with data science tools (e.g. Spark, Kafka, Zeppelin, JupyterHub, H2O, RStudio).

    -Create environments once and run them everywhere – on-premises or on AWS – with Docker containers.

    -Provide enterprise-grade security, monitoring, and auditing for your data pipelines.

    Don’t miss watching this webinar. Learn about data science operations – including key roles, tools, and tips for success.
  • Big Data Analytics on AWS: Getting Started with Big-Data-as-a-Service Recorded: Dec 14 2016 64 mins
    Anant Chintamaneni, Vice President, Products, BlueData; Tom Phelan, Chief Architect, BlueData
    So you want to use Cloudera, Hortonworks, and MapR on AWS. Or maybe Spark with Jupyter or Zeppelin; plus Kafka and Cassandra. Now you can, all from one easy-to-use interface. Best of all, it doesn't require DevOps or AWS expertise.

    In this webinar, we discussed:

    -Onboarding multiple teams onto AWS, with security and cost controls in a multi-tenant architecture

    -Accelerating the creation of data pipelines, with instant clusters for Spark, Hadoop, Kafka, and Cassandra

    -Providing data scientists with choice and flexibility for their preferred Big Data frameworks, distributions, and tools

    -Running analytics using data in Amazon S3 and on-premises storage, with pre-built integration and connectors

    Don’t miss watching this webinar on how to quickly and easily deploy Spark, Hadoop, and more on AWS – without DevOps or AWS-specific skills.
  • Distributed Data Science and Machine Learning - With Python, R, Spark, & More Recorded: Nov 2 2016 63 mins
    Nanda Vijaydev, Director of Solutions Management, BlueData; and Anant Chintamaneni, VP of Products, BlueData
    Implementing data science and machine learning at scale is challenging for developers, data engineers, and data analysts. Methods used on a single laptop need to be redesigned for a distributed pipeline with multiple users and multi-node clusters. So how do you make it work?

    In this on-demand webinar, hear a real-world use case and learn about:

    - Requirements and tools such as R, Python, Spark, H2O, and others

    - Infrastructure complexity, gaps in skill sets, and other challenges

    - Tips for getting data engineers, SQL developers, and data scientists to collaborate

    - How to provide a user-friendly, scalable, and elastic platform for distributed data science

    Learn how to get started with a large-scale distributed platform for data science and machine learning.
  • DevOps and Big Data: Rapid Prototyping for Data Science and Analytics Recorded: Sep 15 2016 61 mins
    Krishna Mayuram, Lead Architect for Big Data, Cisco; Anant Chintamaneni, VP of Products, BlueData
    Watch this on-demand webinar with Cisco and BlueData to learn how to deliver greater agility and flexibility for Big Data analytics with Big-Data-as-a-Service.

    Your data scientists and developers want the latest Big Data tools for iterative prototyping and dev/test environments. Your IT teams need to keep up with the constant evolution of new tools including Hadoop, Spark, Kafka, and other frameworks.

    The DevOps approach is helping to bridge this gap between other developers and IT teams. Can DevOps agility and automation be applied to Big Data?

    In this webinar, we discussed:

    -A way to extend the benefits of DevOps to Big Data, using Docker containers to provide Big-Data-as-a-Service.

    -How data scientists and developers can spin up instant self-service clusters for Hadoop, Spark, and other Big Data tools.

    -The need for next-generation, composable infrastructure to deliver Big-Data-as-a-Service in an on-premises deployment.

    -How BlueData and Cisco UCS can help accelerate time-to-deployment and bring DevOps agility to your Big Data initiative.
  • Running Hadoop and Spark on Docker: Challenges and Lessons Learned Recorded: Aug 18 2016 62 mins
    Tom Phelan, Chief Architect, BlueData; Anant Chintamaneni, VP of Products, BlueData
    Watch this on-demand webinar to learn how to run Hadoop and Spark on Docker in an enterprise deployment.

    Today, most applications can be “Dockerized”. However, there are unique challenges when deploying a Big Data framework such as Spark or Hadoop on Docker containers in a large-scale production environment.

    In this webinar, we discussed:

    -Practical tips on how to deploy multi-node Hadoop and Spark workloads using Docker containers

    -Techniques for multi-host networking, secure isolation, QoS controls, and high availability with containers

    -Best practices to achieve optimal I/O performance for Hadoop and Spark using Docker

    -How a container-based deployment can deliver greater agility, cost savings, and ROI for your Big Data initiative

    Don’t miss watching this webinar on how to "Dockerize" your Big Data applications in a reliable, secure, and high-performance environment.
  • Big-Data-as-a-Service: On-Demand Elastic Infrastructure for Hadoop and Spark Recorded: Jun 22 2016 56 mins
    Kris Applegate, Big Data Solution Architect, Dell; Tom Phelan, Chief Architect, BlueData
    Watch this on-demand webinar to learn about Big-Data-as-a-Service from experts at Dell and BlueData.

    Enterprises have been using both Big Data and Cloud Computing technologies for years. Until recently, the two have not been combined.

    Now the agility and efficiency benefits of self-service elastic infrastructure are being extended to big data initiatives – whether on-premises or in the public cloud.

    In this webinar, you’ll learn about:

    -The benefits of Big-Data-as-a-Service – including agility, cost-savings, and separation of compute from storage

    -Innovations that enable an on-demand cloud operating model for on-premises Hadoop and Spark deployments

    -The use of container technology to deliver equivalent performance to bare-metal for Big Data workloads

    -Tradeoffs, requirements, and key considerations for Big-Data-as-a-Service in the enterprise
Big-Data-as-a-Service
BlueData is transforming how enterprises deploy Big Data analytics and machine learning. BlueData’s Big-Data-as-a-Service software platform leverages Docker container technology to make it easier, faster, and more cost-effective for enterprises to innovate with Big Data and AI technologies -- either on-premises, in the public cloud, or in a hybrid architecture. With BlueData, our customers can spin up containerized environments within minutes, providing their data scientists with on-demand access to the applications, data, and infrastructure they need. Founded in 2012 by VMware veterans and headquartered in Santa Clara, California, BlueData is backed by investors including Amplify Partners, Atlantic Bridge, Dell Technologies Capital, Ignition Partners, and Intel Capital.

Embed in website or blog

Successfully added emails: 0
Remove all
  • Title: Case Study in Big Data and Data Science: University of Georgia
  • Live at: May 11 2016 5:00 pm
  • Presented by: Shannon Quinn, Assistant Professor at University of Georgia; and Nanda Vijaydev, Director of Solutions Management at BlueData
  • From:
Your email has been sent.
or close