The cloud has the potential to deliver on the promise of big data processing for machine learning and analytics to help organizations become more data-driven, however, it presents its own set of challenges.
This webinar covers best practices in areas such as.
- Using automation in the cloud to derive more value from big data by delivering self-service access to data lakes for machine learning and analytics
- Enabling collaboration among data engineers, data scientists, and analysts for end-to-end data processing
- Implementing financial governance to ensure a sustainable program
- Managing security and compliance
- Realizing business value through more users and use cases
In addition, this webinar provides an overview of Qubole’s cloud-native data platform’s capabilities in areas described above.
About Our Speaker:
James Curtis is a Senior Analyst for the Data, AI & Analytics Channel at 451 Research. He has had experience covering the BI reporting and analytics sector and currently covers Hadoop, NoSQL and related analytic and operational database technologies.
James has over 20 years' experience in the IT and technology industry, serving in a number of senior roles in marketing and communications, touching a broad range of technologies. At iQor, he served as a VP for an upstart analytics group, overseeing marketing for custom, advanced analytic solutions. He also worked at Netezza and later at IBM, where he was a senior product marketing manager with responsibility for Hadoop and big data products. In addition, James has worked at Hewlett-Packard managing global programs and as a case editor at Harvard Business School.
James holds a bachelor's degree in English from Utah State University, a master's degree in writing from Northeastern University in Boston, and an MBA from Texas A&M University.
Once you’ve done analysis and found insights in your data, determining the right visualization is crucial to telling the right story. Even if you’ve done the work of selecting the right dataset, asking the right initial question, following up with the right additional queries and using your personal expertise to pinpoint valuable information, all of that analytical work loses value if the charting process is not done correctly.
Choosing the right chart doesn’t need to be intimidating. Similar to data exploration, you just need to ask a series of straightforward questions.
Join Britton Stamper, sales engineer at Periscope Data, when he addresses:
- The first question to ask when building a data visualization
- The different types of data relationships
- When to use each type of chart
- What variables to be aware of when building your chart
This is the third webinar in our “Citizen Data Science Empowerment” series. The first webinar “Getting Started with Citizen Data Scientists” covered the importance of citizen data scientists, how to get them enabled and how to empower them. The second webinar, “How to Do Data Right,” covered the importance of making sure that you are doing your due diligence when evaluating the data to visualize.
Data lakes are centralized data repositories. Data needed by data scientists is physically copied to a data lake which serves as a one storage environment. This way, data scientists can access all the data from only one entry point – a one-stop shop to get the right data. However, such an approach is not always feasible for all the data and limits it’s use to solely data scientists, making it a single-purpose system.
So, what’s the solution?
A multi-purpose data lake allows a broader and deeper use of the data lake without minimizing the potential value for data science and without making it an inflexible environment.
Attend this session to learn:
• Disadvantages and limitations that are weakening or even killing the potential benefits of a data lake.
• Why a multi-purpose data lake is essential in building a universal data delivery system.
• How to build a logical multi-purpose data lake using data virtualization.
Do not miss this opportunity to make your data lake project successful and beneficial.
How do you avoid your enterprise data lake turning into a so-called data swamp? The explosion of structured, unstructured and streaming data can be overwhelming for data lake users, and make it unmanageable for IT. Without scalable, repeatable, and intelligent mechanisms for cataloguing and curating data, the advantages of data lakes diminish. The key to solving the problem of data swamps is Informatica’s metadata driven approach which leverages intelligent methods to automatically discover, profile and infer relationships about data assets. Enabling business analysts and citizen integrators to quickly find, understand and prepare the data they are looking for.Read more >
Consumers are engaging with brands across multiple touchpoints, channels, and devices, generating massive amounts of valuable data. Organizations are quickly adopting a number of solutions to keep up with this explosion of customer data and better capture and correlate user behavior.
Two common solutions brands are leveraging to house and analyze all of this customer data are Enterprise Data Warehouses (EDW) and Data Lakes. Register now for this 30-minute webinar and learn:
- Key benefits of each and which is best for your brand
- Why pairing your enterprise data storage solution with customer data initiatives makes your tech stack even more powerful
- How an automated data supply chain fits in a modern EDW and data lake environment
- And more!
The webinar will conclude with a live Q&A Chat with questions from the audience on all things enterprise data storage.
„You can´t use it, if you can´t find it” – Heutzutage werden in Unternehmen mehr Daten denn je gesammelt, gespeichert und genutzt. Studien und Umfragen zeigen jedoch, dass lieber gesammelt als genutzt wird.
Woran liegt das? Eine Ursache hierfür ist, dass Unternehmen gar nicht wissen, welche Daten gesammelt werden, wo sie gespeichert werden und wie diese genutzt werden können. Es fehlt an Transparenz und Struktur.
In unserem 45-minütigen Webinar möchten wir Ihnen anhand einiger Praxisbeispiele zeigen, wie Sie mit einem Data Catalog Informationen über alle Datentöpfe hinweg zentral zur Verfügung stellen können. Erfahren Sie, wie unsere Kunden davon profitieren und welche Herausforderungen mit einem Data Catalog gemeistert werden können.
Achieving actionable insights from data is the goal of any organization. To help in this regard, data catalogs are being deployed to build an inventory of data assets that provides both business and IT users a way to discover, organize and describe enterprise data assets. This is a good first step that helps all types of users easily find relevant data to extract insights from.
Increasingly, end users want to take the next step in provisioning or procuring this data into a sandbox or analytics environment for further use. Attend this session to see how organizations are looking to build actionable data catalogs via a data marketplace, that allow self-service access to data without sacrificing data governance and security policies.
Learn how to provide governed access and visibility to the data lake while still staying on track and within budget. Join Scott Gidley, Zaloni’s Vice President of Product, as he discusses:
- Architecting your data lake to support next-gen data catalogs
- Rightsizing governance for self-service data
- Where a data catalog falls short and how to address
- Success use cases
This is the second webinar in our “Citizen Data Science” series. Our first webinar, “Getting Started with Citizen Data Scientists,” covered the importance of citizen data scientists, how to get them enabled and how to empower them.
To some citizen data scientists, data is a new language that they’re unfamiliar with translating into insights. For others, they’re eager to dive in and curious to explore, but reluctant to communicate their findings because they don’t have experience translating data into tactical operations.
These new business-focused analysts need to be confident that they’re utilizing data properly. To empower these new data dives, Kyle Dempsey, senior professional services engineer at Periscope Data pulled together a collection of tips to make sure first-time analysts are doing data right.
Join Kyle on September 27th, 2018 at 10 a.m. PDT as he walks through how to:
- Define your analysis and outcomes
- Understand what data is available and know how to ask for more
- Ask questions in a way that allows data to answer them
- Use data to inform decisions
- Enable collaboration between technical and nontechnical teams
Is Your Data Ready for GDPR?
As the deadline for GDPR approaches, it is time to get practical about protecting personal data.
We break down the steps for turning a data lake into a data hub with appropriate data management and governance activities: from capturing and reconciling personal data to providing for consent management, data anomyzation, and the rights of the data subject.
A smart approach to GDPR compliance lays a foundation for personalized and profitable customer and employee relations.
Watch, as experts from MAPR and Talend show you how to:
- Diagnose the maturity of your GDPR compliance;
- Set up milestones and priorities to reach compliance;
- Create a foundation to manage personal data through a data lake;
- Master compliance operations - from data inventory to data transfers to individual rights management.
Anyone who's ever analyzed data knows the pain of digging in only to find that it is poorly structured, full of inaccuracies, or just plain incomplete. But "dirty data" isn't just a pain point for analysts; it can have a major financial and cultural impact on an organization.
Attend this live webinar to learn four actionable ways to overcome common data preparation issues including how to establish a company standard for "clean data" and how to democratize data prep across your organization.
- Andy Cotgreave, Technical Evangelism Director, Tableau
- Jason Harmer, Data Analytics and Visualization Consultant, Nationwide Insurance
- Gordon Strodel, Information Management and Analytics Consultant, Slalom
Today's enterprises need broader access to data for a wider array of use cases to derive more value from data and get to business insights faster. However, it is critical that companies also ensure the proper controls are in place to safeguard data privacy and comply with regulatory requirements.
What does this look like? What are best practices to create a modern, scalable data infrastructure that can support this business challenge?
Zaloni partnered with industry-leading insurance company AIG to implement a data lake to tackle this very problem successfully. During this webcast, AIG's VP of Global Data Platforms, Carlos Matos, and Zaloni CEO, Ben Sharma will share insights from their real-world experience and discuss:
- Best practices for architecture, technology, data management and governance to enable centralized data services
- How to address lineage, data quality and privacy and security, and data lifecycle management
- Strategies for developing an enterprise-wide data lake service for advanced analytics that can bridge the gaps between different lines of business, financial systems and drive shared data insights across the organization
The shelf life of data is shrinking. A streaming shift is taking place and use cases such as IoT connected cars, real-time fraud detection and predictive maintenance using streaming analytics are becoming commonplace. You too can switch to the fast data lane with Informatica, leveraging Kafka and other big data technologies. So shift gears and change lanes with us while we take you on a journey into the world of streaming data.Read more >
Data is the new currency for most organizations and data volumes are continuing to grow at an explosive rate. While the advantage of collection of such large volumes of data is obvious, protection of this data from cybercriminals and malicious actors is becoming increasingly difficult. Conventional security mechanisms are failing and large-scale security breaches, despite increasing security spends, are becoming commonplace. This, along with increasingly stringent regulatory requirements and privacy laws have brought “Data Security” – protection of the data itself whether in motion, in use or in transit into strong focus. While there has been an increasing focus on Data Centric Security, the solution landscape is fractured and enterprises are still struggling to identify and deploy long term solutions with minimal disruption to existing investments and processes. This talk will focus on the current state of Data security and offer pointers to how organizations can embark on a long-term Data Centric journey which truly adds business value.Read more >
If you feel like you don’t trust your data, there’s probably a good reason. It happens all the time; companies implement analytics, customize their solutions and don’t audit the implementation to ensure ongoing data accuracy. This leads to multiple inaccuracies, gaps in tracking, and — even worse — information that’s simply missing. Inaccurate data can send a brand down the wrong path, leading to bad decisions and additional costs for tools and resources that could have otherwise been avoided.
Join and learn:
- What data quality is and why it’s critical to an organization's overall success
- Why your data is a mess and how to identify the warning signs of poor data quality
- Best practices to ensure clean and quality data and how to take back control
- And so much more!
The webinar will conclude with a Fireside Chat with live questions from the audience on all things data quality.
As data is growing at an exponential rate, organizations are increasingly looking to leverage streaming data from mobile devices, wearable technology and sensors for real-time processing and analytics. Gartner estimates that, “By 2020, 70% of organizations will adopt data streaming to enable real-time analytics.” However, implementing real-time data ingest, processing and delivering insights at scale requires infrastructure with zero latency and easy access to information when it is required.
In the webinar, we’ll discuss:
- Adopting Modern Data Lake with the Hortonworks Data Platform (HDP)
- Accelerating real-time data analytics with Hortonworks Data Flow (HDFTM) and Attunity to build a data lake
- Solving challenges with real-time data ingest and managing data in motion workloads
Join subject matter experts from IBM and Hortonworks for a joint webcast to help you accelerate real-time data analytics and manage your data workloads efficiently.
As data analytics becomes more embedded within organizations, as an enterprise business practice, the methods and principles of agile processes must also be employed.
Agile includes DataOps, which refers to the tight coupling of data science model-building and model deployment. Agile can also refer to the rapid integration of new data sets into your big data environment for "zero-day" discovery, insights, and actionable intelligence.
The Data Lake is an advantageous approach to implementing an agile data environment, primarily because of its focus on "schema-on-read", thereby skipping the laborious, time-consuming, and fragile process of database modeling, refactoring, and re-indexing every time a new data set is ingested.
Another huge advantage of the data lake approach is the ability to annotate data sets and data granules with intelligent, searchable, reusable, flexible, user-generated, semantic, and contextual metatags. This tag layer makes your data "smart" -- and that makes your agile big data environment smart also!
Im zweiten Webinar unserer „Big Data im Fokus“ Serie wird es pragmatisch, praktisch, gut. Wir zeigen Ihnen, wie Sie die richtigen Daten nicht nur finden können, sondern auch direkt für Ihre Anwendungsfälle sauber und nachvollziehbar aufbereiten können. Mit unserer Self-Service Data Preparation Lösung nehmen wir Rohdaten in die Hand, bringen diese in Form, stellen die Qualität sicher und erstellen uns eine perfekte Datenbasis für analytische Anwendungsfälle. Anhand eines praktischen Beispiels sehen Sie, wie ein Fachanwender Schritt für Schritt ans Ziel kommt, ohne dabei eine Zeile Programmcode zu schreiben.Read more >
With just a few weeks to the UK's largest data & analytics event, we've gathered some of the elite speakers who will be taking the stage to debate the latest trends, hottest solutions and the biggest opportunities (and challenges) for businesses in a data-driven world.
* Fast Data & DataOps
* Self-Service Analytics
* Artificial Intelligence
* Customer Experience
* Data Governance
What will they be talking about at The Olympia, London, on the 13-14 November 2018, what do they want to hear about, what are they looking forward to?
Join this panel discussion and arm yourself for excellence in this brave new data-driven world.
Richard Corderoy, Chief Data Officer, Oakland Data and Analytics
Andy Mott, Senior Consultant, Arcadia Data
A majority of the data collected by organizations today is wasted. Whether through poor analytics, lack of resources, or just having too damn much of it. So how can organizations turn this around and actually start utilizing their data for powerful results?
By leveraging an X-360 initiative, companies are able to take their customer, product, patient, or other data and provide a 360-degree view using a governed and actionable data lake. By breaking down the silos associated with traditional data located in disjointed systems and databases, companies are finding new ways to improve loyalty programs, product development, marketing campaigns, and even find a new source of revenue from their data.
Join Jatin Hansoty, Director of Solutions Architecture at Zaloni, as he dives into real-world use cases from several of the world’s top companies. Learn from their architecture and the results they achieved.
Topics covered include:
- Best practices
- Common pitfalls to avoid
- Real-world use cases
- Future-proof architecture
Jupyter notebooks are transforming the way we look at computing, coding and problem solving. But is this the only “data scientist experience” that this technology can provide?
In this webinar, Natalino will sketch how you could use Jupyter to create interactive and compelling data science web applications and provide new ways of data exploration and analysis. In the background, these apps are still powered by well understood and documented Jupyter notebooks.
They will present an architecture which is composed of four parts: a jupyter server-only gateway, a Scala/Spark Jupyter kernel, a Spark cluster and a angular/bootstrap web application.
Within the energy industry, Exploration, Production, Refining, and Marketing each present unique business challenges. Analyzing the disparate data created and used upstream, downstream, and beyond poses unique IT challenges. With siloed data assets forming an often-untapped reservoir of business opportunity and insight, energy companies who successfully leverage them can gain a significant competitive advantage.
Watch this joint webinar with TIBCO and Cadeon, a data and business analytics expert in the Oil & Gas, Utilities and Mining industries to learn about:
•How to federate, abstract, and simplify data using TIBCO Data Virtualization for faster analytic insights
•How to overcome data and analytic challenges to gain tangible business benefits
•How to do this with all your data, no matter where it resides — upstream, downstream or anywhere across the globe.