Welcome to the big data and data management community on BrightTALK. Join thousands of data quality engineers, data scientists, database administrators and other professionals to find more information about the hottest topics affecting your data. Subscribe now to learn about efficiently storing, optimizing a complex infrastructure, developing governing policies, ensuring data quality and analyzing data to make better informed decisions. Join the conversation by watching live and on-demand webinars and take the opportunity to interact with top experts and thought leaders in the field.
The storage market has transformed in recent years, with the standalone storage array being attacked from all sides by ambitious rivals vying for customer attention in a world demand more and more from beleaguered IT departments. Today, storage comes in all kinds of forms, from the aforementioned standalone arrays to systems built atop powerful software platforms to fully integrated offerings. We now live in a brave new world in which “buying storage” carries with it far more nuance than in days past. On top of the new choices available in form factor come new choices in media. Flash is now almost a requirement for many products, but that’s not enough for many, either. NVMe is now here to stay and promises to reimagine the data interface in a way that hasn’t been done in a generation.
With both current and emerging flash storage systems, companies are gaining benefits that were only dreams in the days of the disk-based SAN. Far beyond just providing crazy levels of performance, flash storage systems have the potential to help IT organizations focus on the business and less on technology. After all, when you have IOPS to spare, other metrics can take center stage. These can include latency, cost per gig, power consumption, gig per rack unit, and many more.
There are a whole lot of vendors in this burgeoning space, with more jumping in every day. Do you truly understand what each of them brings to the table and how they differ? This ActualTech Media 2018 Storage Flash, and NVMe Storage EcoCast is your opportunity to get deep understanding of this technology and gain knowledge about the leading vendors in this transformational market.
Don’t miss this special event hosted by David Davis and Scott D Lowe!
For research data to be truly useful, it must be easy to access, share and manage without requiring expensive, custom infrastructure. What organizations need is turnkey storage that won't break the bank, with a unified interface for fast, reliable data transfer and sharing.
This webinar introduces Globus for ActiveScale, a cost-effective solution for on-premise object storage that’s simple to deploy and use. With Globus for ActiveScale, researchers have access to advanced capabilities for managing data across a broad range of systems, while administrators gain a cost-effective, scalable, and durable solution they can deploy quickly to help their researchers innovate faster.
In this webinar, attendees will:
- Learn how to deploy and use Globus for ActiveScale
- See a product demonstration
- Engage in a live Q&A session with the Globus Chief Customer Officer
An interactive webinar featuring Actian Vector analytics database. Walt will provide a product overview and demo. Submit your question during the event for the live Q & A. Download the Evaluation Edition at https://www.actian.com/lp/try-vector/
Heard a lot about blockchain and the various cryptocurrencies out there but not sure what it actually means? Want to know what’s the actual difference in the different blockchains out there like bitcoin, Ethereum, NEM, etc.? Are you wondering why a lot of companies are going to ICOs and what does that really mean?
What you will learn:
-What is Bitcoin?
-What is a Blockchain?
-What are the different types of Blockchains?
-What is Ethereum? What is ether?
-What is NEM?
-How is Bitcoin, Ethereum, NEM different from each other?
-What is a smart contract?
-What is an ICO?
-What are the steps for conducting an ICO?
-What are the different ICOs happening currently?
Join Matt Aslett of 451 Research for a briefing on the current big data analytics trends that are driving customers to utilize fast big data applications for increased customer engagement, reduced risk, and greater operational efficiency. After which, Nathan Trueblood will share DataTorrent's direct experiences working with enterprise organizations who are deploying fast big data apps to accelerate business outcomes TODAY and why they believe their customers' use of these applications will be the difference between success or failure in the future.
To be successful at Data Architecture, organisations have always needed to understand what data they have, why they have it, how they use it, where it is, how it got there, and what happened to it along the way - forming a 'pipeline' of information. Very often, this pipeline is managed and documented using a mish-mash of tools and methodologies, often resulting in conflicting and contradictory information, making it difficult for Data Governance to be effective.
In this webcast, George McGeachie will demonstrate the key features of SAP PowerDesigner that support the pipeline. A set of requirements and business rules, and an existing Conceptual Data Model, will be used to create Logical and Physical Data Models. This will include the generation of JSON structures from both Logical and Physical Data Models.
Some of what you will see is not out-of-the-box - it has been built using PowerDesigner's powerful customisation features.
Muchos departamentos de TI no se dan cuenta de que más del 70% de sus datos vive en almacenamiento secundario. Pero conocen de primera mano los dolores de cabeza asociados con el enfoque scale-up para lidiar con el crecimiento de ese almacenamiento y la necesidad de migrar a sistemas cada vez más grandes a medida que se expanden los datos.
La alternativa es una arquitectura scale-out que le permite simplemente expandir la capacidad a medida que la necesita. Además de resolver el problema de escalabilidad, este enfoque es más barato, más fiable y ofrece un coste de rendimiento predecible.
Gartner dice que el 30%1 de los datos ya se ha movido a una infraestructura scale-out, principalmente para almacenamiento primario. Su éxito ahora está impulsando el interés en aplicar los mismos principios al almacenamiento secundario.
En este webinar examinaremos cómo la infraestructura Scale-Out:
•Simplifica la gestión del almacenamiento
•Reduce los costes de infraestructura
•Mejora la fiabilidad y disponibilidad
•Dirige el rendimiento
•Expande el almacenamiento sin fisuras
1Gartner Magic Quadrant for Distributed File Systems and Object Storgae, October 2017
Dell EMC commissioned Forrester Consulting to conduct a Total Economic Impact™ (TEI) study and examine the potential return on investment (ROI) enterprises may realize by deploying Dell EMC All-Flash storage. The purpose of this study is to provide readers with a framework to evaluate the potential financial impact of Dell EMC All-Flash storage solutions on their organizations. To better understand the investment benefits, costs, and risks, Forrester interviewed four current customers who have deployed Dell EMC All-Flash storage solutions in their organizations. Join us to review the findings.
Despite tremendous progress, there remain critically important areas, including multi-tenancy, performance optimization, and workflow monitoring where the DevOps team still requires management help. In this webinar, presenter Kirk lewis discusses the ways that big data clusters slow down, how to fix them, and how to keep them running at an optimal level. He also presents an overview of Pepperdata operation performance management (OPM) solutions. In this online webinar followed by a live Q and A, Field Engineer Kirk Lewis discusses:
• How Pepperdata Cluster Analyzer helps operators overcome Hadoop and Spark performance limitations by monitoring all facets of cluster performance in real time, including CPU, RAM, disk I/O, and network usage by user, job, and task.
• How Pepperdata Capacity Optimizer increases capacity utilization by 30-50% without adding new hardware
• How Pepperdata adaptively and automatically tunes the cluster based on real-time resource utilization with performance improvement results that cannot be achieved through manual tuning.
Kirk Lewis joined Pepperdata in 2015. Previously, he was a Solutions Engineer at StackVelocity. Before that he was the lead technical architect for big data production platforms at American Express. Kirk has a strong background in big data.
Exploring big data, live data, and unstructured text fields is exciting to a business analyst, but can generate significant anxiety for the data protection officer. Learn how to open up data to those who need it and protect what needs protecting.
Learn how everything we offer is available for your organization to embed, customize, extend, and white label as your own. We know it's Zoomdata under the covers. But your customers don't need to know.
You're in for a treat when you first learn about Search Data. Think of it as a mashup of Google and your traditional BI tool. You could use this technology to learn more from customer surveys, performance reviews, or anything else that has an open text field.
Are you struggling to turn big data into useful, accurate, consistent, and impactful insights? Organizations are rapidly adopting new technologies like Hadoop, Spark, and Kafka to help lines of business develop new insights. However, with the sheer quantity of data available and complexity of new technologies, the opportunity to drive operational efficiencies and grow sales can feel overwhelming, inflexible, manual, and complex. Informatica's customers have successfully pursued digital transformation strategies through more systematic approaches for getting the right data, in the right place, at the right time to the right people.
In this webinar, learn how to:
• Design an architecture for data management with Hadoop and big data that can scale for large organizations
• Understand best practices for data integration, data quality, data cataloging, and data preparation
• Exploit automated approaches to data management to get more trusted data assets every time
Join experts from Cloudera and Informatica to learn about an enterprise approach to data management that enables big data to be turned into useful, accurate, consistent, and impactful business insights.
How can ServiceNow transform your HR Service Delivery? Start simplifying HR processes, reducing time spent on routine tasks, and modernizing the employee service experience.
- The Big Picture: Why are we all investing in HR?
- Real World Examples: What's possible in HRSD
- Build your Business Case: Alignment, Impact, and your 4 steps to a killer pitch
- Questions to expect: New demands and changes we see with Acorio clients as ServiceNow HRSD expands
Criminal cases have long been thought of as an arena devoid of electronic discovery issues. But attorneys who regularly handle criminal cases know that’s not the case. So, are there any guidelines and best practices for handling eDiscovery in criminal cases? This CLE-approved* webcast session will take a look at the history of the development of standards and highlight specific issues of importance regarding eDiscovery in criminal matters. Topics include:
+ Overview of Rules for Criminal Matters
+ How Data is Acquired in Criminal Matters
+ Common Data Types and Examples of Cases
+ Data Exchange Formats and Protocols
+ Time Issues Specific to Criminal ESI
+ Working with Social Media as Evidence
+ Border Entry Considerations and Examples
+ Resources for More Information
Presentation Leader: Doug Austin
Doug is the VP of Products and Professional Services for CloudNine. At CloudNine, Doug manages professional services consulting projects for CloudNine clients. Doug has over 25 years of experience providing legal technology consulting, technical project management and software development services to numerous commercial and government clients.
Special Consultant to CloudNine: Tom O'Connor
Tom O’Connor is a nationally known consultant, speaker, and writer in the field of computerized litigation support systems. Tom’s consulting experience is primarily in complex litigation matters.
Vertica Eon Mode Beta, a separation of compute and storage architecture option, addresses the common need to rapidly scale your cluster to variable workloads. It also accommodates an almost endless data warehousing capability. Join us as we preview the benefits of this new architecture and walk through a step-by-step demonstration of its capabilities.
Pure, as one of the leading CRIS (or research information management RIM) systems, has facilitated the emergence of exciting opportunities for research organisations to transform their support for researchers, with many research libraries taking an increasingly important role. Broadly defined, RIM is the aggregation, curation, and utilization of information about institutional research activities, and as such, intersects with many aspects of traditional library services in discovery, acquisition, dissemination, and analysis of scholarly activities.
OCLC Research has been working with members of its international OCLC Research Library Partnership, including the University of St Andrews in Scotland, on a publication to help libraries and other institutional stakeholders to understand developing RIM practices and, in particular, their relevance for service and staff development purposes.
In this presentation, we will provide an overview of the OCLC position paper and provide a case study from the University of St Andrews.
Viele sehen das so, denn es bedarf meist mehrerer Schichten teurer Hardware und Software sowie ein kontinuierliches Change-Management. Kein Wunder also, dass „DR-Test” nicht gerade das Lieblingsthema der IT-ler ist.
Die Cloud macht es einfacher, Disaster-Recovery zu planen und zu implementieren. Aber es braucht ein effizientes Datenmanagement, um tatsächlich DR-Zeit und -Kosten zu sparen. Die erfolgreiche Wiederherstellung von virtuellen Maschinen, Anwendungen und dem Geschäftsbetrieb ganz allgemein erfordert, dass Daten effizient in die Cloud verlagert und dort auch effizient verwaltet werden können.
Investieren Sie 45 Minuten und lernen Sie in diesem Webinar, wie Sie…
* für diverse Typen an Daten, VMs und Anwendungen die richtige Strategie der Wiederherstellung wählen
* durch eine einheitliche Oberfläche und Plattform für alle Belange rund um die Datensicherung die Komplexität der Disaster-Recovery verringern
* Risiken minimieren mithilfe von integrierten Benachrichtigungen, Kontrollen und Berichten zu Ihren SLAs
* menschliche Fehler und manuelle Eingriffe reduzieren, indem komplexe und ungewöhnliche DR-Szenarien entwickelt werden
* DR-relevante Aktionen wie z.B. DR-Tests oder (un-)geplante Ausfallsicherungen optimieren
Wie zeigen Ihnen, wie Commvault Sie professionell unterstützt, Ihre Daten in der Cloud für Backup, Disaster-Recovery und vieles mehr zu nutzen.
One of the many benefits of best-of-breed object storage solutions is the ability to scale on demand and pay as you grow. How do you predict your capacity needs and determine your roadmap for scaling your storage? VP of Product Tony Barbagallo and Sr. Consultant John Bell will provide concrete guidance to help you predict your needs so you can accurately map out and budget for your organization’s data growth.
Today, modern, Customer Support teams use thoughtful, intuitive technology to solve customer issues at light speed. In some cases, they have even transformed from a cost center to a juiced up revenue producer.
In our March Lunch and Learn our Customer Service experts will show how to do just that, with real-world examples from companies winning Customer Service right now, and a live demo.
This 1-hour webinar from GigaOm Research brings together leading minds in cloud data analytics, featuring GigaOm analyst Andrew Brust, joined by guests from cloud big data platform pioneer Qubole and cloud data warehouse juggernaut Snowflake Computing. The roundtable discussion will focus on enabling Enterprise ML and AI by bringing together data from different platforms, with efficiency and common sense.
In this 1-hour webinar, you will discover:
- How the elasticity and storage economics of the cloud have made AI, ML and data analytics on high-volume data feasible, using a variety of technologies.
- That the key to success in this new world of analytics is integrating platforms, so they can work together and share data
- How this enables building accurate, business-critical machine leaning models and produces the data-driven insights that customers need and the industry has promised
- How to make the lake, the warehouse, ML and AI technologies and the cloud work together, technically and strategically.
Register now to join GigaOm Research, Qubole and Snowflake for this free expert webinar.
Die Digitalisierung im Handel stellt Unternehmen vor gänzlich neue Herausforderungen.
Notwendig ist eine Digitalisierungsstrategie, die Ihre Geschäftsprozesse effizient und bedarfsgerecht digital gestaltet.
Genau hier unterstützt Product Information Management (PIM).
Mit der führenden PIM Lösung von Informatica können alle weiterführenden Prozesse auf Ihre Unternehmensstrategie und –ziele abgestimmt werden. Ihre Metriken und KPI´s können definiert und gemessen werden und tragen so automatisch zur Optimierung Ihrer gesamten Wertschöpfung bei.
Ihre Organisation und Prozesse werden ganzheitlich und von Beginn an in die digitale Umsetzung gebracht.
In unserem 30-minütigen Webinar, das wir gemeinsam mit unserem Partner Bayard Consulting durchführen, zeigen wir typische Herausforderungen und Best Practices sowie passende Lösungen für Ihre Digitalisierungsstrategie.
While a number of patterns and architectural guidelines exist for cloud-native applications, a discussion about data often leads to more questions than answers. For example, what are some of the typical data problems encountered, why are they different, and how can they be overcome?
Join Prasad Radhakrishnan from Pivotal and Dave Nielsen from Redis Labs as they discuss:
- Expectations and requirements of cloud-native data
- Common faux pas and strategies on how you can avoid them
Market research shows that around 70% of the self-service initiatives fare “average” or below. Denodo 7.0 information self-service tool will offer data analysts, business users and app developers searching and browsing capability of data and metadata in a business friendly manner for self-service exploration and analytics.
Attend this session to learn:
• How business users will be able to use Denodo Platform integrated google-like search for both content and catalog
• With web based query UI how business users can refine queries without SQL knowledge
• With tags and business categorization, how to standardize business / canonical views while decoupling development artifacts from the business users
• The role of information self-service tool
• Product demonstration
• Summary & Next Steps
Web applications are a prime target for hackers. According to the Verizon DBIR 2017 report, web application attacks were the #1 source of data breaches in the past year. With increasing attacker sophistication and a growing web application footprint for most businesses, investing in the cutting edge of web application vulnerability detection and remediation is crucial.
Two technologies leading the charge are NetSparker and Brinqa.Netsparker is a scalable and dead accurate enterprise web application security solutions that enables businesses to easily identify vulnerabilities in web applications and APIs.Brinqa is a cyber risk management platform that helps security professionals triage and remediate application vulnerabilities in context of likelihood of weaponization and impact to business.
In this webinar we discuss how these two innovative technologies work together to create a comprehensive web application security program that helps AppSec programs identify, prioritize, remediate and report the most imminent and impactful vulnerabilities in your organization's software infrastructure.
May 25th, 2018. We have all had it circled on our calendars for two years. The General Data Protection Regulation is scheduled to go into effect on this date. But is it the beginning of the end or an opportunity to take your customer relations to the next level? Join us for this informative webinar where our team of experts, along with Philip Russom from TDWI Research, will explore the possibilities.
The data contained in the data lake is too valuable to restrict its use to just data scientists. It would make the investment in a data lake more worthwhile if the target audience can be enlarged without hindering the original users. However, this is not the case today, most data lakes are single-purpose. Also, the physical nature of data lakes have potential disadvantages and limitations weakening the benefits and possibly even killing a data lake project entirely.
A multi-purpose data lake allows a broader and greater use of the data lake investment without minimizing the potential value for data science or for making it a less flexible environment. Multi-purpose data lakes are data delivery environments architected to support a broad range of users, from traditional self-service BI users to sophisticated data scientists.
Attend this session to learn:
* The challenges of a physical data lake
* How to create an architecture that makes a physical data lake more flexible
* How to drive the adoption of the data lake by a larger audience
Omnichannel ist niemals ausgeschöpft. Jetzt kommt mit „Conversational Commerce“ ein weiterer Kanal dazu und damit die Chance das Customer Engagement noch persönlicher und schneller zu gestalten. Experten sagen dazu „own your customer experience“, an Stelle die aus der Hand zu geben an andere Kanäle oder Anbieter. Die Kombination von einem 360-Messaging mit der 360-Grad-Sicht auf Kunden, Produkte, Services, Locations, Mitarbeiter und viele mehr leutet ein neue Zeitalter in der Customer Experience ein.
In diesem Webinar erfahren Sie:
* Was ist „Conversational Commerce“
* Wie Messaging Ihren Kundenservice verbessert
* Wie Sie Mehr Omnichannel Umsatz damit realisieren
* Wie Sie die richtigen Voraussetzungen Ihrer Masterdaten schaffen für diese neue Möglichkeit der Customer Experience
Choosing the right chart type can be the difference between a good visualisation that answers the questions you have of your data, or a great visualisation that provokes new questions and discovers even more actionable insights.
In this live webinar, data visualisation specialist Andy Kirk will share his principles for choosing the chart type that delivers the most insights based upon your type of data and the questions you want to answer. Join us to discover how you can become more scientific in your chart selection.
Using research from his crowdsourced project - The Chartmaker Directory, Andy will analyse a number of different chart types so that you can learn:
-What each chart type reveals
-When a chart should and shouldn’t be used
-How to make your charts beautiful
Your first backup appliance was great, but how did you feel about your fifth? As data sets explode into the realm of petabytes, most backup appliances can’t scale to meet the capacity needs of a digital company. The lack of scalability forces companies to buy multiple backup appliances, creating the ‘backup appliance sprawl’. The more backup systems companies have to manage, the more time it takes to balance, monitor and maintain.
Join members of Western Digital and StorReduce to learn about a scale-out deduplication software that enables primary backups to be stored directly on object storage based private or hybrid cloud, eliminating backup appliance sprawl. In this webinar you’ll learn how this technology allows you to:
•Save up to 70% or more for backup environment TCO while retaining your existing backup applications
•Reduce hardware refresh cycle for backup appliances
•Scale as you need without migrations or rip-n-replace hardware changes
•Build a “Data Forever” environment with extreme data durability
Just because Easter is around the corner doesn’t mean your backup appliances should be multiplying like rabbits. Register today!
With new technologies such as Hive LLAP or Spark SQL, do you still need a data warehouse or can you just put everything in a data lake and report off of that? No! In the presentation, James will discuss why you still need a relational data warehouse and how to use a data lake and an RDBMS data warehouse to get the best of both worlds.
James will go into detail on the characteristics of a data lake and its benefits and why you still need data governance tasks in a data lake. He'll also discuss using Hadoop as the data lake, data virtualization, and the need for OLAP in a big data solution, and he will put it all together by showing common big data architectures.
If you’re like most companies, Salesforce is one of your most important cloud investments and you’ve likely noticed that it has grown to become one of the most powerful application ecosystems today. With over 10 million customers, Camping World specializes in selling recreational vehicles, recreational vehicle parts and services, and camping supplies.
Join Terry Britt, Enterprise Architect at Camping World, as he takes you through their journey leveraging Informatica Intelligent Cloud Services to integrate their Salesforce environment with multiple applications running on-premises, such as Oracle EBS, and in the cloud with Heroku.
During this webinar, learn how Camping World achieved:
• Real time application integration, eliminated data duplication, and improved data quality, security, and visualization
• A single view of customer, accurate and real time historical interactions, and multi-channel customer support with a single user interface
• Significantly improved customer experiences through multi-cloud and hybrid data and application integration
Big data is here to stay, and enterprises are determined to transform their data into strategic assets to inform operations and planning. Yet despite years of focus, investment, and effort, CDOs haven't reached a transformational level of maturity in data and analytics. What's holding us back?
Leveraging data from multiple years of results spanning over 5,500 survey responses from 429 companies, the Definitive Big Data Maturity Survey provides the industry's only comprehensive look at where big data and analytics are heading. Join this webinar as AtScale explains the six trends driving enterprise transformation in big data. The webinar will cover:
*The future of the cloud in big data
*BI tools winners and losers
*Strategies enterprises use to win with big data
You won't want to miss this in-depth examination of what's driving big data today.
Who Should Attend:
*Strategic Data Executives
*Business Intelligence and Analytics Leaders
Storage Spaces brings advanced storage management technology into the heart of Microsoft Windows Server. You can now scale locally attached storage at a lower cost than traditional SAN or NAS arrays. By combining Storage Spaces with award-winning Caringo FileFly, you can now scale from TBs to 100s of PBs at a fraction of the cost. FileFly will also protect all of that data, further decreasing costs by reducing your reliance on backup.
Product Manager Glen Olsen and VP of Marketing Adrian "AJ" Herrera present strategies, a live demo and use cases for re-architecting your storage by combining a secure tier of storage that is easy to expand and manage, and that has continuous data protection built in.
Your IT department has its hands full keeping your day-to-day operations in check while working on new projects. They don’t have the time or often the expertise to integrate a new acquisition into your current IT environment. Ask yourself – could your company’s IT department completely integrate a new acquisition in less than 30 days?
In this quick 30-minute webinar, you’ll learn Accudata’s proven formula for success – and understand how IT can enable your M&A business goals. We will cover:
• Why IT departments struggle with business acquisitions
• How to prioritize IT integration tasks
• Accudata’s five-step proven process to make M&A less complicated
• Review a customer example that includes integrating 35 new sites and 600+ users into an existing organization in less than 30 days
Business intelligence (BI) has been at the forefront of business decision-making for more than two decades. Then along came Big Data and it was thought that traditional BI technologies could never handle the volumes and performance issues associated with this unusual source of data.
So what do you do? Cast aside this critical form of analysis? Hardly a good answer. The better answer is to look for BI technologies that can keep up with Big Data, provide the same level of performance regardless of the volume or velocity of the data being analyzed, yet give the BI-savvy business users the familiar interface and multi-dimensionality they have come to know and love.
This webinar will present the findings from a recent survey of Big Data and the challenges and value many organizations have received from their implementations. In addition, the survey will supply a fascinating look into what Big Data technologies are most commonly used, the types of workloads supported, the most important capabilities for these platforms, the value and operational insights derived from the analytics performed in the environment, and the common use cases.
Attendees will also learn about a new BI technology built to handle Big Data queries with superior levels of scalability, performance and support for concurrent users. BI on Big Data platforms enables organizations to provide self-service and interactive on big data for all of their users across the enterprise.
Curating, creating, and delivering fresh content pushes your marketing game to the next level. Your goal: Get the right content to the right people, when and where they want to see it.
Optimization is the key, and analyzing user data, whether by leveraging AI or including engagement and other metrics in your efforts, helps you uncover not just which content is most competitive but what characteristics are pushing it to the top of the list, and even help you fine-tune your message in real time, keeping it fresh and commanding consumer attention. Once that message is refined and optimized, your data can help you home in on consumer intent, seize site-specific demand opportunities, gauge opportunity sizing — and then make sure your message is discovered with predictive intelligence.
To learn more about how to take the educated guesswork out of content creation, make your pitch more persuasive, and magnify your message, don’t miss this VB Live event.
By attending thIs webinar, you'll learn how optimizing content can:
* Increase relevance for consumers across channels
* Leverage data to increase consumer engagement with content
* Increase the volume and scope of content, netting stronger conversions.
* Seize customer journey opportunities when they appear
* Lisa Gibbs, AP Director of News Partnerships, Associated Press
* Ari Goldberg, Founder, Barber Surgeon’s Guild
* Peggy Anne Salz, Principal, Mobile Groove Consulting
* Rachael Brownell, Moderator, VentureBeat
Learn the basics of Apache Spark™ on Azure Databricks. Designed by Databricks, in collaboration with Microsoft, Azure Databricks combines the best of Databricks and Azure to help customers accelerate innovation with one-click set up, streamlined workflows and an interactive workspace that enables collaboration between data scientists, data engineers, and business analysts.
This webinar will cover the following topics:
· RDDs, DataFrames, Datasets, and other fundamentals of Apache Spark.
· How to quickly setup Azure Databricks, relieving you of DataOps duties.
· How to use the Databricks interactive notebooks, which provide a collaborative space for your entire analytics team, and how you can schedule notebooks, immediately putting your work into production.
In Part 1, we examined what GDPR is, what the requirements are and how organizations will be impacted. In Part 2, our panel will discuss more on the potential impacts of GDPR across a typical organization (including assessments, encryption, audit & controls and the impact to each department, from finance to marketing) and what the organization should be doing to plan for GDPR. Join (ISC)² on March 27, 2018 at 1:00PM Eastern, as (ISC)² discusses these topics and answers questions from the audience about this important and looming regulation.
As data analytics becomes more embedded within organizations, as an enterprise business practice, the methods and principles of agile processes must also be employed.
Agile includes DataOps, which refers to the tight coupling of data science model-building and model deployment. Agile can also refer to the rapid integration of new data sets into your big data environment for "zero-day" discovery, insights, and actionable intelligence.
The Data Lake is an advantageous approach to implementing an agile data environment, primarily because of its focus on "schema-on-read", thereby skipping the laborious, time-consuming, and fragile process of database modeling, refactoring, and re-indexing every time a new data set is ingested.
Another huge advantage of the data lake approach is the ability to annotate data sets and data granules with intelligent, searchable, reusable, flexible, user-generated, semantic, and contextual metatags. This tag layer makes your data "smart" -- and that makes your agile big data environment smart also!
When monitoring an increasing number of machines, the infrastructure and tools need to be rethinked. A new tool, ExDeMon, for detecting anomalies and raising actions, has been developed to perform well on this growing infrastructure. Considerations of the development and implementation will be shared.
Daniel has been working at CERN for more than 3 years as Big Data developer, he has been implementing different tools for monitoring the computing infrastructure in the organisation.