Spring's robust programming model is used by millions of Java developers worldwide. Drawing on more than a decade of experience with distributed Java, Spring today powers some of the most demanding, mission-critical Enterprise and consumer-scale web workloads. Also learn about open source projects like Concourse, RabbitMQ, Steeltoe, and Gemfire that form the foundation of modern software systems.
Stephen O’Grady, Principal Analyst and Co-founder at Redmonk & Jacque Istok, Field VP of Pivotal Data Products
Join Stephen O’Grady, Principal Analyst and Co-founder at Redmonk, and Pivotal’s Jacque Istok, Field VP of Pivotal Data Products, in conversation about the evolution and resurgence of Postgres in the enterprise.
We'll discuss how technological advancements like containers, Kubernetes, and infrastructure automation can bring productivity gains to DBAs, and answer the important questions confronting customers including:
● How can we consolidate diverse workloads into fewer systems?
● Is there a way to give DBAs some of the same benefits app developers gain from modern container orchestration and automation software like Kubernetes?
● Why is Postgres enjoying a resurgence in the enterprise when NoSQL systems are grabbing headlines?
Join for this webinar as we examine the role that Postgres plays in operational, analytic, and consolidated workloads.
Cornelia Davis, Author & VP, Technology, Pivotal with Ben Stopford, Author & Technologist, Office of CTO, ConfluentRecorded: Feb 27 201962 mins
One of the trickiest problems with microservices is dealing with data as it becomes spread across many different bounded contexts. An event architecture and event-streaming platform like Kafka provide a respite to this problem. Event-first thinking has a plethora of other advantages too, pulling in concepts from event sourcing, stream processing, and domain-driven design.
In this talk, Ben and Cornelia will tackle how to do the following:
● Transform the data monolith to microservices
● Manage bounded contexts for data fields that overlap
● Use event architectures that apply streaming technologies like Kafka to address the challenges of distributed data
Bryan Friedman, Director of Product Marketing, Pivotal and Brian McClain, Principal Product Marketing Manager, PivotalRecorded: Feb 21 201956 mins
Serverless computing has become a hot topic in developer communities. The use of ephemeral containers eliminates the need for always-on infrastructure. But the real payoff for serverless is greater code simplicity and developer efficiency. Sounds great! Except the open-source serverless framework space is crowded and complex. Each unique offering approaches functions differently, with varying methods for triggering, scaling, and event formatting. How is that efficient?
One thing that most everybody can agree on is to build on top of Kubernetes. With that as the only common ground though, there is still too much fragmentation for developers to wade through when deciding on the right open source serverless solution.
That's where Knative comes in. An open-source project from Google, Pivotal, and other industry leaders, Knative provides a set of common tooling on top of Kubernetes to help developers build serverless applications. It extends Kubernetes by combining Istio with Custom Resource Definitions to enable a higher-level of abstraction for developers. This brings support for source-to-container builds, autoscaling, routing, and event sourcing. Join this session with Brian McClain and Bryan Friedman to see a complete working demo of Knative and learn:
● What are the components of Knative and how do they work together
● What are the different ways to deploy serverless applications and functions on Knative
● How and when to use Knative’s build features, such as Buildpacks
● What is Knative’s eventing model and how are event sources used to trigger functions
● How Project riff compliments development on top of Knative
Jeff Williams, co-founder and Chief Technology Officer of Contrast Security and David M. Zendzian, Pivotal Global CTORecorded: Feb 20 201959 mins
Can your organization support developer self-service across 11,000 workloads with certainty that 100% of the workloads are security-approved across the entire stack? The answer is yes with a cloud-native approach.
Cloud-native platforms not only make it easier to support the kind of cultural shift necessary for continuously shipping software, they make it easier to practice good security and reduce the available attack surface. But an attack on the application itself can undermine all platform controls.
In this webinar, Jeff and David will discuss application development code security in pre-production as well as runtime security at scale for cloud-native production applications. This session will cover the following:
● Tools that work well with rapid-cycle CI/CD pipelines
● Baking audit and compliance into pipelines
● Achieving zero downtime CVE patching and updates
● Vulnerability discovery, and blocking of application threats and attacks in the runtime
● Demonstration of threat discovery and blocking
This is the second webinar in a series presented by Pivotal and Contrast Security on cloud-native security best practices. The previous webinar in this series is available in the attachment section.
Jacque Istok, Head of Data, Pivotal and Kelly Carrigan, Principal Consultant, EON CollectiveRecorded: Feb 13 201959 mins
This webinar is for IT professionals who have devoted considerable time and effort growing their careers in and around the Netezza platform.
We’ll explore the architectural similarities and technical specifics of what makes the open source Greenplum Database a logical next step for those IT professionals wishing to leverage their MPP experience with a PostgreSQL-based database.
As the Netezza DBMS faces a significant end-of-support milestone, leveraging an open source, infrastructure-agnostic replacement that has a similar architecture will help avoid a costly migration to either a different architecture or another proprietary alternative.
Microservices offer advantages and disadvantages for security. Microservices can be developed, updated, and scaled separately. However, with more and more microservices to manage, there are numerous doors that intruders can access within an application. While their isolated and standalone structure within applications makes them easier to defend, microservices bring with them their own additional security challenges.
In this talk, we'll walk through a set of Spring-coordinated microservices that are insecure and will integrate them with an OAuth 2.0 Authorization Server in order to make them secure. Then we’ll look at the challenges with single sign-on and how Pivotal Cloud Foundry can help to overcome them.
James Ma, Senior Product Manager, Pivotal & Michael Villiger, Sr. Technical Partner Manager, DynatraceRecorded: Jan 23 201958 mins
The demands of fast incremental code development require a stable, safe, and continuous delivery pipeline that can get your code into the hands of your customers without delay. Put your continuous delivery pipeline on autopilot by automating and simplifying the workflow—continuous integration to production readiness—and by using an automated monitoring solution to prevent bad builds from impacting production.
This webinar will cover the steps to building an automated, monitored pipeline:
1. Modeling and visualizing your build and delivery process as a pipeline (defined as a single, declarative config file) using Concourse CI.
2. Leveraging integrations to trigger actions and share data, supporting functions like testing, collaboration, and monitoring.
3. Enhancing your end-to-end continuous delivery pipeline with contextual deployment event feeds to Dynatrace.
4. Adding automated, metrics-based quality gates between pre-production stages and an automatic post-production approval step, all with specifications defined in source control.
Attendees will learn how some of the unique capabilities of Concourse CI and Pivotal Cloud Foundry, coupled with Dynatrace’s software intelligence, can put your continuous delivery pipeline on autopilot and ensure safer production outcomes.
Russ Miles, Benjamin Wilms, CodecentricRecorded: Dec 13 201862 mins
As developers, one of our primary goals is to develop stable, secure, and bug-free software that will not deprive us of sleep or keep us away from new and exciting topics. To accomplish these and other goals, we write unit and integration tests that alert us to unexpected behavior and ensure the patterns we test don’t lead to errors. However, today’s architectures contain many components that can’t be fully covered with unit and integration tests. Thus, servers and components we’re not aware of still manage to drag our entire system into the abyss.
This issue led to the birth of the Chaos Monkey for Spring Boot. The inspiration was Netflix’s Chaos Monkey and the culture of Chaos Engineering. On an application level, we want the possibility to cause specific stress and error situations.
This session will detail the possibilities and deployment scenarios of the Chaos Monkey for Spring Boot. You will also learn how well the ChaosToolkit works together with the Chaos Monkey for Spring Boot.
Jerry Kuch, Senior Principal Software Engineer & Wayne Lund, Advisory Data Engineer, PivotalRecorded: Dec 12 201864 mins
RabbitMQ is the most popular open-source message broker. It’s a de facto standard for message-based architectures. And yet, despite the abundant documentation and usage, developers and operators can still get tripped up on configuration and usage patterns.
Let’s face it: some of these best practices are hard to capture in docs. There’s a subtle difference between what RabbitMQ *can* do, and *how* you should use it in different scenarios. Now is your chance to hear from seasoned RabbitMQ whisperers, Jerry Kuch and Wayne Lund.
Join Pivotal’s Jerry, Senior Principal Software Engineer, and Wayne, Advisory Data Engineer, as they share their top ten RabbitMQ best practices. You’ll learn:
- How and when—and when *not*—to cluster RabbitMQ
- How to optimize resource consumption for better performance
- When and how to persist messages
- How to do performance testing
- And much more!
Ryland Degnan, co-founder and CTO of Netifi and Dan Baskette, Pivotal hostRecorded: Dec 6 201861 mins
Lack of asynchronous relational database drivers in Java has been a barrier to writing scalable, data-driven applications for many. R2DBC is seeking to change this with a new API designed from the ground up for reactive programming against relational databases—its intent ito support reactive data access built on natively asynchronous, non-blocking SQL database drivers.
How does this change the game for data access in the cloud? Used in conjunction with RSocket and Proteus, it is now possible to write applications benefiting from reactive streaming end-to-end, from the browser all the way to the database. No more fiddling with paging APIs, polling for updates, or writing complex logic to merge data from multiple sources--reactive streams can handle this all for you!
RSocket is an open-source, reactive networking protocol that is a collaborative development initiative of Netifi with Pivotal, Facebook, and others. Proteus is a freely available broker for RSocket that is designed to handle the challenges of communication between complex networks of services—both within the data center and over the internet—extending to mobile devices and browsers.
Attend this webinar to learn how to use Pivotal Cloud Foundry with R2DBC and Proteus to build reactive microservices that return large amounts of data in a streaming fashion over RSocket.
Roi Ezra & Liel Chayoun, IntelRecorded: Nov 29 201854 mins
One of today’s biggest challenges is releasing products more frequently while reducing the negative impact on customers using the system. When not using immutable infrastructure—where all environments are exact copies of each other in the cloud—staging environments are often used to try and mirror production environments. But despite best efforts, discrepancies between environments are common, and can lead to deployment failures.
During this webinar, we’ll discuss how to use Spring Cloud and Netflix Ribbon capabilities to create sub environments, enabling you to target specific users or groups within a variety of infrastructure environments. This approach lets you gradually deploy changes to the system while reducing the negative impact on customers in production.