Katherine Stanley is a Software Engineer in the IBM Event Streams team based in the UK. Through her work on IBM Event Streams she has gained experience running Apache Kafka on Kubernetes and running enterprise Kafka applications. In her previous role she specialised in cloud native Java applications and microservices architectures. Katherine has co-authored an IBM Redbook on Java microservices and has contributed to the open source microservice project Game On. She enjoys sharing her experiences and has presented at conferences around the world, including the Kafka Summits in New York and London, JavaLand in Germany and JFokus in Sweden.
We now live in a world with data at its heart. The amount of data being produced every day is growing exponentially and a large amount of this data is in the form of events. Whether it be updates from sensors, clicks on a website or even tweets, applications are bombarded with a never-ending stream of new events. So, how can we architect our applications to be more reactive and resilient to these fluctuating loads and better manage our thirst for data? In this session explore how Kafka and Reactive application architecture can be combined in applications to better handle our modern data needs.
Scheduled on Wednesday from 17:50 to 18:40 in Room 9
The rise of Apache Kafka as the de-facto standard for event streaming has coincided with the rise of Kubernetes for cloud-native applications. While Kubernetes is a great choice for any distributed system, that doesn't mean it is easy to deploy and maintain a Kafka cluster running on it. At IBM we have hands-on experience with running Kafka in Kubernetes and in this session I will share our top tips for a smooth ride. I will show an example deployment of Kafka on Kubernetes and step through the system to explain the common pitfalls and how to avoid them. This will include the Kubernetes objects to use, resource considerations and connecting applications to the cluster. I will also discuss useful Kafka metrics to include in Kubernetes liveness and readiness probes. Finally I will introduce some of the tools available to help automate the management of your Kafka deployment.
Scheduled on Wednesday from 12:00 to 12:50 in Room 8
The amount of data the world produces is growing exponentially every year and many companies are realising the potential of harnessing this data. A lot of this is generated in the form of a never ending stream of events, with publishers creating the events and subscribers consuming them in many different ways. This is where Apache Kafka comes in, Kafka isn't just a messaging system - it's an event streaming platform. This session will introduce Kafka and explain concepts such as topic partitioning, consumer groups and exactly-once semantics. This session will give you the knowledge you need to start your journey with the event streaming platform that everyone is talking about.
Scheduled on Wednesday from 13:30 to 13:45 in Room 9
This session dives deep into the architecture of Apache Kafka and how to use it in practice. Kafka offers a new take on publish/subscribe messaging that is very different from other messaging systems. But Kafka isn't just a messaging system - it's an event streaming platform. Become an expert on Kafka concepts such as topic partitioning, consumer groups and exactly-once semantics. Find out best practices for using the APIs to achieve performance and reliability. Learn about advanced concepts such as stream processing using the Kafka Streams and integrating with other services using Kafka Connect.
Scheduled on Tuesday from 09:30 to 12:30 in Room 8
|Talks by tracks||Talks by session types||List of Speakers||Schedule|