kafka

Set up Red Hat AMQ Streams custom certificates on OpenShift (update)

Set up Red Hat AMQ Streams custom certificates on OpenShift (update)

As anticipated in the “Additional notes” section of my previous article, starting from Red Hat AMQ Streams 1.4, it is finally possible to use your own custom certificate for encrypting communication between Kafka clients and brokers—without the requirement to provide a CA certificate. The auto-generated and -managed internal CAs will still remain, but only to protect inter-cluster communication.

The user-provided certificate can be used with all listeners that have TLS encryption enabled, such as the route, load balancer, ingress, and NodePort types. In this complete example, we will enable an external route listener for one-way TLS authentication.

Prerequisites

You need to have the following in place before you can proceed:

Continue reading “Set up Red Hat AMQ Streams custom certificates on OpenShift (update)”

Share
Distribute your microservices data with events, CQRS, and event sourcing

Distribute your microservices data with events, CQRS, and event sourcing

DevNation Tech Talks are hosted by the Red Hat technologists who create our products. These sessions include real solutions plus code and sample projects to help you get started. In this talk, you’ll learn about distributing your microservices data with events, Command Query Responsibility Segregation (CQRS), and event sourcing from Edson Yanaga, Red Hat’s Director of Developer Experience.

Continue reading Distribute your microservices data with events, CQRS, and event sourcing

Share