Companies need a data-processing solution that increases the speed of business agility, not one that is complicated by too many technology requirements. This requires a system that delivers continuous/real-time data-processing capabilities for the new business reality.
Stream processing is a hot topic right now, especially for any organization looking to provide insights faster. But what does it mean for users of Java applications, microservices, and in-memory computing?
In this webinar, we will cover the evolution of stream processing and in-memory related to big data technologies and why it is the logical next step for in-memory processing projects.
Setting up servers and configuring software can get in the way of the problems you are trying to solve. With Hazelcast Cloud we take all of those pain points away.
Watch this webinar to learn how you can instantly fire up and then work with Hazelcast Cloud from anywhere in the world. With our auto-generated client stubs for Java, Go, Node.js, Python and .NET, we can have you connected and coding in less than a minute!
Get a 30-day free trial.
Get started today with the
industry’s leading in-memory computing platform.
The in-memory speed you count on, with the convenience and scalability of cloud.
In this guide, you will learn how to use Hazelcast distributed caching with MicroProfile and deploy to a local Kubernetes cluster. You will then create a Kubernetes Service which load balances between containers and verify that you can share data between microservices.
The microservice you will deploy is called hazelcast-microprofile. The hazelcast-microprofile microservice simply helps you write (“put”) data and read it back. As the Kubernetes Service sends the request to a different pod each time you initiate the request, the data will be served by the shared hazelcast cluster between hazelcast-microprofile pods.
You will use a local single-node Kubernetes cluster. However, you can deploy this application on any Kubernetes distribution.