Get Hazelcast


Operating in today’s always-on, high-volume, high-speed, high-expectation world requires a different level of processing enablement. When microseconds can mean the difference between success and failure, Hazelcast in-memory solutions can deliver blinding speed with scalable and flexible data caching.

New levels of performance delivered to the world’s most demanding companies

New levels of performance delivered to the world’s most demanding companies

Caching puts actively used data in memory, where it can be accessed significantly more quickly. While this sounds simple, it can become very complex, as real-world systems are wildly diverse and constantly changing. Through meticulous engineering, deep caching expertise, and a focused commitment to customer needs, Hazelcast handles that diversity with a robust in-memory computing platform, delivering the benefits of distributed caching where high-speed, high-volume systems need it most.

Blazing Speed
Blazing Speed

Hazelcast’s relentless pursuit of speed has made our in-memory data store the fastest distributed cache available. As a fully in-memory data store, Hazelcast can transform and ingest data at blinding speeds, often shrinking milliseconds into microseconds. Because Hazelcast is built from the ground up as a distributed technology, it leverages the power of distributed processing while effectively eliminating the impact of network latency.


Hazelcast in-memory caching is designed to scale quickly and smoothly. And nodes cluster automatically, without production interruptions. A multitude of scaling options are offered, including High-Density Memory Store for large-scale enterprise use cases. Naturally, the Hazelcast Platform is cloud-ready, and elastic scaling makes it a perfect fit.


Hazelcast offers a wealth of development, deployment, and operational choices. Customers can take advantage of the powerful and easy-to-use JCache standard API through client-server or cluster-only architectures. Our platform also integrates seamlessly with popular frameworks and libraries like Spring, Memcached, and Hibernate.


The Hazelcast Platform puts the power of cache management within reach and keeps the total cost of ownership low by focusing on ease of development, implementation, and operation. The platform is surprisingly simple to set up and configure, offering remarkable resilience and high availability.

Speed and Scalability

Speed and Scalability

Speed in information processing is great, regardless of the industry or application. Speed at scale is a whole different level of opportunity. Speed at scale while maintaining stability? That’s what you need if you’re going to drive transformative change in your organization, and keep happy customers coming back.

The Hazelcast Platform delivers world-class, in-memory caching solutions, based on a distributed architecture that is wildly fast and seamlessly scalable.


Hazelcast offers sophisticated features with nearly instantaneous response

Management Center / Cache Management

Cache management enables the monitoring and management of members running Hazelcast. In addition to monitoring the overall state of clusters, The Hazelcast Management Center can also analyze and browse data structures in detail, update map configurations, and take thread dump from members.

High-Density Caching

Hazelcast Enterprise scales up data sizes using High-Density Memory Store, a way for Java applications to access multiple terabytes of memory per cluster member without long and unpredictable garbage collection pauses. This memory store provides the benefits of “off-heap” memory using a variety of proprietary high-performance memory management techniques.

WAN Replication

Allows multiple Hazelcast clusters to be kept in sync by replicating their state over WAN environments such as the Internet. Hazelcast supports both Active-Passive and Active-Active modes for most common scenarios of WAN Replication.

Cache Hot Restart Store

A high-performance option included in Hazelcast Enterprise in addition to the basic reliability options (write-through and write-behind). This persistence feature of Hot Restart Store provides fast cluster restarts by storing the states of the cluster members on disk in a format specially designed for restart performance and to work in concert with SSDs.

Security Suite

A rich set of JAAS-based security features that authenticate cluster members and clients in order to perform authorization checks on client operations. Socket Interceptor, Security Interceptor, Encryption, SSL, Credentials and ClusterLoginModule, Cluster Member Security, and Native Client Security are the main components of the security suite.

Rolling Upgrades

Allows an upgrade to a running Hazelcast cluster without service disruption, data loss, or interruption to execution. Hazelcast members and clients can be upgraded in a running cluster without interrupting operations.

Enterprise PaaS Deployments

Certified native images enable Hazelcast Enterprise to run in leading enterprise PaaS cloud container environments including Pivotal Cloud Foundry and Red Hat OpenShift Container Platform.

Cache Access Patterns

Hazelcast enables caching when connected to a persistent data store such as a relational database. The most common access patterns are read-through, write-through, and write-behind.

Read-Through Cache

In a read-through pattern, applications request data directly from the caching system, and if the data exists in the cache, that data is returned to the application. If the data does not exist (a “cache miss”), then the system retrieves the data from the underlying backend store. The system then loads that data into the cache and returns the data to the backend store.

Hazelcast handles the entire process so the application does not need to coordinate reads from the underlying store upon cache misses. To establish the link between Hazelcast and the backend store, application developers write pluggable query code that is executed by Hazelcast in the event of a cache miss.

Write-Through Cache

In a write-through pattern, applications can directly update data in the cache, and whenever that is done, the updated data is synchronously and automatically persisted to the backend data store. This pattern is about ensuring the cache and backend store are synchronized and is not intended to address performance (performance is addressed with the write-behind cache pattern, described below) since the backend store is still the bottleneck in an update process.

Write-Behind Cache

In a write-behind pattern, applications can update data in the cache similarly to the write-through cache pattern, except the automatic updates to the backend store are asynchronous. This means this pattern offers a performance advantage since updates to the data do not have to wait for writes to the backend store. Updates to data in the cache are acknowledged quickly since only in-memory data is updated, and Hazelcast will later push the data updates to the backend store.

Near Cache

Near cache is highly recommended for data structures that are mostly read. When an object is fetched from a remote Hazelcast member, it is inserted into the local cache, so subsequent requests are handled by a local member retrieving from the local cache.

Near cache gives Hazelcast users a huge advantage over NoSQL solutions like Redis, Cassandra, MongoDB, etc., which do not have near caching capabilities. In benchmark comparisons, Hazelcast is already 56% faster than Redis without using near cache. Enabling near cache makes Hazelcast 5 times faster than Redis. If Hazelcast near cache features in client-server deployments are being used, microsecond response times can be achieved with the right serialization.

Free Hazelcast Online Training Center

Whether you're interested in learning the basics of in-memory systems, or you're looking for advanced, real-world production examples and best practices, we've got you covered.