Understanding Java Memory Model and Thread Safety

Threads running

In today’s world of multiple-core processors, it is vital to know how Java manages memory and threads. This understanding helps in creating strong, high — performance applications. We often overlook basic understanding threading mechanisms as developers. But when issues arise — and they will knowing the core concepts can save your hours of debugging.

Java Memory Model: More than heap and stack

The Java Memory Model (JMM) defines how threads interact through memory. It’s not just about how we organize memory. It’s also about the rules that control visibility, atomicity, and the order of memory operations.

When Java code runs, each thread gets its own stack, storing local variables and method call information. But all threads share the heap, which contains all objects your program allocates. This shared nature is strong but risky. It lets threads communicate, but it can lead to race conditions.

The JMM, introduced in Java 5 (JSR — 133), provides guarantees about when changes made by one thread become visible to other. Before we dive into the details, let’s understand why this matters.

Consider this code that appears to be harmless:

class SharedData {
  private boolean ready = false;
  private int value = 0;

  //Thread A
  public void producer() {
    value = 42;
    ready = true;
  }

  // Thread B
  public void consumer() {
    while(!ready) {
      //wait
    }
    assert value == 42; //This might fail!
  }
}

This should work base on common understanding. Thread A set value to 42, then signals ready . Thread B waits until ready is true, then reads value . But in reality, this code is flawed for several reasons:

  1. Reordering: The JVM or processor might reordered instructions, setting ready = true , before value = 42 .
  2. Visibility: Even if instructions execute in order, Thread B might not see the updated value.
  3. Infinite loop: Thread B might nnever see the update to ready and loop forever.

Memory visibility and the “happens-before” relationship

The Java Memory Model introduces concept of “happens — before” relationships. If operation A happens before operation B, then B can see A’s effect. Without this connection we can’t be sure about visibility.

Some ways to establish a happens — before relation include:

  1. Program order: Operations in the same thread follows happens — before ordering.
  2. Monitor locks: Using a monitor happens — before acquiring the same monitor.
  3. Volatile variables: Writing to a volatile variables happens — before reading it.
  4. Thread operations: Starting a thread happens — before any actions in that thread.

Let’s fix our previous example using volatile variable

class SharedData {
  private volatile boolean ready = false;
  private int value = 0;

  //Thread A
  public void producer() {
    value = 42;
    ready = true;
  }

  // Thread B
  public void consumer() {
    while(!ready) {
      //wait
    }
    assert value == 42; //This might fail!
  }
}

The volatile keyword does two things:

  • It prevents instruction reordering around the volatile access.
  • It ensures memory visibility of variables modified before a volatile write.

Atomic operations and race conditions

One common misconception is that all basic operations in Java are atomic. They’re not. For example, the operations count++ , which appears innocent, consist of three distinct operations.

  1. Read the current value.
  2. Increment it.
  3. Write it back.

This creates a window for race conditions. If two threads execute count++ concurrently:

Thread A: Read count (value: 5)
Therad B: Read count (value: 5)
Thread A: Increment to 6
Thread B: Increment to 6
Thread A: Write 6
Thread B: Write 6

The final value is 6, not 7 as expected. To fix this, we need atomic operations or synchronization. Java provides several ways to handle atomic operations:

  1. Synchronization: Using synchronized blocks or methods.
  2. Atomic classes: Using classes like AtomicInteger or AtomicReference .
  3. Lock interfaces: Using ReentrantLock , ReadWriteLock , etc.

Here’s an example with AtomicInteger :

import java.util.concurrent.atomic.AtomicInteger;

class Counter {
  private AtomicInteger count = new AtomicInteger(0);

  public void increment() {
    count.incrementAndGet();
  }

  public int getCount() {
    return count.get();
  }
}

Synchronization and locks

The synchronized keyword is Java’s built-in mechanism for mutual exclusion. It ensures that only one thread can execute a synchronized blocks or method at a time.

class Counter {
  private int count = 0;
  
  public synchronized void increment() {
    count++;
  }

  public synchronized int getCount() {
    return count;
  }

}

Every object in Java has intrinsic lock (monitor). When a thread enters a synchronized block, it acquires this lock. If another thread already hold the lock, the current thread waits.

Synchronization provides more than just mutal exclusion — it aslo establishes memory visiblity. When a thread acquires a lock, it sees all changes made by previous holders of that lock.

Beyond basic synchronization: Locks API

Since Java 5, the language offers more flexible locking through the java.util.concurrent.locks package. The Lock interface offers several features. You can try to acquire locks without blocking using tryLock() . It also supports interruptible locks and timed lock waits.

import java.util.concurrent.locks.ReentrantLock;

class Counter {
  private int count = 0;
  private final ReentrantLock lock = new ReentrantLock();

  public void increment() {
    lock.lock
    try {
      count++;
    } finally {
      lock.unlock();
    }
  }
}

For workloads with a lot of reading, ReadWriteLock lets many readers access data at the time. But it ensures that only one write can access the data only.

import java.util.concurrent.locks.ReadWriteLock;
import java.util.concurrent.locks.ReentrantReadWriteLock;

class SharedData {
  private int value = 0;
  private final ReadWriteLock lock = new ReentrantReadWriteLock();

  public void write(int newValue) {
    lock.writeLock().lock();
    try {
      value = newValue;
    } finally {
      lock.writeLock().unlock();
    }
 }

  public int read() {
    lock.readLock().lock();
    try {
      return value;
    } finally {
      lock.readLock().unlock();
    }
  }
}

Thread confinement and ThreadLocal

Sometimes the best way to handle thread safety is to avoid sharing in the first place. Thread confinement ensures that data is exclusively accessed by one thread.

Java provides ThreadLocal for elegant thread confinement:

class Context {
  private static final ThreadLocal<UserSession> userSession = ThreadLocal.withInitial(() -> new UserSession());

  public static UserSession getUserSession() {
    return userSession.get();
  }
} 

Each thread accessing userSession.get() get its own isolated copy. This is really useful for per — thread data. This includes thing like transaction contexts and user authentication info.

Immutable objects — the ultimate thread safety

Immutable objects are inherently thread — safe. Since their state cannot change after construction, there is no risk of race conditions or visibility issues.

To make class immutable:

  1. Make all fields final.
  2. Don’t provide method that modify state.
  3. Ensure proper construction (no leaking this reference).
  4. Properly handle mutable components (defensive copies).
public final class ImmutablePoint {
  private final int x;
  private final int y;

  public ImmutablePoint(int x, int y) {
    this.x = x;
    this.y = y;
  }

  public int getX() { return x; }
  public int getY() { return y; }

  // No setters!

  //Instead of modifying, create a new instances
  public ImmutablePoint translate(int dx, int dy) {
    return new ImmutablePoint(x + dx, y + dy);
  }
}

Immutable objects eliminate an entire class of bugs and signifcantly simplify your code. Use them whenever possible.

Common concurrency issues to watch out for

1. Deadlocks

Deadlock occur when two or more threads each hold a resources the other needs, creating a circular wait. For example:

// Thread A
synchronized (lock A) {
  // Do something
  synchronized (lock B) {
    //Do something else
  }
}

// Thread B
synchronized (lock B) {
  // Do something
  synchronized (lock A) {
    // Do something else 
  }
}

To prevent deadlock:

  • Establish a consistent order for acquiring locks.
  • Use timed lock acquisition (tryLock with timeout).
  • Be aware of nested locking.

2. Liveness hazards

Beyond deadlocks, other liveness issues include:

  • Starvation: Threads are perpetually denied access to resources they need.
  • Livelock: Threads respond to each other’s action but make no progress.
  • Thread leaks: Threads are started but never properly terminated.

3. Performance issues

The most common performance issues with concurrency include:

  • Lock contention: Threads frequently competing for the same lock.
  • Context switching overhead: Too many thread causing excessive CPU switching.
  • Memory trashing: Poor localiy of reference due to thread interactions.

Best practices for thread — safe code

  1. Minimize sharing: The less data shared between threads, the fewer synchronization issues. 
  2. Prefer immutability: Immutable objects eliminate many concurrency problmes.
  3. Use thread — safe collections: ConcurrentHashhMap, CopyOnWriteArrayList , etc.
  4. Consider concurrent algorithm: Often faster than synchronized sequential algorithms.
  5. Be explicit about thread safety: Document thread — safety guarantees for your classes.
  6. Use high — level concurrency tools: Choose executors, concurrent collections, and atomic variables. Avoid raw threads and synchronization.
  7. Test for cocnrrency: Use tools like stress testing and race conditions analysis.

Practical example

Let’s combine several concepts to build a simple thread — safe cache.

import java.util.concurrent.ConcurrentHashMap;
import java.util.concurrent.locks.ReadWriteLock;
import java.util.concurrent.locks.ReentrantReadWriteLock;
import java.util.function.Function;

public class SimpleCache<K, V> {

    private final ConcurrentHashMap<K, V> cache = new ConcurrentHashMap<>();

    private final ReadWriteLock refreshLock = new ReentrantReadWriteLock();

    public V get(K key, Function<K, V> computeFunction) {
        // Fast path - no locking for cache hits
        V value = cache.get(key);
        if (value != null) {
            return value;
        }

        // Slow path with read lock for cache miss
        refreshLock.readLock().lock();

        try {
            // Check again in case another thread updated while we waited
            value = cache.get(key);
            if (value != null) {
                return value;
            }
        } finally {
            refreshLock.readLock().unlock();
        }

        // Need to compute value - acquire write lock

        refreshLock.writeLock().lock();

        try {
            // Triple check to avoid duplicate computation
            value = cache.get(key);
            if (value == null) {
                value = computeFunction.apply(key);
                cache.put(key, value);
            }

            return value;
        } finally {
            refreshLock.writeLock().unlock();
        }
    }

    public void invalidate(K key) {
        cache.remove(key);
    }

    public void clear() {
        cache.clear();
    }
  • Use ConcurrenthashMap for the underlying storage.
  • Employs a read — write lock to minimize contention.
  • Implements a double — checked locking pattern for efficiency.
  • Avoid unnecessary synchroization for cache hits.

Conclusion

The Java Memory Model helps with multi — threaded programming. Its concurrency tools are powerful, but you need to understand and apply them carefully. Understand memory visiblity, atomicity, and synchronization. This knowledge helps your write thread — safe code. It also improve performance and reduced tricky bugs.

Thread safety isn’t about preventing crashes. It’s also about making sure everything works correctly, not matter how threads interleave. This needs careful design and clear grasp of how Java ensures memory consistency between threads.

As multi — core processors are now common, understanding these concepts is key for today’s Java developers.

References

  1. Brian Goetz, et al. “Java Concurrency in Practice” — The definitive resource on Java threading and concurrency.
  2. Doug Lea, “Concurrent Programming in Java” — From the designer of Java’s concurrency utilities.
  3. JSR-133 Java Memory Model — The formal specification.
  4. What volatile means in Java
  5. The Java Language Specification — Chapter on Threads and Locks
  6. Java Concurrency Utilities

Subscribe to Egor Voronianskii | Java Development and whatsoever

Sign up now to get access to the library of members-only issues.
Jamie Larson
Subscribe