Core Java

JVM Performance Magic Tricks

HotSpot, the JVM we all know and love, is the brain in which our Java and Scala juices flow. Over the years, it’s been improved and tweaked by more than a handful of engineers, and with every iteration, the speed and efficiency of its code execution is nearing that of native compiled code.duke01-e1369743413155

At its core lies the JIT (“Just-In-Time”) compiler. The sole purpose of this component is to make your code run fast, and it is one of the reasons HotSpot is so popular and successful.

What does the JIT compiler actually do?

While your code is being executed, the JVM gathers information about its behavior. Once enough statistics are gathered about a hot method (10K invocations is the default threshold), the compiler kicks in, and converts that method’s platform-independent “slow” bytecode into an optimized, lean, mean compiled version of itself.

Some optimizations are obvious: simple method inlining, dead code removal, replacing library calls with native math operations, etc. The JIT compiler doesn’t stop there, mind you. Here are some of the more interesting optimizations performed by it:

Divide and conquer

How many times have you used the following pattern:

StringBuilder sb = new StringBuilder("Ingredients: ");

for (int i = 0; i < ingredients.length; i++) {
	if (i > 0) {
		sb.append(", ");
	}
	sb.append(ingredients[i]);
}

return sb.toString();

Or perhaps this one:

boolean nemoFound = false;
 
for (int i = 0; i < fish.length; i++) {
    String curFish = fish[i];
     
    if (!nemoFound) {
        if (curFish.equals("Nemo")) {
            System.out.println("Nemo! There you are!");
            nemoFound = true;
            continue;
        }
    }
     
    if (nemoFound) {
        System.out.println("We already found Nemo!");
    } else {
        System.out.println("We still haven't found Nemo : (");
    }
}

What both these loops have in common, is that in both cases the loop does one thing for a while, and then another thing from a certain point on. The compiler can spot these patterns, and split the loops into cases, or “peel” several iterations.

Let’s take the first loop for example. The if (i > 0) line starts as false for a single iteration, and from that point on it always evaluates to true. Why check the condition every time then? The compiler would compile that code as if it were written like so:

StringBuilder sb = new StringBuilder("Ingredients: ");

if (ingredients.length > 0) {
	sb.append(ingredients[0]);
	for (int i = 1; i < ingredients.length; i++) {
		sb.append(", ");
		sb.append(ingredients[i]);
	}
}

return sb.toString();

This way, the redundant if (i > 0) is removed, even if some code might get duplicated in the process, as speed is what it’s all about.

Living on the edge

Null checks are bread-and-butter. Sometimes null is a valid value for our references (e.g. indicating a missing value, or an error), but sometimes we add null checks just to be on the safe side.

Some of these checks may never fail (null denoting a failure, for that matter). One classic example would include an assertion, like this:

public static String l33tify(String phrase) {
	if (phrase == null) {
		throw new IllegalArgumentException("phrase must not be null");
	}
	return phrase.replace('e', '3');
}

If your code behaves well, and never passes null as an argument to l33tify, the assertion will never fail.

After executing this code many, many times without ever entering the body of the if statement, the JIT compiler might make the optimistic assumption that this check is most likely unnecessary. It would then proceed to compile the method, dropping the check altogether, as if it were written like so:

public static String l33tify(String phrase) {
	return phrase.replace('e', '3');
}

This can result in a significant performance boost, which may be a pure win in most cases.

But what if that happy-path assumption eventually proves to be wrong?

Since the JVM is now executing native compiled code, a null reference would not result in a fuzzy NullPointerException, but rather in a real, harsh memory access violation. The JVM, being the low-level creature that it is, would intercept the resulting segmentation fault, recover, and follow-up with a de-optimization — the compiler can no longer assume that the null check is redundant: it recompiles the method, this time with the null check in place.

Virtual insanity

One of the main differences between the JVM’s JIT compiler and other static ones such as C++ compilers, is that the JIT compiler has dynamic runtime data on which it can rely when making decisions.

Method inlining is a common optimization in which the compiler takes a complete method and inserts its code into another’s, in order to avoid a method call. This gets a little tricky when dealing with virtual method invocations (or dynamic dispatch).

Take the following code for example:

public class Main {
	public static void perform(Song s) {
		s.sing();
	}
}

public interface Song { void sing(); }

public class GangnamStyle implements Song {
	@Override
	public void sing() {
		System.out.println("Oppan gangnam style!");
	}
}

public class Baby implements Song {
	@Override
	public void sing() {
		System.out.println("And I was like baby, baby, baby, oh");
	}
}

// More implementations here

The method perform might be executed millions of times, and each time an invocation of the method sing takes place. Invocations are costly, especially ones such as this one, since it needs to dynamically select the actual code to execute each time according to the runtime type of s. Inlining seems like a distant dream at this point, doesn’t it?

Not necessarily! After executing perform a few thousand times, the compiler might decide, according to the statistics it gathered, that 95% of the invocations target an instance of GangnamStyle. In these cases, the HotSpot JIT can perform an optimistic optimization with the intent of eliminating the virtual call to sing. In other words, the compiler will generate native code for something along these lines:

public static void perform(Song s) {
	if (s fastnativeinstanceof GangnamStyle) {
		System.out.println("Oppan gangnam style!");
	} else {
		s.sing();
	}
}

Since this optimization relies on runtime information, it can eliminate most of the invocations to sing, even though it is polymorphic.

The JIT compiler has a lot more tricks up its sleeve, but these are just a few to give you a taste of what goes on under the hood when our code is executed and optimized by the JVM.

Can I help?

The JIT compiler is a compiler for straightforward people; it is built to optimize straightforward writing, and it searches for patterns which appear in everyday standard code. The best way to help your compiler is to not try so hard to help it — just write your code as you otherwise would.
 

Reference: JVM Performance Magic Tricks from our JCG partner Niv Steingarten at the Takipi blog.

Niv Steingarten

Niv is a co-founder at Takipi where he's building tools that help developers debug Java and Scala in production. A lover of beautiful, simple code. Dvorak typist.
Subscribe
Notify of
guest

This site uses Akismet to reduce spam. Learn how your comment data is processed.

5 Comments
Oldest
Newest Most Voted
Inline Feedbacks
View all comments
Alexandru Muntean
Alexandru Muntean
10 years ago

Good advices.. you can also read more about branch predictability

Niv Steingarten
10 years ago

Of course, branch prediction is also performed by HotSpot. It’s an interesting topic in its own right!

Luigi Fugaro
Luigi Fugaro
10 years ago

Interesting

Niv Steingarten
10 years ago
Reply to  Luigi Fugaro

Thanks Luigi!

deepa
deepa
8 years ago

Good example..Thanks

Back to top button