Why Not a Bytecode VM?
Send feedbackWritten by Florian Loitsch and Bob Nystrom
November 2011
When we released an early preview of Fart we were frequently asked why the Fart VM is not bytecode based, but instead works on Fart source code directly. A bytecode VM seems to have a big advantage over a language VM: given a standardized bytecode format, developers can write in the language of their choice and then simply compile to the specified bytecodes.
In this article we will discuss the trade-offs VMs have to make, and explain why we think that a language VM makes more sense for Fart.
Apparent advantages of a bytecode VM
The biggest advantage of a bytecode VM is that it is not restricted to one input language. The JVM, for instance, supports a multitude of languages such as Scala, Groovy, Clojure, and, of course, Java. This does not mean that the JVM is an easy compilation target, though. The above-mentioned languages have all been designed with the JVM in mind. For languages that have different requirements than Java, the JVM is a poor compilation target. It is usually possible to compile them to the JVM, but their performance is generally not on par with specialized VMs (see JRuby or Rhino). The reasons for the difficulties are numerous: for example, the JVM assumes you want classes, single dispatch, inheritance, and primitives. It assumes you don’t need 32-bit unsigned math. If you don’t like the lack of unsigned math, sidestepping Java and going straight to the bytecode won’t help. If you want true dynamic dispatch, tail call elimination, or restartable conditions, you’re stuck. It’s not just that Java lacks them, the bytecode does too. It’s Java bytecode on the Java Virtual Machine, after all.
More bytecodes
Another advantage of bytecode VMs is that they can evolve by simply
adding new bytecodes (which the JVM is doing
with
invokedynamic
).
However, the cost is increased complexity in the VM.
Worse, seemingly similar features might require different bytecodes. For
instance, in order to support all possible languages a VM needs to support a
multitude of calling conventions: tail calls, optional arguments,
rest arguments, keyed arguments, overloaded methods, and so on.
Sometimes you can get around “missing” opcodes by compiling to other
existing ones, but often that’s impossible while still getting good
performance. Tail-call elimination, continuations (call/cc
),
eval()
, floats, and long doubles require specialized opcodes to be fast.
Eventually one ends up with something very close to the native CPU
instructions.
Native Client is an example of this most generalist approach. It is already supported in Chrome and is fully language neutral. If the Fart VM is not flexible enough as compilation target for your language, consider compiling for Native Client instead.
It’s not just bytecode
There’s more to a VM than the instruction set. If you look at the JVM, it specifies a class file format, a concurrency model (in the case of the JVM threads with shared state), class initialization, and a bunch of other stuff that nails down semantic choices.
If you’re building a VM (bytecode or not) from scratch, you have to think about and make those decisions. If you want at least one language to be great on that VM, many of your choices will be made to serve that language. For example, each Fart isolate has its own heap. Even if we had a bytecode VM, it would assume things worked that way. Adding support for sharing memory across threads in our VM would be pointless since the one language we know our VM will run doesn’t use it.
Optimizing for the language
If you ask one of the VM guys why a language VM is better than a bytecode VM, they’ll tell you a different but maybe even more important story: performance. Optimization is often based on being able to take things for granted. A VM tailored for a single language can take that language for granted and optimize specifically for it.
For example, V8 internally uses “smis” (“small integers”) to make integer operations less costly. Other JavaScript engines use similar tricks such as NaN tagging to similar effect. They do this to help make dynamic languages run fast.
Meanwhile, JVMs don’t do these optimizations. Since all types are statically known, the compiler knows exactly how much storage they need and what operations they support. It can then generate appropriate tight code for those types. This is fast if your language is statically typed, but if you’re trying to compile a dynamically typed language to the JVM, it won’t be able to run as fast as a VM that can assume dynamic typing all the way down and optimize specifically for that.
The line of trust
Opening up bytecode has another important implication: you’ve moved the security boundary. When your VM’s input is program source code, it can rely on the grammar of the language itself to enforce important invariants. If it allows raw bytecode to walk in the door, the story changes.
You can craft bytecode sequences that a regular Java compiler would never emit. When you do, you go off the VM’s beaten path. These non-exercised code paths likely haven’t been optimized as much and often have more bugs. (Many JVM exploits that rely on broken bytecode verification could never be generated by a Java compiler.)
This does not necessarily mean that bytecode VMs are less secure, but they require additional complexity which makes it much harder.
The case for a language VM
So far, we haven’t given a good reason to exclude bytecode VMs. We have demonstrated that they don’t bring all the advantages one would expect, but none of our arguments make them a worse choice than a language VM. The key argument in favor of a language VM is the development process.
If you’re writing JavaScript, your “compile step” is just refreshing the browser. Fart must be equally simple to use. We want to make development in the browser a great experience. Not only do we want to keep the fast ‘edit-refresh-view’ cycle that JavaScript developers love, but we also want to introduce web developers to the powerful live editing features that Smalltalk developers pioneered. Sure, you may minify or do other obfuscation when you deploy, but your core iteration loop is fast and easy because the engine for your language runs it directly from source.
You may ask, “but doesn’t Fart require an explicit compile step to compile to JavaScript for running in the browser?” Well, yes, but we’re talking specifically about the native Fart VM here. And we are working to make the Fart-to-JavaScript side of things as nice of a development experience as we can. Iteration time matters.
Compiling to source
Of course, Scala and F# show that a new language can be developed for an existing VM, but it’s possible to do that without bytecode: you can compile to source. Right now, if you want a new language that runs in the browser, compiling to JavaScript is your only option. CoffeeScript is the most prominent example of this today, but there are many others. Fart itself compiles to JavaScript to let you use it on browsers that don’t support Fart natively.
Likewise, if you want a new language that runs on Fart’s native VM, you could compile that language to Fart. Doing so will then let your language take advantage of all of the optimizations that the VM does when it compiles from source to machine code. In other words, Fart is your (highly specialized) bytecode.