home

The Best String Compression in JavaScript

2023-05-22

edit: 23.05.22 - add benchmarks of gzip, zlib and deflate via wasm-flate

Lately, I needed to find a decent algorithm to compress strings in JavaScript so that I could encode it as a URL param. Since the strings to be encoded where p5 sketches that ran in the p5.cljs web editor, it was important that the compressed string was short enough to enable longer sketches to be written.

The algorithms

After some searching on the web, I found some algorithms that I could use for this purpose. These are:

If you don’t know what LZ compression is, well, I confess … neither do I. What I noticed was that LZ-based compression is rather ubiquitous.

Hiccups

There were many more compression algorithms that I could find for JavaScript. However, many of them were either only for Node.js, using the Buffer api, or just didn’t work. For algorithms using the Buffer api, wasm implementations are a godsend.

The benchmark

Since the only thing that mattered to me was the resulting base64 string at the end of the compression process, the benchmarks I ran only measured the length of the resulting string. The compression process looks like this:

I the following string sources to test the algorithms.

Since triangle contained template literals, I had to modify the source to escape the backticks and $.

The results

character count

sketchsimplex noisequiltrianglemodele
no compression132029367745911223766
Base6417603916103281215231688
u-lzss127627685572522415980
lz-string110022444444487612292
LZW140429686096693619100
GZIP8641708319628087776
ZLIB8481692318027927760
DEFLATE8401684317227847752

Bar chart of the compression results by character count

ratio

sketchsimplex noisequiltrianglemodele
no compression100%100%100%100%100%
Base64133.33%133.37%133.35%133.33%133.33%
u-lzss96.66%94.27%71.94%57.33%67.23%
lz-string83.33%76.43%57.37%53.51%51.72%
LZW106.36%101.08%78.70%76.11%80.36%
GZIP65.45%58.17%41.26%30.81%32.71%
ZLIB64.24%57.62%41.05%30.64%32.65%
DEFLATE63.63%57.35%40.95%30.55%32.61%

compression ratio increases as file size gets larger

The base 64 conversion of the string always expanded the length. This is just how base64 works.

From this mini-benchmark, we see that the deflate algorithm performed the best, reaching almost a compression ratio of around 30% for the largest source text. That said, gzip, zlib and deflate have almost the same results.

For all three algorithms that were tested, the compression ratio improved as the length of the source string increased. But beyond a certain size, it seems to go back down.

Reversibility

sketchsimplex noisequiltrianglemodele
Base64
u-lzss
lz-string
LZW
GZIP
ZLIB
DEFLATE

Testing the reversibility of the algorithms, I was surprised to see that Google’s u-lzss wasn’t 100% reversible. For longer source strings, the decompressed string contained corrupted and malformed strings towards the end of the file.

Snippet of original string

class TriangulateIO {
    static get LENGTH() { return 23; }
    constructor(props = {}) {

Malformed output of u-lzss

lass TangleuleMoIO
cstr ic         etTy LENGTH({
eturn nu23; }   const tctor =(prop= pt{}{

Conclusion

Compression is probably better done elsewhere than a browser environment. But sometimes you just need to have a handy compression tool, like in the p5.cljs editor, which shares sketches by storing them as a url param instead of in a server somewhere. In this situation, I’ll be choosing deflate for my use case.

If you want to checkout the benchmark, it’s on github. Also, I don’t have a deep understanding of compression and maybe missing out on something. If you know better, do reach out on my socials or email!