The cost of JavaScript in 2019 - 12 minutes read


The cost of JavaScript in 2019 · V8

One large change to the cost of JavaScript over the last few years has been an improvement in how fast browsers can parse and compile script. In 2019, the dominant costs of processing scripts are now download and CPU execution time.

User interaction can be delayed if the browser’s main thread is busy executing JavaScript, so optimizing bottlenecks with script execution time and network can be impactful.

What does this mean for web developers? Parse & compile costs are no longer as slow as we once thought. The three things to focus on for JavaScript bundles are:

Why is it important to optimize download and execution times? Download times are critical for low-end networks. Despite the growth in 4G (and even 5G) across the world, our effective connection types remain inconsistent with many of us running into speeds that feel like 3G (or worse) when we’re on the go.

JavaScript execution time is important for phones with slow CPUs. Due to differences in CPU, GPU, and thermal throttling, there are huge disparities between the performance of high-end and low-end phones. This matters for the performance of JavaScript, as execution is CPU-bound.

In fact, of the total time a page spends loading in a browser like Chrome, anywhere up to 30% of that time can be spent in JavaScript execution. Below is a page load from a site with a pretty typical workload (Reddit.com) on a high-end desktop machine:

On mobile, it takes 3–4× longer for a median phone (Moto G4) to execute Reddit’s JavaScript compared to a high-end device (Pixel 3), and over 6× as long on a low-end device (the <$100 Alcatel 1X):

When you’re trying to optimize JavaScript execution time, keep an eye out for Long Tasks that might be monopolizing the UI thread for long periods of time. These can block critical tasks from executing even if the page looks visually ready. Break these up into smaller tasks. By splitting up your code and prioritizing the order in which it is loaded, you can get pages interactive faster and hopefully have lower input latency.

Raw JavaScript parsing speed in V8 has increased 2× since Chrome 60. At the same time, raw parse (and compile) cost has become less visible/important due to other optimization work in Chrome that parallelizes it.

V8 has reduced the amount of parsing and compilation work on the main thread by an average of 40% (e.g. 46% on Facebook, 62% on Pinterest) with the highest improvement being 81% (YouTube), by parsing and compiling on a worker thread. This is in addition to the existing off-main-thread streaming parse/compile.

We can also visualize the CPU time impact of these changes across different versions of V8 across Chrome releases. In the same amount of time it took Chrome 61 to parse Facebook’s JS, Chrome 75 can now parse both Facebook’s JS AND 6 times Twitter’s JS.

Let’s dive into how these changes were unlocked. In short, script resources can be streaming-parsed and-compiled on a worker thread, meaning:

The not-so-short explanation is… Much older versions of Chrome would download a script in full before beginning to parse it, which is a straightforward approach but it doesn’t fully utilize the CPU. Between versions 41 and 68, Chrome started parsing async and deferred scripts on a separate thread as soon as the download begins.

In Chrome 71, we moved to a task-based setup where the scheduler could parse multiple async/deferred scripts at once. The impact of this change was a ~20% reduction in main thread parse time, yielding an overall ~2% improvement in TTI/FID as measured on real-world websites.

In Chrome 72, we switched to using streaming as the main way to parse: now also regular synchronous scripts are parsed that way (not inline scripts though). We also stopped canceling task-based parsing if the main thread needs it, since that just unnecessarily duplicates any work already done.

Previous versions of Chrome supported streaming parsing and compilation where the script source data coming in from the network had to make its way to Chrome’s main thread before it would be forwarded to the streamer.

This often resulted in the streaming parser waiting for data that arrived from the network already, but had not yet been forwarded to the streaming task as it was blocked by other work on the main thread (like HTML parsing, layout, or JavaScript execution).

We are now experimenting with starting parsing on preload, and the main-thread-bounce was a blocker for this beforehand.

Leszek Swirski’s BlinkOn presentation goes into more detail:

In addition to the above, there was an issue in DevTools that rendered the entire parser task in a way that hints that it’s using CPU (full block). However, the parser blocks whenever it’s starved for data (that needs to go over the main thread). Since we moved from a single streamer thread to streaming tasks, this became really obvious. Here’s what you’d use to see in Chrome 69:

The “parse script” task is shown to take 1.08 seconds. However, parsing JavaScript isn’t really that slow! Most of that time is spent doing nothing except waiting for data to go over the main thread.

In general, the DevTools performance pane is great for getting a high-level overview of what’s happening on your page. For detailed V8-specific metrics such as JavaScript parse and compile times, we recommend using Chrome Tracing with Runtime Call Stats (RCS). In RCS results, and tell you how much time was spent parsing and compiling JavaScript off the main thread, whereas and captures the main thread metrics.

Let’s look at some examples of real-world sites and how script streaming applies.

Reddit.com has several 100 kB+ bundles which are wrapped in outer functions causing lots of lazy compilation on the main thread. In the above chart, the main thread time is all that really matters because keeping the main thread busy can delay interactivity. Reddit spends most of its time on the main thread with minimum usage of the Worker/Background thread.

They’d benefit from splitting up some of their larger bundles into smaller ones (e.g 50 kB each) without the wrapping to maximize parallelization — so that each bundle could be streaming-parsed + compiled separately and reduce main thread parse/compile during start-up.

We can also look at a site like Facebook.com. Facebook loads ~6MB of compressed JS across ~292 requests, some of it async, some preloaded, and some fetched with a lower priority. A lot of their scripts are very small and granular — this can help with overall parallelization on the Background/Worker thread as these smaller scripts can be streaming-parsed/compiled at the same time.

Note, you’re probably not Facebook and likely don’t have a long-lived app like Facebook or Gmail where this much script may be justifiable on desktop. However, in general, keep your bundles coarse and only load what you need.

Although most JavaScript parsing and compilation work can happen in a streaming fashion on a background thread, some work still has to happen on the main thread. When the main thread is busy, the page can’t respond to user input. Do keep an eye on the impact both downloading and executing code has on your UX.

Because the JSON grammar is much simpler than JavaScript’s grammar, JSON can be parsed more efficiently than JavaScript. This knowledge can be applied to improve start-up performance for web apps that ship large JSON-like configuration object literals (such as inline Redux stores). Instead of inlining the data as a JavaScript object literal, like so:

…it can be represented in JSON-stringified form, and then JSON-parsed at runtime:

As long as the JSON string is only evaluated once, the approach is much faster compared to the JavaScript object literal, especially for cold loads.

There’s an additional risk when using plain object literals for large amounts of data: they could be parsed twice!

The first pass can’t be avoided. Luckily, the second pass can be avoided by placing the object literal at the top-level, or within a PIFE.

V8’s (byte)code-caching optimization can help. When a script is first requested, Chrome downloads it and gives it to V8 to compile. It also stores the file in the browser’s on-disk cache. When the JS file is requested a second time, Chrome takes the file from the browser cache and once again gives it to V8 to compile. This time, however, the compiled code is serialized, and is attached to the cached script file as metadata.

The third time, Chrome takes both the file and the file’s metadata from the cache, and hands both to V8. V8 deserializes the metadata and can skip compilation. Code caching kicks in if the first two visits happen within 72 hours. Chrome also has eager code caching if a service worker is used to cache scripts. You can read more about code caching in code caching for web developers.

Download and execution time are the primary bottlenecks for loading scripts in 2019. Aim for a small bundle of synchronous (inline) scripts for your above-the-fold content with one or more deferred scripts for the rest of the page. Break down your large bundles so you focus on only shipping code the user needs when they need it. This maximizes parallelization in V8.

On mobile, you’ll want to ship a lot less script because of network, memory consumption and execution time for slower CPUs. Balance latency with cacheability to maximize the amount of parsing and compilation work that can happen off the main thread.

Source: V8.dev

Powered by NewsAPI.org

Keywords:

JavaScriptV8 (JavaScript engine)JavaScriptWeb browserScripting languageData processingScripting languageDownloadCentral processing unitRun time (program lifecycle phase)Human–computer interactionWeb browserEntry pointThread (computing)Execution (computing)JavaScriptProgram optimizationScripting languageRun time (program lifecycle phase)Computer networkWeb developmentParsingCompilerJavaScriptComputer network4G5G3GUSB On-The-GoJavaScriptMobile phoneCentral processing unitCentral processing unitGraphics processing unitThermal design powerComputer performanceJavaScriptExecution (computing)CPU-boundWeb pageWeb browserGoogle ChromeJavaScriptWeb pageRedditGraphical user interfaceMachineMobile phone4XMobile phoneMoto G4RedditJavaScriptComputer hardwarePixelAlcatel-LucentJavaScriptUser interfaceWeb pageReady... BreakSource codeWeb pageJavaScriptParsingV8 (JavaScript engine)Google ChromeParsingCompilerMathematical optimizationGoogle ChromeV8 (JavaScript engine)ParsingCompilerFacebookPinterestYouTubeParsingThread (computing)Thread (computing)ParsingCompilerSoftware versioningV8 (JavaScript engine)Google ChromeGoogle ChromeFacebookJavaScriptGoogle ChromeFacebookJavaScriptTwitterJavaScriptScripting languageStreaming mediaParsingThread (computing)Google ChromeDigital distributionScripting languageCentral processing unitSoftware versioningChrome OSScripting languageThread (computing)DownloadChrome OSScheduling (computing)ParsingGoogle ChromeSynchronization (computer science)Scripting languageScripting languageParsingThread (computing)Shadow CopyGraphical user interfaceParsingCompilerScripting languageComputer networkGraphical user interfaceEntry pointThread (computing)Tape driveParsingComputer networkThread (computing)HTMLParsingJavaScriptExecution (computing)ParsingThread (computing)ParsingCentral processing unitParsingThread (computing)Thread (computing)Graphical user interfaceParsingScripting languageParsingJavaScriptDataThread (computing)High-level programming languageV8 (JavaScript engine)Software metricJavaScriptParsingCompilerGraphical user interfaceTracing (software)Run time (program lifecycle phase)SubroutineStatisticsRevision Control SystemRevision Control SystemParsingCompilerJavaScriptEntry pointThread (computing)Thread (computing)Scripting languageRedditKilobyteSubroutineCompilerThread (computing)All That Really MattersRedditThread (computing)KilobyteParallel computingProduct bundlingStreaming mediaParsingThread (computing)ParsingCompilerBootingFacebookFacebookData compressionJavaScriptScheduling (computing)Scripting languageParallel computingThread (computing)ParsingCompilerFacebookMobile appFacebookGmailDesktop computerGlory (Britney Spears album)JavaScriptParsingCompilerThread (computing)Thread (computing)Thread (computing)Input/outputSource codeUser experience designJSONJavaScriptJSONJavaScriptKnowledge representation and reasoningComputer performanceWeb applicationJSONConfiguration managementObject-oriented programmingLiteral (computer programming)DataJavaScriptObject (computer science)Literal (computer programming)JSONJSONParsingRun time (program lifecycle phase)JSONString (computer science)JavaScriptObject (computer science)Literal (computer programming)Object (computer science)Literal (computer programming)Object (computer science)Literal (computer programming)V8 (JavaScript engine)ByteSource codeCache (computing)Program optimizationScripting languageChrome OSComputer fileWeb browserPage cacheJavaScriptComputer fileGoogle ChromeComputer fileWeb cacheCompilerSource codeSerializationWeb cacheShell scriptMetadataGoogle ChromeMetadataCache (computing)V8 (JavaScript engine)V8 (JavaScript engine)MetadataCompilerSource codeCache (computing)Google ChromeSource codeCache (computing)Scripting languageSource codeCache (computing)Source codeCache (computing)Web developmentDigital distributionRun time (program lifecycle phase)Scripting languageAbove the foldScripting languagePage breakSource codeUser (computing)Parallel computingV8 (JavaScript engine)Mobile phoneScripting languageComputer networkComputer memoryRun time (program lifecycle phase)Central processing unitLatency (engineering)ParsingCompilerEntry pointThread (computing)