[Update: In July 2015 I gave a talk based on this post at the Seattle Node.js meetup. Here’s the screencast. The slides and demos are on Github.]
It’s not all roses in the world of Node.js. In fact, more than just about every other platform I’ve used, it can feel like a precarious mountain path with a long cliff below. Yes, it features some very beautiful vistas, but one wrong step and it’s all over. Let’s cover four of those simple mistakes with substantial, unexpected consequences, as well as some potential guardrails.
The simplest of these mistakes is the classic crash. My code has done this so very many times. It’s easy; just try to get a key from an null object, or try to call a string like a function, and you’ve crashed.
var x = 5;
x.getData();
// TypeError: Object 5 has no method 'getData'
var y;
y.left = 6;
// TypeError: Cannot set property 'left' of undefined
The difference in Node.js is the event loop. When you ask for a file from disk, your provided callback will be called directly by the system on a future turn of the event loop. A crash in that callback will immediately take down the entire process. I’ll say that again - in an asynchronous callback, a crash will immediately take down the entire process. Ah, brings me back to those core dumps I got all the time when I was writing C++…
So, the general approach in the world of Node.js is to let the process crash. Just restart it with apps like forever
, nodemon
, pm2
, or OS-level systems like upstart
, runit
, systemd
, etc. Maybe capture the error with its callstack via process.on('uncaughtException')
and report it somewhere. This feels a bit weird at first, especially if you’re used to the standard application server approach: catching the error, serving an error page, and moving on.
Sadly, even quick restarts will result in socket hang-ups for your clients. In the worst case, if you have a multi-stage async operation in progress, it can even result in corrupt or inconsistent data. What else do you expect if you immediately take down the process?
thehelp-cluster
is a library I wrote to shut down your express
server gracefully when a crash happens. It first responds with an error to the original request (via your express
error handler), then closes all open keepalive connections, and finally takes the process down when all outstanding requests are complete. After a crash your process is in an unknown state, so I thought it prudent to shut down as soon as possible.hapi
is also resilient to crashes in endpoint handlers. It keeps the server alive after crashes, which could result in unexpected behavior or memory leaks.domain
. Sadly, domains are deprecated despite having no good replacement solution.Last November one of my node.js apps started taking a really long time to respond. It was responding in about a minute, where I was expecting it to be far less than a second. First, I looked at the application logs on the app node the slow requests were coming from, and found no responses that took longer than a few ms.
Confused, I looked at my load balancer’s logs and discovered that the app node which actually served me the response I saw was the second one tried, and 60 seconds was the load balancer’s timeout. After some poking of the problematic first app node, I discovered that my node processes were hanging when they requested session data from redis.
Node is really, really good at doing a lot of async operations in parallel. Promises and libraries like async
help you kick off several at once and then do something when they’re all complete. But what happens when one of those async calls never returns? In Rails, for example, if a page takes too long to gather data and render, an error is raised. In the Node.js world, it just looks like the user cancelled the request.
Happily, a lot of libraries have timeouts and will return errors in these kinds of cases. superagent
and request
will return a timeout error even if the underlying web request never returns. redis
, however, can get into a bad state and hang on all async requests (be sure to listen for both ‘error’ and ‘end’ events on your redis client object). This is a perfect example of why we all need to understand our dependencies well.
But even if you’ve fully vetted all your third-party libraries, it’s still easy to cause this yourself by forgetting a res.send()
or next()
call somewhere. Or even easier, in a multi-branch async method, forgetting to call the callback in an infrequently-hit branch.
morgan
express logger, which installs a high-resolution timer you can access via the :response-time
token. I’ve found that on hangs it outputs not a duration but a dash, but this can also mean that the user cancelled the action. Either way, lots of dashes for :response-time
is a bad sign.hapi
doesn’t seem to output anything to the logs on a hang, even with the good-console
plugin in place. :0(express
and hapi
will both keep a connection open until the remote client gives up, and I’ve seen Chrome to be something like two minutes, and nginx
is 60 seconds by default. hapi
does have a built-in route timeout option you’ll want to investigate. express
is a little trickier. You can use middleware like timeout
, but you will need watch for your long-running handlers attempting to send responses after the timeout error has already been sent to the client.My first Node.js app backed a site with heavy client-side interactivity, and delivered the required data in three big chunks. Some of that data went back as far as a year to enable all the interactivity we wanted in the charts. More and more data built up every day, adding to the size. Things just didn’t seem snappy with all that data flying around. It became clear that getting the data from the data store, then transforming it and sending it to the client, took long enough to slow down the event loop.
With the event loop rapidly processing through many small tasks, a Node.js process is healthy. You keep this in mind as you implement a standard web application endpoint: you validate an incoming request’s user and parameters, then make a request for some data from a database, transform it slightly, and send it back to the client.
What could be wrong with that? Well, if the database query doesn’t include any kind of limit, we could get an order of magnitude more records than expected. The transformation step would then take more than a couple milliseconds.
For all that data transformation time, the process is unresponsive to other requests. Callbacks from your setInterval
calls are dropped. Even if other in-progress responses would take just a couple milliseconds, they will all need to wait until their turn on the single-threaded event loop. Every in-flight request will have that wait time added to its response time. All of your users will see delays, not just the one that requested the large data set.
streams
are the way to handle large amounts of data without taking down the process. Put the time in to learn how to use them effectively.morgan
support for response times. The good-console
plugin for hapi
includes response times by default. Because one too-long synchronous code block will delay everything else, you’ll see very clear patterns in the data.For my last contract I needed to write a program to process 11.6 gigabytes of JSON data, transforming it before injecting it into a database. Happily, JSONStream was there to assist in handling the data little bits at a time, because I certainly couldn’t load it all into memory at once. I had it implemented in short order, then started my first run.
It was working, but the JSON processing it was just a little bit faster than the import into the database, so memory usage increased as the database imports fell further and further behind. The solution? Periodically call stream.pause()
, then wait on all outstanding database import calls, then call stream.resume()
. With that key application of backpressure, my program was free to rip through all that data in about 45 minutes, with constant memory usage.
Node.js is really good at kicking off a lot of concurrent tasks. But like anything, it’s good to think about the long-term in addition to the short term. Otherwise, it’s really easy to get into a state where you’ve bitten off more than you can chew. Or worse, perhaps you’re fine, but a service you depend on is starting to buckle under the pressure of your application’s highly-concurrent onslaught.
toobusy-js
to monitor your event loop latency. At its most basic, it will detect a slow event loop, which can be a sign of high load or long synchronous processing. You can also use it to turn away requests under high latency, avoiding runaway failure as more incoming requests add load to an already-struggling process.agent
, keep-alive and sockets. It’s impolite and slow to close and reopen lots of sockets if you don’t need to.Of course, testing is the first line of defense to prevent all of these things. But a production service needs resiliency. These mitigations are defense in depth techniques, so if an issue does get to production, its effect will be minimized.
With all this in place, and our service humming along in production, we can relax for a bit. Maybe even enjoy the view.
If you’re a frequent visitor at scottnonnenberg.com (and of course you should be), you may have noticed an unusual entry on the work page: the Social Security Administration (SSA). Yes, I recently... Read more »
Happy New Year! This time last year I posted an entry about my system for managing the day-to-day; something a little different from the standard list of resolutions. For 2015 I’m announcing the... Read more »