Wide logging: Stripe's canonical log line pattern

(blog.alcazarsec.com)

19 points | by alcazar 3 hours ago

7 comments

  • bradleybuda 1 hour ago
    This is all good advice but it's not what I would traditionally think of as "logging" - it's something more like what the industry calls APM or tracing.

    > For a web service, that usually means one log event at the end of every request.

    The point of a log, as I have seen it used, is to understand what's going on during a request. Yes, some of that is timing / success / failure information, but you also need:

    - Logical tracing - at a high level, what decisions did we make (authz, business logic) when serving this request

    - Partial failures or warnings - the request may have succeeded, but we may have failed or timed out a call to a dependent service and delivered a partially degraded response as a result

    - Tracepoints to support active development - how often do we hit a certain codepath (to help developers roll out or deprecate features)

    It's useful to distinguish between "context" (things we know about a request), timing, success, and plain old logging (using human language to describe to an operator what the system is doing). Context should be included with every line - I strongly agree with OP's proposal of wide events, don't make operators ever JOIN if you can avoid it - but it won't always be the same because we learn things about the request as we process it. Yes, if you wait until the end of a request, you will have maximal context, but at the loss of fine-grained detail during processing.

    • alcazar 1 hour ago
      I don't believe this should replace traditional logging but rather complement it. It transforms breadcrumbs into unified records that are easier to query during an incident.

      I completely agree with your plain-old logging list. I would add canonical log lines in addition to the normal log traces. They serve as the summary at the end.

      • t-writescode 1 hour ago
        I’ve worked at a company where all we had were large, unified, post-request logs and honestly those logs were orders of magnitude easier to work with than having to coallate lots of logs into a single concept. ELK liked those giant logs more, too.

        It does help that the system was built from the ground up to use the big logs.

        I think the best feature was that we would extract and build a ton of keys into the full-size log, so you could find all transactions related to X really, really easily, where X was a wide variety of things.

        Every place I’ve been elsewhere the logs have always been way harder to parse, process or even just find. It would always take several disparate queries to get everything.

        E.g. “oh, we need to find something related to X? Well okay, that gives us 50 different separate calls, now let’s do another search on each of those calls’ unique transaction id and review each whole transaction individually” vs “I need X. Cool, here’s all 50, complete calls.”

        Edit: to be clear, throughout the code it was just “context.info()”, over and over again, like regular logging, it was just shoved into a big object at the end of all non-application-crashing exceptions. And the application was built to not fully crash, like, ever. (And this worked)

  • zamalek 1 hour ago
    This is just structured logging. Slog in go, tracing in rust, serilog in C#, etc. You should be combining this with the likes of otel or datadog, which will keep those fields structured and filterable.
  • deskamess 1 hour ago
    Are there any restrictions on how short the error_slug should be? The meat of some of my errors can be pretty long (for example an ffmpeg error). There are also many phases to a job - call them tasks. Can a canonical log line be a collection of task log lines?
    • alcazar 1 hour ago
      You should avoid dumping the raw error entirely. The idea is that error_slug is a stable grouping key.

      The idea is to consolidate all that can be grouped into one logical unit. So you would do one long log line at the end, after all tasks are done.

  • stephantul 1 hour ago
    I’ve never been so conflicted about an article. It has clearly been generated by an llm, but still has useful content. It’s a good article, but…
  • joaonmatos 1 hour ago
    This is not much different than CloudWatch metric events. At Amazon we output a service log or Embedded Metrics Format event with these things.
  • formerly_proven 2 hours ago
    It's strange how undercooked logging is (not just in standard libraries, but in general) given everyone and every app does it.
    • alcazar 2 hours ago
      It's one of those things that keeps getting reinvented over and over. Most people don't want to put in the effort to perfect it.

      For the longest time, I've done the bare minimum and just used console.log/print for whatever I thought I needed in that particular piece of code; without any coherent system.

  • whorleater 1 hour ago
    Honestly I'm kinda down on most of FB's tech, but them pioneering Scuba was pretty sick. It has all the wide logging/apm/tracing/etc, and it's hard to describe how useful it is if you haven't seen it in action.