paint-brush
Java Backend Observability with OpenTelemetry Traces and Minimal Codeby@apanasevich
328 reads
328 reads

Java Backend Observability with OpenTelemetry Traces and Minimal Code

by Dmitriy ApanasevichNovember 15th, 2024
Read on Terminal Reader
Read this story w/o Javascript

Too Long; Didn't Read

How we integrated the OpenTelemetry framework into our Java backend, getting tracing with minimal coding.
featured image - Java Backend Observability with OpenTelemetry Traces and Minimal Code
Dmitriy Apanasevich HackerNoon profile picture

Hello everyone! I'm Dmitriy Apanasevich, Java Developer at MY.GAMES, working on the game Rush Royale, and I'd like to share our experience integrating the OpenTelemetry framework into our Java backend. There’s quite a bit to cover here: We’ll cover necessary code changes required to implement it, as well as the new components we needed to install and configure – and, of course, we’ll share some of our results.

Our goal: achieving system observability

Let’s give some more context to our case. As developers, we want to create software that’s easy to monitor, evaluate, and understand (and this is precisely the purpose of implementing OpenTelemetry — to maximize system observability).


Traditional methods for gathering insights into application performance often involve manually logging events, metrics, and errors:



Of course, there are many frameworks that allow us to work with logs, and I’m sure that everyone reading this article has a configured system for collecting, storing and analyzing logs.


Logging was also fully configured for us, so we did not use the capabilities provided by OpenTelemetry for working with logs.


Another common way to monitor the system is by leveraging metrics:


We also had a fully configured system for collecting and visualizing metrics, so here too we ignored the capabilities of OpenTelemetry in terms of working with metrics.


But a less common tool for obtaining and analyzing this kind of system data are traces.


A trace represents the path a request takes through our system during its lifetime, and it typically begins when the system receives a request and ends with the response. Traces consist of multiple spans, each representing a specific unit of work determined by the developer or their library of choice. These spans form a hierarchical structure that helps visualize how the system processes the request.


For this discussion, we'll concentrate on the tracing aspect of OpenTelemetry.

Some more background on OpenTelemetry

Let’s also shed some light on the OpenTelemetry project, which came about by merging the OpenTracing and OpenCensus projects.


OpenTelemetry now provides a comprehensive range of components based on a standard that defines a set of APIs, SDKs, and tools for various programming languages, and the project’s primary goal is to generate, collect, manage, and export data.


That said, OpenTelemetry does not offer a backend for data storage or visualization tools.


Since we were only interested in tracing, we explored the most popular open-source solutions for storing and visualizing traces:

  • Jaeger
  • Zipkin
  • Grafana Tempo


Ultimately, we chose Grafana Tempo due to its impressive visualization capabilities, rapid development pace, and integration with our existing Grafana setup for metrics visualization. Having a single, unified tool was also a significant advantage.

OpenTelemetry components

Let’s also dissect the components of OpenTelemetry a bit.


The specification:

  • API — types of data, operations, enums

  • SDK — specification implementation, APIs on different programming languages.  A different language means a different SDK state, from alpha to stable.

  • Data protocol (OTLP) and semantic conventions


The Java API the SDK:

  • Code instrumentation libraries
  • Exporters — tools for exporting generated traces to the backend
  • Cross Service Propagators — a tool for transferring execution context outside the process (JVM)


The OpenTelemetry Collector is an important component, a proxy that receives data, processes it, and passes it on – let's take a closer look.

OpenTelemetry Collector

For high-load systems handling thousands of requests per second, managing the data volume is crucial. Trace data often surpasses business data in volume, making it essential to prioritize what data to collect and store. This is where our data processing and filtering tool comes in and enables you to determine which data is worth storing. Typically, teams want to store traces that meet specific criteria, such as:


  • Traces with response times exceeding a certain threshold.
  • Traces that encountered errors during processing.
  • Traces that contain specific attributes, such as those that passed through a certain microservice or were flagged as suspicious in the code.
  • A random selection of regular traces that provide a statistical snapshot of the system's normal operations, helping you understand typical behavior and identify trends.

Here are the two main sampling methods used to determine which traces to save and which to discard:

  • Head sampling — decides at the start of a trace whether to keep it or not
  • Tail sampling — decides only after the complete trace is available. This is necessary when the decision depends on data that appears later in the trace. For example, data including error spans. These cases cannot be handled by head sampling since they require analyzing the entire trace first


The OpenTelemetry Collector helps configure the data collection system so that it will save only the necessary data. We will discuss its configuration later, but for now, let's move on to the question of what needs to be changed in the code so that it starts generating traces.

Zero-code instrumentation

Getting trace generation really required minimal coding – it was just necessary to launch our applications with a java-agent, specifying the configuration:


-javaagent:/opentelemetry-javaagent-1.29.0.jar

-Dotel.javaagent.configuration-file=/otel-config.properties


OpenTelemetry supports a huge number of libraries and frameworks, so after launching the application with the agent, we immediately received traces with data on the stages of processing requests between services, in the DBMS, and so on.


In our agent configuration, we disabled the libraries we’re using whose spans we didn’t want to see in the traces, and to get data about how our code worked, we marked it with annotations:


@WithSpan("acquire locks")
public CompletableFuture<Lock> acquire(SortedSet<Object> source) {
  var traceLocks = source.stream().map(Object::toString).collect(joining(", "));
  Span.current().setAttribute("locks", traceLocks);

  return CompletableFuture.supplyAsync(() -> /* async job */);
}


In this example, the @WithSpan annotation is used for the method, which signals the need to create a new span named "acquire locks", and the "locks" attribute is added to the created span in the method body.


When the method finishes working, the span is closed, and it is important to pay attention to this detail for asynchronous code. If you need to obtain data related to the work of asynchronous code in lambda functions called from an annotated method, you need to separate these lambdas into separate methods and mark them with an additional annotation.

Our trace collection setup

Now, let's talk about how to configure the entire trace collection system. All our JVM applications are launched with a Java agent that sends data to the OpenTelemetry collector.


However, a single collector cannot handle a large data flow and this part of the system must be scaled. If you launch a separate collector for each JVM application, tail sampling will break, because trace analysis must occur on one collector, and if the request goes through several JVMs, the spans of one trace will end up on different collectors and their analysis will be impossible.


Here, a collector configured as a balancer comes to the rescue.


As a result, we get the following system: Each JVM application sends data to the same balancer collector, whose only task is to distribute data received from different applications, but related to a given trace, to the same collector-processor. Then, the collector-processor sends data to Grafana Tempo.



Let's take a closer look at the configuration of the components in this system.

Load balancing collector

In the collector-balancer configuration, we’ve configured the following main parts:


receivers:

  otlp:

    protocols:

      grpc:

exporters:

  loadbalancing:

    protocol:

      otlp:

        tls:

          insecure: true

    resolver:

      static:

        hostnames:

          - collector-1.example.com:4317

          - collector-2.example.com:4317

          - collector-3.example.com:4317

service:

  pipelines:

    traces:

      receivers: [otlp]

      exporters: [loadbalancing]


  • Receivers — where the methods (via which data can be received by the collector) are configured. We've configured data reception solely in the OTLP format. (It is possible to configure the reception of data via many other protocols, for example Zipkin, Jaeger.)
  • Exporters — the part of the configuration where data balancing is configured. Among the collectors-processors specified in this section, the data is distributed depending on the hash calculated from the trace identifier.
  • The Service section specifies the configuration of how the service will work: only with traces, using the OTLP receiver configured on top and transmitting data as a balancer, i.e. without processing.

The collector with data processing

The configuration of collectors-processors is more complicated, so let’s take a look there:


receivers:

  otlp:

    protocols:

      grpc:

        endpoint: 0.0.0.0:14317

processors:

  tail_sampling:

    decision_wait: 10s

    num_traces: 100

    expected_new_traces_per_sec: 10

    policies:

      [

          {

            name: latency500-policy,

            type: latency,

            latency: {threshold_ms: 500}

          },

          {

            name: error-policy,

            type: string_attribute,

            string_attribute: {key: error, values: [true, True]}

          },

          {

            name: probabilistic10-policy,

            type: probabilistic,

            probabilistic: {sampling_percentage: 10}

          }

      ]

  resource/delete:

    attributes:

      - key: process.command_line

        action: delete

      - key: process.executable.path

        action: delete

      - key: process.pid

        action: delete

      - key: process.runtime.description

        action: delete

      - key: process.runtime.name

        action: delete

      - key: process.runtime.version

        action: delete

exporters:

  otlp:

    endpoint: tempo:4317

    tls:

      insecure: true

service:

  pipelines:

    traces:

      receivers: [otlp]

      exporters: [otlp]


Similar to the collector-balancer configuration, the processing configuration consists of Receivers, Exporters, and Service sections. However, we'll focus on the Processors section, which explains how data is processed.


First, the tail_sampling section demonstrates a configuration that allows filtering the data needed for storage and analysis:


  • latency500-policy: this rule selects traces with a latency exceeding 500 milliseconds.

  • error-policy: this rule selects traces that encountered errors during processing. It searches for a string attribute named "error" with values "true" or "True" in the trace spans.

  • probabilistic10-policy: this rule randomly selects 10% of all traces to provide insights into normal application operation, errors, and long request processing.


In addition to tail_sampling, this example shows the resource/delete section to delete unnecessary attributes not required for data analysis and storage.

Results

The resulting Grafana trace search window enables you to filter data by various criteria. In this example, we simply display a list of traces received from the lobby service, which processes game metadata. The configuration allows for future filtering by attributes like latency, errors, and random sampling.


The trace view window displays the execution timeline of the lobby service, including the various spans that make up the request.


As you can see from the picture, the sequence of events is as follows — locks are acquired, then objects are retrieved from the cache, followed by the execution of a transaction that processes the requests, after which the objects are stored in the cache again and the locks are released.


The spans related to database requests were automatically generated due to the instrumentation of standard libraries. In contrast, the spans related to lock management, cache operations, and transaction initiation were manually added to the business code using the aforementioned annotations.



When viewing a span, you can see attributes that allow you to better understand what happened during processing, for example, see a query in the database.



One of the interesting features of Grafana Tempo is the service graph, which graphically displays all services exporting traces, the connections between them, the rate and latency of requests:


Wrapping up

As we’ve seen, working with OpenTelemetry tracing has enhanced our observation abilities quite nicely. With minimal code changes and a well-structured collector setup, we got deep insights – plus, we saw how Grafana Tempo's visualization capabilities further complemented our setup. Thanks for reading!