Tracing and Monitoring IPFS Cluster

Tracing and Monitoring IPFS Cluster

IPFS Cluster supports exposing a Prometheus endpoint for metric-scraping as well as submitting trace information to Jaeger.

These are configured in the observations section of the configuration and can be enabled from there or by starting a cluster peer with:

$ ipfs-cluster-service daemon --stats --tracing

Development setup for tracing and metrics

The following section shows how to:

  • Configure and run Jaeger and Prometheus services locally using Docker
  • Configure IPFS Cluster to send traces to Jaeger and metrics to Prometheus
This section shows how to deploy and configure  tracing and metrics on a local development environment. Production  deployment of either Jaeger or Prometheus is beyond the scope of what is  being covered here.

Jaeger

First, pull down the Jaeger all-in-one image:

$ docker pull jaegertracing/all-in-one:1.9

Once the image has been downloaded, run the image with the following configuration:

$ docker run -d --name jaeger \
  -e COLLECTOR_ZIPKIN_HTTP_PORT=9411 \
  -p 5775:5775/udp \
  -p 6831:6831/udp \
  -p 6832:6832/udp \
  -p 5778:5778 \
  -p 16686:16686 \
  -p 14268:14268 \
  -p 9411:9411 \
  jaegertracing/all-in-one:1.9

Of particular note are the following ports on the Jaeger container:  - 6831 is default agent endpoint used by IPFS Cluster  - 16686 exposes the web UI of the Jaeger service, where you can query and search collected traces

Prometheus

To configure Prometheus, we create a prometheus.yml file, such as the following:

global:
  scrape_interval:     15s
  evaluation_interval: 15s

scrape_configs:
  - job_name: ipfs-cluster-daemon
    scrape_interval:     2s
    static_configs:
      - targets: ['localhost:8888']

The target address specified matches the default address in the  metrics configuration in IPFS Cluster, but feel to change it to  something more suitable to your environment, just make sure to update  your ~/.ipfs-cluster/service.json to match.

In order to run prometheus, pull the following Docker image:

$ docker pull prom/prometheus

Then run the Prometheus container, making sure to mount the configuration file we just created:

$ docker run --network host -v /tmp/prometheus.yml:/etc/prometheus/prometheus.yml --name promy prom/prometheus

Note that to have Prometheus reach the metrics endpoint exposed by  IPFS Cluster, it requires that the container be run on the host’s  network, this done via the --network host flag in the run command above.

IPFS Cluster configuration

Configure the observations section in the service.json file as follows:

{
  "metrics": {
    "enable_stats": true,
    "prometheus_endpoint": "/ip4/0.0.0.0/tcp/8888",
    "reporting_interval": "2s"
  },
  "tracing": {
    "enable_tracing": true,
    "jaeger_agent_endpoint": "/ip4/0.0.0.0/udp/6831",
    "sampling_prob": 0.3,
    "service_name": "cluster-daemon"
  }
}

For local development tracing, it is advised to change the observations.tracing.sampling_prob to 1, so that every action in the system is recorded and sent to Jaeger.

Running the cluster peer with the configuration above should provide  an endpoint for Prometheus to collect metrics and will push traces to  Jaeger.

Once cluster peer has started, go to http://localhost:9090/targets to confirm that Prometheus has been able to beginning scraping metrics from IPFS Cluster.

To confirm that tracing is functioning correctly, we will add a file  and pin to IPFS Cluster in one step by using the IPFS Cluster add command and then search for its trace in Jaeger.

$ echo 'test tracing file' > test.file
$ ipfs-cluster-ctl add test.file

Go to https://localhost:16686 and you should see a trace, it may be labelled <trace-without-root-span> due to an issue with how Jaeger creates/determines root spans, but all  the information is still inside. If there is nothing there, give it  sometime to flush the traces to the Jaeger Collector as it isn’t  instantaneous.

After having run a few commands to get some traces, it is a good time to go check out the graph page of Prometheus, which is prefilled with a histogram of the request latencies of the gorpc calls between IPFS Cluster components. There are plenty of other  metrics configured for collection and they can be found in the drop-down  next to the Execute button.

Hopefully, this tooling enables you to better understand how IPFS Cluster operates and performs.

Subscribe to SimpleAsWater

Don’t miss out on the latest issues. Sign up now to get access to the library of members-only issues.
jamie@example.com
Subscribe