Google is committed to advancing racial equity for Black communities. See how.

Converting and visualizing a trace

When you record a trace with the Fuchsia trace system, the data can be stored in various file formats. For every data format, there is a specific tool to visualize the data.


Before you attempt to convert or analyze a trace file, make sure you have done the following:

Fuchsia trace file formats

The following types of file formats can store Fuchsia trace data:

  • FXT, or Fuchsia Trace Format, is a binary format that is a direct encoding of the original trace data that is produced by the various programs. For information on how to view this trace format, see FXT trace.
  • JSON, or Chrome Trace Format. For information on how to view this trace format, see JSON trace.
  • HTML, a standalone file that includes both the viewer and trace data. For information on how to view this data, see HTML trace.

Convert Fuchsia trace files

You can convert one or more files from FXT to JSON with the following:

fx traceutil convert FILE

Visualize a trace

There are different ways to visualize a trace based on the format of that trace:

FXT trace

To visualize an FXT trace, you can use the Perfetto Trace Viewer, which also allows you to use SQL to query your trace data.

JSON trace

To visualize this data, you can use Chromium's Trace-Viewer. The viewer is built into Chrome and can be loaded with chrome://tracing. For more information on how to use Chrome's trace viewer, see The Trace Event Profiling Tool.

HTML trace

To generate an HTML trace (viewer bundled with trace data) you'll need to run trace2html from the Chromium Catapult Repository. From the catapult repository, run:

./tracing/bin/trace2html JSON_TRACE_FILE

Analyze a trace file

To analyze a JSON trace file, open a new tab in Google Chrome and navigate to chrome://tracing. Click the Load button and open your JSON trace file.

As there is a lot of information in a trace file, there are some useful keyboard shortcuts that you can use:

  • w and s: Zoom in and zoom out, respectively. The zoom function is based on the current position of your mouse.
  • W and S: Zoom in and zoom out at a larger scale, respectively. The zoom function is based on the current position of your mouse.
  • a and d: Pan left and right, respectively.
  • A and D: Pan left and right at a larger scale, respectively.
  • You can deselect specific process rows to remove processes that aren't important for your current trace. To deselect a specific process row, click the x in the right corner of the process row.

Interpret the trace data

This example shows a trace of what the system is doing while running the du command. The du command shows disk usage.

Before you can record trace data, you must start a Fuchsia instance. From your host, if you don't have a Fuchsia target device, you can start a Fuchsia emulator with networking:

fx emu -N

This command configures and runs Fuchsia.

From a new terminal, run traceutil to record a trace of du:

fx traceutil record --buffer-size=64 \
    --categories=all --spawn \
    /boot/bin/sh -c "'\
        sleep 2 ;\
        i=0 ;\
        while [ \$i -lt 10 ] ;\
        do /bin/du /boot ;\
            i=\$(( \$i + 1 )) ;\

This command runs du in a loop, sets a recording buffer size of 64 megabytes, records all tracing categories, and launches du with fdio_spawn().

Once the command finishes running, it creates an HTML file:


A trace file has a lot of information including a time scale near the top of the trace. In this example, the whole trace lasted about 2.5 seconds.

CPU usage

The region marked by the yellow circle shows the CPU usage area where you can see the overall CPU usage on all CPU cores.

Program execution

The region marked by the green circle shows the program execution.

In this example, you can see 10 invocations of the du program, which is expected since the trace was recorded during a loop of du. Therefore, you can see 10 different du process IDs, one after the other.

blobfs CPU usage

The region marked by the blue circle shows the CPU usage to write to the blobstore filesystem (blobFS).

In this example, you can see little bursts of CPU time that are each related to an invocation of du.

At this high level, it can be difficult to determine the exact correlation between the CPU usage and the filesystem:

  • Is the CPU usage caused by the loading of du from the filesystem?
  • Is the CPU usage caused by the execution du as it runs through the target filesystem to see how much space is in use?

You can zoom in on specific areas of this region to determine the correlation between the CPU usage and the filesystem.


In this example, you can see just two du executions (the first is marked with a green circle). The first blobfs CPU burst actually consists of three main clusters and some smaller spikes. Subsequent blobfs CPU bursts have two clusters.

From analyzing this example, you can see that the blobfs bursts happen before the du program is executed. This information shows that the blobfs bursts are not due to the du program reading the filesystem. Instead, it shows that the bursts are due to loading the du program.

You are now ready to dive further into what is causing the blobs bursts.


In this example, notice the time scale that spans a time period from 2,023,500 microseconds to just past 2,024,500 which indicated a time scale of about 1 millisecond.

During that millisecond, blobfs executed code, starting with a process identified as FileReadAt, which then called Blob::Read, which then called Blob::ReadInternal.

To correlate this information with the code, you can click on parts of the report for more detailed information about a specific object.

If you click on FileReadAt, you can see the following information:


This information tells you the following:

  • The trace category for FileReadAt is vfs.
  • The length of time of the function execution.

If you click on Blob::Read, you can see the following information:


The code for Blob::Read is:

zx_status_t Blob::Read(void* data,
                       size_t len,
                       size_t off,
                       size_t* out_actual) {
    TRACE_DURATION("blobfs", "Blob::Read", "len", len, "off", off);
    LatencyEvent event(&blobfs_->GetMutableVnodeMetrics()->read,

    return ReadInternal(data, len, off, out_actual);

This code calls calls the TRACE_DURATION() macro with the category of blobfs, a name of Blob::Read, and a length and offset key and value pairs. All of this information is recorded in the trace file.

After analyzing these examples, you can still see additional objects that are being traced.