Link Search Menu Expand Document Documentation Menu

Summary report

At the end of each run, OpenSearch Benchmark shows a summary report based on the metric keys defined in the workload. This page gives details on each line of the summary report and that line’s associated metric key.

Cumulative indexing time of primary shards

Corresponding metrics key: indexing_total_time

The cumulative time used for indexing as reported by the Index Stats API. Note that this is not wall-clock time, for example, if M indexing threads ran for N minutes, report M * N minutes, not N minutes.

Cumulative indexing time across primary shards

Corresponding metrics key: indexing_total_time (property: per-shard)

The minimum, median, and maximum cumulative time used for indexing across primary shards as reported by the Index Stats API.

Cumulative indexing throttle time of primary shards

Corresponding metrics key: indexing_throttle_time

The cumulative time that the indexing has been throttled as reported by the Index Stats API. Note that this is not wall-clock time, for example, if M indexing threads ran for N minutes, report M * N minutes, not N minutes.

Cumulative indexing throttle time across primary shards

Corresponding metrics key: indexing_throttle_time (property: per-shard)

The minimum, median, and maximum cumulative time used that indexing has been throttled across primary shards as reported by the Index Stats API.

Cumulative merge time of primary shards

Corresponding metrics key: merges_total_time

The cumulative runtime of merges of primary shards, as reported by the index stats API. Note that this is not wall-clock time.

Cumulative merge count of primary shards

Corresponding metrics key: merges_total_count

The cumulative number of merges of primary shards, as reported by the Index Stats API under _all/primaries.

Cumulative merge time across primary shards

Corresponding metrics key: merges_total_time (property: per-shard)

The minimum, median, and maximum cumulative time of merges across primary shards as reported by the Index Stats API.

Cumulative refresh time of primary shards

Corresponding metrics key: refresh_total_time

The cumulative time used for index refresh of primary shards as reported by the Index Stats API. Note that this is not wall-clock time.

Cumulative refresh count of primary shards

Corresponding metrics key: refresh_total_count

The cumulative number of refreshes of primary shards as reported by the Index Stats API under _all/primaries.

Cumulative refresh time across primary shards

Corresponding metrics key: refresh_total_time (property: per-shard)

The minimum, median, and maximum cumulative time for index refresh across primary shards as reported by the Index Stats API.

Cumulative flush time of primary shards

Corresponding metrics key: flush_total_time

The cumulative time used for index flush of primary shards as reported by the Index Stats API. Note that this is not wall-clock time.

Cumulative flush count of primary shards

Corresponding metrics key: flush_total_count

The cumulative number of flushes of primary shards as reported by the Index Stats API under _all/primaries.

Cumulative flush time across primary shards

Corresponding metrics key: flush_total_time (property: per-shard)

The minimum, median, and maximum time for index flush across primary shards as reported by the Index Stats API.

Cumulative merge throttle time of primary shards

Corresponding metrics key: merges_total_throttled_time

The cumulative time within merges that have been throttled as reported by the Index Stats API. Note that this is not wall-clock time.

Cumulative merge throttle time across primary shards

Corresponding metrics key: merges_total_throttled_time (property: per-shard)

The minimum, median, and maximum cumulative time that merges have been throttled across primary shards as reported by the Index Stats API.

ML processing time

Corresponding metrics key: ml_processing_time

The minimum, mean, median, and maximum time in milliseconds that a machine learning (ML) job has spent processing a single bucket.

Total young gen GC time

Corresponding metrics key: node_total_young_gen_gc_time

The total runtime of the young generation (gen) garbage collector (GC) across the whole cluster as reported by the Node Stats API.

Total young gen GC count

Corresponding metrics key: node_total_young_gen_gc_count

The total number of young gen GCs across the whole cluster as reported by the Node Stats API.

Total old gen GC time

Corresponding metrics key: node_total_old_gen_gc_time

The total runtime of the old gen GC across the whole cluster as reported by the Node Stats API.

Total old gen GC count

Corresponding metrics key: node_total_old_gen_gc_count

The total number of old gen GCs across the whole cluster as reported by the Node Stats API.

Total ZGC cycles GC time

Corresponding metrics key: node_total_zgc_cycles_gc_count

The total number of garbage collections performed by the Z garbage collector (ZGC) across the whole cluster as reported by the Node Stats API.

Total ZGC pauses GC time

Corresponding metrics key: node_total_zgc_pauses_gc_time

The total time spent in stop-the-world pauses by the ZGC across the whole cluster as reported by the Node Stats API.

Total ZGC pauses GC count

Corresponding metrics key: node_total_zgc_pauses_gc_count

The total number of stop-the-world pauses performed by the ZGC across the whole cluster as reported by the Node Stats API.

Store size

Corresponding metrics key: store_size_in_bytes

The index size in bytes (excluding the translog) as reported by the Index Stats API.

Translog size

Corresponding metrics key: translog_size_in_bytes

The translog size in bytes as reported by the Index Stats API.

Heap used for X

Corresponding metrics keys: segments_*_in_bytes

The number of bytes used for the corresponding item as reported by the Index Stats API. The item may be any of the following:

  • Doc values
  • Terms
  • Norms
  • Points
  • Stored fields

Segments count

Corresponding metrics key: segments_count

The total number of segments as reported by the Index Stats API.

Total ingest pipeline count

Corresponding metrics key: ingest_pipeline_cluster_count

The total number of documents ingested by all nodes within the cluster over the race duration.

Total ingest pipeline time

Corresponding metrics key: ingest_pipeline_cluster_time

The total time in milliseconds spent preprocessing ingest documents by all nodes within the cluster over the race duration.

Total ingest pipeline failed

Corresponding metrics key: ingest_pipeline_cluster_failed

The total number of failed ingest operations by all nodes within the cluster over the race duration.

Throughput

Corresponding metrics key: throughput

Reports the minimum, mean, median, and maximum throughput for each task.

The number of operations that OpenSearch can perform within a certain time period per second. The report includes the minimum, mean, median, and maximum throughput for each task.

Latency

Corresponding metrics key: latency

The time period between submission of a request and receiving the complete response. It includes the wait time the request spends waiting before it is processed by OpenSearch. OpenSearch reports several percentile numbers for each task. Which percentiles are shown depends on how many requests OpenSearch can capture during the latency period.

Service time

Corresponding metrics key: service_time

The time period between sending a request and receiving the corresponding response. It does not include waiting time. While many load testing tools refer to this metric as latency, it is not the same. OpenSearch reports several percentile numbers for each task. Which percentiles are shown depends on how many requests OpenSearch can capture during the latency period.

Processing time

Processing time is only reported if the setting output.processingtime is set to true in the OpenSearch Benchmark configuration file.

Corresponding metrics key: processing_time

The time period between start of request processing and retrieval of the complete response. Unlike service_time, this metric includes OpenSearch’s client-side processing overhead. The larger the difference between service_time and processing_time, the higher the overhead in the client. Depending on your processing goals, this can point to a potential client-side bottleneck that requires investigation.

Error rate

Corresponding metrics key: service_time. Each service_time record has a meta.success flag.

The ratio of erroneous responses relative to the total number of responses. Any exception thrown by the Python OpenSearch client is considered erroneous, for example, HTTP response codes 4xx, 5xx, or network errors (network unreachable). You can investigate the root cause by inspecting OpenSearch and OpenSearch Benchmark logs and rerunning the benchmark.

Disk usage

Corresponding metrics keys: disk_usage_total Metric metadata: index and field

The total number of bytes that a single field uses on disk. Recorded for each field returned by the Disk Usage API even if the total is 0.