Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

release-22.2: util/metric: change preset buckets from NetworkLatencyBuckets to IOLatencyBuckets #107102

Merged
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
6 changes: 3 additions & 3 deletions pkg/ccl/sqlproxyccl/connector_test.go
Original file line number Diff line number Diff line change
Expand Up @@ -383,7 +383,7 @@ func TestConnector_dialTenantCluster(t *testing.T) {
Mode: metric.HistogramModePrometheus,
Metadata: metaDialTenantLatency,
Duration: time.Millisecond,
Buckets: metric.NetworkLatencyBuckets,
Buckets: metric.IOLatencyBuckets,
}),
DialTenantRetries: metric.NewCounter(metaDialTenantRetries),
}
Expand Down Expand Up @@ -468,7 +468,7 @@ func TestConnector_dialTenantCluster(t *testing.T) {
Mode: metric.HistogramModePreferHdrLatency,
Metadata: metaDialTenantLatency,
Duration: time.Millisecond,
Buckets: metric.NetworkLatencyBuckets,
Buckets: metric.IOLatencyBuckets,
}),
DialTenantRetries: metric.NewCounter(metaDialTenantRetries),
}
Expand Down Expand Up @@ -502,7 +502,7 @@ func TestConnector_dialTenantCluster(t *testing.T) {
Mode: metric.HistogramModePreferHdrLatency,
Metadata: metaDialTenantLatency,
Duration: time.Millisecond,
Buckets: metric.NetworkLatencyBuckets,
Buckets: metric.IOLatencyBuckets,
}),
DialTenantRetries: metric.NewCounter(metaDialTenantRetries),
}
Expand Down
6 changes: 3 additions & 3 deletions pkg/ccl/sqlproxyccl/metrics.go
Original file line number Diff line number Diff line change
Expand Up @@ -235,7 +235,7 @@ func makeProxyMetrics() metrics {
Mode: metric.HistogramModePreferHdrLatency,
Metadata: metaConnMigrationAttemptedCount,
Duration: base.DefaultHistogramWindowInterval(),
Buckets: metric.NetworkLatencyBuckets,
Buckets: metric.IOLatencyBuckets,
}),
AuthFailedCount: metric.NewCounter(metaAuthFailedCount),
ExpiredClientConnCount: metric.NewCounter(metaExpiredClientConnCount),
Expand All @@ -244,7 +244,7 @@ func makeProxyMetrics() metrics {
Mode: metric.HistogramModePreferHdrLatency,
Metadata: metaDialTenantLatency,
Duration: base.DefaultHistogramWindowInterval(),
Buckets: metric.NetworkLatencyBuckets},
Buckets: metric.IOLatencyBuckets},
),
DialTenantRetries: metric.NewCounter(metaDialTenantRetries),
// Connection migration metrics.
Expand All @@ -256,7 +256,7 @@ func makeProxyMetrics() metrics {
Mode: metric.HistogramModePreferHdrLatency,
Metadata: metaConnMigrationAttemptedLatency,
Duration: base.DefaultHistogramWindowInterval(),
Buckets: metric.NetworkLatencyBuckets,
Buckets: metric.IOLatencyBuckets,
}),
ConnMigrationTransferResponseMessageSize: metric.NewHistogram(metric.HistogramOptions{
Metadata: metaConnMigrationTransferResponseMessageSize,
Expand Down
4 changes: 2 additions & 2 deletions pkg/kv/kvprober/kvprober.go
Original file line number Diff line number Diff line change
Expand Up @@ -258,15 +258,15 @@ func NewProber(opts Opts) *Prober {
Mode: metric.HistogramModePreferHdrLatency,
Metadata: metaReadProbeLatency,
Duration: opts.HistogramWindowInterval,
Buckets: metric.NetworkLatencyBuckets,
Buckets: metric.IOLatencyBuckets,
}),
WriteProbeAttempts: metric.NewCounter(metaWriteProbeAttempts),
WriteProbeFailures: metric.NewCounter(metaWriteProbeFailures),
WriteProbeLatency: metric.NewHistogram(metric.HistogramOptions{
Mode: metric.HistogramModePreferHdrLatency,
Metadata: metaWriteProbeLatency,
Duration: opts.HistogramWindowInterval,
Buckets: metric.NetworkLatencyBuckets,
Buckets: metric.IOLatencyBuckets,
}),
WriteProbeQuarantineOldestDuration: metric.NewFunctionalGauge(
metaWriteProbeQuarantineOldestDuration,
Expand Down
2 changes: 1 addition & 1 deletion pkg/kv/kvserver/liveness/liveness.go
Original file line number Diff line number Diff line change
Expand Up @@ -331,7 +331,7 @@ func NewNodeLiveness(opts NodeLivenessOptions) *NodeLiveness {
Mode: metric.HistogramModePreferHdrLatency,
Metadata: metaHeartbeatLatency,
Duration: opts.HistogramWindowInterval,
Buckets: metric.NetworkLatencyBuckets,
Buckets: metric.IOLatencyBuckets,
}),
}
nl.mu.nodes = make(map[roachpb.NodeID]Record)
Expand Down
2 changes: 1 addition & 1 deletion pkg/kv/kvserver/metrics.go
Original file line number Diff line number Diff line change
Expand Up @@ -2231,7 +2231,7 @@ func newStoreMetrics(histogramWindow time.Duration) *StoreMetrics {
Mode: metric.HistogramModePreferHdrLatency,
Metadata: metaLeaseRequestLatency,
Duration: histogramWindow,
Buckets: metric.NetworkLatencyBuckets,
Buckets: metric.IOLatencyBuckets,
}),
LeaseTransferSuccessCount: metric.NewCounter(metaLeaseTransferSuccessCount),
LeaseTransferErrorCount: metric.NewCounter(metaLeaseTransferErrorCount),
Expand Down
68 changes: 0 additions & 68 deletions pkg/util/metric/histogram_buckets.go
Original file line number Diff line number Diff line change
Expand Up @@ -78,74 +78,6 @@ var IOLatencyBuckets = []float64{
9999999999.999969, // 9.999999999s
}

// NetworkLatencyBuckets are prometheus histogram buckets suitable for a histogram
// that records a quantity (nanosecond-denominated) in which most measurements
// behave like network latencies, i.e. most measurements are in the ms to sub-second
// range during normal operation.
var NetworkLatencyBuckets = []float64{
// Generated via TestHistogramBuckets/NetworkLatencyBuckets.
500000.000000, // 500µs
568747.715565, // 568.747µs
646947.927922, // 646.947µs
735900.312190, // 735.9µs
837083.242884, // 837.083µs
952178.364257, // 952.178µs
1083098.538963, // 1.083098ms
1232019.639535, // 1.232019ms
1401416.711034, // 1.401416ms
1594105.105912, // 1.594105ms
1813287.274717, // 1.813287ms
2062605.990318, // 2.062605ms
2346204.890209, // 2.346204ms
2668797.343109, // 2.668797ms
3035744.784401, // 3.035744ms
3453145.822334, // 3.453145ms
3927937.595933, // 3.927937ms
4468011.069141, // 4.468011ms
5082342.177389, // 5.082342ms
5781141.006222, // 5.781141ms
6576021.481300, // 6.576021ms
7480194.389996, // 7.480194ms
8508686.942589, // 8.508686ms
9678592.522117, // 9.678592ms
11009354.773683, // 11.009354ms
12523090.754761, // 12.52309ms
14244958.517175, // 14.244958ms
16203575.229933, // 16.203575ms
18431492.792031, // 18.431492ms
20965738.839853, // 20.965738ms
23848432.140611, // 23.848432ms
27127482.599575, // 27.127482ms
30857387.515093, // 30.857387ms
35100137.315047, // 35.100137ms
39926245.827925, // 39.926245ms
45415922.211464, // 45.415922ms
51660404.016126, // 51.660404ms
58763473.538708, // 58.763473ms
66843182.667648, // 66.843182ms
76033814.886682, // 76.033814ms
86488117.045035, // 86.488117ms
98379837.985822, // 98.379837ms
111906616.224248, // 111.906616ms
127293264.668375, // 127.293264ms
144795506.973983, // 144.795506ms
164704227.631154, // 164.704227ms
187350306.418342, // 187.350306ms
213110117.571795, // 213.110117ms
242411785.065635, // 242.411785ms
275742297.964389, // 275.742297ms
313655604.103963, // 313.655604ms
356781816.616787, // 356.781816ms
405837686.312094, // 405.837686ms
461638513.960647, // 461.638513ms
525111700.464186, // 525.1117ms
597312160.111267, // 597.31216ms
679439853.085354, // 679.439853ms
772859728.612681, // 772.859728ms
879124410.201811, // 879.12441ms
1000000000.000001, // 1s
}

// BatchProcessLatencyBuckets are prometheus histogram buckets suitable for a
// histogram that records a quantity (nanosecond-denominated) in which most
// measurements are in the seconds to minutes range during normal operation.
Expand Down
5 changes: 0 additions & 5 deletions pkg/util/metric/histogram_buckets_test.go
Original file line number Diff line number Diff line change
Expand Up @@ -52,11 +52,6 @@ func TestHistogramBuckets(t *testing.T) {
verifyAndPrint(t, exp, IOLatencyBuckets, LATENCY)
})

t.Run("NetworkLatencyBuckets", func(t *testing.T) {
exp := prometheus.ExponentialBucketsRange(500e3, 1e9, 60)
verifyAndPrint(t, exp, NetworkLatencyBuckets, LATENCY)
})

t.Run("BatchProcessLatencyBuckets", func(t *testing.T) {
exp := prometheus.ExponentialBucketsRange(500e6, 300e9, 60)
verifyAndPrint(t, exp, BatchProcessLatencyBuckets, LATENCY)
Expand Down