From bd38ee691ca5687a1666c5aba9bf270953a11204 Mon Sep 17 00:00:00 2001 From: Shawn Poulson Date: Mon, 11 Mar 2024 12:09:47 -0400 Subject: [PATCH] Fix tests. --- functional_test.go | 72 ++++++++++++++++++++++++++++++--------------- gubernator.go | 4 +-- peer_client_test.go | 8 +++-- 3 files changed, 56 insertions(+), 28 deletions(-) diff --git a/functional_test.go b/functional_test.go index dfa12ee7..348c3199 100644 --- a/functional_test.go +++ b/functional_test.go @@ -1526,6 +1526,8 @@ func TestResetRemaining(t *testing.T) { } func TestHealthCheck(t *testing.T) { + name := t.Name() + key := randomKey() client, err := guber.DialV1Server(cluster.DaemonAt(0).GRPCListeners[0].Addr().String(), nil) require.NoError(t, err) @@ -1539,8 +1541,8 @@ func TestHealthCheck(t *testing.T) { _, err = client.GetRateLimits(context.Background(), &guber.GetRateLimitsReq{ Requests: []*guber.RateLimitReq{ { - Name: "test_health_check", - UniqueKey: "account:12345", + Name: name, + UniqueKey: key, Algorithm: guber.Algorithm_TOKEN_BUCKET, Behavior: guber.Behavior_BATCHING, Duration: guber.Second * 3, @@ -1589,19 +1591,32 @@ func TestHealthCheck(t *testing.T) { ctx, cancel := context.WithTimeout(context.Background(), clock.Second*15) defer cancel() require.NoError(t, cluster.Restart(ctx)) + + // wait for every peer instance to come back online + numPeers := int32(len(cluster.GetPeers())) + for _, peer := range cluster.GetPeers() { + peerClient, err := guber.DialV1Server(peer.GRPCAddress, nil) + require.NoError(t, err) + testutil.UntilPass(t, 10, 300*clock.Millisecond, func(t testutil.TestingT) { + healthResp, err = peerClient.HealthCheck(context.Background(), &guber.HealthCheckReq{}) + assert.Equal(t, "healthy", healthResp.GetStatus()) + assert.Equal(t, numPeers, healthResp.PeerCount) + }) + } } func TestLeakyBucketDivBug(t *testing.T) { defer clock.Freeze(clock.Now()).Unfreeze() - + name := t.Name() + key := randomKey() client, err := guber.DialV1Server(cluster.GetRandomPeer(cluster.DataCenterNone).GRPCAddress, nil) require.NoError(t, err) resp, err := client.GetRateLimits(context.Background(), &guber.GetRateLimitsReq{ Requests: []*guber.RateLimitReq{ { - Name: "test_leaky_bucket_div", - UniqueKey: "account:12345", + Name: name, + UniqueKey: key, Algorithm: guber.Algorithm_LEAKY_BUCKET, Duration: guber.Millisecond * 1000, Hits: 1, @@ -1619,8 +1634,8 @@ func TestLeakyBucketDivBug(t *testing.T) { resp, err = client.GetRateLimits(context.Background(), &guber.GetRateLimitsReq{ Requests: []*guber.RateLimitReq{ { - Name: "test_leaky_bucket_div", - UniqueKey: "account:12345", + Name: name, + UniqueKey: key, Algorithm: guber.Algorithm_LEAKY_BUCKET, Duration: guber.Millisecond * 1000, Hits: 100, @@ -1644,6 +1659,8 @@ func TestMultiRegion(t *testing.T) { } func TestGRPCGateway(t *testing.T) { + name := t.Name() + key := randomKey() address := cluster.GetRandomPeer(cluster.DataCenterNone).HTTPAddress resp, err := http.DefaultClient.Get("http://" + address + "/v1/HealthCheck") require.NoError(t, err) @@ -1663,8 +1680,8 @@ func TestGRPCGateway(t *testing.T) { payload, err := json.Marshal(&guber.GetRateLimitsReq{ Requests: []*guber.RateLimitReq{ { - Name: "requests_per_sec", - UniqueKey: "account:12345", + Name: name, + UniqueKey: key, Duration: guber.Millisecond * 1000, Hits: 1, Limit: 10, @@ -1692,6 +1709,7 @@ func TestGRPCGateway(t *testing.T) { } func TestGetPeerRateLimits(t *testing.T) { + name := t.Name() ctx := context.Background() peerClient, err := guber.NewPeerClient(guber.PeerConfig{ Info: cluster.GetRandomPeer(cluster.DataCenterNone), @@ -1701,6 +1719,7 @@ func TestGetPeerRateLimits(t *testing.T) { t.Run("Stable rate check request order", func(t *testing.T) { // Ensure response order matches rate check request order. // Try various batch sizes. + requestTime := epochMillis(clock.Now()) testCases := []int{1, 2, 5, 10, 100, 1000} for _, n := range testCases { @@ -1711,13 +1730,14 @@ func TestGetPeerRateLimits(t *testing.T) { } for i := 0; i < n; i++ { req.Requests[i] = &guber.RateLimitReq{ - Name: "Foobar", - UniqueKey: fmt.Sprintf("%08x", i), - Hits: 0, - Limit: 1000 + int64(i), - Duration: 1000, - Algorithm: guber.Algorithm_TOKEN_BUCKET, - Behavior: guber.Behavior_BATCHING, + Name: name, + UniqueKey: randomKey(), + Hits: 0, + Limit: 1000 + int64(i), + Duration: 1000, + Algorithm: guber.Algorithm_TOKEN_BUCKET, + Behavior: guber.Behavior_BATCHING, + RequestTime: &requestTime, } } @@ -1743,16 +1763,18 @@ func TestGetPeerRateLimits(t *testing.T) { func TestGlobalBehavior(t *testing.T) { const limit = 1000 broadcastTimeout := 400 * time.Millisecond + requestTime := epochMillis(clock.Now()) makeReq := func(name, key string, hits int64) *guber.RateLimitReq { return &guber.RateLimitReq{ - Name: name, - UniqueKey: key, - Algorithm: guber.Algorithm_TOKEN_BUCKET, - Behavior: guber.Behavior_GLOBAL, - Duration: guber.Minute * 3, - Hits: hits, - Limit: limit, + Name: name, + UniqueKey: key, + Algorithm: guber.Algorithm_TOKEN_BUCKET, + Behavior: guber.Behavior_GLOBAL, + Duration: guber.Minute * 3, + Hits: hits, + Limit: limit, + RequestTime: &requestTime, } } @@ -2368,3 +2390,7 @@ func sendHit(t *testing.T, d *guber.Daemon, req *guber.RateLimitReq, expectStatu func randomKey() string { return fmt.Sprintf("%016x", rand.Int()) } + +func epochMillis(t time.Time) int64 { + return t.UnixNano() / 1_000_000 +} diff --git a/gubernator.go b/gubernator.go index 9542e7da..e931c41c 100644 --- a/gubernator.go +++ b/gubernator.go @@ -188,7 +188,7 @@ func (s *V1Instance) GetRateLimits(ctx context.Context, r *GetRateLimitsReq) (*G "Requests.RateLimits list too large; max size is '%d'", maxBatchSize) } - requestTime := EpochMillis(clock.Now()) + requestTime := epochMillis(clock.Now()) resp := GetRateLimitsResp{ Responses: make([]*RateLimitResp, len(r.Requests)), } @@ -795,6 +795,6 @@ func isDeadlineExceeded(err error) bool { return errors.Is(err, context.DeadlineExceeded) } -func EpochMillis(t time.Time) int64 { +func epochMillis(t time.Time) int64 { return t.UnixNano() / 1_000_000 } diff --git a/peer_client_test.go b/peer_client_test.go index d739f40a..926eb2da 100644 --- a/peer_client_test.go +++ b/peer_client_test.go @@ -37,6 +37,7 @@ func TestPeerClientShutdown(t *testing.T) { } const threads = 10 + requestTime := epochMillis(clock.Now()) cases := []test{ {"No batching", gubernator.Behavior_NO_BATCHING}, @@ -71,9 +72,10 @@ func TestPeerClientShutdown(t *testing.T) { wg.Go(func() error { ctx := context.Background() _, err := client.GetPeerRateLimit(ctx, &gubernator.RateLimitReq{ - Hits: 1, - Limit: 100, - Behavior: c.Behavior, + Hits: 1, + Limit: 100, + Behavior: c.Behavior, + RequestTime: &requestTime, }) if err != nil {