/* * * Copyright 2020 gRPC authors. * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. * */ package client import ( "testing" "time" "google.golang.org/grpc/xds/internal/testutils" ) type clusterUpdateErr struct { u ClusterUpdate err error } // TestClusterWatch covers the cases: // - an update is received after a watch() // - an update for another resource name // - an upate is received after cancel() func (s) TestClusterWatch(t *testing.T) { v2ClientCh, cleanup := overrideNewXDSV2Client() defer cleanup() c, err := New(clientOpts(testXDSServer)) if err != nil { t.Fatalf("failed to create client: %v", err) } defer c.Close() // TODO: add a timeout to this recv. // Note that this won't be necessary if we finish the TODO below to call // Client directly instead of v2Client.r. v2Client := <-v2ClientCh clusterUpdateCh := testutils.NewChannel() cancelWatch := c.WatchCluster(testCDSName, func(update ClusterUpdate, err error) { clusterUpdateCh.Send(clusterUpdateErr{u: update, err: err}) }) if _, err := v2Client.addWatches[cdsURL].Receive(); err != nil { t.Fatalf("want new watch to start, got error %v", err) } wantUpdate := ClusterUpdate{ServiceName: testEDSName} // This is calling v2Client.r to send the update, but r is set to Client, so // this is same as calling Client to update. The one thing this covers is // that `NewXDSV2Client` is called with the right parent. // // TODO: in a future cleanup, this (and the same thing in other tests) can // be changed call Client directly. v2Client.r.newCDSUpdate(map[string]ClusterUpdate{ testCDSName: wantUpdate, }) if u, err := clusterUpdateCh.Receive(); err != nil || u != (clusterUpdateErr{wantUpdate, nil}) { t.Errorf("unexpected clusterUpdate: %v, error receiving from channel: %v", u, err) } // Another update, with an extra resource for a different resource name. v2Client.r.newCDSUpdate(map[string]ClusterUpdate{ testCDSName: wantUpdate, "randomName": {}, }) if u, err := clusterUpdateCh.Receive(); err != nil || u != (clusterUpdateErr{wantUpdate, nil}) { t.Errorf("unexpected clusterUpdate: %+v, %v, want channel recv timeout", u, err) } // Cancel watch, and send update again. cancelWatch() v2Client.r.newCDSUpdate(map[string]ClusterUpdate{ testCDSName: wantUpdate, }) if u, err := clusterUpdateCh.TimedReceive(chanRecvTimeout); err != testutils.ErrRecvTimeout { t.Errorf("unexpected clusterUpdate: %v, %v, want channel recv timeout", u, err) } } // TestClusterTwoWatchSameResourceName covers the case where an update is received // after two watch() for the same resource name. func (s) TestClusterTwoWatchSameResourceName(t *testing.T) { v2ClientCh, cleanup := overrideNewXDSV2Client() defer cleanup() c, err := New(clientOpts(testXDSServer)) if err != nil { t.Fatalf("failed to create client: %v", err) } defer c.Close() v2Client := <-v2ClientCh var clusterUpdateChs []*testutils.Channel const count = 2 var cancelLastWatch func() for i := 0; i < count; i++ { clusterUpdateCh := testutils.NewChannel() clusterUpdateChs = append(clusterUpdateChs, clusterUpdateCh) cancelLastWatch = c.WatchCluster(testCDSName, func(update ClusterUpdate, err error) { clusterUpdateCh.Send(clusterUpdateErr{u: update, err: err}) }) if _, err := v2Client.addWatches[cdsURL].Receive(); i == 0 && err != nil { t.Fatalf("want new watch to start, got error %v", err) } } wantUpdate := ClusterUpdate{ServiceName: testEDSName} v2Client.r.newCDSUpdate(map[string]ClusterUpdate{ testCDSName: wantUpdate, }) for i := 0; i < count; i++ { if u, err := clusterUpdateChs[i].Receive(); err != nil || u != (clusterUpdateErr{wantUpdate, nil}) { t.Errorf("i=%v, unexpected clusterUpdate: %v, error receiving from channel: %v", i, u, err) } } // Cancel the last watch, and send update again. cancelLastWatch() v2Client.r.newCDSUpdate(map[string]ClusterUpdate{ testCDSName: wantUpdate, }) for i := 0; i < count-1; i++ { if u, err := clusterUpdateChs[i].Receive(); err != nil || u != (clusterUpdateErr{wantUpdate, nil}) { t.Errorf("i=%v, unexpected clusterUpdate: %v, error receiving from channel: %v", i, u, err) } } if u, err := clusterUpdateChs[count-1].TimedReceive(chanRecvTimeout); err != testutils.ErrRecvTimeout { t.Errorf("unexpected clusterUpdate: %v, %v, want channel recv timeout", u, err) } } // TestClusterThreeWatchDifferentResourceName covers the case where an update is // received after three watch() for different resource names. func (s) TestClusterThreeWatchDifferentResourceName(t *testing.T) { v2ClientCh, cleanup := overrideNewXDSV2Client() defer cleanup() c, err := New(clientOpts(testXDSServer)) if err != nil { t.Fatalf("failed to create client: %v", err) } defer c.Close() v2Client := <-v2ClientCh var clusterUpdateChs []*testutils.Channel const count = 2 // Two watches for the same name. for i := 0; i < count; i++ { clusterUpdateCh := testutils.NewChannel() clusterUpdateChs = append(clusterUpdateChs, clusterUpdateCh) c.WatchCluster(testCDSName+"1", func(update ClusterUpdate, err error) { clusterUpdateCh.Send(clusterUpdateErr{u: update, err: err}) }) if _, err := v2Client.addWatches[cdsURL].Receive(); i == 0 && err != nil { t.Fatalf("want new watch to start, got error %v", err) } } // Third watch for a different name. clusterUpdateCh2 := testutils.NewChannel() c.WatchCluster(testCDSName+"2", func(update ClusterUpdate, err error) { clusterUpdateCh2.Send(clusterUpdateErr{u: update, err: err}) }) if _, err := v2Client.addWatches[cdsURL].Receive(); err != nil { t.Fatalf("want new watch to start, got error %v", err) } wantUpdate1 := ClusterUpdate{ServiceName: testEDSName + "1"} wantUpdate2 := ClusterUpdate{ServiceName: testEDSName + "2"} v2Client.r.newCDSUpdate(map[string]ClusterUpdate{ testCDSName + "1": wantUpdate1, testCDSName + "2": wantUpdate2, }) for i := 0; i < count; i++ { if u, err := clusterUpdateChs[i].Receive(); err != nil || u != (clusterUpdateErr{wantUpdate1, nil}) { t.Errorf("i=%v, unexpected clusterUpdate: %v, error receiving from channel: %v", i, u, err) } } if u, err := clusterUpdateCh2.Receive(); err != nil || u != (clusterUpdateErr{wantUpdate2, nil}) { t.Errorf("unexpected clusterUpdate: %v, error receiving from channel: %v", u, err) } } // TestClusterWatchAfterCache covers the case where watch is called after the update // is in cache. func (s) TestClusterWatchAfterCache(t *testing.T) { v2ClientCh, cleanup := overrideNewXDSV2Client() defer cleanup() c, err := New(clientOpts(testXDSServer)) if err != nil { t.Fatalf("failed to create client: %v", err) } defer c.Close() v2Client := <-v2ClientCh clusterUpdateCh := testutils.NewChannel() c.WatchCluster(testCDSName, func(update ClusterUpdate, err error) { clusterUpdateCh.Send(clusterUpdateErr{u: update, err: err}) }) if _, err := v2Client.addWatches[cdsURL].Receive(); err != nil { t.Fatalf("want new watch to start, got error %v", err) } wantUpdate := ClusterUpdate{ServiceName: testEDSName} v2Client.r.newCDSUpdate(map[string]ClusterUpdate{ testCDSName: wantUpdate, }) if u, err := clusterUpdateCh.Receive(); err != nil || u != (clusterUpdateErr{wantUpdate, nil}) { t.Errorf("unexpected clusterUpdate: %v, error receiving from channel: %v", u, err) } // Another watch for the resource in cache. clusterUpdateCh2 := testutils.NewChannel() c.WatchCluster(testCDSName, func(update ClusterUpdate, err error) { clusterUpdateCh2.Send(clusterUpdateErr{u: update, err: err}) }) if n, err := v2Client.addWatches[cdsURL].Receive(); err == nil { t.Fatalf("want no new watch to start (recv timeout), got resource name: %v error %v", n, err) } // New watch should receives the update. if u, err := clusterUpdateCh2.Receive(); err != nil || u != (clusterUpdateErr{wantUpdate, nil}) { t.Errorf("unexpected clusterUpdate: %v, error receiving from channel: %v", u, err) } // Old watch should see nothing. if u, err := clusterUpdateCh.TimedReceive(chanRecvTimeout); err != testutils.ErrRecvTimeout { t.Errorf("unexpected clusterUpdate: %v, %v, want channel recv timeout", u, err) } } // TestClusterWatchExpiryTimer tests the case where the client does not receive // an CDS response for the request that it sends out. We want the watch callback // to be invoked with an error once the watchExpiryTimer fires. func (s) TestClusterWatchExpiryTimer(t *testing.T) { oldWatchExpiryTimeout := defaultWatchExpiryTimeout defaultWatchExpiryTimeout = 500 * time.Millisecond defer func() { defaultWatchExpiryTimeout = oldWatchExpiryTimeout }() v2ClientCh, cleanup := overrideNewXDSV2Client() defer cleanup() c, err := New(clientOpts(testXDSServer)) if err != nil { t.Fatalf("failed to create client: %v", err) } defer c.Close() v2Client := <-v2ClientCh clusterUpdateCh := testutils.NewChannel() c.WatchCluster(testCDSName, func(u ClusterUpdate, err error) { clusterUpdateCh.Send(clusterUpdateErr{u: u, err: err}) }) if _, err := v2Client.addWatches[cdsURL].Receive(); err != nil { t.Fatalf("want new watch to start, got error %v", err) } u, err := clusterUpdateCh.TimedReceive(defaultWatchExpiryTimeout * 2) if err != nil { t.Fatalf("failed to get clusterUpdate: %v", err) } uu := u.(clusterUpdateErr) if uu.u != (ClusterUpdate{}) { t.Errorf("unexpected clusterUpdate: %v, want %v", uu.u, ClusterUpdate{}) } if uu.err == nil { t.Errorf("unexpected clusterError: , want error watcher timeout") } } // TestClusterWatchExpiryTimerStop tests the case where the client does receive // an CDS response for the request that it sends out. We want no error even // after expiry timeout. func (s) TestClusterWatchExpiryTimerStop(t *testing.T) { oldWatchExpiryTimeout := defaultWatchExpiryTimeout defaultWatchExpiryTimeout = 500 * time.Millisecond defer func() { defaultWatchExpiryTimeout = oldWatchExpiryTimeout }() v2ClientCh, cleanup := overrideNewXDSV2Client() defer cleanup() c, err := New(clientOpts(testXDSServer)) if err != nil { t.Fatalf("failed to create client: %v", err) } defer c.Close() v2Client := <-v2ClientCh clusterUpdateCh := testutils.NewChannel() c.WatchCluster(testCDSName, func(u ClusterUpdate, err error) { clusterUpdateCh.Send(clusterUpdateErr{u: u, err: err}) }) if _, err := v2Client.addWatches[cdsURL].Receive(); err != nil { t.Fatalf("want new watch to start, got error %v", err) } wantUpdate := ClusterUpdate{ServiceName: testEDSName} v2Client.r.newCDSUpdate(map[string]ClusterUpdate{ testCDSName: wantUpdate, }) if u, err := clusterUpdateCh.Receive(); err != nil || u != (clusterUpdateErr{wantUpdate, nil}) { t.Errorf("unexpected clusterUpdate: %v, error receiving from channel: %v", u, err) } // Wait for an error, the error should never happen. u, err := clusterUpdateCh.TimedReceive(defaultWatchExpiryTimeout * 2) if err != testutils.ErrRecvTimeout { t.Fatalf("got unexpected: %v, %v, want recv timeout", u.(clusterUpdateErr).u, u.(clusterUpdateErr).err) } } // TestClusterResourceRemoved covers the cases: // - an update is received after a watch() // - another update is received, with one resource removed // - this should trigger callback with resource removed error // - one more update without the removed resource // - the callback (above) shouldn't receive any update func (s) TestClusterResourceRemoved(t *testing.T) { v2ClientCh, cleanup := overrideNewXDSV2Client() defer cleanup() c, err := New(clientOpts(testXDSServer)) if err != nil { t.Fatalf("failed to create client: %v", err) } defer c.Close() v2Client := <-v2ClientCh clusterUpdateCh1 := testutils.NewChannel() c.WatchCluster(testCDSName+"1", func(update ClusterUpdate, err error) { clusterUpdateCh1.Send(clusterUpdateErr{u: update, err: err}) }) if _, err := v2Client.addWatches[cdsURL].Receive(); err != nil { t.Fatalf("want new watch to start, got error %v", err) } // Another watch for a different name. clusterUpdateCh2 := testutils.NewChannel() c.WatchCluster(testCDSName+"2", func(update ClusterUpdate, err error) { clusterUpdateCh2.Send(clusterUpdateErr{u: update, err: err}) }) if _, err := v2Client.addWatches[cdsURL].Receive(); err != nil { t.Fatalf("want new watch to start, got error %v", err) } wantUpdate1 := ClusterUpdate{ServiceName: testEDSName + "1"} wantUpdate2 := ClusterUpdate{ServiceName: testEDSName + "2"} v2Client.r.newCDSUpdate(map[string]ClusterUpdate{ testCDSName + "1": wantUpdate1, testCDSName + "2": wantUpdate2, }) if u, err := clusterUpdateCh1.Receive(); err != nil || u != (clusterUpdateErr{wantUpdate1, nil}) { t.Errorf("unexpected clusterUpdate: %v, error receiving from channel: %v", u, err) } if u, err := clusterUpdateCh2.Receive(); err != nil || u != (clusterUpdateErr{wantUpdate2, nil}) { t.Errorf("unexpected clusterUpdate: %v, error receiving from channel: %v", u, err) } // Send another update to remove resource 1. v2Client.r.newCDSUpdate(map[string]ClusterUpdate{ testCDSName + "2": wantUpdate2, }) // watcher 1 should get an error. if u, err := clusterUpdateCh1.Receive(); err != nil || ErrType(u.(clusterUpdateErr).err) != ErrorTypeResourceNotFound { t.Errorf("unexpected clusterUpdate: %v, error receiving from channel: %v, want update with error resource not found", u, err) } // watcher 2 should get the same update again. if u, err := clusterUpdateCh2.Receive(); err != nil || u != (clusterUpdateErr{wantUpdate2, nil}) { t.Errorf("unexpected clusterUpdate: %v, error receiving from channel: %v", u, err) } // Send one more update without resource 1. v2Client.r.newCDSUpdate(map[string]ClusterUpdate{ testCDSName + "2": wantUpdate2, }) // watcher 1 should get an error. if u, err := clusterUpdateCh1.Receive(); err != testutils.ErrRecvTimeout { t.Errorf("unexpected clusterUpdate: %v, want receiving from channel timeout", u) } // watcher 2 should get the same update again. if u, err := clusterUpdateCh2.Receive(); err != nil || u != (clusterUpdateErr{wantUpdate2, nil}) { t.Errorf("unexpected clusterUpdate: %v, error receiving from channel: %v", u, err) } }