fix: flaky tests

This commit is contained in:
Ludovic Fernandez 2022-12-07 10:56:05 +01:00 committed by GitHub
parent abd569701f
commit a8df674dcf
No known key found for this signature in database
GPG key ID: 4AEE18F83AFDEB23
3 changed files with 71 additions and 61 deletions

View file

@ -58,7 +58,7 @@ func (p *mockProvider) Init() error {
func TestNewConfigurationWatcher(t *testing.T) { func TestNewConfigurationWatcher(t *testing.T) {
routinesPool := safe.NewPool(context.Background()) routinesPool := safe.NewPool(context.Background())
defer routinesPool.Stop() t.Cleanup(routinesPool.Stop)
pvd := &mockProvider{ pvd := &mockProvider{
messages: []dynamic.Message{{ messages: []dynamic.Message{{
@ -115,7 +115,6 @@ func TestNewConfigurationWatcher(t *testing.T) {
func TestWaitForRequiredProvider(t *testing.T) { func TestWaitForRequiredProvider(t *testing.T) {
routinesPool := safe.NewPool(context.Background()) routinesPool := safe.NewPool(context.Background())
defer routinesPool.Stop()
pvdAggregator := &mockProvider{ pvdAggregator := &mockProvider{
wait: 5 * time.Millisecond, wait: 5 * time.Millisecond,
@ -151,7 +150,9 @@ func TestWaitForRequiredProvider(t *testing.T) {
}) })
watcher.Start() watcher.Start()
defer watcher.Stop()
t.Cleanup(watcher.Stop)
t.Cleanup(routinesPool.Stop)
// give some time so that the configuration can be processed // give some time so that the configuration can be processed
time.Sleep(20 * time.Millisecond) time.Sleep(20 * time.Millisecond)
@ -162,7 +163,6 @@ func TestWaitForRequiredProvider(t *testing.T) {
func TestIgnoreTransientConfiguration(t *testing.T) { func TestIgnoreTransientConfiguration(t *testing.T) {
routinesPool := safe.NewPool(context.Background()) routinesPool := safe.NewPool(context.Background())
defer routinesPool.Stop()
config := &dynamic.Configuration{ config := &dynamic.Configuration{
HTTP: th.BuildConfiguration( HTTP: th.BuildConfiguration(
@ -190,7 +190,9 @@ func TestIgnoreTransientConfiguration(t *testing.T) {
}) })
watcher.Start() watcher.Start()
defer watcher.Stop()
t.Cleanup(watcher.Stop)
t.Cleanup(routinesPool.Stop)
watcher.allProvidersConfigs <- dynamic.Message{ watcher.allProvidersConfigs <- dynamic.Message{
ProviderName: "mock", ProviderName: "mock",
@ -243,7 +245,6 @@ func TestIgnoreTransientConfiguration(t *testing.T) {
func TestListenProvidersThrottleProviderConfigReload(t *testing.T) { func TestListenProvidersThrottleProviderConfigReload(t *testing.T) {
routinesPool := safe.NewPool(context.Background()) routinesPool := safe.NewPool(context.Background())
defer routinesPool.Stop()
pvd := &mockProvider{ pvd := &mockProvider{
wait: 10 * time.Millisecond, wait: 10 * time.Millisecond,
@ -274,7 +275,9 @@ func TestListenProvidersThrottleProviderConfigReload(t *testing.T) {
}) })
watcher.Start() watcher.Start()
defer watcher.Stop()
t.Cleanup(watcher.Stop)
t.Cleanup(routinesPool.Stop)
// Give some time so that the configuration can be processed. // Give some time so that the configuration can be processed.
time.Sleep(100 * time.Millisecond) time.Sleep(100 * time.Millisecond)
@ -287,7 +290,6 @@ func TestListenProvidersThrottleProviderConfigReload(t *testing.T) {
func TestListenProvidersSkipsEmptyConfigs(t *testing.T) { func TestListenProvidersSkipsEmptyConfigs(t *testing.T) {
routinesPool := safe.NewPool(context.Background()) routinesPool := safe.NewPool(context.Background())
defer routinesPool.Stop()
pvd := &mockProvider{ pvd := &mockProvider{
messages: []dynamic.Message{{ProviderName: "mock"}}, messages: []dynamic.Message{{ProviderName: "mock"}},
@ -299,7 +301,9 @@ func TestListenProvidersSkipsEmptyConfigs(t *testing.T) {
}) })
watcher.Start() watcher.Start()
defer watcher.Stop()
t.Cleanup(watcher.Stop)
t.Cleanup(routinesPool.Stop)
// give some time so that the configuration can be processed // give some time so that the configuration can be processed
time.Sleep(100 * time.Millisecond) time.Sleep(100 * time.Millisecond)
@ -307,7 +311,6 @@ func TestListenProvidersSkipsEmptyConfigs(t *testing.T) {
func TestListenProvidersSkipsSameConfigurationForProvider(t *testing.T) { func TestListenProvidersSkipsSameConfigurationForProvider(t *testing.T) {
routinesPool := safe.NewPool(context.Background()) routinesPool := safe.NewPool(context.Background())
defer routinesPool.Stop()
message := dynamic.Message{ message := dynamic.Message{
ProviderName: "mock", ProviderName: "mock",
@ -331,7 +334,9 @@ func TestListenProvidersSkipsSameConfigurationForProvider(t *testing.T) {
}) })
watcher.Start() watcher.Start()
defer watcher.Stop()
t.Cleanup(watcher.Stop)
t.Cleanup(routinesPool.Stop)
// give some time so that the configuration can be processed // give some time so that the configuration can be processed
time.Sleep(100 * time.Millisecond) time.Sleep(100 * time.Millisecond)
@ -340,7 +345,6 @@ func TestListenProvidersSkipsSameConfigurationForProvider(t *testing.T) {
func TestListenProvidersDoesNotSkipFlappingConfiguration(t *testing.T) { func TestListenProvidersDoesNotSkipFlappingConfiguration(t *testing.T) {
routinesPool := safe.NewPool(context.Background()) routinesPool := safe.NewPool(context.Background())
defer routinesPool.Stop()
configuration := &dynamic.Configuration{ configuration := &dynamic.Configuration{
HTTP: th.BuildConfiguration( HTTP: th.BuildConfiguration(
@ -374,7 +378,9 @@ func TestListenProvidersDoesNotSkipFlappingConfiguration(t *testing.T) {
}) })
watcher.Start() watcher.Start()
defer watcher.Stop()
t.Cleanup(watcher.Stop)
t.Cleanup(routinesPool.Stop)
// give some time so that the configuration can be processed // give some time so that the configuration can be processed
time.Sleep(100 * time.Millisecond) time.Sleep(100 * time.Millisecond)
@ -407,7 +413,6 @@ func TestListenProvidersDoesNotSkipFlappingConfiguration(t *testing.T) {
func TestListenProvidersIgnoreSameConfig(t *testing.T) { func TestListenProvidersIgnoreSameConfig(t *testing.T) {
routinesPool := safe.NewPool(context.Background()) routinesPool := safe.NewPool(context.Background())
defer routinesPool.Stop()
configuration := &dynamic.Configuration{ configuration := &dynamic.Configuration{
HTTP: th.BuildConfiguration( HTTP: th.BuildConfiguration(
@ -453,8 +458,7 @@ func TestListenProvidersIgnoreSameConfig(t *testing.T) {
configurationReloads++ configurationReloads++
lastConfig = conf lastConfig = conf
// Allows next configurations to be sent by the mock provider // Allows next configurations to be sent by the mock provider as soon as the first configuration message is applied.
// as soon as the first configuration message is applied.
once.Do(func() { once.Do(func() {
pvd.first <- struct{}{} pvd.first <- struct{}{}
// Wait for all configuration messages to pile in // Wait for all configuration messages to pile in
@ -463,7 +467,9 @@ func TestListenProvidersIgnoreSameConfig(t *testing.T) {
}) })
watcher.Start() watcher.Start()
defer watcher.Stop()
t.Cleanup(watcher.Stop)
t.Cleanup(routinesPool.Stop)
// Wait long enough // Wait long enough
time.Sleep(50 * time.Millisecond) time.Sleep(50 * time.Millisecond)
@ -498,7 +504,6 @@ func TestListenProvidersIgnoreSameConfig(t *testing.T) {
func TestApplyConfigUnderStress(t *testing.T) { func TestApplyConfigUnderStress(t *testing.T) {
routinesPool := safe.NewPool(context.Background()) routinesPool := safe.NewPool(context.Background())
defer routinesPool.Stop()
watcher := NewConfigurationWatcher(routinesPool, &mockProvider{}, []string{"defaultEP"}, "") watcher := NewConfigurationWatcher(routinesPool, &mockProvider{}, []string{"defaultEP"}, "")
@ -525,15 +530,16 @@ func TestApplyConfigUnderStress(t *testing.T) {
}) })
watcher.Start() watcher.Start()
defer watcher.Stop()
t.Cleanup(watcher.Stop)
t.Cleanup(routinesPool.Stop)
time.Sleep(100 * time.Millisecond) time.Sleep(100 * time.Millisecond)
// Ensure that at least two configurations have been applied // Ensure that at least two configurations have been applied
// if we simulate being spammed configuration changes by the // if we simulate being spammed configuration changes by the provider(s).
// provider(s). // In theory, checking at least one would be sufficient,
// In theory, checking at least one would be sufficient, but // but checking for two also ensures that we're looping properly,
// checking for two also ensures that we're looping properly,
// and that the whole algo holds, etc. // and that the whole algo holds, etc.
t.Log(configurationReloads) t.Log(configurationReloads)
assert.GreaterOrEqual(t, configurationReloads, 2) assert.GreaterOrEqual(t, configurationReloads, 2)
@ -541,7 +547,6 @@ func TestApplyConfigUnderStress(t *testing.T) {
func TestListenProvidersIgnoreIntermediateConfigs(t *testing.T) { func TestListenProvidersIgnoreIntermediateConfigs(t *testing.T) {
routinesPool := safe.NewPool(context.Background()) routinesPool := safe.NewPool(context.Background())
defer routinesPool.Stop()
configuration := &dynamic.Configuration{ configuration := &dynamic.Configuration{
HTTP: th.BuildConfiguration( HTTP: th.BuildConfiguration(
@ -596,7 +601,9 @@ func TestListenProvidersIgnoreIntermediateConfigs(t *testing.T) {
}) })
watcher.Start() watcher.Start()
defer watcher.Stop()
t.Cleanup(watcher.Stop)
t.Cleanup(routinesPool.Stop)
// Wait long enough // Wait long enough
time.Sleep(500 * time.Millisecond) time.Sleep(500 * time.Millisecond)
@ -631,7 +638,6 @@ func TestListenProvidersIgnoreIntermediateConfigs(t *testing.T) {
func TestListenProvidersPublishesConfigForEachProvider(t *testing.T) { func TestListenProvidersPublishesConfigForEachProvider(t *testing.T) {
routinesPool := safe.NewPool(context.Background()) routinesPool := safe.NewPool(context.Background())
defer routinesPool.Stop()
configuration := &dynamic.Configuration{ configuration := &dynamic.Configuration{
HTTP: th.BuildConfiguration( HTTP: th.BuildConfiguration(
@ -656,7 +662,9 @@ func TestListenProvidersPublishesConfigForEachProvider(t *testing.T) {
}) })
watcher.Start() watcher.Start()
defer watcher.Stop()
t.Cleanup(watcher.Stop)
t.Cleanup(routinesPool.Stop)
// give some time so that the configuration can be processed // give some time so that the configuration can be processed
time.Sleep(100 * time.Millisecond) time.Sleep(100 * time.Millisecond)
@ -695,7 +703,6 @@ func TestListenProvidersPublishesConfigForEachProvider(t *testing.T) {
func TestPublishConfigUpdatedByProvider(t *testing.T) { func TestPublishConfigUpdatedByProvider(t *testing.T) {
routinesPool := safe.NewPool(context.Background()) routinesPool := safe.NewPool(context.Background())
defer routinesPool.Stop()
pvdConfiguration := dynamic.Configuration{ pvdConfiguration := dynamic.Configuration{
TCP: &dynamic.TCPConfiguration{ TCP: &dynamic.TCPConfiguration{
@ -725,12 +732,14 @@ func TestPublishConfigUpdatedByProvider(t *testing.T) {
watcher.AddListener(func(configuration dynamic.Configuration) { watcher.AddListener(func(configuration dynamic.Configuration) {
publishedConfigCount++ publishedConfigCount++
// Update the provider configuration published in next dynamic Message which should trigger a new publish. // Update the provider configuration published in next dynamic Message which should trigger a new publishing.
pvdConfiguration.TCP.Routers["bar"] = &dynamic.TCPRouter{} pvdConfiguration.TCP.Routers["bar"] = &dynamic.TCPRouter{}
}) })
watcher.Start() watcher.Start()
defer watcher.Stop()
t.Cleanup(watcher.Stop)
t.Cleanup(routinesPool.Stop)
// give some time so that the configuration can be processed. // give some time so that the configuration can be processed.
time.Sleep(100 * time.Millisecond) time.Sleep(100 * time.Millisecond)
@ -740,7 +749,6 @@ func TestPublishConfigUpdatedByProvider(t *testing.T) {
func TestPublishConfigUpdatedByConfigWatcherListener(t *testing.T) { func TestPublishConfigUpdatedByConfigWatcherListener(t *testing.T) {
routinesPool := safe.NewPool(context.Background()) routinesPool := safe.NewPool(context.Background())
defer routinesPool.Stop()
pvd := &mockProvider{ pvd := &mockProvider{
wait: 10 * time.Millisecond, wait: 10 * time.Millisecond,
@ -774,13 +782,15 @@ func TestPublishConfigUpdatedByConfigWatcherListener(t *testing.T) {
watcher.AddListener(func(configuration dynamic.Configuration) { watcher.AddListener(func(configuration dynamic.Configuration) {
publishedConfigCount++ publishedConfigCount++
// Modify the provided configuration. This should not modify the configuration stored in the configuration // Modify the provided configuration.
// watcher and cause a new publish. // This should not modify the configuration stored in the configuration watcher and therefore there will be no new publishing.
configuration.TCP.Routers["foo@mock"].Rule = "bar" configuration.TCP.Routers["foo@mock"].Rule = "bar"
}) })
watcher.Start() watcher.Start()
defer watcher.Stop()
t.Cleanup(watcher.Stop)
t.Cleanup(routinesPool.Stop)
// give some time so that the configuration can be processed. // give some time so that the configuration can be processed.
time.Sleep(100 * time.Millisecond) time.Sleep(100 * time.Millisecond)

View file

@ -48,18 +48,13 @@ func TestShutdownTCP(t *testing.T) {
require.NoError(t, err) require.NoError(t, err)
err = router.AddRoute("HostSNI(`*`)", 0, tcp.HandlerFunc(func(conn tcp.WriteCloser) { err = router.AddRoute("HostSNI(`*`)", 0, tcp.HandlerFunc(func(conn tcp.WriteCloser) {
for {
_, err := http.ReadRequest(bufio.NewReader(conn)) _, err := http.ReadRequest(bufio.NewReader(conn))
if err != nil {
if errors.Is(err, io.EOF) || (err != nil && errors.Is(err, net.ErrClosed)) {
return return
} }
require.NoError(t, err)
resp := http.Response{StatusCode: http.StatusOK} resp := http.Response{StatusCode: http.StatusOK}
err = resp.Write(conn) _ = resp.Write(conn)
require.NoError(t, err)
}
})) }))
require.NoError(t, err) require.NoError(t, err)
@ -89,6 +84,7 @@ func testShutdown(t *testing.T, router *tcprouter.Router) {
conn, err := startEntrypoint(entryPoint, router) conn, err := startEntrypoint(entryPoint, router)
require.NoError(t, err) require.NoError(t, err)
t.Cleanup(func() { _ = conn.Close() })
epAddr := entryPoint.listener.Addr().String() epAddr := entryPoint.listener.Addr().String()
@ -97,14 +93,14 @@ func testShutdown(t *testing.T, router *tcprouter.Router) {
time.Sleep(100 * time.Millisecond) time.Sleep(100 * time.Millisecond)
// We need to do a write on the conn before the shutdown to make it "exist". // We need to do a write on conn before the shutdown to make it "exist".
// Because the connection indeed exists as far as TCP is concerned, // Because the connection indeed exists as far as TCP is concerned,
// but since we only pass it along to the HTTP server after at least one byte is peeked, // but since we only pass it along to the HTTP server after at least one byte is peeked,
// the HTTP server (and hence its shutdown) does not know about the connection until that first byte peeked. // the HTTP server (and hence its shutdown) does not know about the connection until that first byte peeked.
err = request.Write(conn) err = request.Write(conn)
require.NoError(t, err) require.NoError(t, err)
reader := bufio.NewReader(conn) reader := bufio.NewReaderSize(conn, 1)
// Wait for first byte in response. // Wait for first byte in response.
_, err = reader.Peek(1) _, err = reader.Peek(1)
require.NoError(t, err) require.NoError(t, err)

View file

@ -32,16 +32,19 @@ func TestShutdownUDPConn(t *testing.T) {
for { for {
b := make([]byte, 1024*1024) b := make([]byte, 1024*1024)
n, err := conn.Read(b) n, err := conn.Read(b)
require.NoError(t, err) if err != nil {
// We control the termination, otherwise we would block on the Read above, until return
// conn is closed by a timeout. Which means we would get an error, and even though }
// we are in a goroutine and the current test might be over, go test would still
// yell at us if this happens while other tests are still running. // We control the termination, otherwise we would block on the Read above,
// until conn is closed by a timeout.
// Which means we would get an error,
// and even though we are in a goroutine and the current test might be over,
// go test would still yell at us if this happens while other tests are still running.
if string(b[:n]) == "CLOSE" { if string(b[:n]) == "CLOSE" {
return return
} }
_, err = conn.Write(b[:n]) _, _ = conn.Write(b[:n])
require.NoError(t, err)
} }
})) }))
@ -68,9 +71,9 @@ func TestShutdownUDPConn(t *testing.T) {
// Packet is accepted, but dropped // Packet is accepted, but dropped
require.NoError(t, err) require.NoError(t, err)
// Make sure that our session is yet again still live. This is specifically to // Make sure that our session is yet again still live.
// make sure we don't create a regression in listener's readLoop, i.e. that we only // This is specifically to make sure we don't create a regression in listener's readLoop,
// terminate the listener's readLoop goroutine by closing its pConn. // i.e. that we only terminate the listener's readLoop goroutine by closing its pConn.
requireEcho(t, "TEST3", conn, time.Second) requireEcho(t, "TEST3", conn, time.Second)
done := make(chan bool) done := make(chan bool)
@ -101,10 +104,11 @@ func TestShutdownUDPConn(t *testing.T) {
} }
} }
// requireEcho tests that the conn session is live and functional, by writing // requireEcho tests that conn session is live and functional,
// data through it, and expecting the same data as a response when reading on it. // by writing data through it,
// It fatals if the read blocks longer than timeout, which is useful to detect // and expecting the same data as a response when reading on it.
// regressions that would make a test wait forever. // It fatals if the read blocks longer than timeout,
// which is useful to detect regressions that would make a test wait forever.
func requireEcho(t *testing.T, data string, conn io.ReadWriter, timeout time.Duration) { func requireEcho(t *testing.T, data string, conn io.ReadWriter, timeout time.Duration) {
t.Helper() t.Helper()