mirror of
https://github.com/vmware-tanzu/pinniped.git
synced 2026-02-06 21:11:47 +00:00
Add initial controller boilerplate and example controller
Signed-off-by: Monis Khan <mok@vmware.com>
This commit is contained in:
170
test/integration/examplecontroller/controller/creating_test.go
Normal file
170
test/integration/examplecontroller/controller/creating_test.go
Normal file
@@ -0,0 +1,170 @@
|
||||
/*
|
||||
Copyright 2020 VMware, Inc.
|
||||
SPDX-License-Identifier: Apache-2.0
|
||||
*/
|
||||
|
||||
package controller
|
||||
|
||||
import (
|
||||
"context"
|
||||
"testing"
|
||||
"time"
|
||||
|
||||
"github.com/stretchr/testify/require"
|
||||
corev1 "k8s.io/api/core/v1"
|
||||
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
|
||||
"k8s.io/apimachinery/pkg/runtime/schema"
|
||||
"k8s.io/client-go/informers"
|
||||
"k8s.io/client-go/kubernetes/fake"
|
||||
coretesting "k8s.io/client-go/testing"
|
||||
"k8s.io/client-go/tools/events"
|
||||
|
||||
"github.com/suzerain-io/placeholder-name/internal/controller"
|
||||
"github.com/suzerain-io/placeholder-name/test/integration/examplecontroller/api"
|
||||
)
|
||||
|
||||
func TestNewExampleCreatingController(t *testing.T) {
|
||||
secretsGVR := schema.GroupVersionResource{Version: "v1", Resource: "secrets"}
|
||||
|
||||
type args struct {
|
||||
services []*corev1.Service
|
||||
secrets []*corev1.Secret
|
||||
secretData string
|
||||
}
|
||||
type keyErr struct {
|
||||
key controller.Key
|
||||
err error
|
||||
}
|
||||
tests := []struct {
|
||||
name string
|
||||
args args
|
||||
wantActions []coretesting.Action
|
||||
wantKeyErrs []keyErr
|
||||
}{
|
||||
{
|
||||
name: "service has annotation but secret does not exist",
|
||||
args: args{
|
||||
services: []*corev1.Service{
|
||||
{
|
||||
ObjectMeta: metav1.ObjectMeta{
|
||||
Namespace: "ns-1",
|
||||
Name: "service-1",
|
||||
Annotations: map[string]string{
|
||||
api.SecretNameAnnotation: "secret-1",
|
||||
},
|
||||
UID: "0001",
|
||||
},
|
||||
},
|
||||
},
|
||||
secretData: "foo-secret-1",
|
||||
},
|
||||
wantKeyErrs: []keyErr{
|
||||
{
|
||||
key: controller.Key{
|
||||
Namespace: "ns-1",
|
||||
Name: "service-1",
|
||||
},
|
||||
err: nil, // we expect no error with this key
|
||||
},
|
||||
},
|
||||
wantActions: []coretesting.Action{
|
||||
coretesting.NewCreateAction(secretsGVR, "ns-1", &corev1.Secret{
|
||||
ObjectMeta: metav1.ObjectMeta{
|
||||
Name: "secret-1",
|
||||
Namespace: "ns-1",
|
||||
Annotations: map[string]string{
|
||||
api.ServiceUIDAnnotation: "0001",
|
||||
api.ServiceNameAnnotation: "service-1",
|
||||
},
|
||||
OwnerReferences: []metav1.OwnerReference{
|
||||
{
|
||||
APIVersion: "v1",
|
||||
Kind: "Service",
|
||||
Name: "service-1",
|
||||
UID: "0001",
|
||||
},
|
||||
},
|
||||
},
|
||||
Type: corev1.SecretTypeOpaque,
|
||||
Data: map[string][]byte{
|
||||
api.SecretDataKey: []byte("foo-secret-1"),
|
||||
},
|
||||
}),
|
||||
},
|
||||
},
|
||||
}
|
||||
for _, tt := range tests {
|
||||
tt := tt
|
||||
|
||||
t.Run(tt.name, func(t *testing.T) {
|
||||
kubeClient := fake.NewSimpleClientset()
|
||||
for i := range tt.args.services {
|
||||
service := tt.args.services[i]
|
||||
err := kubeClient.Tracker().Add(service)
|
||||
require.NoError(t, err)
|
||||
}
|
||||
for i := range tt.args.secrets {
|
||||
secret := tt.args.secrets[i]
|
||||
err := kubeClient.Tracker().Add(secret)
|
||||
require.NoError(t, err)
|
||||
}
|
||||
|
||||
recorder := events.NewEventBroadcasterAdapter(kubeClient).NewRecorder("example-controller")
|
||||
kubeInformers := informers.NewSharedInformerFactory(kubeClient, 0)
|
||||
|
||||
creatingController := NewExampleCreatingController(
|
||||
kubeInformers.Core().V1().Services(),
|
||||
kubeInformers.Core().V1().Secrets(),
|
||||
kubeClient.CoreV1(),
|
||||
recorder,
|
||||
tt.args.secretData,
|
||||
)
|
||||
|
||||
keyErrs := make(chan keyErr)
|
||||
controller.TestWrap(t, creatingController, func(syncer controller.Syncer) controller.Syncer {
|
||||
return controller.SyncFunc(func(ctx controller.Context) error {
|
||||
err := syncer.Sync(ctx)
|
||||
|
||||
keyErrs <- keyErr{
|
||||
key: ctx.Key,
|
||||
err: err,
|
||||
}
|
||||
|
||||
return err
|
||||
})
|
||||
})
|
||||
|
||||
// a different approach would be to use TestSync and run each iteration manually:
|
||||
//
|
||||
// err := controller.TestSync(t, c, ...)
|
||||
|
||||
ctx, cancel := context.WithCancel(context.Background())
|
||||
defer cancel()
|
||||
|
||||
kubeInformers.Start(ctx.Done())
|
||||
go creatingController.Run(ctx, 5) // TODO maybe only use one worker?
|
||||
|
||||
var actualKeyErrs []keyErr
|
||||
done:
|
||||
for {
|
||||
select {
|
||||
case key := <-keyErrs:
|
||||
actualKeyErrs = append(actualKeyErrs, key)
|
||||
|
||||
case <-time.After(3 * time.Second):
|
||||
// this assumes that calls to Sync are never more than three seconds apart
|
||||
// we have five workers so there is little chance they all hang around doing nothing for that long
|
||||
break done
|
||||
}
|
||||
}
|
||||
|
||||
// TODO: Figure out how to capture actions from informers
|
||||
// TODO: I think we need some more fancy order independent equal comparison here
|
||||
|
||||
require.Equal(t, tt.wantKeyErrs, actualKeyErrs)
|
||||
|
||||
// ignore the discovery call from the event recorder and the list/watch from both informers (first five events)
|
||||
require.Equal(t, tt.wantActions, kubeClient.Actions()[5:])
|
||||
})
|
||||
}
|
||||
}
|
||||
Reference in New Issue
Block a user