mirror of
https://github.com/vmware-tanzu/velero.git
synced 2026-01-03 11:45:20 +00:00
Added tracking for deleted namespace status check in restore flow (#8233)
* Added tracking for deleted namespace status check in restore flow Signed-off-by: sangitaray2021 <sangitaray@microsoft.com> fixed unittest Signed-off-by: sangitaray2021 <sangitaray@microsoft.com> refactored tracker execution and caller Signed-off-by: sangitaray2021 <sangitaray@microsoft.com> added change log Signed-off-by: sangitaray2021 <sangitaray@microsoft.com> Author: sangitaray2021 <sangitaray@microsft.com> Author: sangitaray2021 <sangitaray@microsoft.com> Date: Thu Sep 19 02:26:14 2024 +0530 Signed-off-by: sangitaray2021 <sangitaray@microsoft.com> * fixed linter issuer Signed-off-by: sangitaray2021 <sangitaray@microsoft.com> * incorporated PR comments Signed-off-by: sangitaray2021 <sangitaray@microsoft.com> * resolved comments Signed-off-by: sangitaray2021 <sangitaray@microsoft.com> --------- Signed-off-by: sangitaray2021 <sangitaray@microsoft.com>
This commit is contained in:
@@ -58,6 +58,7 @@ import (
|
||||
uploadermocks "github.com/vmware-tanzu/velero/pkg/podvolume/mocks"
|
||||
"github.com/vmware-tanzu/velero/pkg/test"
|
||||
"github.com/vmware-tanzu/velero/pkg/types"
|
||||
"github.com/vmware-tanzu/velero/pkg/util/kube"
|
||||
kubeutil "github.com/vmware-tanzu/velero/pkg/util/kube"
|
||||
. "github.com/vmware-tanzu/velero/pkg/util/results"
|
||||
)
|
||||
@@ -2292,10 +2293,11 @@ func TestShouldRestore(t *testing.T) {
|
||||
h := newHarness(t)
|
||||
|
||||
ctx := &restoreContext{
|
||||
log: h.log,
|
||||
dynamicFactory: client.NewDynamicFactory(h.DynamicClient),
|
||||
namespaceClient: h.KubeClient.CoreV1().Namespaces(),
|
||||
resourceTerminatingTimeout: time.Millisecond,
|
||||
log: h.log,
|
||||
dynamicFactory: client.NewDynamicFactory(h.DynamicClient),
|
||||
namespaceClient: h.KubeClient.CoreV1().Namespaces(),
|
||||
resourceTerminatingTimeout: time.Millisecond,
|
||||
resourceDeletionStatusTracker: kube.NewResourceDeletionStatusTracker(),
|
||||
}
|
||||
|
||||
for _, resource := range tc.apiResources {
|
||||
@@ -3711,9 +3713,10 @@ func newHarness(t *testing.T) *harness {
|
||||
fileSystem: test.NewFakeFileSystem(),
|
||||
|
||||
// unsupported
|
||||
podVolumeRestorerFactory: nil,
|
||||
podVolumeTimeout: 0,
|
||||
kbClient: kbClient,
|
||||
podVolumeRestorerFactory: nil,
|
||||
podVolumeTimeout: 0,
|
||||
kbClient: kbClient,
|
||||
resourceDeletionStatusTracker: kube.NewResourceDeletionStatusTracker(),
|
||||
},
|
||||
log: log,
|
||||
}
|
||||
@@ -3900,9 +3903,10 @@ func TestIsAlreadyExistsError(t *testing.T) {
|
||||
h := newHarness(t)
|
||||
|
||||
ctx := &restoreContext{
|
||||
log: h.log,
|
||||
dynamicFactory: client.NewDynamicFactory(h.DynamicClient),
|
||||
namespaceClient: h.KubeClient.CoreV1().Namespaces(),
|
||||
log: h.log,
|
||||
dynamicFactory: client.NewDynamicFactory(h.DynamicClient),
|
||||
namespaceClient: h.KubeClient.CoreV1().Namespaces(),
|
||||
resourceDeletionStatusTracker: kube.NewResourceDeletionStatusTracker(),
|
||||
}
|
||||
|
||||
if test.apiResource != nil {
|
||||
@@ -4019,7 +4023,8 @@ func TestHasCSIVolumeSnapshot(t *testing.T) {
|
||||
h := newHarness(t)
|
||||
|
||||
ctx := &restoreContext{
|
||||
log: h.log,
|
||||
log: h.log,
|
||||
resourceDeletionStatusTracker: kube.NewResourceDeletionStatusTracker(),
|
||||
}
|
||||
|
||||
if tc.vs != nil {
|
||||
@@ -4119,9 +4124,10 @@ func TestHasSnapshotDataUpload(t *testing.T) {
|
||||
h := newHarness(t)
|
||||
|
||||
ctx := &restoreContext{
|
||||
log: h.log,
|
||||
kbClient: h.restorer.kbClient,
|
||||
restore: tc.restore,
|
||||
log: h.log,
|
||||
kbClient: h.restorer.kbClient,
|
||||
restore: tc.restore,
|
||||
resourceDeletionStatusTracker: kube.NewResourceDeletionStatusTracker(),
|
||||
}
|
||||
|
||||
if tc.duResult != nil {
|
||||
|
||||
Reference in New Issue
Block a user