mirror of
https://github.com/ceph/ceph-csi.git
synced 2024-12-23 13:30:23 +00:00
68588dc7df
Unit-testing often fails due to a race condition while writing the clusterMappingConfigFile from multiple go-routines at the same time. Failures from `make containerized-test` look like this: === CONT TestGetClusterMappingInfo/site2-storage_cluster-id_mapping cluster_mapping_test.go:153: GetClusterMappingInfo() = <nil>, expected data &[{map[site1-storage:site2-storage] [map[1:3]] [map[11:5]]} {map[site3-storage:site2-storage] [map[8:3]] [map[10:5]]}] === CONT TestGetClusterMappingInfo/site3-storage_cluster-id_mapping cluster_mapping_test.go:153: GetClusterMappingInfo() = <nil>, expected data &[{map[site3-storage:site2-storage] [map[8:3]] [map[10:5]]}] --- FAIL: TestGetClusterMappingInfo (0.01s) --- PASS: TestGetClusterMappingInfo/mapping_file_not_found (0.00s) --- PASS: TestGetClusterMappingInfo/mapping_file_found_with_empty_data (0.00s) --- PASS: TestGetClusterMappingInfo/cluster-id_mapping_not_found (0.00s) --- FAIL: TestGetClusterMappingInfo/site2-storage_cluster-id_mapping (0.00s) --- FAIL: TestGetClusterMappingInfo/site3-storage_cluster-id_mapping (0.00s) --- PASS: TestGetClusterMappingInfo/site1-storage_cluster-id_mapping (0.00s) By splitting the public GetClusterMappingInfo() function into an internal getClusterMappingInfo() that takes a filename, unit-testing can use different files for each go-routine, and testing becomes more predictable. Signed-off-by: Niels de Vos <ndevos@redhat.com>
242 lines
7.7 KiB
Go
242 lines
7.7 KiB
Go
/*
|
|
Copyright 2021 The Ceph-CSI Authors.
|
|
|
|
Licensed under the Apache License, Version 2.0 (the "License");
|
|
you may not use this file except in compliance with the License.
|
|
You may obtain a copy of the License at
|
|
|
|
http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
Unless required by applicable law or agreed to in writing, software
|
|
distributed under the License is distributed on an "AS IS" BASIS,
|
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
See the License for the specific language governing permissions and
|
|
limitations under the License.
|
|
*/
|
|
|
|
package util
|
|
|
|
import (
|
|
"encoding/json"
|
|
"fmt"
|
|
"io/ioutil"
|
|
"reflect"
|
|
"testing"
|
|
)
|
|
|
|
func TestGetClusterMappingInfo(t *testing.T) {
|
|
t.Parallel()
|
|
mappingBasePath := t.TempDir()
|
|
|
|
// clusterID,poolID on site1
|
|
clusterIDOfSite1 := "site1-storage"
|
|
rbdPoolIDOfSite1 := "1"
|
|
cephfsFscIDOfSite1 := "11"
|
|
|
|
// clusterID,poolID on site2
|
|
clusterIDOfSite2 := "site2-storage"
|
|
rbdPoolIDOfSite2 := "3"
|
|
cephfsFscIDOfSite2 := "5"
|
|
|
|
// clusterID,poolID on site3
|
|
clusterIDOfSite3 := "site3-storage"
|
|
rbdPoolIDOfSite3 := "8"
|
|
cephfsFscIDOfSite3 := "10"
|
|
|
|
clusterMappingInfos := make([]ClusterMappingInfo, 2)
|
|
|
|
// create mapping between site1 and site2
|
|
clusterIDmappingOfSite1To2 := make(map[string]string)
|
|
clusterIDmappingOfSite1To2[clusterIDOfSite1] = clusterIDOfSite2
|
|
rbdMappingOfSite1To2 := make([]map[string]string, 1)
|
|
rbdMappingOfSite1To2[0] = map[string]string{rbdPoolIDOfSite1: rbdPoolIDOfSite2}
|
|
cephFSMappingOfSite1To2 := make([]map[string]string, 1)
|
|
cephFSMappingOfSite1To2[0] = map[string]string{cephfsFscIDOfSite1: cephfsFscIDOfSite2}
|
|
mappingOfSite1To2 := ClusterMappingInfo{
|
|
clusterIDmappingOfSite1To2,
|
|
rbdMappingOfSite1To2,
|
|
cephFSMappingOfSite1To2,
|
|
}
|
|
|
|
// create mapping between site3 and site2
|
|
clusterIDmappingOfSite3To2 := make(map[string]string)
|
|
clusterIDmappingOfSite3To2[clusterIDOfSite3] = clusterIDOfSite2
|
|
rbdMappingOfSite3To2 := make([]map[string]string, 1)
|
|
rbdMappingOfSite3To2[0] = map[string]string{rbdPoolIDOfSite3: rbdPoolIDOfSite2}
|
|
cephFSMappingOfSite3To2 := make([]map[string]string, 1)
|
|
cephFSMappingOfSite3To2[0] = map[string]string{cephfsFscIDOfSite3: cephfsFscIDOfSite2}
|
|
mappingOfSite3To2 := ClusterMappingInfo{
|
|
clusterIDmappingOfSite3To2,
|
|
rbdMappingOfSite3To2,
|
|
cephFSMappingOfSite3To2,
|
|
}
|
|
|
|
clusterMappingInfos[0] = mappingOfSite1To2
|
|
clusterMappingInfos[1] = mappingOfSite3To2
|
|
|
|
mappingFileContent, err := json.Marshal(clusterMappingInfos)
|
|
if err != nil {
|
|
t.Errorf("failed to marshal mapping info %v", err)
|
|
}
|
|
// expected output of mapping
|
|
expectedSite2Data := clusterMappingInfos
|
|
expectedSite1To2Data := clusterMappingInfos[:1]
|
|
expectedSite3To2Data := clusterMappingInfos[1:]
|
|
|
|
tests := []struct {
|
|
name string
|
|
clusterID string
|
|
mappingFilecontent []byte
|
|
expectedData *[]ClusterMappingInfo
|
|
expectErr bool
|
|
}{
|
|
{
|
|
name: "mapping file not found",
|
|
clusterID: "site-a-clusterid",
|
|
mappingFilecontent: []byte{},
|
|
expectedData: nil,
|
|
expectErr: false,
|
|
},
|
|
{
|
|
name: "mapping file found with empty data",
|
|
clusterID: "site-a-clusterid",
|
|
mappingFilecontent: []byte{},
|
|
expectedData: nil,
|
|
expectErr: false,
|
|
},
|
|
{
|
|
name: "cluster-id mapping not found",
|
|
clusterID: "site-a-clusterid",
|
|
mappingFilecontent: mappingFileContent,
|
|
expectedData: nil,
|
|
expectErr: false,
|
|
},
|
|
{
|
|
name: "site2-storage cluster-id mapping",
|
|
clusterID: clusterIDOfSite2,
|
|
mappingFilecontent: mappingFileContent,
|
|
expectedData: &expectedSite2Data,
|
|
expectErr: false,
|
|
},
|
|
{
|
|
name: "site1-storage cluster-id mapping",
|
|
clusterID: clusterIDOfSite1,
|
|
mappingFilecontent: mappingFileContent,
|
|
expectedData: &expectedSite1To2Data,
|
|
expectErr: false,
|
|
},
|
|
{
|
|
name: "site3-storage cluster-id mapping",
|
|
clusterID: clusterIDOfSite3,
|
|
mappingFilecontent: mappingFileContent,
|
|
expectedData: &expectedSite3To2Data,
|
|
expectErr: false,
|
|
},
|
|
}
|
|
for i, tt := range tests {
|
|
currentI := i
|
|
currentTT := tt
|
|
t.Run(currentTT.name, func(t *testing.T) {
|
|
t.Parallel()
|
|
mappingConfigFile := fmt.Sprintf("%s/mapping-%d.json", mappingBasePath, currentI)
|
|
if len(currentTT.mappingFilecontent) != 0 {
|
|
err = ioutil.WriteFile(mappingConfigFile, currentTT.mappingFilecontent, 0o600)
|
|
if err != nil {
|
|
t.Errorf("failed to write to %q, error = %v", mappingConfigFile, err)
|
|
}
|
|
}
|
|
data, mErr := getClusterMappingInfo(currentTT.clusterID, mappingConfigFile)
|
|
if (mErr != nil) != currentTT.expectErr {
|
|
t.Errorf("getClusterMappingInfo() error = %v, expected Error %v", mErr, currentTT.expectErr)
|
|
}
|
|
if !reflect.DeepEqual(data, currentTT.expectedData) {
|
|
t.Errorf("getClusterMappingInfo() = %v, expected data %v", data, currentTT.expectedData)
|
|
}
|
|
})
|
|
}
|
|
|
|
clusterMappingConfigFile = fmt.Sprintf("%s/mapping.json", mappingBasePath)
|
|
err = ioutil.WriteFile(clusterMappingConfigFile, mappingFileContent, 0o600)
|
|
if err != nil {
|
|
t.Errorf("failed to write mapping content error = %v", err)
|
|
}
|
|
|
|
// validate site-3 to site-2 and site-1 to site-2 mappings when failover to
|
|
// site-2.
|
|
// The volumeId's from site-1 looks like `0001-0013-site1-storage-xyz` we
|
|
// need to have validate `site1-storage` to `site2-storage` mapping exists
|
|
// The volumeId's from site-3 looks like `0001-0013-site3-storage-xyz` we
|
|
// need to have validate `site3-storage` to `site2-storage` mapping exists
|
|
mappedClusterCount := 2
|
|
err = validateMapping(t, clusterIDOfSite2, rbdPoolIDOfSite2, cephfsFscIDOfSite2, mappedClusterCount)
|
|
if err != nil {
|
|
t.Error(err)
|
|
}
|
|
// validate site-2 and site-1 mappings when failback to site-1.
|
|
// The volumeId's from site-2 looks like `0001-0013-site2-storage-xyz` we
|
|
// need to have validate `site2-storage` to `site3-storage` mapping exists
|
|
mappedClusterCount = 1
|
|
err = validateMapping(t, clusterIDOfSite1, rbdPoolIDOfSite1, cephfsFscIDOfSite1, mappedClusterCount)
|
|
if err != nil {
|
|
t.Error(err)
|
|
}
|
|
// validate site-2 and site-3 mappings when failback to site-3
|
|
// The volumeId's from site-2 looks like `0001-0013-site2-storage-xyz` we
|
|
// need to have validate `site2-storage` to `site3-storage` mapping exists
|
|
mappedClusterCount = 1
|
|
err = validateMapping(t, clusterIDOfSite3, rbdPoolIDOfSite3, cephfsFscIDOfSite3, mappedClusterCount)
|
|
if err != nil {
|
|
t.Error(err)
|
|
}
|
|
}
|
|
|
|
func validateMapping(t *testing.T, clusterID, rbdPoolID, cephFSPoolID string, mappingCount int) error {
|
|
t.Helper()
|
|
|
|
mapping, err := GetClusterMappingInfo(clusterID)
|
|
if err != nil {
|
|
return fmt.Errorf("failed to retrieve mapping %w", err)
|
|
}
|
|
// verify we are able to retrieve both site-1:site2 and site-2:site3 mapping
|
|
if mapping == nil || len(*mapping) != mappingCount {
|
|
return fmt.Errorf(
|
|
"clusterID mapping got length=%v, expected length=%v",
|
|
len(*mapping),
|
|
mappingCount)
|
|
}
|
|
// check mapping rbd pool mapping exists in mapping
|
|
foundRBDPoolMappingCount := 0
|
|
foundCephFSPoolMappingCount := 0
|
|
for _, c := range *mapping {
|
|
for _, rp := range c.RBDpoolIDMappingInfo {
|
|
for k, v := range rp {
|
|
if k == rbdPoolID || v == rbdPoolID {
|
|
foundRBDPoolMappingCount++
|
|
}
|
|
}
|
|
}
|
|
|
|
for _, cp := range c.CephFSFscIDMappingInfo {
|
|
for k, v := range cp {
|
|
if k == cephFSPoolID || v == cephFSPoolID {
|
|
foundCephFSPoolMappingCount++
|
|
}
|
|
}
|
|
}
|
|
}
|
|
if foundRBDPoolMappingCount != mappingCount {
|
|
return fmt.Errorf(
|
|
"rbd pool mapping got length= %v, expected length=%v",
|
|
foundRBDPoolMappingCount,
|
|
mappingCount)
|
|
}
|
|
if foundCephFSPoolMappingCount != mappingCount {
|
|
return fmt.Errorf(
|
|
"cephFS filesystem mapping got length= %v, expected length=%v",
|
|
foundCephFSPoolMappingCount,
|
|
mappingCount)
|
|
}
|
|
|
|
return nil
|
|
}
|