1
0
Fork 0
mirror of https://github.com/arangodb/kube-arangodb.git synced 2024-12-14 11:57:37 +00:00
kube-arangodb/pkg/storage/pv_creator.go

386 lines
13 KiB
Go

//
// DISCLAIMER
//
// Copyright 2016-2023 ArangoDB GmbH, Cologne, Germany
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.
//
// Copyright holder is ArangoDB GmbH, Cologne, Germany
//
package storage
import (
"context"
"crypto/sha1"
"fmt"
"net"
"path/filepath"
"sort"
"strconv"
"strings"
"time"
"github.com/dchest/uniuri"
core "k8s.io/api/core/v1"
storage "k8s.io/api/storage/v1"
"k8s.io/apimachinery/pkg/api/resource"
meta "k8s.io/apimachinery/pkg/apis/meta/v1"
"github.com/arangodb/kube-arangodb/pkg/apis/shared"
api "github.com/arangodb/kube-arangodb/pkg/apis/storage/v1alpha"
"github.com/arangodb/kube-arangodb/pkg/deployment/features"
"github.com/arangodb/kube-arangodb/pkg/storage/provisioner"
resources "github.com/arangodb/kube-arangodb/pkg/storage/resources"
"github.com/arangodb/kube-arangodb/pkg/util"
"github.com/arangodb/kube-arangodb/pkg/util/constants"
"github.com/arangodb/kube-arangodb/pkg/util/errors"
"github.com/arangodb/kube-arangodb/pkg/util/k8sutil"
"github.com/arangodb/kube-arangodb/pkg/util/k8sutil/kerrors"
)
const (
defaultVolumeSize = int64(8 * 1024 * 1024 * 1024) // 8GB
// AnnProvisionedBy is the external provisioner annotation in PV object
AnnProvisionedBy = "pv.kubernetes.io/provisioned-by"
)
var (
// name of the annotation containing the node name
nodeNameAnnotation = api.SchemeGroupVersion.Group + "/node-name"
)
// createPVs creates a given number of PersistentVolume's.
func (ls *LocalStorage) createPVs(ctx context.Context, apiObject *api.ArangoLocalStorage, unboundClaims []core.PersistentVolumeClaim) (bool, error) {
// Fetch StorageClass name
var bm = storage.VolumeBindingImmediate
var reclaimPolicy = core.PersistentVolumeReclaimRetain
if sc, err := ls.deps.Client.Kubernetes().StorageV1().StorageClasses().Get(ctx, ls.apiObject.Spec.StorageClass.Name, meta.GetOptions{}); err == nil {
// We are able to fetch storageClass
if b := sc.VolumeBindingMode; b != nil {
bm = *b
}
if c := sc.ReclaimPolicy; c != nil {
reclaimPolicy = *c
}
}
// Find provisioner clients
clients, err := ls.createProvisionerClients(ctx)
if err != nil {
return false, errors.WithStack(err)
}
if len(clients) == 0 {
// No provisioners available
return false, errors.WithStack(errors.Newf("No ready provisioner endpoints found"))
}
for i, claim := range unboundClaims {
// Find deployment name & role in the claim (if any)
deplName, role, enforceAniAffinity := getDeploymentInfo(claim)
allowedClients := clients
if deplName != "" {
// Select nodes to choose from such that no volume in group lands on the same node
var err error
allowedClients, err = ls.filterAllowedNodes(clients, deplName, role)
if err != nil {
ls.log.Err(err).Warn("Failed to filter allowed nodes")
continue // We'll try this claim again later
}
if !enforceAniAffinity && len(allowedClients) == 0 {
// No possible nodes found that have no other volume (in same group) on it.
// We don't have to enforce separate nodes, so use all clients.
allowedClients = clients
}
}
// Find size of PVC
volSize := defaultVolumeSize
if reqStorage, ok := claim.Spec.Resources.Requests[core.ResourceStorage]; ok {
if v, ok := reqStorage.AsInt64(); ok && v > 0 {
volSize = v
}
}
if bm == storage.VolumeBindingWaitForFirstConsumer {
podList, err := resources.ListPods(ctx, ls.deps.Client.Kubernetes().CoreV1().Pods(claim.GetNamespace()))
if err != nil {
ls.log.Err(err).Warn("Unable to list pods")
continue
}
podList = podList.FilterByPVCName(claim.GetName())
nodeList, err := resources.ListNodes(ctx, ls.deps.Client.Kubernetes().CoreV1().Nodes())
if err != nil {
ls.log.Err(err).Warn("Unable to list nodes")
continue
}
nodeList = nodeList.FilterSchedulable().FilterPodsTaints(podList).SortBySchedulablePodsTaints(podList)
allowedClients = allowedClients.Filter(func(node string, client provisioner.API) bool {
for _, n := range nodeList {
if n.GetName() == node {
return true
}
}
return false
})
}
if len(allowedClients) == 0 {
ls.log.Info("PVC Cannot be created on any node")
continue
}
// Create PV
if err := ls.createPV(ctx, apiObject, allowedClients, i, volSize, claim, reclaimPolicy, deplName, role); err != nil {
ls.log.Err(err).Error("Failed to create PersistentVolume")
}
return true, nil
}
return false, nil
}
// createPV creates a PersistentVolume.
func (ls *LocalStorage) createPV(ctx context.Context, apiObject *api.ArangoLocalStorage, clients Clients, clientsOffset int, volSize int64, claim core.PersistentVolumeClaim, storageClassReclaimPolicy core.PersistentVolumeReclaimPolicy, deploymentName, role string) error {
// Try clients
keys := clients.Keys()
for clientIdx := 0; clientIdx < len(keys); clientIdx++ {
client := clients[keys[(clientsOffset+clientIdx)%len(keys)]]
// Try local path within client
for _, localPathRoot := range apiObject.Spec.LocalPath {
log := ls.log.Str("local-path-root", localPathRoot)
info, err := client.GetInfo(ctx, localPathRoot)
if err != nil {
log.Err(err).Error("Failed to get client info")
continue
}
if info.Available < volSize {
ls.log.Error("Not enough available size")
continue
}
// Ok, prepare a directory
name := strings.ToLower(uniuri.New())
localPath := filepath.Join(localPathRoot, name)
log = ls.log.Str("local-path", localPath)
if err := client.Prepare(ctx, localPath); err != nil {
log.Err(err).Error("Failed to prepare local path")
continue
}
reclaimPolicy := core.PersistentVolumeReclaimRetain
if features.LocalStorageReclaimPolicyPass().Enabled() {
reclaimPolicy = storageClassReclaimPolicy
}
// Create a volume
pvName := strings.ToLower(apiObject.GetName() + "-" + shortHash(info.NodeName) + "-" + name)
volumeMode := core.PersistentVolumeFilesystem
nodeSel := createNodeSelector(info.NodeName)
pv := &core.PersistentVolume{
ObjectMeta: meta.ObjectMeta{
Name: pvName,
Annotations: map[string]string{
AnnProvisionedBy: storageClassProvisioner,
nodeNameAnnotation: info.NodeName,
},
Labels: map[string]string{
k8sutil.LabelKeyArangoDeployment: deploymentName,
k8sutil.LabelKeyRole: role,
},
Finalizers: util.BoolSwitch(features.LocalStorageReclaimPolicyPass().Enabled(), []string{
FinalizerPersistentVolumeCleanup,
}, nil),
},
Spec: core.PersistentVolumeSpec{
Capacity: core.ResourceList{
core.ResourceStorage: *resource.NewQuantity(volSize, resource.BinarySI),
},
PersistentVolumeReclaimPolicy: reclaimPolicy,
PersistentVolumeSource: core.PersistentVolumeSource{
Local: &core.LocalVolumeSource{
Path: localPath,
},
},
AccessModes: []core.PersistentVolumeAccessMode{
core.ReadWriteOnce,
},
StorageClassName: apiObject.Spec.StorageClass.Name,
VolumeMode: &volumeMode,
ClaimRef: &core.ObjectReference{
Kind: "PersistentVolumeClaim",
APIVersion: "",
Name: claim.GetName(),
Namespace: claim.GetNamespace(),
UID: claim.GetUID(),
},
NodeAffinity: &core.VolumeNodeAffinity{
Required: nodeSel,
},
},
}
// Attach PV to ArangoLocalStorage
pv.SetOwnerReferences(append(pv.GetOwnerReferences(), apiObject.AsOwner()))
if _, err := ls.deps.Client.Kubernetes().CoreV1().PersistentVolumes().Create(context.Background(), pv, meta.CreateOptions{}); err != nil {
log.Err(err).Error("Failed to create PersistentVolume")
continue
}
log.
Str("name", pvName).
Str("node-name", info.NodeName).
Debug("Created PersistentVolume")
// Bind claim to volume
if err := ls.bindClaimToVolume(claim, pv.GetName()); err != nil {
// Try to delete the PV now
if features.LocalStorageReclaimPolicyPass().Enabled() {
ls.removePVObjectWithLog(pv)
} else {
if err := ls.deps.Client.Kubernetes().CoreV1().PersistentVolumes().Delete(context.Background(), pv.GetName(), meta.DeleteOptions{}); err != nil {
log.Err(err).Error("Failed to delete PV after binding PVC failed")
}
}
return errors.WithStack(err)
}
return nil
}
}
return errors.WithStack(errors.Newf("No more nodes available"))
}
// createValidEndpointList convers the given endpoints list into
// valid addresses.
func createValidEndpointList(list *core.EndpointsList) []string {
result := make([]string, 0, len(list.Items))
for _, ep := range list.Items {
for _, subset := range ep.Subsets {
for _, ip := range subset.Addresses {
addr := net.JoinHostPort(ip.IP, strconv.Itoa(provisioner.DefaultPort))
result = append(result, addr)
}
}
}
sort.Strings(result)
return result
}
// createNodeAffinity creates a node affinity serialized to string.
func createNodeSelector(nodeName string) *core.NodeSelector {
return &core.NodeSelector{
NodeSelectorTerms: []core.NodeSelectorTerm{
core.NodeSelectorTerm{
MatchExpressions: []core.NodeSelectorRequirement{
core.NodeSelectorRequirement{
Key: shared.TopologyKeyHostname,
Operator: core.NodeSelectorOpIn,
Values: []string{nodeName},
},
},
},
},
}
}
// getDeploymentInfo returns the name of the deployment that created the given claim,
// the role of the server that the claim is used for and the value for `enforceAntiAffinity`.
// If not found, empty strings are returned.
// Returns deploymentName, role, enforceAntiAffinity.
func getDeploymentInfo(pvc core.PersistentVolumeClaim) (string, string, bool) {
deploymentName := pvc.GetLabels()[k8sutil.LabelKeyArangoDeployment]
role := pvc.GetLabels()[k8sutil.LabelKeyRole]
enforceAntiAffinity, _ := strconv.ParseBool(pvc.GetAnnotations()[constants.AnnotationEnforceAntiAffinity]) // If annotation empty, this will yield false.
return deploymentName, role, enforceAntiAffinity
}
// filterAllowedNodes returns those clients that do not yet have a volume for the given deployment name & role.
func (ls *LocalStorage) filterAllowedNodes(clients Clients, deploymentName, role string) (Clients, error) {
// Find all PVs for given deployment & role
list, err := ls.deps.Client.Kubernetes().CoreV1().PersistentVolumes().List(context.Background(), meta.ListOptions{
LabelSelector: fmt.Sprintf("%s=%s,%s=%s", k8sutil.LabelKeyArangoDeployment, deploymentName, k8sutil.LabelKeyRole, role),
})
if err != nil {
return nil, errors.WithStack(err)
}
excludedNodes := make(map[string]struct{})
for _, pv := range list.Items {
nodeName := pv.GetAnnotations()[nodeNameAnnotation]
excludedNodes[nodeName] = struct{}{}
}
return clients.Filter(func(node string, client provisioner.API) bool {
_, ok := excludedNodes[node]
return !ok
}), nil
}
// bindClaimToVolume tries to bind the given claim to the volume with given name.
// If the claim has been updated, the function retries several times.
func (ls *LocalStorage) bindClaimToVolume(claim core.PersistentVolumeClaim, volumeName string) error {
log := ls.log.Str("pvc-name", claim.GetName()).Str("volume-name", volumeName)
pvcs := ls.deps.Client.Kubernetes().CoreV1().PersistentVolumeClaims(claim.GetNamespace())
for attempt := 0; attempt < 10; attempt++ {
// Backoff if needed
time.Sleep(time.Millisecond * time.Duration(10*attempt))
// Fetch latest version of claim
updated, err := pvcs.Get(context.Background(), claim.GetName(), meta.GetOptions{})
if kerrors.IsNotFound(err) {
return errors.WithStack(err)
} else if err != nil {
log.Err(err).Warn("Failed to load updated PersistentVolumeClaim")
continue
}
// Check claim. If already bound, bail out
if !pvcNeedsVolume(*updated) {
if updated.Spec.VolumeName == volumeName {
log.Info("PersistentVolumeClaim already bound to PersistentVolume")
return nil
}
return errors.WithStack(errors.Newf("PersistentVolumeClaim '%s' no longer needs a volume", claim.GetName()))
}
// Try to bind
updated.Spec.VolumeName = volumeName
if _, err := pvcs.Update(context.Background(), updated, meta.UpdateOptions{}); kerrors.IsConflict(err) {
// Claim modified already, retry
log.Err(err).Debug("PersistentVolumeClaim has been modified. Retrying.")
} else if err != nil {
log.Err(err).Error("Failed to bind PVC to volume")
return errors.WithStack(err)
}
ls.log.Debug("Bound volume to PersistentVolumeClaim")
return nil
}
log.Error("All attempts to bind PVC to volume failed")
return errors.WithStack(errors.Newf("All attempts to bind PVC to volume failed"))
}
// shortHash creates a 6 letter hash of the given name.
func shortHash(name string) string {
h := sha1.Sum([]byte(name))
return fmt.Sprintf("%0x", h)[:6]
}