mirror of
https://github.com/kyverno/kyverno.git
synced 2025-03-05 15:37:19 +00:00
fix: improve shutdown gracefulness (#5107)
* fix: improve shutdown gracefulness Signed-off-by: Charles-Edouard Brétéché <charles.edouard@nirmata.com> * fix Signed-off-by: Charles-Edouard Brétéché <charles.edouard@nirmata.com> * fix Signed-off-by: Charles-Edouard Brétéché <charles.edouard@nirmata.com> * fix Signed-off-by: Charles-Edouard Brétéché <charles.edouard@nirmata.com> * fix Signed-off-by: Charles-Edouard Brétéché <charles.edouard@nirmata.com> * fix Signed-off-by: Charles-Edouard Brétéché <charles.edouard@nirmata.com> --------- Signed-off-by: Charles-Edouard Brétéché <charles.edouard@nirmata.com> Co-authored-by: shuting <shuting@nirmata.com>
This commit is contained in:
parent
6f7f06f0d4
commit
fc453b1faa
6 changed files with 48 additions and 27 deletions
|
@ -206,10 +206,12 @@ func main() {
|
||||||
logging.WithName("EventGenerator"),
|
logging.WithName("EventGenerator"),
|
||||||
)
|
)
|
||||||
// this controller only subscribe to events, nothing is returned...
|
// this controller only subscribe to events, nothing is returned...
|
||||||
|
var wg sync.WaitGroup
|
||||||
policymetricscontroller.NewController(
|
policymetricscontroller.NewController(
|
||||||
metricsConfig,
|
metricsConfig,
|
||||||
kyvernoInformer.Kyverno().V1().ClusterPolicies(),
|
kyvernoInformer.Kyverno().V1().ClusterPolicies(),
|
||||||
kyvernoInformer.Kyverno().V1().Policies(),
|
kyvernoInformer.Kyverno().V1().Policies(),
|
||||||
|
&wg,
|
||||||
)
|
)
|
||||||
engine := engine.NewEngine(
|
engine := engine.NewEngine(
|
||||||
configuration,
|
configuration,
|
||||||
|
@ -225,7 +227,7 @@ func main() {
|
||||||
os.Exit(1)
|
os.Exit(1)
|
||||||
}
|
}
|
||||||
// start event generator
|
// start event generator
|
||||||
go eventGenerator.Run(signalCtx, 3)
|
go eventGenerator.Run(signalCtx, 3, &wg)
|
||||||
// setup leader election
|
// setup leader election
|
||||||
le, err := leaderelection.New(
|
le, err := leaderelection.New(
|
||||||
logger.WithName("leader-election"),
|
logger.WithName("leader-election"),
|
||||||
|
@ -280,6 +282,7 @@ func main() {
|
||||||
for {
|
for {
|
||||||
select {
|
select {
|
||||||
case <-signalCtx.Done():
|
case <-signalCtx.Done():
|
||||||
|
wg.Wait()
|
||||||
return
|
return
|
||||||
default:
|
default:
|
||||||
le.Run(signalCtx)
|
le.Run(signalCtx)
|
||||||
|
|
|
@ -332,6 +332,7 @@ func main() {
|
||||||
logger.Error(err, "Failed to create openapi manager")
|
logger.Error(err, "Failed to create openapi manager")
|
||||||
os.Exit(1)
|
os.Exit(1)
|
||||||
}
|
}
|
||||||
|
var wg sync.WaitGroup
|
||||||
certRenewer := tls.NewCertRenewer(
|
certRenewer := tls.NewCertRenewer(
|
||||||
kubeClient.CoreV1().Secrets(config.KyvernoNamespace()),
|
kubeClient.CoreV1().Secrets(config.KyvernoNamespace()),
|
||||||
secretLister,
|
secretLister,
|
||||||
|
@ -353,6 +354,7 @@ func main() {
|
||||||
metricsConfig,
|
metricsConfig,
|
||||||
kyvernoInformer.Kyverno().V1().ClusterPolicies(),
|
kyvernoInformer.Kyverno().V1().ClusterPolicies(),
|
||||||
kyvernoInformer.Kyverno().V1().Policies(),
|
kyvernoInformer.Kyverno().V1().Policies(),
|
||||||
|
&wg,
|
||||||
)
|
)
|
||||||
// log policy changes
|
// log policy changes
|
||||||
genericloggingcontroller.NewController(
|
genericloggingcontroller.NewController(
|
||||||
|
@ -415,7 +417,7 @@ func main() {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
// start event generator
|
// start event generator
|
||||||
go eventGenerator.Run(signalCtx, 3)
|
go eventGenerator.Run(signalCtx, 3, &wg)
|
||||||
// setup leader election
|
// setup leader election
|
||||||
le, err := leaderelection.New(
|
le, err := leaderelection.New(
|
||||||
logger.WithName("leader-election"),
|
logger.WithName("leader-election"),
|
||||||
|
@ -476,7 +478,6 @@ func main() {
|
||||||
os.Exit(1)
|
os.Exit(1)
|
||||||
}
|
}
|
||||||
// start non leader controllers
|
// start non leader controllers
|
||||||
var wg sync.WaitGroup
|
|
||||||
for _, controller := range nonLeaderControllers {
|
for _, controller := range nonLeaderControllers {
|
||||||
controller.Run(signalCtx, logger.WithName("controllers"), &wg)
|
controller.Run(signalCtx, logger.WithName("controllers"), &wg)
|
||||||
}
|
}
|
||||||
|
|
|
@ -310,7 +310,8 @@ func main() {
|
||||||
os.Exit(1)
|
os.Exit(1)
|
||||||
}
|
}
|
||||||
// start event generator
|
// start event generator
|
||||||
go eventGenerator.Run(ctx, 3)
|
var wg sync.WaitGroup
|
||||||
|
go eventGenerator.Run(ctx, 3, &wg)
|
||||||
eng := engine.NewEngine(
|
eng := engine.NewEngine(
|
||||||
configuration,
|
configuration,
|
||||||
dClient,
|
dClient,
|
||||||
|
@ -386,6 +387,7 @@ func main() {
|
||||||
for {
|
for {
|
||||||
select {
|
select {
|
||||||
case <-ctx.Done():
|
case <-ctx.Done():
|
||||||
|
wg.Wait()
|
||||||
return
|
return
|
||||||
default:
|
default:
|
||||||
le.Run(ctx)
|
le.Run(ctx)
|
||||||
|
|
|
@ -2,6 +2,7 @@ package policy
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"context"
|
"context"
|
||||||
|
"sync"
|
||||||
|
|
||||||
kyvernov1 "github.com/kyverno/kyverno/api/kyverno/v1"
|
kyvernov1 "github.com/kyverno/kyverno/api/kyverno/v1"
|
||||||
"github.com/kyverno/kyverno/pkg/autogen"
|
"github.com/kyverno/kyverno/pkg/autogen"
|
||||||
|
@ -25,10 +26,17 @@ type controller struct {
|
||||||
// listers
|
// listers
|
||||||
cpolLister kyvernov1listers.ClusterPolicyLister
|
cpolLister kyvernov1listers.ClusterPolicyLister
|
||||||
polLister kyvernov1listers.PolicyLister
|
polLister kyvernov1listers.PolicyLister
|
||||||
|
|
||||||
|
waitGroup *sync.WaitGroup
|
||||||
}
|
}
|
||||||
|
|
||||||
// TODO: this is a strange controller, it only processes events, this should be changed to a real controller.
|
// TODO: this is a strange controller, it only processes events, this should be changed to a real controller.
|
||||||
func NewController(metricsConfig metrics.MetricsConfigManager, cpolInformer kyvernov1informers.ClusterPolicyInformer, polInformer kyvernov1informers.PolicyInformer) {
|
func NewController(
|
||||||
|
metricsConfig metrics.MetricsConfigManager,
|
||||||
|
cpolInformer kyvernov1informers.ClusterPolicyInformer,
|
||||||
|
polInformer kyvernov1informers.PolicyInformer,
|
||||||
|
waitGroup *sync.WaitGroup,
|
||||||
|
) {
|
||||||
meterProvider := global.MeterProvider()
|
meterProvider := global.MeterProvider()
|
||||||
meter := meterProvider.Meter(metrics.MeterName)
|
meter := meterProvider.Meter(metrics.MeterName)
|
||||||
policyRuleInfoMetric, err := meter.Float64ObservableGauge(
|
policyRuleInfoMetric, err := meter.Float64ObservableGauge(
|
||||||
|
@ -43,6 +51,7 @@ func NewController(metricsConfig metrics.MetricsConfigManager, cpolInformer kyve
|
||||||
ruleInfo: policyRuleInfoMetric,
|
ruleInfo: policyRuleInfoMetric,
|
||||||
cpolLister: cpolInformer.Lister(),
|
cpolLister: cpolInformer.Lister(),
|
||||||
polLister: polInformer.Lister(),
|
polLister: polInformer.Lister(),
|
||||||
|
waitGroup: waitGroup,
|
||||||
}
|
}
|
||||||
controllerutils.AddEventHandlers(cpolInformer.Informer(), c.addPolicy, c.updatePolicy, c.deletePolicy)
|
controllerutils.AddEventHandlers(cpolInformer.Informer(), c.addPolicy, c.updatePolicy, c.deletePolicy)
|
||||||
controllerutils.AddEventHandlers(polInformer.Informer(), c.addNsPolicy, c.updateNsPolicy, c.deleteNsPolicy)
|
controllerutils.AddEventHandlers(polInformer.Informer(), c.addNsPolicy, c.updateNsPolicy, c.deleteNsPolicy)
|
||||||
|
@ -111,16 +120,24 @@ func (c *controller) reportPolicy(ctx context.Context, policy kyvernov1.PolicyIn
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func (c *controller) startRountine(routine func()) {
|
||||||
|
c.waitGroup.Add(1)
|
||||||
|
go func() {
|
||||||
|
defer c.waitGroup.Done()
|
||||||
|
routine()
|
||||||
|
}()
|
||||||
|
}
|
||||||
|
|
||||||
func (c *controller) addPolicy(obj interface{}) {
|
func (c *controller) addPolicy(obj interface{}) {
|
||||||
p := obj.(*kyvernov1.ClusterPolicy)
|
p := obj.(*kyvernov1.ClusterPolicy)
|
||||||
// register kyverno_policy_changes_total metric concurrently
|
// register kyverno_policy_changes_total metric concurrently
|
||||||
go c.registerPolicyChangesMetricAddPolicy(context.TODO(), logger, p)
|
c.startRountine(func() { c.registerPolicyChangesMetricAddPolicy(context.TODO(), logger, p) })
|
||||||
}
|
}
|
||||||
|
|
||||||
func (c *controller) updatePolicy(old, cur interface{}) {
|
func (c *controller) updatePolicy(old, cur interface{}) {
|
||||||
oldP, curP := old.(*kyvernov1.ClusterPolicy), cur.(*kyvernov1.ClusterPolicy)
|
oldP, curP := old.(*kyvernov1.ClusterPolicy), cur.(*kyvernov1.ClusterPolicy)
|
||||||
// register kyverno_policy_changes_total metric concurrently
|
// register kyverno_policy_changes_total metric concurrently
|
||||||
go c.registerPolicyChangesMetricUpdatePolicy(context.TODO(), logger, oldP, curP)
|
c.startRountine(func() { c.registerPolicyChangesMetricUpdatePolicy(context.TODO(), logger, oldP, curP) })
|
||||||
}
|
}
|
||||||
|
|
||||||
func (c *controller) deletePolicy(obj interface{}) {
|
func (c *controller) deletePolicy(obj interface{}) {
|
||||||
|
@ -130,19 +147,19 @@ func (c *controller) deletePolicy(obj interface{}) {
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
// register kyverno_policy_changes_total metric concurrently
|
// register kyverno_policy_changes_total metric concurrently
|
||||||
go c.registerPolicyChangesMetricDeletePolicy(context.TODO(), logger, p)
|
c.startRountine(func() { c.registerPolicyChangesMetricDeletePolicy(context.TODO(), logger, p) })
|
||||||
}
|
}
|
||||||
|
|
||||||
func (c *controller) addNsPolicy(obj interface{}) {
|
func (c *controller) addNsPolicy(obj interface{}) {
|
||||||
p := obj.(*kyvernov1.Policy)
|
p := obj.(*kyvernov1.Policy)
|
||||||
// register kyverno_policy_changes_total metric concurrently
|
// register kyverno_policy_changes_total metric concurrently
|
||||||
go c.registerPolicyChangesMetricAddPolicy(context.TODO(), logger, p)
|
c.startRountine(func() { c.registerPolicyChangesMetricAddPolicy(context.TODO(), logger, p) })
|
||||||
}
|
}
|
||||||
|
|
||||||
func (c *controller) updateNsPolicy(old, cur interface{}) {
|
func (c *controller) updateNsPolicy(old, cur interface{}) {
|
||||||
oldP, curP := old.(*kyvernov1.Policy), cur.(*kyvernov1.Policy)
|
oldP, curP := old.(*kyvernov1.Policy), cur.(*kyvernov1.Policy)
|
||||||
// register kyverno_policy_changes_total metric concurrently
|
// register kyverno_policy_changes_total metric concurrently
|
||||||
go c.registerPolicyChangesMetricUpdatePolicy(context.TODO(), logger, oldP, curP)
|
c.startRountine(func() { c.registerPolicyChangesMetricUpdatePolicy(context.TODO(), logger, oldP, curP) })
|
||||||
}
|
}
|
||||||
|
|
||||||
func (c *controller) deleteNsPolicy(obj interface{}) {
|
func (c *controller) deleteNsPolicy(obj interface{}) {
|
||||||
|
@ -152,5 +169,5 @@ func (c *controller) deleteNsPolicy(obj interface{}) {
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
// register kyverno_policy_changes_total metric concurrently
|
// register kyverno_policy_changes_total metric concurrently
|
||||||
go c.registerPolicyChangesMetricDeletePolicy(context.TODO(), logger, p)
|
c.startRountine(func() { c.registerPolicyChangesMetricDeletePolicy(context.TODO(), logger, p) })
|
||||||
}
|
}
|
||||||
|
|
|
@ -2,13 +2,13 @@ package event
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"context"
|
"context"
|
||||||
|
"sync"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
"github.com/go-logr/logr"
|
"github.com/go-logr/logr"
|
||||||
kyvernov1informers "github.com/kyverno/kyverno/pkg/client/informers/externalversions/kyverno/v1"
|
kyvernov1informers "github.com/kyverno/kyverno/pkg/client/informers/externalversions/kyverno/v1"
|
||||||
kyvernov1listers "github.com/kyverno/kyverno/pkg/client/listers/kyverno/v1"
|
kyvernov1listers "github.com/kyverno/kyverno/pkg/client/listers/kyverno/v1"
|
||||||
"github.com/kyverno/kyverno/pkg/clients/dclient"
|
"github.com/kyverno/kyverno/pkg/clients/dclient"
|
||||||
"github.com/kyverno/kyverno/pkg/controllers"
|
|
||||||
corev1 "k8s.io/api/core/v1"
|
corev1 "k8s.io/api/core/v1"
|
||||||
errors "k8s.io/apimachinery/pkg/api/errors"
|
errors "k8s.io/apimachinery/pkg/api/errors"
|
||||||
"k8s.io/apimachinery/pkg/runtime"
|
"k8s.io/apimachinery/pkg/runtime"
|
||||||
|
@ -48,8 +48,8 @@ type generator struct {
|
||||||
|
|
||||||
// Controller interface to generate event
|
// Controller interface to generate event
|
||||||
type Controller interface {
|
type Controller interface {
|
||||||
controllers.Controller
|
|
||||||
Interface
|
Interface
|
||||||
|
Run(context.Context, int, *sync.WaitGroup)
|
||||||
}
|
}
|
||||||
|
|
||||||
// Interface to generate event
|
// Interface to generate event
|
||||||
|
@ -84,13 +84,11 @@ func NewEventGenerator(
|
||||||
// Add queues an event for generation
|
// Add queues an event for generation
|
||||||
func (gen *generator) Add(infos ...Info) {
|
func (gen *generator) Add(infos ...Info) {
|
||||||
logger := gen.log
|
logger := gen.log
|
||||||
|
|
||||||
logger.V(3).Info("generating events", "count", len(infos))
|
logger.V(3).Info("generating events", "count", len(infos))
|
||||||
if gen.maxQueuedEvents == 0 || gen.queue.Len() > gen.maxQueuedEvents {
|
if gen.maxQueuedEvents == 0 || gen.queue.Len() > gen.maxQueuedEvents {
|
||||||
logger.V(2).Info("exceeds the event queue limit, dropping the event", "maxQueuedEvents", gen.maxQueuedEvents, "current size", gen.queue.Len())
|
logger.V(2).Info("exceeds the event queue limit, dropping the event", "maxQueuedEvents", gen.maxQueuedEvents, "current size", gen.queue.Len())
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
for _, info := range infos {
|
for _, info := range infos {
|
||||||
if info.Name == "" {
|
if info.Name == "" {
|
||||||
// dont create event for resources with generateName
|
// dont create event for resources with generateName
|
||||||
|
@ -103,15 +101,18 @@ func (gen *generator) Add(infos ...Info) {
|
||||||
}
|
}
|
||||||
|
|
||||||
// Run begins generator
|
// Run begins generator
|
||||||
func (gen *generator) Run(ctx context.Context, workers int) {
|
func (gen *generator) Run(ctx context.Context, workers int, waitGroup *sync.WaitGroup) {
|
||||||
logger := gen.log
|
logger := gen.log
|
||||||
defer utilruntime.HandleCrash()
|
|
||||||
|
|
||||||
logger.Info("start")
|
logger.Info("start")
|
||||||
defer logger.Info("shutting down")
|
defer logger.Info("shutting down")
|
||||||
|
defer utilruntime.HandleCrash()
|
||||||
|
defer gen.queue.ShutDown()
|
||||||
for i := 0; i < workers; i++ {
|
for i := 0; i < workers; i++ {
|
||||||
go wait.UntilWithContext(ctx, gen.runWorker, time.Second)
|
waitGroup.Add(1)
|
||||||
|
go func() {
|
||||||
|
defer waitGroup.Done()
|
||||||
|
wait.UntilWithContext(ctx, gen.runWorker, time.Second)
|
||||||
|
}()
|
||||||
}
|
}
|
||||||
<-ctx.Done()
|
<-ctx.Done()
|
||||||
}
|
}
|
||||||
|
@ -135,7 +136,6 @@ func (gen *generator) handleErr(err error, key interface{}) {
|
||||||
gen.queue.AddRateLimited(key)
|
gen.queue.AddRateLimited(key)
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
gen.queue.Forget(key)
|
gen.queue.Forget(key)
|
||||||
if !errors.IsNotFound(err) {
|
if !errors.IsNotFound(err) {
|
||||||
logger.Error(err, "failed to generate event", "key", key)
|
logger.Error(err, "failed to generate event", "key", key)
|
||||||
|
@ -147,7 +147,6 @@ func (gen *generator) processNextWorkItem() bool {
|
||||||
if shutdown {
|
if shutdown {
|
||||||
return false
|
return false
|
||||||
}
|
}
|
||||||
|
|
||||||
defer gen.queue.Done(obj)
|
defer gen.queue.Done(obj)
|
||||||
var key Info
|
var key Info
|
||||||
var ok bool
|
var ok bool
|
||||||
|
@ -158,7 +157,6 @@ func (gen *generator) processNextWorkItem() bool {
|
||||||
}
|
}
|
||||||
err := gen.syncHandler(key)
|
err := gen.syncHandler(key)
|
||||||
gen.handleErr(err, obj)
|
gen.handleErr(err, obj)
|
||||||
|
|
||||||
return true
|
return true
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -56,9 +56,10 @@ func newControllerMetrics(logger logr.Logger, controllerName string) *controller
|
||||||
|
|
||||||
func Run(ctx context.Context, logger logr.Logger, controllerName string, period time.Duration, queue workqueue.RateLimitingInterface, n, maxRetries int, r reconcileFunc, routines ...func(context.Context, logr.Logger)) {
|
func Run(ctx context.Context, logger logr.Logger, controllerName string, period time.Duration, queue workqueue.RateLimitingInterface, n, maxRetries int, r reconcileFunc, routines ...func(context.Context, logr.Logger)) {
|
||||||
logger.Info("starting ...")
|
logger.Info("starting ...")
|
||||||
defer runtime.HandleCrash()
|
|
||||||
defer logger.Info("stopped")
|
defer logger.Info("stopped")
|
||||||
var wg sync.WaitGroup
|
var wg sync.WaitGroup
|
||||||
|
defer wg.Wait()
|
||||||
|
defer runtime.HandleCrash()
|
||||||
metric := newControllerMetrics(logger, controllerName)
|
metric := newControllerMetrics(logger, controllerName)
|
||||||
func() {
|
func() {
|
||||||
ctx, cancel := context.WithCancel(ctx)
|
ctx, cancel := context.WithCancel(ctx)
|
||||||
|
@ -68,8 +69,8 @@ func Run(ctx context.Context, logger logr.Logger, controllerName string, period
|
||||||
wg.Add(1)
|
wg.Add(1)
|
||||||
go func(logger logr.Logger) {
|
go func(logger logr.Logger) {
|
||||||
logger.Info("starting worker")
|
logger.Info("starting worker")
|
||||||
defer wg.Done()
|
|
||||||
defer logger.Info("worker stopped")
|
defer logger.Info("worker stopped")
|
||||||
|
defer wg.Done()
|
||||||
wait.UntilWithContext(ctx, func(ctx context.Context) { worker(ctx, logger, metric, queue, maxRetries, r) }, period)
|
wait.UntilWithContext(ctx, func(ctx context.Context) { worker(ctx, logger, metric, queue, maxRetries, r) }, period)
|
||||||
}(logger.WithName("worker").WithValues("id", i))
|
}(logger.WithName("worker").WithValues("id", i))
|
||||||
}
|
}
|
||||||
|
@ -77,15 +78,14 @@ func Run(ctx context.Context, logger logr.Logger, controllerName string, period
|
||||||
wg.Add(1)
|
wg.Add(1)
|
||||||
go func(logger logr.Logger, routine func(context.Context, logr.Logger)) {
|
go func(logger logr.Logger, routine func(context.Context, logr.Logger)) {
|
||||||
logger.Info("starting routine")
|
logger.Info("starting routine")
|
||||||
defer wg.Done()
|
|
||||||
defer logger.Info("routine stopped")
|
defer logger.Info("routine stopped")
|
||||||
|
defer wg.Done()
|
||||||
routine(ctx, logger)
|
routine(ctx, logger)
|
||||||
}(logger.WithName("routine").WithValues("id", i), routine)
|
}(logger.WithName("routine").WithValues("id", i), routine)
|
||||||
}
|
}
|
||||||
<-ctx.Done()
|
<-ctx.Done()
|
||||||
}()
|
}()
|
||||||
logger.Info("waiting for workers to terminate ...")
|
logger.Info("waiting for workers to terminate ...")
|
||||||
wg.Wait()
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func worker(ctx context.Context, logger logr.Logger, metric *controllerMetrics, queue workqueue.RateLimitingInterface, maxRetries int, r reconcileFunc) {
|
func worker(ctx context.Context, logger logr.Logger, metric *controllerMetrics, queue workqueue.RateLimitingInterface, maxRetries int, r reconcileFunc) {
|
||||||
|
|
Loading…
Add table
Reference in a new issue