1
0
Fork 0
mirror of https://github.com/prometheus-operator/prometheus-operator.git synced 2025-04-15 16:56:24 +00:00
prometheus-operator/Documentation/proposals/202405-agent-daemonset.md
2024-07-02 20:33:43 +05:30

11 KiB
Raw Blame History

DaemonSet deployment for Prometheus Agent

This proposal is about designing and implementing the deployment of Prometheus Agent as DaemonSet. Currently, Prometheus Agent can only be deployed as StatefulSet, which could be considered as “cluster-wide” strategy, meaning one or several high-availability Prometheus Agents are responsible for scraping metrics of the whole cluster. Though this works well for many use cases, some use cases may indeed prefer a “node-specific” strategy (DaemonSet), where Prometheus Agent pods scale with the nodes and only scrape the metrics from the targets located on the same node.

1. Why

When deploying Prometheus Agent in Kubernetes, two of the biggest users concerns are load distribution and scalability.

DaemonSet deployment is a good solution for these:

  • Load distribution: Each Prometheus Agent pod will only scrape the targets located on the same node. Even though the targets on some nodes may produce more metrics than other nodes, the load distribution is reliable enough. This has been proven in Google Cloud Managed Service for Prometheus (GMP)'s operator which follows a similar approach.
  • Automatic scalability: When new nodes are added to the cluster, new Prometheus Agent pods will be automatically deployed to those nodes. Similarly, at some durations when some nodes are not needed and removed from the cluster, Prometheus Agent pods on those nodes will also be removed. Users can also select which set of nodes they want to deploy Prometheus Agent and the priority of Prometheus Agent pods compared to other pods on the same node.

DaemonSet deployment is especially more suitable for Prometheus Agent than Prometheus because the Agent mode is customized for collect-and-forward approach, so it's more lightweight. In specific, it requires (currently around 20 to 30%) less memory, does not produce TSDB blocks on disks, and naturally blocks querying APIs and rule distribution.

This deployment mode has been implemented and proven in the Google Cloud Kubernetes Engine (GKE) with Google Cloud Managed Service for Prometheus (GMP)'s operator, so we can learn from their cases and collaborate for shared improvements.

2. Pitfalls of the current solution

The key pitfalls of managing load distribution and scalability with the current StatefulSet deployment are:

  • They are done on the cluster scope. In other words, since one or several high-availability Prometheus Agents are responsible for scraping metrics of the whole cluster, users need to calculate/estimate the load and scalability of the whole cluster to decide on replicas and sharding strategies. Estimating cluster-wide load and scalability is a much harder task than estimating node-wide load and scalability.
  • StatefulSet inherently doesn't scale alongside the scale of nodes. So there may be situations where more or less existing Prometheus Agent pods are needed for the scraped load. Even though users can use helping tools like Horizontal Pod Autoscaler (HPA), that's still additional complexity.

This is not to say that DaemonSet is superior to StatefulSet. StatefulSet also has its own advantages, such as easier storage handling. This is to say that DaemonSet can solve some of the existing problems in StatefulSet, and vice versa. So DaemonSet is a good deployment option to implement besides StatefulSet. Users can choose to use one or both of them, depending on their cases.

3. Audience

Users with use cases where scraped load is very large or hard to estimate and/or scalability is hard to predict, so they need a simple way to manage load distribution and scalability. They may apply DaemonSet deployment on the whole cluster or only some nodes according to their needs.

An example of audience is expressed here.

4. Goals

Provide an MVP version of the DaemonSet deployment of Prometheus Agent to the Audience. In specific, the MVP will need to:

  • Allow users to deploy one Prometheus Agent pod per node.
  • Allow users to restrict which set of nodes they want to deploy Prometheus Agent, if desired.
  • Allow users to set the priority of Prometheus Agent pod compared to other pods on the same node, if desired.
  • Allow each Prometheus Agent pod to only scrape from the pods from PodMonitor that run on the same node.

5. Non-Goals

This proposal only aims at Prometheus Agent, not Prometheus.

Other non-goals are the features that are not easy to implement and require more investigation. We will need to investigate whether there are actual user needs for them, if yes, then how to best implement them. We can also learn from similar projects such as OpenTelemetry Operator (they have DaemonSet mode) and Grafana Agent on how they approach these problems and what we can apply for our cases. Well handle these after the MVP. Those (currently) non-goals features are:

  • ServiceMonitor support: There's a performance issue regarding this feature. Since each Prometheus Agent running on a node requires one watch, making all Prometheus Agent pods watch all endpoints will put a huge stress on Kubernetes API server. This is the main reason why GMP hasnt supported this, even though there are user needs stated in some issues (#362, #192). However, as discussed with Danny from GMP here, ServiceMonitor support based on EndpointSlice seems like a viable approach. Well investigate this further after the MVP.
  • Storage: We will need to spend time studying more about the WAL, different storage solutions provided by Kubernetes, and how to gracefully handle storage in different cases of crashes. For example, theres an issue in Prometheus showing that samples may be lost if remote write didnt flush cleanly. Well investigate these further after the MVP.

In the MVP version, we will not allow users to directly switch from a live StatefulSet to DaemonSet deployment. Reasons are explained in the CRD subsection in the How section.

6. How

The MVP version of DaemonSet deployment will be put behind a feature flag.

6.1. CRD:

Currently, we already have a PrometheusAgent CRD that supports StatefulSet deployment. Well add new field(s) in this CRD to enable DaemonSet deployment.

The reason for enhancing existing CRD (instead of introducing a new CRD) is it would take less time to finish the MVP. Well let users experiment with the MVP, and in case users report a separate CRD is needed, well separate the logic of DaemonSet deployment into a new CRD later.

The current PrometheusAgent CRD already has sufficient fields for the DaemonSet deployment. The DaemonSet deployment can use all the existing fields in the CRD except the ones related to:

  • Selectors for service, probe, ScrapeConfig
  • Replica
  • Shard
  • Storage

We will add a new mode field that accepts either StatefulSet or DaemonSet, with StatefulSet being the default. If the DaemonSet mode is activated (mode: DaemonSet), all the unrelated fields listed above will not be accepted. In the MVP, we will simply fail the reconciliation if any of those fields are set. We will prevent users to directly switch from a live StatefulSet setup to DaemonSet, because that might break their workload if they forget to unset the unsupported fields. Following up, we will leverage validation rules with Kubernetes' Common Expression Language (CEL). Only then, we will allow switching from a live StatefulSet setup to DaemonSet. We already have an issue for CEL here.

6.2. Node detecting:

As pointed out in Danny from GMPs talk, to make Prometheus Agent DaemonSet know which node its on, we can use Kubernetes downward API. In config-reloader container, we can mount the node name as an environment variable like this:

containers:
- name: config-reloader
  env:
  - name: NODE_NAME
    valueFrom:
      fieldRef:
        apiVersion: v1
        fieldPath: spec.nodeName

6.3. Targets filtering for pods (PodMonitor support):

To filter pod targets, Dannys talk has pointed out the best option is to use field selector:

kubernetes_sd_configs:
- role: pod
  selectors:
  - role: pod
    field: spec.nodeName=$NODE_NAME

We'll go with this option, because it filters targets right at discovery time, and also because Kubernetes API server watch cache indexes pods by node name (as we can see in Kubernetes codebase).

We've also considered using relabel config that filters pods by __meta_kubernetes_pod_node_name label. However, we didn't choose to go with this option because it filters pods only after discovering all the pods from PodMonitor, which increases load on Kubernetes API server.

7. Action Plan

For the implementation, well do what we detailed in the How section. The common logics between StatefulSet and DaemonSet modes will be extracted into one place. We will have a separate daemonset.go for the separate logic of the DaemonSet mode.

For the test, we will have unit tests covering new logic, and integration tests covering the basic user cases, which are:

  • Users cannot switch directly from StatefulSet to DaemonSet.
  • Prometheus Agent DaemonSet is created/deleted successfully.
  • Prometheus Agent DaemonSet is installed on the right nodes.
  • Prometheus Agent DaemonSet selects correctly the pods from PodMonitor on the same node. Currently we only set up a Kind cluster of one node for integration tests. Since the test cases for DaemonSet deployment requires at least two nodes, we will need to modify the Kind cluster config for that.

Well also need a new user guide explaining how to use this new mode.

8. Follow-ups

After the Goals of this proposal have been met, well reevaluate the features in the Non-goals section and see if any of them should/can be addressed.

We will also work on enhancements, such as leveraging validation rules with Kubernetes' Common Expression Language (CEL) for the fields in Prometheus Agent CRD for DaemonSet mode.