From fcfc8f13fb11075007a416220a37e6ca80295c88 Mon Sep 17 00:00:00 2001 From: HadhemiDD <43783545+HadhemiDD@users.noreply.github.com> Date: Tue, 17 Sep 2024 15:43:26 +0200 Subject: [PATCH 01/34] Kubeflow update dashboard and readme (#18600) * update readme * update readme * [Release] Bumped kubeflow version to 1.0.0 * revert release --- kubeflow/README.md | 46 +++++++++++++++++++ kubeflow/assets/dashboards/overview.json | 43 ++++++++++++----- .../changelog.d/18391.added | 0 3 files changed, 78 insertions(+), 11 deletions(-) rename kubeflow/{ => datadog_checks}/changelog.d/18391.added (100%) diff --git a/kubeflow/README.md b/kubeflow/README.md index 0d9df6cb33b3b..634017875995c 100644 --- a/kubeflow/README.md +++ b/kubeflow/README.md @@ -27,6 +27,52 @@ For the Agent to start collecting metrics, the `kubeflow` pods need to be annota Kubeflow has metrics endpoints that can be accessed on port `9090`. +To enable metrics exposure in kubeflow through prometheus, you might need to enable the prometheus service monitoring for the component in question. + +You can use Kube-Prometheus-Stack or a custom Prometheus installation. + +##### How to install Kube-Prometheus-Stack: +1. Add Helm Repository: +``` +helm repo add prometheus-community https://prometheus-community.github.io/helm-charts +helm repo update +``` + +2. Install the Chart: +``` +helm install prometheus-stack prometheus-community/kube-prometheus-stack +``` + +3. Expose Prometheus service externally: +``` +kubectl port-forward prometheus-stack 9090:9090 +``` +##### Set Up ServiceMonitors for Kubeflow Components: + +You need to configure ServiceMonitors for Kubeflow components to expose their Prometheus metrics. +If your Kubeflow component exposes Prometheus metrics by default. You'll just need to configure Prometheus to scrape these metrics. + +The ServiceMonitor would look like this: + +```yaml +apiVersion: monitoring.coreos.com/v1 +kind: ServiceMonitor +metadata: + name: -monitor + labels: + release: prometheus-stack +spec: + selector: + matchLabels: + app: + endpoints: + - port: http + path: /metrics +``` + +Where `` is to be replaced by `pipelines`, `kserve` or `katib` and `` is to be replaced by `ml-pipeline`, `kserve` or `katib`. + + **Note**: The listed metrics can only be collected if they are available(depending on the version). Some metrics are generated only when certain actions are performed. The only parameter required for configuring the `kubeflow` check is `openmetrics_endpoint`. This parameter should be set to the location where the Prometheus-formatted metrics are exposed. The default port is `9090`. In containerized environments, `%%host%%` should be used for [host autodetection][3]. diff --git a/kubeflow/assets/dashboards/overview.json b/kubeflow/assets/dashboards/overview.json index 8b77b0122d965..b573e5f477156 100644 --- a/kubeflow/assets/dashboards/overview.json +++ b/kubeflow/assets/dashboards/overview.json @@ -46,7 +46,7 @@ }, "id": 4610707819074916, "layout": { - "height": 3, + "height": 4, "width": 3, "x": 0, "y": 3 @@ -67,7 +67,7 @@ }, "id": 8366490141273904, "layout": { - "height": 3, + "height": 4, "width": 2, "x": 3, "y": 3 @@ -77,7 +77,7 @@ }, "id": 3375620455700908, "layout": { - "height": 7, + "height": 8, "width": 5, "x": 0, "y": 0 @@ -93,8 +93,8 @@ "widgets": [ { "definition": { - "background_color": "blue", - "content": "The service checks show the Kubeflow OpenMetrics/Prometheus endpoint status.\n\nThe monitor summary shows you any active alerts for the most crucial Metrics. ", + "background_color": "pink", + "content": "If many widgets are empty, you are using a version of Kubeflow that does not expose certain metrics. Refer to the metadata.csv file for metrics list. \n\nReach out to support to indicate version incompatibilities.", "font_size": "14", "has_padding": true, "show_tick": true, @@ -112,6 +112,27 @@ "y": 0 } }, + { + "definition": { + "background_color": "blue", + "content": "The service checks show the Kubeflow OpenMetrics/Prometheus endpoint status.\n\nThe monitor summary shows you any active alerts for the most crucial Metrics. ", + "font_size": "14", + "has_padding": true, + "show_tick": true, + "text_align": "left", + "tick_edge": "left", + "tick_pos": "50%", + "type": "note", + "vertical_align": "center" + }, + "id": 6145599891700518, + "layout": { + "height": 1, + "width": 7, + "x": 0, + "y": 2 + } + }, { "definition": { "color_preference": "text", @@ -134,14 +155,14 @@ "height": 4, "width": 7, "x": 0, - "y": 2 + "y": 3 } } ] }, "id": 3510698085005998, "layout": { - "height": 7, + "height": 8, "width": 7, "x": 5, "y": 0 @@ -273,7 +294,7 @@ "height": 5, "width": 12, "x": 0, - "y": 7 + "y": 8 } }, { @@ -988,7 +1009,7 @@ "is_column_break": true, "width": 12, "x": 0, - "y": 12 + "y": 13 } }, { @@ -1224,7 +1245,7 @@ "height": 8, "width": 12, "x": 0, - "y": 28 + "y": 29 } }, { @@ -1424,7 +1445,7 @@ "height": 7, "width": 12, "x": 0, - "y": 36 + "y": 37 } } ] diff --git a/kubeflow/changelog.d/18391.added b/kubeflow/datadog_checks/changelog.d/18391.added similarity index 100% rename from kubeflow/changelog.d/18391.added rename to kubeflow/datadog_checks/changelog.d/18391.added From 69cffd9e02ffbdc3015294a2c39c9215c0df637a Mon Sep 17 00:00:00 2001 From: Kyle Neale Date: Tue, 17 Sep 2024 10:43:07 -0400 Subject: [PATCH 02/34] Bump Python 3 to 3.12 (#18212) * bump py version to 3.12 * update dbm dev container * update builders and build deps * fix python string * Build lxml dependencies for linux x64 (#18467) * fix rmq version * fix gunicorn tests * Fix lock file generation * comment out if clause to force dep resolution * bump python to 3.12.6 * fix python on Windows sha * remove DD_PYTHON2 unbound variable * fix hatch errors * remove setup.py files * fix CI * revert datadog_checks_test_helper --------- Co-authored-by: Alex Lopez Co-authored-by: Ilia Kurenkov Co-authored-by: Alex Lopez --- .builders/images/linux-aarch64/Dockerfile | 4 +-- .builders/images/linux-x86_64/Dockerfile | 34 +++++++++++++++++-- .builders/images/windows-x86_64/Dockerfile | 12 +++---- .builders/lock.py | 25 +++++++++++--- .devcontainer/Dockerfile | 2 +- .devcontainer/dbm/Dockerfile | 2 +- .devcontainer/dbm/devcontainer.json | 2 +- .devcontainer/devcontainer.json | 2 +- .github/workflows/build-ddev.yml | 2 +- .github/workflows/build-deps.yml | 8 ++--- .github/workflows/cache-shared-deps.yml | 2 +- .github/workflows/compute-matrix.yml | 2 +- .github/workflows/docs.yml | 2 +- .github/workflows/pr-quick-check.yml | 2 +- .github/workflows/release-base.yml | 2 +- .github/workflows/release-dev.yml | 2 +- .github/workflows/release-hash-check.yml | 2 +- .github/workflows/run-validations.yml | 2 +- .github/workflows/test-target.yml | 4 +-- .github/workflows/update-agent-changelog.yml | 2 +- .github/workflows/update-dependencies.yml | 2 +- .gitlab/tagger/Dockerfile | 2 +- active_directory/changelog.d/18207.added | 1 + active_directory/hatch.toml | 2 +- active_directory/pyproject.toml | 2 +- activemq/changelog.d/18207.added | 1 + activemq/hatch.toml | 4 +-- activemq/pyproject.toml | 2 +- activemq_xml/changelog.d/18207.added | 1 + activemq_xml/hatch.toml | 2 +- activemq_xml/pyproject.toml | 2 +- aerospike/changelog.d/18207.added | 1 + aerospike/hatch.toml | 2 +- aerospike/pyproject.toml | 2 +- airflow/changelog.d/18207.added | 1 + airflow/hatch.toml | 2 +- airflow/pyproject.toml | 2 +- amazon_msk/changelog.d/18207.added | 1 + amazon_msk/hatch.toml | 2 +- amazon_msk/pyproject.toml | 2 +- ambari/changelog.d/18207.added | 1 + ambari/hatch.toml | 2 +- ambari/pyproject.toml | 2 +- apache/changelog.d/18207.added | 1 + apache/hatch.toml | 2 +- apache/pyproject.toml | 2 +- arangodb/changelog.d/18207.added | 1 + arangodb/hatch.toml | 2 +- arangodb/pyproject.toml | 2 +- argo_rollouts/changelog.d/18207.added | 1 + argo_rollouts/hatch.toml | 2 +- argo_rollouts/pyproject.toml | 2 +- argo_workflows/changelog.d/18207.added | 1 + argo_workflows/hatch.toml | 2 +- argo_workflows/pyproject.toml | 2 +- argocd/changelog.d/18207.added | 1 + argocd/hatch.toml | 2 +- argocd/pyproject.toml | 2 +- aspdotnet/changelog.d/18207.added | 1 + aspdotnet/hatch.toml | 2 +- aspdotnet/pyproject.toml | 2 +- avi_vantage/changelog.d/18207.added | 1 + avi_vantage/hatch.toml | 2 +- avi_vantage/pyproject.toml | 2 +- aws_neuron/hatch.toml | 2 +- aws_neuron/pyproject.toml | 2 +- azure_iot_edge/changelog.d/18207.added | 1 + azure_iot_edge/hatch.toml | 2 +- azure_iot_edge/pyproject.toml | 2 +- boundary/changelog.d/18207.added | 1 + boundary/hatch.toml | 2 +- boundary/pyproject.toml | 2 +- btrfs/changelog.d/18207.added | 1 + btrfs/hatch.toml | 2 +- btrfs/pyproject.toml | 2 +- cacti/changelog.d/18207.added | 1 + cacti/hatch.toml | 2 +- cacti/pyproject.toml | 2 +- calico/changelog.d/18207.added | 1 + calico/hatch.toml | 2 +- calico/pyproject.toml | 2 +- cassandra/changelog.d/18207.added | 1 + cassandra/hatch.toml | 2 +- cassandra/pyproject.toml | 2 +- cassandra_nodetool/changelog.d/18207.added | 1 + cassandra_nodetool/hatch.toml | 2 +- cassandra_nodetool/pyproject.toml | 2 +- ceph/changelog.d/18207.added | 1 + ceph/hatch.toml | 2 +- ceph/pyproject.toml | 2 +- cert_manager/changelog.d/18207.added | 1 + cert_manager/hatch.toml | 2 +- cert_manager/pyproject.toml | 2 +- cilium/changelog.d/18207.added | 1 + cilium/hatch.toml | 4 +-- cilium/pyproject.toml | 2 +- cisco_aci/changelog.d/18207.added | 1 + cisco_aci/hatch.toml | 2 +- cisco_aci/pyproject.toml | 2 +- citrix_hypervisor/changelog.d/18207.added | 1 + citrix_hypervisor/hatch.toml | 2 +- citrix_hypervisor/pyproject.toml | 2 +- clickhouse/changelog.d/18207.added | 1 + clickhouse/hatch.toml | 2 +- clickhouse/pyproject.toml | 2 +- cloud_foundry_api/changelog.d/18207.added | 1 + cloud_foundry_api/hatch.toml | 2 +- cloud_foundry_api/pyproject.toml | 2 +- cloudera/changelog.d/18207.added | 1 + cloudera/hatch.toml | 2 +- cloudera/pyproject.toml | 2 +- cockroachdb/changelog.d/18207.added | 1 + cockroachdb/hatch.toml | 2 +- cockroachdb/pyproject.toml | 2 +- confluent_platform/changelog.d/18207.added | 1 + confluent_platform/hatch.toml | 2 +- confluent_platform/pyproject.toml | 2 +- consul/changelog.d/18207.added | 1 + consul/hatch.toml | 2 +- consul/pyproject.toml | 2 +- coredns/changelog.d/18207.added | 1 + coredns/hatch.toml | 2 +- coredns/pyproject.toml | 2 +- couch/changelog.d/18207.added | 1 + couch/hatch.toml | 2 +- couch/pyproject.toml | 2 +- couchbase/changelog.d/18207.added | 1 + couchbase/hatch.toml | 2 +- couchbase/pyproject.toml | 2 +- crio/changelog.d/18207.added | 1 + crio/hatch.toml | 2 +- crio/pyproject.toml | 2 +- datadog_checks_base/changelog.d/18207.added | 1 + datadog_checks_base/hatch.toml | 2 +- datadog_checks_base/pyproject.toml | 2 +- .../changelog.d/18207.added | 1 + .../pyproject.toml | 2 +- datadog_checks_dev/changelog.d/18207.added | 1 + .../integration/check/{check_name}/hatch.toml | 2 +- .../check/{check_name}/pyproject.toml | 4 +-- .../integration/jmx/{check_name}/hatch.toml | 2 +- .../jmx/{check_name}/pyproject.toml | 4 +-- .../logs/{check_name}/pyproject.toml | 2 +- datadog_checks_dev/hatch.toml | 6 ++-- datadog_checks_dev/pyproject.toml | 2 +- .../changelog.d/18207.added | 1 + datadog_checks_downloader/hatch.toml | 2 +- datadog_checks_downloader/pyproject.toml | 2 +- datadog_cluster_agent/changelog.d/18207.added | 1 + datadog_cluster_agent/hatch.toml | 2 +- datadog_cluster_agent/pyproject.toml | 2 +- dcgm/changelog.d/18207.added | 1 + dcgm/hatch.toml | 2 +- dcgm/pyproject.toml | 2 +- ddev/changelog.d/18207.added | 1 + ddev/hatch.toml | 2 +- ddev/pyproject.toml | 6 ++-- ddev/src/ddev/repo/constants.py | 2 +- directory/changelog.d/18207.added | 1 + directory/hatch.toml | 2 +- directory/pyproject.toml | 2 +- disk/changelog.d/18207.added | 1 + disk/hatch.toml | 2 +- disk/pyproject.toml | 2 +- dns_check/changelog.d/18207.added | 1 + dns_check/hatch.toml | 2 +- dns_check/pyproject.toml | 2 +- docs/developer/setup.md | 12 +++---- dotnetclr/changelog.d/18207.added | 1 + dotnetclr/hatch.toml | 2 +- dotnetclr/pyproject.toml | 2 +- druid/changelog.d/18207.added | 1 + druid/hatch.toml | 2 +- druid/pyproject.toml | 2 +- ecs_fargate/changelog.d/18207.added | 1 + ecs_fargate/hatch.toml | 2 +- ecs_fargate/pyproject.toml | 2 +- eks_fargate/changelog.d/18207.added | 1 + eks_fargate/hatch.toml | 2 +- eks_fargate/pyproject.toml | 2 +- elastic/changelog.d/18207.added | 1 + elastic/hatch.toml | 4 +-- elastic/pyproject.toml | 2 +- envoy/changelog.d/18207.added | 1 + envoy/hatch.toml | 2 +- envoy/pyproject.toml | 2 +- esxi/changelog.d/18207.added | 1 + esxi/hatch.toml | 4 +-- esxi/pyproject.toml | 2 +- etcd/changelog.d/18207.added | 1 + etcd/hatch.toml | 2 +- etcd/pyproject.toml | 2 +- exchange_server/changelog.d/18207.added | 1 + exchange_server/hatch.toml | 2 +- exchange_server/pyproject.toml | 2 +- external_dns/changelog.d/18207.added | 1 + external_dns/hatch.toml | 2 +- external_dns/pyproject.toml | 2 +- flink/changelog.d/18207.added | 1 + flink/pyproject.toml | 2 +- fluentd/changelog.d/18207.added | 1 + fluentd/hatch.toml | 2 +- fluentd/pyproject.toml | 2 +- fluxcd/changelog.d/18207.added | 1 + fluxcd/hatch.toml | 2 +- fluxcd/pyproject.toml | 2 +- fly_io/hatch.toml | 4 +-- fly_io/pyproject.toml | 2 +- foundationdb/changelog.d/18207.added | 1 + foundationdb/hatch.toml | 4 +-- foundationdb/pyproject.toml | 2 +- gearmand/changelog.d/18207.added | 1 + gearmand/hatch.toml | 2 +- gearmand/pyproject.toml | 2 +- gitlab/changelog.d/18207.added | 1 + gitlab/hatch.toml | 2 +- gitlab/pyproject.toml | 2 +- gitlab_runner/changelog.d/18207.added | 1 + gitlab_runner/hatch.toml | 2 +- gitlab_runner/pyproject.toml | 2 +- glusterfs/changelog.d/18207.added | 1 + glusterfs/hatch.toml | 2 +- glusterfs/pyproject.toml | 2 +- go_expvar/changelog.d/18207.added | 1 + go_expvar/hatch.toml | 2 +- go_expvar/pyproject.toml | 2 +- gunicorn/changelog.d/18207.added | 1 + gunicorn/hatch.toml | 9 ++--- gunicorn/pyproject.toml | 2 +- haproxy/changelog.d/18207.added | 1 + haproxy/hatch.toml | 4 +-- haproxy/pyproject.toml | 2 +- harbor/changelog.d/18207.added | 1 + harbor/hatch.toml | 2 +- harbor/pyproject.toml | 2 +- hazelcast/changelog.d/18207.added | 1 + hazelcast/hatch.toml | 2 +- hdfs_datanode/changelog.d/18207.added | 1 + hdfs_datanode/hatch.toml | 2 +- hdfs_datanode/pyproject.toml | 2 +- hdfs_namenode/changelog.d/18207.added | 1 + hdfs_namenode/hatch.toml | 2 +- hdfs_namenode/pyproject.toml | 2 +- hive/changelog.d/18207.added | 1 + hive/hatch.toml | 2 +- hive/pyproject.toml | 2 +- hivemq/changelog.d/18207.added | 1 + hivemq/hatch.toml | 2 +- hivemq/pyproject.toml | 2 +- http_check/changelog.d/18207.added | 1 + http_check/hatch.toml | 2 +- http_check/pyproject.toml | 2 +- hudi/changelog.d/18207.added | 1 + hudi/hatch.toml | 2 +- hudi/pyproject.toml | 2 +- hyperv/changelog.d/18207.added | 1 + hyperv/hatch.toml | 2 +- hyperv/pyproject.toml | 2 +- ibm_ace/changelog.d/18207.added | 1 + ibm_ace/hatch.toml | 2 +- ibm_ace/pyproject.toml | 2 +- ibm_db2/changelog.d/18207.added | 1 + ibm_db2/hatch.toml | 2 +- ibm_db2/pyproject.toml | 2 +- ibm_i/changelog.d/18207.added | 1 + ibm_i/hatch.toml | 2 +- ibm_i/pyproject.toml | 2 +- ibm_mq/changelog.d/18207.added | 1 + ibm_mq/hatch.toml | 4 +-- ibm_mq/pyproject.toml | 2 +- ibm_was/changelog.d/18207.added | 1 + ibm_was/hatch.toml | 2 +- ibm_was/pyproject.toml | 2 +- ignite/changelog.d/18207.added | 1 + ignite/hatch.toml | 2 +- ignite/pyproject.toml | 2 +- iis/changelog.d/18207.added | 1 + iis/hatch.toml | 2 +- iis/pyproject.toml | 2 +- impala/changelog.d/18207.added | 1 + impala/hatch.toml | 2 +- impala/pyproject.toml | 2 +- istio/changelog.d/18207.added | 1 + istio/hatch.toml | 2 +- istio/pyproject.toml | 2 +- jboss_wildfly/changelog.d/18207.added | 1 + jboss_wildfly/hatch.toml | 2 +- jboss_wildfly/pyproject.toml | 2 +- journald/changelog.d/18207.added | 1 + journald/pyproject.toml | 2 +- kafka/changelog.d/18207.added | 1 + kafka/hatch.toml | 2 +- kafka/pyproject.toml | 2 +- kafka_consumer/changelog.d/18207.added | 1 + kafka_consumer/hatch.toml | 4 +-- kafka_consumer/pyproject.toml | 2 +- karpenter/changelog.d/18207.added | 1 + karpenter/hatch.toml | 2 +- karpenter/pyproject.toml | 2 +- kong/changelog.d/18207.added | 1 + kong/hatch.toml | 2 +- kong/pyproject.toml | 2 +- .../changelog.d/18207.added | 1 + kube_apiserver_metrics/hatch.toml | 2 +- kube_apiserver_metrics/pyproject.toml | 2 +- .../changelog.d/18207.added | 1 + kube_controller_manager/hatch.toml | 2 +- kube_controller_manager/pyproject.toml | 2 +- kube_dns/changelog.d/18207.added | 1 + kube_dns/hatch.toml | 2 +- kube_dns/pyproject.toml | 2 +- kube_metrics_server/changelog.d/18207.added | 1 + kube_metrics_server/hatch.toml | 2 +- kube_metrics_server/pyproject.toml | 2 +- kube_proxy/changelog.d/18207.added | 1 + kube_proxy/hatch.toml | 2 +- kube_proxy/pyproject.toml | 2 +- kube_scheduler/changelog.d/18207.added | 1 + kube_scheduler/hatch.toml | 2 +- kube_scheduler/pyproject.toml | 2 +- kubeflow/pyproject.toml | 2 +- kubelet/changelog.d/18207.added | 1 + kubelet/hatch.toml | 2 +- kubelet/pyproject.toml | 2 +- .../changelog.d/18207.added | 1 + kubernetes_cluster_autoscaler/hatch.toml | 2 +- kubernetes_cluster_autoscaler/pyproject.toml | 2 +- kubernetes_state/changelog.d/18207.added | 1 + kubernetes_state/hatch.toml | 2 +- kubernetes_state/pyproject.toml | 2 +- kubevirt_api/hatch.toml | 2 +- kubevirt_api/pyproject.toml | 2 +- kubevirt_controller/hatch.toml | 2 +- kubevirt_controller/pyproject.toml | 2 +- kyototycoon/changelog.d/18207.added | 1 + kyototycoon/hatch.toml | 2 +- kyototycoon/pyproject.toml | 2 +- kyverno/changelog.d/18207.added | 1 + kyverno/hatch.toml | 2 +- kyverno/pyproject.toml | 2 +- lighttpd/changelog.d/18207.added | 1 + lighttpd/hatch.toml | 2 +- lighttpd/pyproject.toml | 2 +- linkerd/changelog.d/18207.added | 1 + linkerd/hatch.toml | 2 +- linkerd/pyproject.toml | 2 +- linux_proc_extras/changelog.d/18207.added | 1 + linux_proc_extras/hatch.toml | 2 +- linux_proc_extras/pyproject.toml | 2 +- mapr/changelog.d/18207.added | 1 + mapr/hatch.toml | 2 +- mapr/pyproject.toml | 2 +- mapreduce/changelog.d/18207.added | 1 + mapreduce/hatch.toml | 2 +- mapreduce/pyproject.toml | 2 +- marathon/changelog.d/18207.added | 1 + marathon/hatch.toml | 2 +- marathon/pyproject.toml | 2 +- marklogic/changelog.d/18207.added | 1 + marklogic/hatch.toml | 2 +- marklogic/pyproject.toml | 2 +- mcache/changelog.d/18207.added | 1 + mcache/hatch.toml | 2 +- mcache/pyproject.toml | 2 +- mesos_master/changelog.d/18207.added | 1 + mesos_master/hatch.toml | 2 +- mesos_master/pyproject.toml | 2 +- mesos_slave/changelog.d/18207.added | 1 + mesos_slave/hatch.toml | 2 +- mesos_slave/pyproject.toml | 2 +- mongo/changelog.d/18207.added | 1 + mongo/hatch.toml | 2 +- mongo/pyproject.toml | 2 +- mysql/changelog.d/18207.added | 1 + mysql/hatch.toml | 6 ++-- mysql/pyproject.toml | 2 +- nagios/changelog.d/18207.added | 1 + nagios/hatch.toml | 2 +- nagios/pyproject.toml | 2 +- network/changelog.d/18207.added | 1 + network/hatch.toml | 2 +- network/pyproject.toml | 2 +- nfsstat/changelog.d/18207.added | 1 + nfsstat/hatch.toml | 2 +- nfsstat/pyproject.toml | 2 +- nginx/changelog.d/18207.added | 1 + nginx/hatch.toml | 2 +- nginx/pyproject.toml | 2 +- .../changelog.d/18207.added | 1 + nginx_ingress_controller/hatch.toml | 2 +- nginx_ingress_controller/pyproject.toml | 2 +- nvidia_triton/changelog.d/18207.added | 1 + nvidia_triton/hatch.toml | 2 +- nvidia_triton/pyproject.toml | 2 +- openldap/changelog.d/18207.added | 1 + openldap/hatch.toml | 2 +- openldap/pyproject.toml | 2 +- openmetrics/changelog.d/18207.added | 1 + openmetrics/hatch.toml | 2 +- openmetrics/pyproject.toml | 2 +- openstack/changelog.d/18207.added | 1 + openstack/hatch.toml | 2 +- openstack/pyproject.toml | 2 +- openstack_controller/changelog.d/18207.added | 1 + openstack_controller/hatch.toml | 10 +++--- openstack_controller/pyproject.toml | 2 +- oracle/changelog.d/18207.added | 1 + oracle/pyproject.toml | 2 +- ossec_security/pyproject.toml | 2 +- pan_firewall/changelog.d/18207.added | 1 + pan_firewall/pyproject.toml | 2 +- pdh_check/changelog.d/18207.added | 1 + pdh_check/hatch.toml | 2 +- pdh_check/pyproject.toml | 2 +- pgbouncer/changelog.d/18207.added | 1 + pgbouncer/hatch.toml | 2 +- pgbouncer/pyproject.toml | 2 +- php_fpm/changelog.d/18207.added | 1 + php_fpm/hatch.toml | 2 +- php_fpm/pyproject.toml | 2 +- ping_federate/pyproject.toml | 2 +- postfix/changelog.d/18207.added | 1 + postfix/hatch.toml | 2 +- postfix/pyproject.toml | 2 +- postgres/changelog.d/18207.added | 1 + postgres/hatch.toml | 2 +- postgres/pyproject.toml | 2 +- powerdns_recursor/changelog.d/18207.added | 1 + powerdns_recursor/hatch.toml | 2 +- powerdns_recursor/pyproject.toml | 2 +- presto/changelog.d/18207.added | 1 + presto/hatch.toml | 2 +- presto/pyproject.toml | 2 +- process/changelog.d/18207.added | 1 + process/hatch.toml | 2 +- process/pyproject.toml | 2 +- prometheus/changelog.d/18207.added | 1 + prometheus/hatch.toml | 2 +- prometheus/pyproject.toml | 2 +- proxysql/changelog.d/18207.added | 1 + proxysql/hatch.toml | 2 +- proxysql/pyproject.toml | 2 +- pulsar/changelog.d/18207.added | 1 + pulsar/hatch.toml | 2 +- pulsar/pyproject.toml | 2 +- rabbitmq/changelog.d/18207.added | 1 + rabbitmq/hatch.toml | 2 +- rabbitmq/pyproject.toml | 2 +- ray/changelog.d/18207.added | 1 + ray/hatch.toml | 2 +- ray/pyproject.toml | 2 +- redisdb/changelog.d/18207.added | 1 + redisdb/hatch.toml | 2 +- redisdb/pyproject.toml | 2 +- rethinkdb/changelog.d/18207.added | 1 + rethinkdb/hatch.toml | 2 +- rethinkdb/pyproject.toml | 2 +- riak/changelog.d/18207.added | 1 + riak/hatch.toml | 2 +- riak/pyproject.toml | 2 +- riakcs/changelog.d/18207.added | 1 + riakcs/hatch.toml | 2 +- riakcs/pyproject.toml | 2 +- sap_hana/changelog.d/18207.added | 1 + sap_hana/hatch.toml | 2 +- sap_hana/pyproject.toml | 2 +- scylla/changelog.d/18207.added | 1 + scylla/hatch.toml | 2 +- scylla/pyproject.toml | 2 +- sidekiq/changelog.d/18207.added | 1 + sidekiq/pyproject.toml | 2 +- silk/changelog.d/18207.added | 1 + silk/hatch.toml | 2 +- silk/pyproject.toml | 2 +- singlestore/changelog.d/18207.added | 1 + singlestore/hatch.toml | 2 +- singlestore/pyproject.toml | 2 +- snmp/changelog.d/18207.added | 1 + snmp/hatch.toml | 2 +- snmp/pyproject.toml | 2 +- snowflake/changelog.d/18207.added | 1 + snowflake/hatch.toml | 2 +- snowflake/pyproject.toml | 2 +- solr/changelog.d/18207.added | 1 + solr/hatch.toml | 2 +- solr/pyproject.toml | 2 +- sonarqube/changelog.d/18207.added | 1 + sonarqube/hatch.toml | 2 +- sonarqube/pyproject.toml | 2 +- spark/changelog.d/18207.added | 1 + spark/hatch.toml | 2 +- spark/pyproject.toml | 2 +- sqlserver/changelog.d/18207.added | 1 + sqlserver/hatch.toml | 4 +-- sqlserver/pyproject.toml | 2 +- squid/changelog.d/18207.added | 1 + squid/hatch.toml | 2 +- squid/pyproject.toml | 2 +- ssh_check/changelog.d/18207.added | 1 + ssh_check/hatch.toml | 2 +- ssh_check/pyproject.toml | 2 +- statsd/changelog.d/18207.added | 1 + statsd/hatch.toml | 2 +- statsd/pyproject.toml | 2 +- strimzi/changelog.d/18207.added | 1 + strimzi/hatch.toml | 2 +- strimzi/pyproject.toml | 2 +- supervisord/changelog.d/18207.added | 1 + supervisord/hatch.toml | 2 +- supervisord/pyproject.toml | 2 +- suricata/pyproject.toml | 2 +- system_core/changelog.d/18207.added | 1 + system_core/hatch.toml | 2 +- system_core/pyproject.toml | 2 +- system_swap/changelog.d/18207.added | 1 + system_swap/hatch.toml | 2 +- system_swap/pyproject.toml | 2 +- tcp_check/changelog.d/18207.added | 1 + tcp_check/hatch.toml | 2 +- tcp_check/pyproject.toml | 2 +- teamcity/changelog.d/18207.added | 1 + teamcity/hatch.toml | 2 +- teamcity/pyproject.toml | 2 +- tekton/changelog.d/18207.added | 1 + tekton/hatch.toml | 2 +- tekton/pyproject.toml | 2 +- teleport/changelog.d/18207.added | 1 + teleport/hatch.toml | 2 +- teleport/pyproject.toml | 2 +- temporal/changelog.d/18207.added | 1 + temporal/hatch.toml | 2 +- temporal/pyproject.toml | 2 +- tenable/changelog.d/18207.added | 1 + tenable/pyproject.toml | 2 +- teradata/changelog.d/18207.added | 1 + teradata/hatch.toml | 6 ++-- teradata/pyproject.toml | 2 +- tibco_ems/hatch.toml | 2 +- tibco_ems/pyproject.toml | 2 +- tls/changelog.d/18207.added | 1 + tls/hatch.toml | 2 +- tls/pyproject.toml | 2 +- tomcat/changelog.d/18207.added | 1 + tomcat/hatch.toml | 2 +- tomcat/pyproject.toml | 2 +- torchserve/changelog.d/18207.added | 1 + torchserve/hatch.toml | 2 +- torchserve/pyproject.toml | 2 +- traefik_mesh/changelog.d/18207.added | 1 + traefik_mesh/hatch.toml | 2 +- traefik_mesh/pyproject.toml | 2 +- traffic_server/changelog.d/18207.added | 1 + traffic_server/hatch.toml | 2 +- traffic_server/pyproject.toml | 2 +- twemproxy/changelog.d/18207.added | 1 + twemproxy/hatch.toml | 2 +- twemproxy/pyproject.toml | 2 +- twistlock/changelog.d/18207.added | 1 + twistlock/hatch.toml | 2 +- twistlock/pyproject.toml | 2 +- varnish/changelog.d/18207.added | 1 + varnish/hatch.toml | 2 +- varnish/pyproject.toml | 2 +- vault/changelog.d/18207.added | 1 + vault/hatch.toml | 2 +- vault/pyproject.toml | 2 +- vertica/changelog.d/18207.added | 1 + vertica/hatch.toml | 2 +- vertica/pyproject.toml | 2 +- vllm/changelog.d/18207.added | 1 + vllm/hatch.toml | 2 +- vllm/pyproject.toml | 2 +- voltdb/changelog.d/18207.added | 1 + voltdb/hatch.toml | 2 +- voltdb/pyproject.toml | 2 +- vsphere/changelog.d/18207.added | 1 + vsphere/hatch.toml | 2 +- vsphere/pyproject.toml | 2 +- weaviate/changelog.d/18207.added | 1 + weaviate/hatch.toml | 2 +- weaviate/pyproject.toml | 2 +- weblogic/changelog.d/18207.added | 1 + weblogic/hatch.toml | 2 +- weblogic/pyproject.toml | 2 +- win32_event_log/changelog.d/18207.added | 1 + win32_event_log/hatch.toml | 2 +- win32_event_log/pyproject.toml | 2 +- .../changelog.d/18207.added | 1 + windows_performance_counters/hatch.toml | 2 +- windows_performance_counters/pyproject.toml | 2 +- windows_service/changelog.d/18207.added | 1 + windows_service/hatch.toml | 2 +- windows_service/pyproject.toml | 2 +- wmi_check/changelog.d/18207.added | 1 + wmi_check/hatch.toml | 2 +- wmi_check/pyproject.toml | 2 +- yarn/changelog.d/18207.added | 1 + yarn/hatch.toml | 2 +- yarn/pyproject.toml | 2 +- zk/changelog.d/18207.added | 1 + zk/hatch.toml | 2 +- zk/pyproject.toml | 2 +- 602 files changed, 696 insertions(+), 459 deletions(-) create mode 100644 active_directory/changelog.d/18207.added create mode 100644 activemq/changelog.d/18207.added create mode 100644 activemq_xml/changelog.d/18207.added create mode 100644 aerospike/changelog.d/18207.added create mode 100644 airflow/changelog.d/18207.added create mode 100644 amazon_msk/changelog.d/18207.added create mode 100644 ambari/changelog.d/18207.added create mode 100644 apache/changelog.d/18207.added create mode 100644 arangodb/changelog.d/18207.added create mode 100644 argo_rollouts/changelog.d/18207.added create mode 100644 argo_workflows/changelog.d/18207.added create mode 100644 argocd/changelog.d/18207.added create mode 100644 aspdotnet/changelog.d/18207.added create mode 100644 avi_vantage/changelog.d/18207.added create mode 100644 azure_iot_edge/changelog.d/18207.added create mode 100644 boundary/changelog.d/18207.added create mode 100644 btrfs/changelog.d/18207.added create mode 100644 cacti/changelog.d/18207.added create mode 100644 calico/changelog.d/18207.added create mode 100644 cassandra/changelog.d/18207.added create mode 100644 cassandra_nodetool/changelog.d/18207.added create mode 100644 ceph/changelog.d/18207.added create mode 100644 cert_manager/changelog.d/18207.added create mode 100644 cilium/changelog.d/18207.added create mode 100644 cisco_aci/changelog.d/18207.added create mode 100644 citrix_hypervisor/changelog.d/18207.added create mode 100644 clickhouse/changelog.d/18207.added create mode 100644 cloud_foundry_api/changelog.d/18207.added create mode 100644 cloudera/changelog.d/18207.added create mode 100644 cockroachdb/changelog.d/18207.added create mode 100644 confluent_platform/changelog.d/18207.added create mode 100644 consul/changelog.d/18207.added create mode 100644 coredns/changelog.d/18207.added create mode 100644 couch/changelog.d/18207.added create mode 100644 couchbase/changelog.d/18207.added create mode 100644 crio/changelog.d/18207.added create mode 100644 datadog_checks_base/changelog.d/18207.added create mode 100644 datadog_checks_dependency_provider/changelog.d/18207.added create mode 100644 datadog_checks_dev/changelog.d/18207.added create mode 100644 datadog_checks_downloader/changelog.d/18207.added create mode 100644 datadog_cluster_agent/changelog.d/18207.added create mode 100644 dcgm/changelog.d/18207.added create mode 100644 ddev/changelog.d/18207.added create mode 100644 directory/changelog.d/18207.added create mode 100644 disk/changelog.d/18207.added create mode 100644 dns_check/changelog.d/18207.added create mode 100644 dotnetclr/changelog.d/18207.added create mode 100644 druid/changelog.d/18207.added create mode 100644 ecs_fargate/changelog.d/18207.added create mode 100644 eks_fargate/changelog.d/18207.added create mode 100644 elastic/changelog.d/18207.added create mode 100644 envoy/changelog.d/18207.added create mode 100644 esxi/changelog.d/18207.added create mode 100644 etcd/changelog.d/18207.added create mode 100644 exchange_server/changelog.d/18207.added create mode 100644 external_dns/changelog.d/18207.added create mode 100644 flink/changelog.d/18207.added create mode 100644 fluentd/changelog.d/18207.added create mode 100644 fluxcd/changelog.d/18207.added create mode 100644 foundationdb/changelog.d/18207.added create mode 100644 gearmand/changelog.d/18207.added create mode 100644 gitlab/changelog.d/18207.added create mode 100644 gitlab_runner/changelog.d/18207.added create mode 100644 glusterfs/changelog.d/18207.added create mode 100644 go_expvar/changelog.d/18207.added create mode 100644 gunicorn/changelog.d/18207.added create mode 100644 haproxy/changelog.d/18207.added create mode 100644 harbor/changelog.d/18207.added create mode 100644 hazelcast/changelog.d/18207.added create mode 100644 hdfs_datanode/changelog.d/18207.added create mode 100644 hdfs_namenode/changelog.d/18207.added create mode 100644 hive/changelog.d/18207.added create mode 100644 hivemq/changelog.d/18207.added create mode 100644 http_check/changelog.d/18207.added create mode 100644 hudi/changelog.d/18207.added create mode 100644 hyperv/changelog.d/18207.added create mode 100644 ibm_ace/changelog.d/18207.added create mode 100644 ibm_db2/changelog.d/18207.added create mode 100644 ibm_i/changelog.d/18207.added create mode 100644 ibm_mq/changelog.d/18207.added create mode 100644 ibm_was/changelog.d/18207.added create mode 100644 ignite/changelog.d/18207.added create mode 100644 iis/changelog.d/18207.added create mode 100644 impala/changelog.d/18207.added create mode 100644 istio/changelog.d/18207.added create mode 100644 jboss_wildfly/changelog.d/18207.added create mode 100644 journald/changelog.d/18207.added create mode 100644 kafka/changelog.d/18207.added create mode 100644 kafka_consumer/changelog.d/18207.added create mode 100644 karpenter/changelog.d/18207.added create mode 100644 kong/changelog.d/18207.added create mode 100644 kube_apiserver_metrics/changelog.d/18207.added create mode 100644 kube_controller_manager/changelog.d/18207.added create mode 100644 kube_dns/changelog.d/18207.added create mode 100644 kube_metrics_server/changelog.d/18207.added create mode 100644 kube_proxy/changelog.d/18207.added create mode 100644 kube_scheduler/changelog.d/18207.added create mode 100644 kubelet/changelog.d/18207.added create mode 100644 kubernetes_cluster_autoscaler/changelog.d/18207.added create mode 100644 kubernetes_state/changelog.d/18207.added create mode 100644 kyototycoon/changelog.d/18207.added create mode 100644 kyverno/changelog.d/18207.added create mode 100644 lighttpd/changelog.d/18207.added create mode 100644 linkerd/changelog.d/18207.added create mode 100644 linux_proc_extras/changelog.d/18207.added create mode 100644 mapr/changelog.d/18207.added create mode 100644 mapreduce/changelog.d/18207.added create mode 100644 marathon/changelog.d/18207.added create mode 100644 marklogic/changelog.d/18207.added create mode 100644 mcache/changelog.d/18207.added create mode 100644 mesos_master/changelog.d/18207.added create mode 100644 mesos_slave/changelog.d/18207.added create mode 100644 mongo/changelog.d/18207.added create mode 100644 mysql/changelog.d/18207.added create mode 100644 nagios/changelog.d/18207.added create mode 100644 network/changelog.d/18207.added create mode 100644 nfsstat/changelog.d/18207.added create mode 100644 nginx/changelog.d/18207.added create mode 100644 nginx_ingress_controller/changelog.d/18207.added create mode 100644 nvidia_triton/changelog.d/18207.added create mode 100644 openldap/changelog.d/18207.added create mode 100644 openmetrics/changelog.d/18207.added create mode 100644 openstack/changelog.d/18207.added create mode 100644 openstack_controller/changelog.d/18207.added create mode 100644 oracle/changelog.d/18207.added create mode 100644 pan_firewall/changelog.d/18207.added create mode 100644 pdh_check/changelog.d/18207.added create mode 100644 pgbouncer/changelog.d/18207.added create mode 100644 php_fpm/changelog.d/18207.added create mode 100644 postfix/changelog.d/18207.added create mode 100644 postgres/changelog.d/18207.added create mode 100644 powerdns_recursor/changelog.d/18207.added create mode 100644 presto/changelog.d/18207.added create mode 100644 process/changelog.d/18207.added create mode 100644 prometheus/changelog.d/18207.added create mode 100644 proxysql/changelog.d/18207.added create mode 100644 pulsar/changelog.d/18207.added create mode 100644 rabbitmq/changelog.d/18207.added create mode 100644 ray/changelog.d/18207.added create mode 100644 redisdb/changelog.d/18207.added create mode 100644 rethinkdb/changelog.d/18207.added create mode 100644 riak/changelog.d/18207.added create mode 100644 riakcs/changelog.d/18207.added create mode 100644 sap_hana/changelog.d/18207.added create mode 100644 scylla/changelog.d/18207.added create mode 100644 sidekiq/changelog.d/18207.added create mode 100644 silk/changelog.d/18207.added create mode 100644 singlestore/changelog.d/18207.added create mode 100644 snmp/changelog.d/18207.added create mode 100644 snowflake/changelog.d/18207.added create mode 100644 solr/changelog.d/18207.added create mode 100644 sonarqube/changelog.d/18207.added create mode 100644 spark/changelog.d/18207.added create mode 100644 sqlserver/changelog.d/18207.added create mode 100644 squid/changelog.d/18207.added create mode 100644 ssh_check/changelog.d/18207.added create mode 100644 statsd/changelog.d/18207.added create mode 100644 strimzi/changelog.d/18207.added create mode 100644 supervisord/changelog.d/18207.added create mode 100644 system_core/changelog.d/18207.added create mode 100644 system_swap/changelog.d/18207.added create mode 100644 tcp_check/changelog.d/18207.added create mode 100644 teamcity/changelog.d/18207.added create mode 100644 tekton/changelog.d/18207.added create mode 100644 teleport/changelog.d/18207.added create mode 100644 temporal/changelog.d/18207.added create mode 100644 tenable/changelog.d/18207.added create mode 100644 teradata/changelog.d/18207.added create mode 100644 tls/changelog.d/18207.added create mode 100644 tomcat/changelog.d/18207.added create mode 100644 torchserve/changelog.d/18207.added create mode 100644 traefik_mesh/changelog.d/18207.added create mode 100644 traffic_server/changelog.d/18207.added create mode 100644 twemproxy/changelog.d/18207.added create mode 100644 twistlock/changelog.d/18207.added create mode 100644 varnish/changelog.d/18207.added create mode 100644 vault/changelog.d/18207.added create mode 100644 vertica/changelog.d/18207.added create mode 100644 vllm/changelog.d/18207.added create mode 100644 voltdb/changelog.d/18207.added create mode 100644 vsphere/changelog.d/18207.added create mode 100644 weaviate/changelog.d/18207.added create mode 100644 weblogic/changelog.d/18207.added create mode 100644 win32_event_log/changelog.d/18207.added create mode 100644 windows_performance_counters/changelog.d/18207.added create mode 100644 windows_service/changelog.d/18207.added create mode 100644 wmi_check/changelog.d/18207.added create mode 100644 yarn/changelog.d/18207.added create mode 100644 zk/changelog.d/18207.added diff --git a/.builders/images/linux-aarch64/Dockerfile b/.builders/images/linux-aarch64/Dockerfile index eca90a702cc87..29a7795e968a6 100644 --- a/.builders/images/linux-aarch64/Dockerfile +++ b/.builders/images/linux-aarch64/Dockerfile @@ -53,11 +53,11 @@ RUN yum install -y perl-IPC-Cmd && \ ldconfig # Compile and install Python 3 -ENV PYTHON3_VERSION=3.11.8 +ENV PYTHON3_VERSION=3.12.6 RUN yum install -y libffi-devel && \ DOWNLOAD_URL="https://python.org/ftp/python/{{version}}/Python-{{version}}.tgz" \ VERSION="${PYTHON3_VERSION}" \ - SHA256="d3019a613b9e8761d260d9ebe3bd4df63976de30464e5c0189566e1ae3f61889" \ + SHA256="85a4c1be906d20e5c5a69f2466b00da769c221d6a684acfd3a514dbf5bf10a66" \ RELATIVE_PATH="Python-{{version}}" \ bash install-from-source.sh \ --prefix=/opt/python/${PYTHON_VERSION} \ diff --git a/.builders/images/linux-x86_64/Dockerfile b/.builders/images/linux-x86_64/Dockerfile index 0ea7bf99315d6..6c87ac897dc5e 100644 --- a/.builders/images/linux-x86_64/Dockerfile +++ b/.builders/images/linux-x86_64/Dockerfile @@ -51,11 +51,11 @@ RUN yum install -y perl-IPC-Cmd && \ ldconfig # Compile and install Python 3 -ENV PYTHON3_VERSION=3.11.8 +ENV PYTHON3_VERSION=3.12.6 RUN yum install -y libffi-devel && \ DOWNLOAD_URL="https://python.org/ftp/python/{{version}}/Python-{{version}}.tgz" \ VERSION="${PYTHON3_VERSION}" \ - SHA256="d3019a613b9e8761d260d9ebe3bd4df63976de30464e5c0189566e1ae3f61889" \ + SHA256="85a4c1be906d20e5c5a69f2466b00da769c221d6a684acfd3a514dbf5bf10a66" \ RELATIVE_PATH="Python-{{version}}" \ bash install-from-source.sh --prefix=/opt/python/${PYTHON_VERSION} --with-ensurepip=yes --enable-ipv6 --with-dbmliborder= ENV PATH="/opt/python/${PYTHON_VERSION}/bin:${PATH}" @@ -81,6 +81,36 @@ RUN \ RELATIVE_PATH="krb5-{{version}}/src" \ bash install-from-source.sh --without-keyutils --without-system-verto --without-libedit --disable-static +# libxml & libxslt for lxml +RUN \ + DOWNLOAD_URL="https://download.gnome.org/sources/libxml2/2.12/libxml2-{{version}}.tar.xz" \ + VERSION="2.12.6" \ + SHA256="889c593a881a3db5fdd96cc9318c87df34eb648edfc458272ad46fd607353fbb" \ + RELATIVE_PATH="libxml2-{{version}}" \ + bash install-from-source.sh \ + --without-iconv \ + --without-python \ + --without-icu \ + --without-debug \ + --without-mem-debug \ + --without-run-debug \ + --without-legacy \ + --without-catalog \ + --without-docbook \ + --disable-static + +RUN \ + DOWNLOAD_URL="https://download.gnome.org/sources/libxslt/1.1/libxslt-{{version}}.tar.xz" \ + VERSION="1.1.39" \ + SHA256="2a20ad621148339b0759c4d4e96719362dee64c9a096dbba625ba053846349f0" \ + RELATIVE_PATH="libxslt-{{version}}" \ + bash install-from-source.sh \ + --without-python \ + --without-crypto \ + --without-profiler \ + --without-debugger \ + --disable-static + # libpq and pg_config as needed by psycopg2 RUN \ DOWNLOAD_URL="https://ftp.postgresql.org/pub/source/v{{version}}/postgresql-{{version}}.tar.bz2" \ diff --git a/.builders/images/windows-x86_64/Dockerfile b/.builders/images/windows-x86_64/Dockerfile index 8838692733e27..03c3c80da91fb 100644 --- a/.builders/images/windows-x86_64/Dockerfile +++ b/.builders/images/windows-x86_64/Dockerfile @@ -73,17 +73,17 @@ RUN Get-RemoteFile ` Approve-File -Path $($Env:USERPROFILE + '\.cargo\bin\rustc.exe') -Hash $Env:RUSTC_HASH # Install Python 3 -ENV PYTHON_VERSION="3.11.7" +ENV PYTHON_VERSION="3.12.6" RUN Get-RemoteFile ` -Uri https://www.python.org/ftp/python/$Env:PYTHON_VERSION/python-$Env:PYTHON_VERSION-amd64.exe ` -Path python-$Env:PYTHON_VERSION-amd64.exe ` - -Hash 'c117c6444494bbe4cc937e8a5a61899d53f7f5c5bc573c5d130304e457d54024'; ` + -Hash '5914748e6580e70bedeb7c537a0832b3071de9e09a2e4e7e3d28060616045e0a'; ` Start-Process -Wait python-$Env:PYTHON_VERSION-amd64.exe -ArgumentList '/quiet', 'InstallAllUsers=1'; ` Remove-Item python-$Env:PYTHON_VERSION-amd64.exe; ` - & 'C:\Program Files\Python311\python.exe' -m pip install --no-warn-script-location --upgrade pip; ` - & 'C:\Program Files\Python311\python.exe' -m pip install --no-warn-script-location virtualenv; ` - & 'C:\Program Files\Python311\python.exe' -m virtualenv 'C:\py3'; ` - Add-ToPath -Append 'C:\Program Files\Python311' + & 'C:\Program Files\Python312\python.exe' -m pip install --no-warn-script-location --upgrade pip; ` + & 'C:\Program Files\Python312\python.exe' -m pip install --no-warn-script-location virtualenv; ` + & 'C:\Program Files\Python312\python.exe' -m virtualenv 'C:\py3'; ` + Add-ToPath -Append 'C:\Program Files\Python312' # Install Python 2 ENV PYTHON_VERSION="2.7.18" diff --git a/.builders/lock.py b/.builders/lock.py index 6899054065121..52d5b1f731605 100644 --- a/.builders/lock.py +++ b/.builders/lock.py @@ -36,6 +36,11 @@ def default_python_version() -> str: return match.group(1) +@cache +def target_python_for_major(python_major: str): + return '2.7' if python_major == '2' else default_python_version() + + def is_compatible_wheel( target_name: str, target_python_major: str, @@ -44,7 +49,7 @@ def is_compatible_wheel( platform: str, ) -> bool: if interpreter.startswith('cp'): - target_python = '2.7' if target_python_major == '2' else default_python_version() + target_python = target_python_for_major(target_python_major) expected_tag = f'cp{target_python_major}' if abi == 'abi3' else f'cp{target_python}'.replace('.', '') if expected_tag not in interpreter: return False @@ -59,8 +64,17 @@ def is_compatible_wheel( return True -def generate_lock_file(requirements_file: Path, lock_file: Path) -> None: - target, _, python_version = lock_file.stem.rpartition('_') +def generate_lock_file( + requirements_file: Path, + lock_file_folder: Path, + target: str, + python_version: str, +) -> None: + python_target = target_python_for_major(python_version) + # The lockfiles contain the major.minor Python version + # so that the Agent can transition safely + lock_file = lock_file_folder / f'{target}_{python_target}.txt' + python_major = python_version[-1] dependencies: dict[str, str] = {} @@ -135,7 +149,10 @@ def main(): for python_version in target.iterdir(): if python_version.name.startswith('py'): generate_lock_file( - python_version / 'frozen.txt', LOCK_FILE_DIR / f'{target.name}_{python_version.name}.txt' + python_version / 'frozen.txt', + LOCK_FILE_DIR, + target.name, + python_version.name.strip('py'), ) if (image_digest_file := target / 'image_digest').is_file(): diff --git a/.devcontainer/Dockerfile b/.devcontainer/Dockerfile index 46b5d17f958be..01960e1f9e162 100644 --- a/.devcontainer/Dockerfile +++ b/.devcontainer/Dockerfile @@ -1,6 +1,6 @@ # This dockerfile is used to build the devcontainer environment. # more info about vscode devcontainer: https://code.visualstudio.com/docs/devcontainers/containers -FROM mcr.microsoft.com/devcontainers/python:1-3.11-bullseye +FROM mcr.microsoft.com/devcontainers/python:1-3.12-bullseye RUN apt update && DEBIAN_FRONTEND=noninteractive apt install -y liblz4-dev libunwind-dev ca-certificates curl gnupg # Docker and docker-compose installation RUN install -m 0755 -d /etc/apt/keyrings diff --git a/.devcontainer/dbm/Dockerfile b/.devcontainer/dbm/Dockerfile index 9fc50f45142b9..98a347726cfc2 100644 --- a/.devcontainer/dbm/Dockerfile +++ b/.devcontainer/dbm/Dockerfile @@ -1,6 +1,6 @@ # This dockerfile is used to build the devcontainer environment. # more info about vscode devcontainer: https://code.visualstudio.com/docs/devcontainers/containers -FROM mcr.microsoft.com/devcontainers/python:1-3.11-bullseye +FROM mcr.microsoft.com/devcontainers/python:1-3.12-bullseye RUN apt update && DEBIAN_FRONTEND=noninteractive apt install -y liblz4-dev libunwind-dev ca-certificates curl gnupg # Docker and docker-compose installation RUN install -m 0755 -d /etc/apt/keyrings diff --git a/.devcontainer/dbm/devcontainer.json b/.devcontainer/dbm/devcontainer.json index 2e841ce8f3231..f136629acad0e 100644 --- a/.devcontainer/dbm/devcontainer.json +++ b/.devcontainer/dbm/devcontainer.json @@ -8,7 +8,7 @@ "privileged": true, "features": { "ghcr.io/devcontainers/features/python:1": { - "version": "3.11" + "version": "3.12" } }, "customizations": { diff --git a/.devcontainer/devcontainer.json b/.devcontainer/devcontainer.json index 935588a034939..c3e36bc9a4802 100644 --- a/.devcontainer/devcontainer.json +++ b/.devcontainer/devcontainer.json @@ -16,7 +16,7 @@ // When updating it, modify both the base Dockerfile and the devcontainer.json reference. // Ref: https://github.com/devcontainers/features/blob/562305d37b97d47331d96306ffc2a0a3cce55e64/src/python/install.sh#L10 "ghcr.io/devcontainers/features/python:1": { - "version": "3.11" + "version": "3.12" } }, diff --git a/.github/workflows/build-ddev.yml b/.github/workflows/build-ddev.yml index 1019ec3bdd0b3..420238b467ecb 100644 --- a/.github/workflows/build-ddev.yml +++ b/.github/workflows/build-ddev.yml @@ -24,7 +24,7 @@ defaults: env: APP_NAME: ddev - PYTHON_VERSION: "3.11" + PYTHON_VERSION: "3.12" PYOXIDIZER_VERSION: "0.24.0" jobs: diff --git a/.github/workflows/build-deps.yml b/.github/workflows/build-deps.yml index 874989502850f..0052055a900e8 100644 --- a/.github/workflows/build-deps.yml +++ b/.github/workflows/build-deps.yml @@ -29,7 +29,7 @@ defaults: env: PYTHONUNBUFFERED: "1" - PYTHON_VERSION: "3.11" + PYTHON_VERSION: "3.12" DIRECT_DEPENDENCY_FILE: agent_requirements.in # https://reproducible-builds.org/specs/source-date-epoch/ SOURCE_DATE_EPOCH: "1580601600" @@ -109,7 +109,7 @@ jobs: if: matrix.job.image == 'linux-aarch64' run: | mkdir -p ~/miniconda3 - wget https://repo.anaconda.com/miniconda/Miniconda3-py311_24.1.2-0-Linux-aarch64.sh -O ~/miniconda3/miniconda.sh + wget https://repo.anaconda.com/miniconda/Miniconda3-py312_24.5.0-0-Linux-aarch64.sh -O ~/miniconda3/miniconda.sh bash ~/miniconda3/miniconda.sh -b -u -p ~/miniconda3 rm -rf ~/miniconda3/miniconda.sh ~/miniconda3/bin/conda init bash @@ -201,7 +201,7 @@ jobs: env: TARGET_NAME: macos-x86_64 OUT_DIR: output/macos-x86_64 - DD_PYTHON3: "/Library/Frameworks/Python.framework/Versions/3.11/bin/python" + DD_PYTHON3: "/Library/Frameworks/Python.framework/Versions/3.12/bin/python" steps: - name: Set up environment @@ -214,7 +214,7 @@ jobs: - name: Set up Python env: # Despite the name, this is built for the macOS 11 SDK on arm64 and 10.9+ on intel - PYTHON3_DOWNLOAD_URL: "https://www.python.org/ftp/python/3.11.5/python-3.11.5-macos11.pkg" + PYTHON3_DOWNLOAD_URL: "https://www.python.org/ftp/python/3.12.6/python-3.12.6-macos11.pkg" run: |- curl "$PYTHON3_DOWNLOAD_URL" -o python3.pkg sudo installer -pkg python3.pkg -target / diff --git a/.github/workflows/cache-shared-deps.yml b/.github/workflows/cache-shared-deps.yml index 05379e5a5ee52..c3dc212dd9fc7 100644 --- a/.github/workflows/cache-shared-deps.yml +++ b/.github/workflows/cache-shared-deps.yml @@ -16,7 +16,7 @@ jobs: os: [ubuntu-22.04, windows-2022] env: - PYTHON_VERSION: "3.11" + PYTHON_VERSION: "3.12" steps: - uses: actions/checkout@v4 diff --git a/.github/workflows/compute-matrix.yml b/.github/workflows/compute-matrix.yml index d2ca1a5625563..a85759012ab2d 100644 --- a/.github/workflows/compute-matrix.yml +++ b/.github/workflows/compute-matrix.yml @@ -15,7 +15,7 @@ defaults: shell: bash env: - PYTHON_VERSION: "3.11" + PYTHON_VERSION: "3.12" MATRIX_SCRIPT: "ddev/src/ddev/utils/scripts/ci_matrix.py" jobs: diff --git a/.github/workflows/docs.yml b/.github/workflows/docs.yml index b79a8b0d046b2..4113adce459bf 100644 --- a/.github/workflows/docs.yml +++ b/.github/workflows/docs.yml @@ -32,7 +32,7 @@ jobs: - name: Set up Python uses: actions/setup-python@v5 with: - python-version: '3.11' + python-version: '3.12' cache: 'pip' - name: Upgrade Python packaging tools diff --git a/.github/workflows/pr-quick-check.yml b/.github/workflows/pr-quick-check.yml index 4165b491ac5c8..40bf4f8e07cf4 100644 --- a/.github/workflows/pr-quick-check.yml +++ b/.github/workflows/pr-quick-check.yml @@ -12,7 +12,7 @@ defaults: shell: bash env: - PYTHON_VERSION: "3.11" + PYTHON_VERSION: "3.12" CHECK_SCRIPT: "ddev/src/ddev/utils/scripts/check_pr.py" jobs: diff --git a/.github/workflows/release-base.yml b/.github/workflows/release-base.yml index ad61dc4c7e715..3d32a198fdfb8 100644 --- a/.github/workflows/release-base.yml +++ b/.github/workflows/release-base.yml @@ -15,7 +15,7 @@ jobs: - name: Set up Python uses: actions/setup-python@v5 with: - python-version: '3.11' + python-version: '3.12' - name: Upgrade Python packaging tools run: pip install --disable-pip-version-check --upgrade pip setuptools wheel diff --git a/.github/workflows/release-dev.yml b/.github/workflows/release-dev.yml index 409c27bd412a0..59d77c92432dc 100644 --- a/.github/workflows/release-dev.yml +++ b/.github/workflows/release-dev.yml @@ -15,7 +15,7 @@ jobs: - name: Set up Python uses: actions/setup-python@v5 with: - python-version: '3.11' + python-version: '3.12' - name: Upgrade Python packaging tools run: pip install --disable-pip-version-check --upgrade pip setuptools wheel diff --git a/.github/workflows/release-hash-check.yml b/.github/workflows/release-hash-check.yml index 8eefbdf48b256..9bc603a3ed1d4 100644 --- a/.github/workflows/release-hash-check.yml +++ b/.github/workflows/release-hash-check.yml @@ -19,7 +19,7 @@ jobs: - name: Set up Python uses: actions/setup-python@v5 with: - python-version: '3.11' + python-version: '3.12' - id: files run: | diff --git a/.github/workflows/run-validations.yml b/.github/workflows/run-validations.yml index ca52dc1d5cda1..0d825da22215c 100644 --- a/.github/workflows/run-validations.yml +++ b/.github/workflows/run-validations.yml @@ -117,7 +117,7 @@ jobs: runs-on: ubuntu-22.04 env: - PYTHON_VERSION: "3.11" + PYTHON_VERSION: "3.12" TARGET: ${{ github.event_name == 'pull_request' && 'changed' || '' }} steps: diff --git a/.github/workflows/test-target.yml b/.github/workflows/test-target.yml index e126b9b00b720..b527a7ee878df 100644 --- a/.github/workflows/test-target.yml +++ b/.github/workflows/test-target.yml @@ -94,8 +94,8 @@ jobs: env: FORCE_COLOR: "1" - PYTHON_VERSION: "${{ inputs.python-version || '3.11' }}" - PYTHON_FILTER: "${{ (inputs.test-py2 && !inputs.test-py3) && '2.7' || (!inputs.test-py2 && inputs.test-py3) && (inputs.python-version || '3.11') || '' }}" + PYTHON_VERSION: "${{ inputs.python-version || '3.12' }}" + PYTHON_FILTER: "${{ (inputs.test-py2 && !inputs.test-py3) && '2.7' || (!inputs.test-py2 && inputs.test-py3) && (inputs.python-version || '3.12') || '' }}" SKIP_ENV_NAME: "${{ (inputs.test-py2 && !inputs.test-py3) && 'py3.*' || (!inputs.test-py2 && inputs.test-py3) && 'py2.*' || '' }}" # Windows E2E requires Windows containers DDEV_E2E_AGENT: "${{ inputs.platform == 'windows' && (inputs.agent-image-windows || 'datadog/agent-dev:master-py3-win-servercore') || inputs.agent-image }}" diff --git a/.github/workflows/update-agent-changelog.yml b/.github/workflows/update-agent-changelog.yml index eeafbd061ae3a..d426d183ca7ea 100644 --- a/.github/workflows/update-agent-changelog.yml +++ b/.github/workflows/update-agent-changelog.yml @@ -18,7 +18,7 @@ jobs: - name: Set up Python uses: actions/setup-python@v5 with: - python-version: "3.11" + python-version: 3.12 - name: Install ddev from local folder run: |- pip install -e ./datadog_checks_dev[cli] diff --git a/.github/workflows/update-dependencies.yml b/.github/workflows/update-dependencies.yml index a06135cc8c54c..78a78aa87a69e 100644 --- a/.github/workflows/update-dependencies.yml +++ b/.github/workflows/update-dependencies.yml @@ -18,7 +18,7 @@ jobs: - name: Set up Python uses: actions/setup-python@v5 with: - python-version: "3.11" + python-version: 3.12 - name: Install ddev from local folder run: |- pip install -e ./datadog_checks_dev[cli] diff --git a/.gitlab/tagger/Dockerfile b/.gitlab/tagger/Dockerfile index b0835e5b71c38..e6cc27674f73b 100644 --- a/.gitlab/tagger/Dockerfile +++ b/.gitlab/tagger/Dockerfile @@ -1,5 +1,5 @@ # Use a separate image to build ddev from source so the final image doesn't need git -FROM python:3.11 AS ddev-python +FROM python:3.12 AS ddev-python COPY . /home RUN python -m pip install build RUN python -m build -s ./home/ddev diff --git a/active_directory/changelog.d/18207.added b/active_directory/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/active_directory/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/active_directory/hatch.toml b/active_directory/hatch.toml index 63c94c51afff5..f9d790e2603e8 100644 --- a/active_directory/hatch.toml +++ b/active_directory/hatch.toml @@ -10,4 +10,4 @@ platforms = [ ] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] diff --git a/active_directory/pyproject.toml b/active_directory/pyproject.toml index 4c890ddf12e26..e3a3d3611e863 100644 --- a/active_directory/pyproject.toml +++ b/active_directory/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/activemq/changelog.d/18207.added b/activemq/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/activemq/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/activemq/hatch.toml b/activemq/hatch.toml index 4dfa61b7f1519..2194aa0cbf91c 100644 --- a/activemq/hatch.toml +++ b/activemq/hatch.toml @@ -1,12 +1,12 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] version = ["2.15.0"] compose-file = ["artemis.yaml"] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] version = ["5.15.9"] compose-file = ["activemq.yaml"] diff --git a/activemq/pyproject.toml b/activemq/pyproject.toml index b42281225f9dc..d2aad008b6ca4 100644 --- a/activemq/pyproject.toml +++ b/activemq/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/activemq_xml/changelog.d/18207.added b/activemq_xml/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/activemq_xml/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/activemq_xml/hatch.toml b/activemq_xml/hatch.toml index cde2e9940b124..15ccd88f34a1e 100644 --- a/activemq_xml/hatch.toml +++ b/activemq_xml/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] version = ["5.11.1"] [envs.default.overrides] diff --git a/activemq_xml/pyproject.toml b/activemq_xml/pyproject.toml index 6280e20a75d14..3c23903fb08c9 100644 --- a/activemq_xml/pyproject.toml +++ b/activemq_xml/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/aerospike/changelog.d/18207.added b/aerospike/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/aerospike/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/aerospike/hatch.toml b/aerospike/hatch.toml index 2e8eaf335531e..addb780e2e411 100644 --- a/aerospike/hatch.toml +++ b/aerospike/hatch.toml @@ -3,7 +3,7 @@ check-types = false [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] version = ["4", "5.0", "5.3", "5.6"] [envs.default.overrides] diff --git a/aerospike/pyproject.toml b/aerospike/pyproject.toml index ff847c5883619..78d14d3e5d834 100644 --- a/aerospike/pyproject.toml +++ b/aerospike/pyproject.toml @@ -24,7 +24,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/airflow/changelog.d/18207.added b/airflow/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/airflow/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/airflow/hatch.toml b/airflow/hatch.toml index 84ee67e30748c..31074a4c6ccf7 100644 --- a/airflow/hatch.toml +++ b/airflow/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] version = ["2.1", "2.6"] [envs.default.overrides] diff --git a/airflow/pyproject.toml b/airflow/pyproject.toml index 8a7582317bd24..9d883d928e13c 100644 --- a/airflow/pyproject.toml +++ b/airflow/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/amazon_msk/changelog.d/18207.added b/amazon_msk/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/amazon_msk/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/amazon_msk/hatch.toml b/amazon_msk/hatch.toml index 821e6ed8e37ca..f62423b83c762 100644 --- a/amazon_msk/hatch.toml +++ b/amazon_msk/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] [envs.default.env-vars] DDEV_SKIP_GENERIC_TAGS_CHECK = "true" diff --git a/amazon_msk/pyproject.toml b/amazon_msk/pyproject.toml index f47548b11d133..475ce4e7031de 100644 --- a/amazon_msk/pyproject.toml +++ b/amazon_msk/pyproject.toml @@ -24,7 +24,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/ambari/changelog.d/18207.added b/ambari/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/ambari/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/ambari/hatch.toml b/ambari/hatch.toml index 001e43ce25414..c85c5f07a7df2 100644 --- a/ambari/hatch.toml +++ b/ambari/hatch.toml @@ -1,4 +1,4 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] diff --git a/ambari/pyproject.toml b/ambari/pyproject.toml index 77806c16a9596..23c506e9b8a0c 100644 --- a/ambari/pyproject.toml +++ b/ambari/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/apache/changelog.d/18207.added b/apache/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/apache/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/apache/hatch.toml b/apache/hatch.toml index d57e8fb4781a8..9b487f51ed6c4 100644 --- a/apache/hatch.toml +++ b/apache/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] version = ["2.4.27"] [envs.default.overrides] diff --git a/apache/pyproject.toml b/apache/pyproject.toml index 977063e134e8c..c5eb9434f534e 100644 --- a/apache/pyproject.toml +++ b/apache/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/arangodb/changelog.d/18207.added b/arangodb/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/arangodb/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/arangodb/hatch.toml b/arangodb/hatch.toml index 63797eda67bf6..cf7caf8f2678b 100644 --- a/arangodb/hatch.toml +++ b/arangodb/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] version = ["3.8"] [envs.default.overrides] diff --git a/arangodb/pyproject.toml b/arangodb/pyproject.toml index 77918dd730a3a..58702df689362 100644 --- a/arangodb/pyproject.toml +++ b/arangodb/pyproject.toml @@ -24,7 +24,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/argo_rollouts/changelog.d/18207.added b/argo_rollouts/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/argo_rollouts/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/argo_rollouts/hatch.toml b/argo_rollouts/hatch.toml index 777aac7b7ec9e..e39921cdb93c2 100644 --- a/argo_rollouts/hatch.toml +++ b/argo_rollouts/hatch.toml @@ -4,7 +4,7 @@ DDEV_SKIP_GENERIC_TAGS_CHECK = "true" [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] version = ["1.6.6"] [envs.default.overrides] diff --git a/argo_rollouts/pyproject.toml b/argo_rollouts/pyproject.toml index 219f13ff22d36..d05e7a5c0c01b 100644 --- a/argo_rollouts/pyproject.toml +++ b/argo_rollouts/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", "Private :: Do Not Upload", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", ] dependencies = [ diff --git a/argo_workflows/changelog.d/18207.added b/argo_workflows/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/argo_workflows/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/argo_workflows/hatch.toml b/argo_workflows/hatch.toml index 001e43ce25414..c85c5f07a7df2 100644 --- a/argo_workflows/hatch.toml +++ b/argo_workflows/hatch.toml @@ -1,4 +1,4 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] diff --git a/argo_workflows/pyproject.toml b/argo_workflows/pyproject.toml index b806ec9b49e37..42bcc168aba65 100644 --- a/argo_workflows/pyproject.toml +++ b/argo_workflows/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", "Private :: Do Not Upload", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", ] dependencies = [ diff --git a/argocd/changelog.d/18207.added b/argocd/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/argocd/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/argocd/hatch.toml b/argocd/hatch.toml index 6c5cba84584d7..e31ecbafc6d6f 100644 --- a/argocd/hatch.toml +++ b/argocd/hatch.toml @@ -4,7 +4,7 @@ DDEV_SKIP_GENERIC_TAGS_CHECK = "true" [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] version = ["2.4.7"] [envs.default.overrides] diff --git a/argocd/pyproject.toml b/argocd/pyproject.toml index 0d6a92f5dcb11..ce1cd506cc3f2 100644 --- a/argocd/pyproject.toml +++ b/argocd/pyproject.toml @@ -24,7 +24,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", ] dependencies = [ diff --git a/aspdotnet/changelog.d/18207.added b/aspdotnet/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/aspdotnet/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/aspdotnet/hatch.toml b/aspdotnet/hatch.toml index 63c94c51afff5..f9d790e2603e8 100644 --- a/aspdotnet/hatch.toml +++ b/aspdotnet/hatch.toml @@ -10,4 +10,4 @@ platforms = [ ] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] diff --git a/aspdotnet/pyproject.toml b/aspdotnet/pyproject.toml index ec447d479c201..c73f179dea07f 100644 --- a/aspdotnet/pyproject.toml +++ b/aspdotnet/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/avi_vantage/changelog.d/18207.added b/avi_vantage/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/avi_vantage/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/avi_vantage/hatch.toml b/avi_vantage/hatch.toml index 001e43ce25414..c85c5f07a7df2 100644 --- a/avi_vantage/hatch.toml +++ b/avi_vantage/hatch.toml @@ -1,4 +1,4 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] diff --git a/avi_vantage/pyproject.toml b/avi_vantage/pyproject.toml index 1f45ae5136866..39fd1d7798e3b 100644 --- a/avi_vantage/pyproject.toml +++ b/avi_vantage/pyproject.toml @@ -26,7 +26,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/aws_neuron/hatch.toml b/aws_neuron/hatch.toml index 001e43ce25414..c85c5f07a7df2 100644 --- a/aws_neuron/hatch.toml +++ b/aws_neuron/hatch.toml @@ -1,4 +1,4 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] diff --git a/aws_neuron/pyproject.toml b/aws_neuron/pyproject.toml index efd983eb0051d..4a7005002add4 100644 --- a/aws_neuron/pyproject.toml +++ b/aws_neuron/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", "Private :: Do Not Upload", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", ] dependencies = [ diff --git a/azure_iot_edge/changelog.d/18207.added b/azure_iot_edge/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/azure_iot_edge/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/azure_iot_edge/hatch.toml b/azure_iot_edge/hatch.toml index 52b6d241c7bf8..f58f23834bfbe 100644 --- a/azure_iot_edge/hatch.toml +++ b/azure_iot_edge/hatch.toml @@ -3,7 +3,7 @@ check-types = false [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] protocol = ["tls"] [envs.default.overrides] diff --git a/azure_iot_edge/pyproject.toml b/azure_iot_edge/pyproject.toml index bf2283ba5c6bf..0080f809a7865 100644 --- a/azure_iot_edge/pyproject.toml +++ b/azure_iot_edge/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/boundary/changelog.d/18207.added b/boundary/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/boundary/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/boundary/hatch.toml b/boundary/hatch.toml index e95adb4f00b0b..7187108da4532 100644 --- a/boundary/hatch.toml +++ b/boundary/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] version = ["0.8"] [envs.default.overrides] diff --git a/boundary/pyproject.toml b/boundary/pyproject.toml index 36ebe626ff783..1874f346a8e66 100644 --- a/boundary/pyproject.toml +++ b/boundary/pyproject.toml @@ -22,7 +22,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/btrfs/changelog.d/18207.added b/btrfs/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/btrfs/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/btrfs/hatch.toml b/btrfs/hatch.toml index fb5625a080ed5..93bc6bef22ef8 100644 --- a/btrfs/hatch.toml +++ b/btrfs/hatch.toml @@ -8,4 +8,4 @@ platforms = [ ] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] diff --git a/btrfs/pyproject.toml b/btrfs/pyproject.toml index 69ac8f8563694..b357e8e8da0ee 100644 --- a/btrfs/pyproject.toml +++ b/btrfs/pyproject.toml @@ -24,7 +24,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/cacti/changelog.d/18207.added b/cacti/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/cacti/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/cacti/hatch.toml b/cacti/hatch.toml index 282320c2bd3dd..c3be24889bf10 100644 --- a/cacti/hatch.toml +++ b/cacti/hatch.toml @@ -6,4 +6,4 @@ dependencies = [ ] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] diff --git a/cacti/pyproject.toml b/cacti/pyproject.toml index 081135d2a36f7..45f68faa8e788 100644 --- a/cacti/pyproject.toml +++ b/cacti/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/calico/changelog.d/18207.added b/calico/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/calico/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/calico/hatch.toml b/calico/hatch.toml index 001e43ce25414..c85c5f07a7df2 100644 --- a/calico/hatch.toml +++ b/calico/hatch.toml @@ -1,4 +1,4 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] diff --git a/calico/pyproject.toml b/calico/pyproject.toml index 56abfd47d08f3..4854fe0f4590a 100644 --- a/calico/pyproject.toml +++ b/calico/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/cassandra/changelog.d/18207.added b/cassandra/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/cassandra/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/cassandra/hatch.toml b/cassandra/hatch.toml index 001e43ce25414..c85c5f07a7df2 100644 --- a/cassandra/hatch.toml +++ b/cassandra/hatch.toml @@ -1,4 +1,4 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] diff --git a/cassandra/pyproject.toml b/cassandra/pyproject.toml index b247564c5d336..1c6714ec2aafb 100644 --- a/cassandra/pyproject.toml +++ b/cassandra/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/cassandra_nodetool/changelog.d/18207.added b/cassandra_nodetool/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/cassandra_nodetool/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/cassandra_nodetool/hatch.toml b/cassandra_nodetool/hatch.toml index c9c6b5d824311..2abdc87cdea11 100644 --- a/cassandra_nodetool/hatch.toml +++ b/cassandra_nodetool/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] version = ["2.1", "3.0"] [envs.default.overrides] diff --git a/cassandra_nodetool/pyproject.toml b/cassandra_nodetool/pyproject.toml index 10d14859d3313..4d41749f33d38 100644 --- a/cassandra_nodetool/pyproject.toml +++ b/cassandra_nodetool/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/ceph/changelog.d/18207.added b/ceph/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/ceph/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/ceph/hatch.toml b/ceph/hatch.toml index c58aa74eaf910..2147947a70f95 100644 --- a/ceph/hatch.toml +++ b/ceph/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] version = ["4.0", "5.0"] [envs.default.overrides] diff --git a/ceph/pyproject.toml b/ceph/pyproject.toml index ecfde2326b5d4..97e3696c4c478 100644 --- a/ceph/pyproject.toml +++ b/ceph/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/cert_manager/changelog.d/18207.added b/cert_manager/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/cert_manager/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/cert_manager/hatch.toml b/cert_manager/hatch.toml index 821e6ed8e37ca..f62423b83c762 100644 --- a/cert_manager/hatch.toml +++ b/cert_manager/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] [envs.default.env-vars] DDEV_SKIP_GENERIC_TAGS_CHECK = "true" diff --git a/cert_manager/pyproject.toml b/cert_manager/pyproject.toml index 1c698a5ad99b7..3855c0dcf5a30 100644 --- a/cert_manager/pyproject.toml +++ b/cert_manager/pyproject.toml @@ -26,7 +26,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/cilium/changelog.d/18207.added b/cilium/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/cilium/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/cilium/hatch.toml b/cilium/hatch.toml index 80639298c8037..68cc58696dfb3 100644 --- a/cilium/hatch.toml +++ b/cilium/hatch.toml @@ -1,12 +1,12 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] impl = ["legacy"] version = ["1.9"] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] version = ["1.9", "1.10", "1.11"] [envs.default.env-vars] diff --git a/cilium/pyproject.toml b/cilium/pyproject.toml index b8a31cc3cc53b..4dd8108d9a079 100644 --- a/cilium/pyproject.toml +++ b/cilium/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/cisco_aci/changelog.d/18207.added b/cisco_aci/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/cisco_aci/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/cisco_aci/hatch.toml b/cisco_aci/hatch.toml index b271cca6d3e45..0169207d8cffb 100644 --- a/cisco_aci/hatch.toml +++ b/cisco_aci/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] [envs.default] dependencies = [ diff --git a/cisco_aci/pyproject.toml b/cisco_aci/pyproject.toml index 703e39f823cf4..eb08546ee1fa2 100644 --- a/cisco_aci/pyproject.toml +++ b/cisco_aci/pyproject.toml @@ -24,7 +24,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/citrix_hypervisor/changelog.d/18207.added b/citrix_hypervisor/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/citrix_hypervisor/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/citrix_hypervisor/hatch.toml b/citrix_hypervisor/hatch.toml index 9d33ad72c8eef..0cc6f15224980 100644 --- a/citrix_hypervisor/hatch.toml +++ b/citrix_hypervisor/hatch.toml @@ -3,5 +3,5 @@ check-types = false [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] diff --git a/citrix_hypervisor/pyproject.toml b/citrix_hypervisor/pyproject.toml index b1efb277e4c16..584263112d453 100644 --- a/citrix_hypervisor/pyproject.toml +++ b/citrix_hypervisor/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/clickhouse/changelog.d/18207.added b/clickhouse/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/clickhouse/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/clickhouse/hatch.toml b/clickhouse/hatch.toml index 7f7517e6e5e86..fa5265c7e434a 100644 --- a/clickhouse/hatch.toml +++ b/clickhouse/hatch.toml @@ -4,7 +4,7 @@ CLICKHOUSE_REPOSITORY = "clickhouse/clickhouse-server" [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] version = ["18", "19", "20", "21.8", "22.7"] [envs.default.overrides] diff --git a/clickhouse/pyproject.toml b/clickhouse/pyproject.toml index 1f16c0f0e5523..0944ab13d5d9d 100644 --- a/clickhouse/pyproject.toml +++ b/clickhouse/pyproject.toml @@ -24,7 +24,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/cloud_foundry_api/changelog.d/18207.added b/cloud_foundry_api/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/cloud_foundry_api/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/cloud_foundry_api/hatch.toml b/cloud_foundry_api/hatch.toml index 84290a4e2ebe1..056785192a3b5 100644 --- a/cloud_foundry_api/hatch.toml +++ b/cloud_foundry_api/hatch.toml @@ -9,4 +9,4 @@ mypy-deps = [ e2e-env = false [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] diff --git a/cloud_foundry_api/pyproject.toml b/cloud_foundry_api/pyproject.toml index 9b7921cbfd05e..d1909b403b19d 100644 --- a/cloud_foundry_api/pyproject.toml +++ b/cloud_foundry_api/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/cloudera/changelog.d/18207.added b/cloudera/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/cloudera/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/cloudera/hatch.toml b/cloudera/hatch.toml index 001e43ce25414..c85c5f07a7df2 100644 --- a/cloudera/hatch.toml +++ b/cloudera/hatch.toml @@ -1,4 +1,4 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] diff --git a/cloudera/pyproject.toml b/cloudera/pyproject.toml index c7f78be036b73..da6cd2de5037e 100644 --- a/cloudera/pyproject.toml +++ b/cloudera/pyproject.toml @@ -27,7 +27,7 @@ classifiers = [ "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", "Private :: Do Not Upload", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", ] dependencies = [ diff --git a/cockroachdb/changelog.d/18207.added b/cockroachdb/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/cockroachdb/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/cockroachdb/hatch.toml b/cockroachdb/hatch.toml index 62b5bcb8475d2..e7ee92c1ab643 100644 --- a/cockroachdb/hatch.toml +++ b/cockroachdb/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] version = ["2.0", "22.1", "23.2"] [envs.default.overrides] diff --git a/cockroachdb/pyproject.toml b/cockroachdb/pyproject.toml index f13e06a527fa9..9118478397f09 100644 --- a/cockroachdb/pyproject.toml +++ b/cockroachdb/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/confluent_platform/changelog.d/18207.added b/confluent_platform/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/confluent_platform/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/confluent_platform/hatch.toml b/confluent_platform/hatch.toml index ebbb590666b02..9271d848e8336 100644 --- a/confluent_platform/hatch.toml +++ b/confluent_platform/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] version = ["5.4", "6.2"] [envs.default.overrides] diff --git a/confluent_platform/pyproject.toml b/confluent_platform/pyproject.toml index b3b5252141166..9f26a28d15eda 100644 --- a/confluent_platform/pyproject.toml +++ b/confluent_platform/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/consul/changelog.d/18207.added b/consul/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/consul/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/consul/hatch.toml b/consul/hatch.toml index 67de49b21a7cc..42b3b74d52300 100644 --- a/consul/hatch.toml +++ b/consul/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] version = ["1.6", "1.9"] [envs.default.overrides] diff --git a/consul/pyproject.toml b/consul/pyproject.toml index 55cf34f8ebb11..0573106958db0 100644 --- a/consul/pyproject.toml +++ b/consul/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/coredns/changelog.d/18207.added b/coredns/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/coredns/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/coredns/hatch.toml b/coredns/hatch.toml index 8d3bcc0f3d099..62fcdb921d5f0 100644 --- a/coredns/hatch.toml +++ b/coredns/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] version = ["1.2", "1.8"] [envs.default.overrides] diff --git a/coredns/pyproject.toml b/coredns/pyproject.toml index ceec99e2da61b..0a768df59a1b6 100644 --- a/coredns/pyproject.toml +++ b/coredns/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/couch/changelog.d/18207.added b/couch/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/couch/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/couch/hatch.toml b/couch/hatch.toml index ee7cc27754fed..b3f69ac5f66bd 100644 --- a/couch/hatch.toml +++ b/couch/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] version = ["1.6", "2.3", "3.1"] [envs.default.overrides] diff --git a/couch/pyproject.toml b/couch/pyproject.toml index 993a987d05f7e..4b7ad5f700d06 100644 --- a/couch/pyproject.toml +++ b/couch/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/couchbase/changelog.d/18207.added b/couchbase/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/couchbase/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/couchbase/hatch.toml b/couchbase/hatch.toml index 32cc203185db6..21a594adf0315 100644 --- a/couchbase/hatch.toml +++ b/couchbase/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] # https://www.couchbase.com/support-policy/enterprise-software version = ["6.6", "7.0", "7.1"] diff --git a/couchbase/pyproject.toml b/couchbase/pyproject.toml index 9b3d7c5557ff7..f1538ebe2d970 100644 --- a/couchbase/pyproject.toml +++ b/couchbase/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/crio/changelog.d/18207.added b/crio/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/crio/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/crio/hatch.toml b/crio/hatch.toml index 001e43ce25414..c85c5f07a7df2 100644 --- a/crio/hatch.toml +++ b/crio/hatch.toml @@ -1,4 +1,4 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] diff --git a/crio/pyproject.toml b/crio/pyproject.toml index 6a1fc4966154a..f5a1bf83a5c00 100644 --- a/crio/pyproject.toml +++ b/crio/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/datadog_checks_base/changelog.d/18207.added b/datadog_checks_base/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/datadog_checks_base/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/datadog_checks_base/hatch.toml b/datadog_checks_base/hatch.toml index f3d334a271abe..f3c5084b6a5d3 100644 --- a/datadog_checks_base/hatch.toml +++ b/datadog_checks_base/hatch.toml @@ -3,7 +3,7 @@ check-types = false [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] [envs.default] features = ["db", "deps", "http", "json", "kube"] diff --git a/datadog_checks_base/pyproject.toml b/datadog_checks_base/pyproject.toml index 6134ffd9228ef..1abe90220249a 100644 --- a/datadog_checks_base/pyproject.toml +++ b/datadog_checks_base/pyproject.toml @@ -23,7 +23,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", ] dynamic = [ diff --git a/datadog_checks_dependency_provider/changelog.d/18207.added b/datadog_checks_dependency_provider/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/datadog_checks_dependency_provider/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/datadog_checks_dependency_provider/pyproject.toml b/datadog_checks_dependency_provider/pyproject.toml index e2c1264517a6a..0463186ed5754 100644 --- a/datadog_checks_dependency_provider/pyproject.toml +++ b/datadog_checks_dependency_provider/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: System Administrators", "Topic :: System :: Monitoring", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", ] dependencies = [ "datadog-checks-base>=11.2.0", diff --git a/datadog_checks_dev/changelog.d/18207.added b/datadog_checks_dev/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/datadog_checks_dev/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/datadog_checks_dev/datadog_checks/dev/tooling/templates/integration/check/{check_name}/hatch.toml b/datadog_checks_dev/datadog_checks/dev/tooling/templates/integration/check/{check_name}/hatch.toml index 001e43ce25414..c85c5f07a7df2 100644 --- a/datadog_checks_dev/datadog_checks/dev/tooling/templates/integration/check/{check_name}/hatch.toml +++ b/datadog_checks_dev/datadog_checks/dev/tooling/templates/integration/check/{check_name}/hatch.toml @@ -1,4 +1,4 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] diff --git a/datadog_checks_dev/datadog_checks/dev/tooling/templates/integration/check/{check_name}/pyproject.toml b/datadog_checks_dev/datadog_checks/dev/tooling/templates/integration/check/{check_name}/pyproject.toml index 1a1a0d5b43ba5..f64b51df1418b 100644 --- a/datadog_checks_dev/datadog_checks/dev/tooling/templates/integration/check/{check_name}/pyproject.toml +++ b/datadog_checks_dev/datadog_checks/dev/tooling/templates/integration/check/{check_name}/pyproject.toml @@ -9,7 +9,7 @@ name = "datadog-{project_name}" description = "The {integration_name} check" readme = "README.md" license = "BSD-3-Clause" -requires-python = ">=3.11" +requires-python = ">=3.12" keywords = [ "datadog", "datadog agent", @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", "Private :: Do Not Upload", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", ] dependencies = [ diff --git a/datadog_checks_dev/datadog_checks/dev/tooling/templates/integration/jmx/{check_name}/hatch.toml b/datadog_checks_dev/datadog_checks/dev/tooling/templates/integration/jmx/{check_name}/hatch.toml index 001e43ce25414..c85c5f07a7df2 100644 --- a/datadog_checks_dev/datadog_checks/dev/tooling/templates/integration/jmx/{check_name}/hatch.toml +++ b/datadog_checks_dev/datadog_checks/dev/tooling/templates/integration/jmx/{check_name}/hatch.toml @@ -1,4 +1,4 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] diff --git a/datadog_checks_dev/datadog_checks/dev/tooling/templates/integration/jmx/{check_name}/pyproject.toml b/datadog_checks_dev/datadog_checks/dev/tooling/templates/integration/jmx/{check_name}/pyproject.toml index 1a1a0d5b43ba5..f64b51df1418b 100644 --- a/datadog_checks_dev/datadog_checks/dev/tooling/templates/integration/jmx/{check_name}/pyproject.toml +++ b/datadog_checks_dev/datadog_checks/dev/tooling/templates/integration/jmx/{check_name}/pyproject.toml @@ -9,7 +9,7 @@ name = "datadog-{project_name}" description = "The {integration_name} check" readme = "README.md" license = "BSD-3-Clause" -requires-python = ">=3.11" +requires-python = ">=3.12" keywords = [ "datadog", "datadog agent", @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", "Private :: Do Not Upload", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", ] dependencies = [ diff --git a/datadog_checks_dev/datadog_checks/dev/tooling/templates/integration/logs/{check_name}/pyproject.toml b/datadog_checks_dev/datadog_checks/dev/tooling/templates/integration/logs/{check_name}/pyproject.toml index 75ccd780b7e7a..bda6b5be9e8e6 100644 --- a/datadog_checks_dev/datadog_checks/dev/tooling/templates/integration/logs/{check_name}/pyproject.toml +++ b/datadog_checks_dev/datadog_checks/dev/tooling/templates/integration/logs/{check_name}/pyproject.toml @@ -24,7 +24,7 @@ classifiers = [ "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", "Private :: Do Not Upload", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", ] dependencies = [ diff --git a/datadog_checks_dev/hatch.toml b/datadog_checks_dev/hatch.toml index ff0f5e90597e4..7bb3c9e11baae 100644 --- a/datadog_checks_dev/hatch.toml +++ b/datadog_checks_dev/hatch.toml @@ -7,13 +7,13 @@ e2e-env = false DDEV_TESTING_PLUGIN = "true" [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] [envs.default.overrides] matrix.python.features = [ - { value = "cli", if = ["3.11"] }, + { value = "cli", if = ["3.12"] }, ] # TODO: remove this when the old CLI is gone matrix.python.pre-install-commands = [ - { value = "python -m pip install --no-deps --disable-pip-version-check {verbosity:flag:-1} -e ../ddev", if = ["3.11"] }, + { value = "python -m pip install --no-deps --disable-pip-version-check {verbosity:flag:-1} -e ../ddev", if = ["3.12"] }, ] diff --git a/datadog_checks_dev/pyproject.toml b/datadog_checks_dev/pyproject.toml index 9c4b6796cb238..e9e8f703c9281 100644 --- a/datadog_checks_dev/pyproject.toml +++ b/datadog_checks_dev/pyproject.toml @@ -28,7 +28,7 @@ classifiers = [ "License :: OSI Approved :: BSD License", "Natural Language :: English", "Operating System :: OS Independent", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", ] dependencies = [ "coverage>=5.0.3", diff --git a/datadog_checks_downloader/changelog.d/18207.added b/datadog_checks_downloader/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/datadog_checks_downloader/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/datadog_checks_downloader/hatch.toml b/datadog_checks_downloader/hatch.toml index ef1e2233f7c8a..185e0e2516cb4 100644 --- a/datadog_checks_downloader/hatch.toml +++ b/datadog_checks_downloader/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] [envs.default] e2e-env = false diff --git a/datadog_checks_downloader/pyproject.toml b/datadog_checks_downloader/pyproject.toml index 89ad5c54ec84d..2974e65abcc60 100644 --- a/datadog_checks_downloader/pyproject.toml +++ b/datadog_checks_downloader/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", ] dynamic = [ diff --git a/datadog_cluster_agent/changelog.d/18207.added b/datadog_cluster_agent/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/datadog_cluster_agent/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/datadog_cluster_agent/hatch.toml b/datadog_cluster_agent/hatch.toml index ce1f02498d72b..8eb7559f584f9 100644 --- a/datadog_cluster_agent/hatch.toml +++ b/datadog_cluster_agent/hatch.toml @@ -1,5 +1,5 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] diff --git a/datadog_cluster_agent/pyproject.toml b/datadog_cluster_agent/pyproject.toml index a26faebb5773b..4592b159c015d 100644 --- a/datadog_cluster_agent/pyproject.toml +++ b/datadog_cluster_agent/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/dcgm/changelog.d/18207.added b/dcgm/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/dcgm/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/dcgm/hatch.toml b/dcgm/hatch.toml index 001e43ce25414..c85c5f07a7df2 100644 --- a/dcgm/hatch.toml +++ b/dcgm/hatch.toml @@ -1,4 +1,4 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] diff --git a/dcgm/pyproject.toml b/dcgm/pyproject.toml index 0f6c724933974..e17cc60b915a6 100644 --- a/dcgm/pyproject.toml +++ b/dcgm/pyproject.toml @@ -26,7 +26,7 @@ classifiers = [ "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", "Private :: Do Not Upload", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", ] dependencies = [ diff --git a/ddev/changelog.d/18207.added b/ddev/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/ddev/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/ddev/hatch.toml b/ddev/hatch.toml index 2f299a9ceb09c..398997183a7e4 100644 --- a/ddev/hatch.toml +++ b/ddev/hatch.toml @@ -5,7 +5,7 @@ mypy-args = [ ] [envs.default] -python = "3.11" +python = "3.12" e2e-env = false dependencies = [ "pyyaml", diff --git a/ddev/pyproject.toml b/ddev/pyproject.toml index 49f5f471453ec..45765f43639e5 100644 --- a/ddev/pyproject.toml +++ b/ddev/pyproject.toml @@ -23,7 +23,7 @@ classifiers = [ "License :: OSI Approved :: BSD License", "Natural Language :: English", "Operating System :: OS Independent", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", ] dependencies = [ "click~=8.1.6", @@ -76,12 +76,12 @@ scripts = ["ddev"] include = '\.pyi?$' line-length = 120 skip-string-normalization = true -target-version = ["py311"] +target-version = ["py312"] extend-exclude = "src/ddev/_version.py" [tool.ruff] exclude = [] -target-version = "py311" +target-version = "py312" line-length = 120 [tool.ruff.lint] diff --git a/ddev/src/ddev/repo/constants.py b/ddev/src/ddev/repo/constants.py index 446b36b2a7a60..d2a40cb3334b5 100644 --- a/ddev/src/ddev/repo/constants.py +++ b/ddev/src/ddev/repo/constants.py @@ -11,4 +11,4 @@ } # This is automatically maintained -PYTHON_VERSION = '3.11' +PYTHON_VERSION = '3.12' diff --git a/directory/changelog.d/18207.added b/directory/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/directory/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/directory/hatch.toml b/directory/hatch.toml index 801832c211847..57ca29df505a9 100644 --- a/directory/hatch.toml +++ b/directory/hatch.toml @@ -1,6 +1,6 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] [envs.bench] diff --git a/directory/pyproject.toml b/directory/pyproject.toml index d86492021fe18..77e70ff27ced5 100644 --- a/directory/pyproject.toml +++ b/directory/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/disk/changelog.d/18207.added b/disk/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/disk/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/disk/hatch.toml b/disk/hatch.toml index 001e43ce25414..c85c5f07a7df2 100644 --- a/disk/hatch.toml +++ b/disk/hatch.toml @@ -1,4 +1,4 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] diff --git a/disk/pyproject.toml b/disk/pyproject.toml index 638bae2a48b34..7e5ed5c3f73b6 100644 --- a/disk/pyproject.toml +++ b/disk/pyproject.toml @@ -24,7 +24,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/dns_check/changelog.d/18207.added b/dns_check/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/dns_check/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/dns_check/hatch.toml b/dns_check/hatch.toml index 61068f0d3abf9..70bd33fcab628 100644 --- a/dns_check/hatch.toml +++ b/dns_check/hatch.toml @@ -3,4 +3,4 @@ check-types = false [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] diff --git a/dns_check/pyproject.toml b/dns_check/pyproject.toml index f9c7f80ae6ca4..afa020b28b601 100644 --- a/dns_check/pyproject.toml +++ b/dns_check/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/docs/developer/setup.md b/docs/developer/setup.md index 14c89b383d8ac..3915fa131a7bd 100644 --- a/docs/developer/setup.md +++ b/docs/developer/setup.md @@ -11,7 +11,7 @@ you intend to work on. ## Python -To work on any integration you must install Python 3.11. +To work on any integration you must install Python 3.12. After installation, restart your terminal and ensure that your newly installed Python comes first in your `PATH`. @@ -25,7 +25,7 @@ After installation, restart your terminal and ensure that your newly installed P then install Python: ``` - brew install python@3.11 + brew install python@3.12 ``` After it completes, check the output to see if it asked you to run any extra commands and if so, execute them. @@ -39,7 +39,7 @@ After installation, restart your terminal and ensure that your newly installed P === "Windows" Windows users have it the easiest. - Download the [Python 3.11 64-bit executable installer](https://www.python.org/downloads/release/python-3115/) and run it. + Download the [Python 3.12 64-bit executable installer](https://www.python.org/downloads/release/python-3124/) and run it. When prompted, be sure to select the option to add to your `PATH`. Also, it is recommended that you choose the per-user installation method. Verify successful `PATH` modification: @@ -51,7 +51,7 @@ After installation, restart your terminal and ensure that your newly installed P === "Linux" Ah, you enjoy difficult things. Are you using Gentoo? - We recommend using either [Miniconda][miniconda-docs] or [pyenv][pyenv-github] to install Python 3.11. Whatever you do, never modify the system Python. + We recommend using either [Miniconda][miniconda-docs] or [pyenv][pyenv-github] to install Python 3.12. Whatever you do, never modify the system Python. Verify successful `PATH` modification: @@ -242,11 +242,11 @@ This is if you cloned [integrations-core][] and want to always use the version b === "ARM" ``` - pipx install -e /path/to/integrations-core/ddev --python /opt/homebrew/opt/python@3.11/bin/python3.11 + pipx install -e /path/to/integrations-core/ddev --python /opt/homebrew/opt/python@3.12/bin/python3.12 ``` === "Intel" ``` - pipx install -e /path/to/integrations-core/ddev --python /usr/local/opt/python@3.11/bin/python3.11 + pipx install -e /path/to/integrations-core/ddev --python /usr/local/opt/python@3.12/bin/python3.12 ``` !!! warning diff --git a/dotnetclr/changelog.d/18207.added b/dotnetclr/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/dotnetclr/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/dotnetclr/hatch.toml b/dotnetclr/hatch.toml index 6f2e122f880de..4f891ddac37fd 100644 --- a/dotnetclr/hatch.toml +++ b/dotnetclr/hatch.toml @@ -9,4 +9,4 @@ platforms = [ ] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] diff --git a/dotnetclr/pyproject.toml b/dotnetclr/pyproject.toml index 72a9a1f23eeb6..f46f5887ec543 100644 --- a/dotnetclr/pyproject.toml +++ b/dotnetclr/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/druid/changelog.d/18207.added b/druid/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/druid/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/druid/hatch.toml b/druid/hatch.toml index 001e43ce25414..c85c5f07a7df2 100644 --- a/druid/hatch.toml +++ b/druid/hatch.toml @@ -1,4 +1,4 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] diff --git a/druid/pyproject.toml b/druid/pyproject.toml index 5748ccdea4708..ac0fce91714fa 100644 --- a/druid/pyproject.toml +++ b/druid/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/ecs_fargate/changelog.d/18207.added b/ecs_fargate/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/ecs_fargate/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/ecs_fargate/hatch.toml b/ecs_fargate/hatch.toml index 821e6ed8e37ca..f62423b83c762 100644 --- a/ecs_fargate/hatch.toml +++ b/ecs_fargate/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] [envs.default.env-vars] DDEV_SKIP_GENERIC_TAGS_CHECK = "true" diff --git a/ecs_fargate/pyproject.toml b/ecs_fargate/pyproject.toml index ee4dcf77128ce..602ff576e1419 100644 --- a/ecs_fargate/pyproject.toml +++ b/ecs_fargate/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/eks_fargate/changelog.d/18207.added b/eks_fargate/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/eks_fargate/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/eks_fargate/hatch.toml b/eks_fargate/hatch.toml index 169926426dae2..87b66d031883f 100644 --- a/eks_fargate/hatch.toml +++ b/eks_fargate/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] [envs.default] e2e-env = false diff --git a/eks_fargate/pyproject.toml b/eks_fargate/pyproject.toml index c0d369b029518..03620d946d5fa 100644 --- a/eks_fargate/pyproject.toml +++ b/eks_fargate/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/elastic/changelog.d/18207.added b/elastic/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/elastic/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/elastic/hatch.toml b/elastic/hatch.toml index cdae4fd101625..8ab01c4ecbb58 100644 --- a/elastic/hatch.toml +++ b/elastic/hatch.toml @@ -2,12 +2,12 @@ base-package-features = ["deps", "http"] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] flavor = ["elasticsearch"] version = ["7.2", "7.7", "7.9", "7.10", "8.8"] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] # Opensearch, compatible with elasticsearch flavor = ["opensearch"] version = ["1.1"] diff --git a/elastic/pyproject.toml b/elastic/pyproject.toml index b81be0f851e5a..4280b54f62c18 100644 --- a/elastic/pyproject.toml +++ b/elastic/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/envoy/changelog.d/18207.added b/envoy/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/envoy/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/envoy/hatch.toml b/envoy/hatch.toml index e2359d03581ea..6e6ec397259d1 100644 --- a/envoy/hatch.toml +++ b/envoy/hatch.toml @@ -6,7 +6,7 @@ dependencies = [ ] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] api-version = ["2", "3"] [envs.default.overrides] diff --git a/envoy/pyproject.toml b/envoy/pyproject.toml index cab901eee95b1..1d549fcae090d 100644 --- a/envoy/pyproject.toml +++ b/envoy/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/esxi/changelog.d/18207.added b/esxi/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/esxi/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/esxi/hatch.toml b/esxi/hatch.toml index 9fad694f9b43b..fc3f5d074ed9a 100644 --- a/esxi/hatch.toml +++ b/esxi/hatch.toml @@ -1,11 +1,11 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] setup = ["lab"] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] setup = ["vcsim"] version = ["6.5", "7.0"] diff --git a/esxi/pyproject.toml b/esxi/pyproject.toml index ce9b56b83225d..cbb6d6aea813e 100644 --- a/esxi/pyproject.toml +++ b/esxi/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", "Private :: Do Not Upload", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", ] dependencies = [ diff --git a/etcd/changelog.d/18207.added b/etcd/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/etcd/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/etcd/hatch.toml b/etcd/hatch.toml index b4788aebb9a34..f3d30b89ee778 100644 --- a/etcd/hatch.toml +++ b/etcd/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] version = ["v3.4.26" , "v3.5.9"] [envs.default.overrides] diff --git a/etcd/pyproject.toml b/etcd/pyproject.toml index 09024a3f45830..1bfbef8e3ae09 100644 --- a/etcd/pyproject.toml +++ b/etcd/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/exchange_server/changelog.d/18207.added b/exchange_server/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/exchange_server/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/exchange_server/hatch.toml b/exchange_server/hatch.toml index 413e6c51fe94c..0f9f06cf58486 100644 --- a/exchange_server/hatch.toml +++ b/exchange_server/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] [envs.default] dependencies = [ diff --git a/exchange_server/pyproject.toml b/exchange_server/pyproject.toml index d48128619c891..88b90d1f9c0f5 100644 --- a/exchange_server/pyproject.toml +++ b/exchange_server/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/external_dns/changelog.d/18207.added b/external_dns/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/external_dns/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/external_dns/hatch.toml b/external_dns/hatch.toml index 001e43ce25414..c85c5f07a7df2 100644 --- a/external_dns/hatch.toml +++ b/external_dns/hatch.toml @@ -1,4 +1,4 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] diff --git a/external_dns/pyproject.toml b/external_dns/pyproject.toml index 6e3e4e253cae1..f6ad12bc045ba 100644 --- a/external_dns/pyproject.toml +++ b/external_dns/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/flink/changelog.d/18207.added b/flink/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/flink/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/flink/pyproject.toml b/flink/pyproject.toml index 2130b12282146..da7755c0c7ac1 100644 --- a/flink/pyproject.toml +++ b/flink/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/fluentd/changelog.d/18207.added b/fluentd/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/fluentd/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/fluentd/hatch.toml b/fluentd/hatch.toml index 7890ff4df450a..351f3b2acfba8 100644 --- a/fluentd/hatch.toml +++ b/fluentd/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] version = ["1.17"] [envs.default.overrides] diff --git a/fluentd/pyproject.toml b/fluentd/pyproject.toml index 4a41be9d1c2f7..cf89e04ec7ed9 100644 --- a/fluentd/pyproject.toml +++ b/fluentd/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/fluxcd/changelog.d/18207.added b/fluxcd/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/fluxcd/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/fluxcd/hatch.toml b/fluxcd/hatch.toml index 001e43ce25414..c85c5f07a7df2 100644 --- a/fluxcd/hatch.toml +++ b/fluxcd/hatch.toml @@ -1,4 +1,4 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] diff --git a/fluxcd/pyproject.toml b/fluxcd/pyproject.toml index 138a357f0ebd1..d3a5dc43fad68 100644 --- a/fluxcd/pyproject.toml +++ b/fluxcd/pyproject.toml @@ -23,7 +23,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/fly_io/hatch.toml b/fly_io/hatch.toml index fb8856ae1da1d..9cde6362848ab 100644 --- a/fly_io/hatch.toml +++ b/fly_io/hatch.toml @@ -1,11 +1,11 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] setup = ["caddy"] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] setup = ["lab"] [envs.default.overrides] diff --git a/fly_io/pyproject.toml b/fly_io/pyproject.toml index b65bd46e1ff08..ae7927991c535 100644 --- a/fly_io/pyproject.toml +++ b/fly_io/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", "Private :: Do Not Upload", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", ] dependencies = [ diff --git a/foundationdb/changelog.d/18207.added b/foundationdb/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/foundationdb/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/foundationdb/hatch.toml b/foundationdb/hatch.toml index 97a77e47030a1..2dfbc037ca63c 100644 --- a/foundationdb/hatch.toml +++ b/foundationdb/hatch.toml @@ -7,10 +7,10 @@ dependencies = [ e2e-env = false [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] protocol = ["tls"] [envs.default.overrides] diff --git a/foundationdb/pyproject.toml b/foundationdb/pyproject.toml index 470049575ebe9..852c3de459e5a 100644 --- a/foundationdb/pyproject.toml +++ b/foundationdb/pyproject.toml @@ -24,7 +24,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/gearmand/changelog.d/18207.added b/gearmand/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/gearmand/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/gearmand/hatch.toml b/gearmand/hatch.toml index f9ae641fd89e2..faa624cac3d33 100644 --- a/gearmand/hatch.toml +++ b/gearmand/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] version = ["1.0", "1.1"] [envs.default.overrides] diff --git a/gearmand/pyproject.toml b/gearmand/pyproject.toml index 6cbb39f22e913..182308e461433 100644 --- a/gearmand/pyproject.toml +++ b/gearmand/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/gitlab/changelog.d/18207.added b/gitlab/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/gitlab/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/gitlab/hatch.toml b/gitlab/hatch.toml index 106c9a86d6018..b7ded317ffcd8 100644 --- a/gitlab/hatch.toml +++ b/gitlab/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] # https://about.gitlab.com/support/statement-of-support/#version-support version = ["13.12", "14.10", "15.10"] diff --git a/gitlab/pyproject.toml b/gitlab/pyproject.toml index 2b18e80e4f688..7d201fcd21a28 100644 --- a/gitlab/pyproject.toml +++ b/gitlab/pyproject.toml @@ -24,7 +24,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/gitlab_runner/changelog.d/18207.added b/gitlab_runner/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/gitlab_runner/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/gitlab_runner/hatch.toml b/gitlab_runner/hatch.toml index f723a4530d931..e10701f1d8874 100644 --- a/gitlab_runner/hatch.toml +++ b/gitlab_runner/hatch.toml @@ -5,7 +5,7 @@ DDEV_SKIP_GENERIC_TAGS_CHECK = "true" GITLAB_IMAGE = "gitlab/gitlab-ce" [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] version = ["10.8.0"] [envs.default.overrides] diff --git a/gitlab_runner/pyproject.toml b/gitlab_runner/pyproject.toml index 6a98e5af356a6..ffd85f719b3bf 100644 --- a/gitlab_runner/pyproject.toml +++ b/gitlab_runner/pyproject.toml @@ -24,7 +24,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/glusterfs/changelog.d/18207.added b/glusterfs/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/glusterfs/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/glusterfs/hatch.toml b/glusterfs/hatch.toml index 6e69a5a11402a..e9d5dffb64861 100644 --- a/glusterfs/hatch.toml +++ b/glusterfs/hatch.toml @@ -6,7 +6,7 @@ mypy-deps = [ ] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] version = ["7.1"] [envs.default.env-vars] diff --git a/glusterfs/pyproject.toml b/glusterfs/pyproject.toml index b6dc2025ac9e5..e70370c454c8e 100644 --- a/glusterfs/pyproject.toml +++ b/glusterfs/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/go_expvar/changelog.d/18207.added b/go_expvar/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/go_expvar/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/go_expvar/hatch.toml b/go_expvar/hatch.toml index 001e43ce25414..c85c5f07a7df2 100644 --- a/go_expvar/hatch.toml +++ b/go_expvar/hatch.toml @@ -1,4 +1,4 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] diff --git a/go_expvar/pyproject.toml b/go_expvar/pyproject.toml index 1ab17cbdbb1cc..0bdea0592f989 100644 --- a/go_expvar/pyproject.toml +++ b/go_expvar/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/gunicorn/changelog.d/18207.added b/gunicorn/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/gunicorn/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/gunicorn/hatch.toml b/gunicorn/hatch.toml index d2fe13c38ff4b..5a9c5432fd90e 100644 --- a/gunicorn/hatch.toml +++ b/gunicorn/hatch.toml @@ -1,13 +1,14 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] -version = ["19.9", "20.1"] +python = ["3.12"] +# v22 of gunicorn is the first that supports python 3.12 +version = ["22", "23"] [envs.default.overrides] matrix.version.env-vars = [ - { key = "GUNICORN_VERSION", value = "19.9.0", if = ["19.9"] }, - { key = "GUNICORN_VERSION", value = "20.1.0", if = ["20.1"] }, + { key = "GUNICORN_VERSION", value = "22.0.0", if = ["22"] }, + { key = "GUNICORN_VERSION", value = "23.0.0", if = ["23"] }, ] [envs.default] diff --git a/gunicorn/pyproject.toml b/gunicorn/pyproject.toml index 36eecd88d930e..3b01aaeedcb86 100644 --- a/gunicorn/pyproject.toml +++ b/gunicorn/pyproject.toml @@ -24,7 +24,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/haproxy/changelog.d/18207.added b/haproxy/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/haproxy/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/haproxy/hatch.toml b/haproxy/hatch.toml index 1ed6a7c98c1b6..4e4fff488c80d 100644 --- a/haproxy/hatch.toml +++ b/haproxy/hatch.toml @@ -1,11 +1,11 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] version = ["2.0", "2.2", "2.4", "2.5", "2.6", "2.7"] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] version = ["2.0"] impl = ["legacy"] diff --git a/haproxy/pyproject.toml b/haproxy/pyproject.toml index 331fd6d769f85..8b78030486b59 100644 --- a/haproxy/pyproject.toml +++ b/haproxy/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/harbor/changelog.d/18207.added b/harbor/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/harbor/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/harbor/hatch.toml b/harbor/hatch.toml index 155d1da3f4ca9..fee63406f4cb5 100644 --- a/harbor/hatch.toml +++ b/harbor/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] version = ["1.10", "2.0", "2.3"] [envs.default.overrides] diff --git a/harbor/pyproject.toml b/harbor/pyproject.toml index cfa46f3e817ba..c5220e2e61b2c 100644 --- a/harbor/pyproject.toml +++ b/harbor/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/hazelcast/changelog.d/18207.added b/hazelcast/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/hazelcast/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/hazelcast/hatch.toml b/hazelcast/hatch.toml index af5357b0b4b14..32f8bc109ed3c 100644 --- a/hazelcast/hatch.toml +++ b/hazelcast/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] version = ["4.0", "5.0"] [envs.default.overrides] diff --git a/hdfs_datanode/changelog.d/18207.added b/hdfs_datanode/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/hdfs_datanode/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/hdfs_datanode/hatch.toml b/hdfs_datanode/hatch.toml index e4c471521e26d..44e0c07abc17b 100644 --- a/hdfs_datanode/hatch.toml +++ b/hdfs_datanode/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] [envs.default.env-vars] HDFS_RAW_VERSION = "3.1.3" diff --git a/hdfs_datanode/pyproject.toml b/hdfs_datanode/pyproject.toml index 3052f61d40d3d..e22554bcdc479 100644 --- a/hdfs_datanode/pyproject.toml +++ b/hdfs_datanode/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/hdfs_namenode/changelog.d/18207.added b/hdfs_namenode/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/hdfs_namenode/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/hdfs_namenode/hatch.toml b/hdfs_namenode/hatch.toml index e4c471521e26d..44e0c07abc17b 100644 --- a/hdfs_namenode/hatch.toml +++ b/hdfs_namenode/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] [envs.default.env-vars] HDFS_RAW_VERSION = "3.1.3" diff --git a/hdfs_namenode/pyproject.toml b/hdfs_namenode/pyproject.toml index 7cff415a294fe..9b56ef48480d6 100644 --- a/hdfs_namenode/pyproject.toml +++ b/hdfs_namenode/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/hive/changelog.d/18207.added b/hive/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/hive/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/hive/hatch.toml b/hive/hatch.toml index ce1f02498d72b..8eb7559f584f9 100644 --- a/hive/hatch.toml +++ b/hive/hatch.toml @@ -1,5 +1,5 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] diff --git a/hive/pyproject.toml b/hive/pyproject.toml index 8fda69785c74c..4b29a04186362 100644 --- a/hive/pyproject.toml +++ b/hive/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/hivemq/changelog.d/18207.added b/hivemq/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/hivemq/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/hivemq/hatch.toml b/hivemq/hatch.toml index 29c96ca15410c..c89132b5629e0 100644 --- a/hivemq/hatch.toml +++ b/hivemq/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] version = ["4.3"] [envs.default.overrides] diff --git a/hivemq/pyproject.toml b/hivemq/pyproject.toml index 36cbeaf9e5072..a55b56328dd44 100644 --- a/hivemq/pyproject.toml +++ b/hivemq/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/http_check/changelog.d/18207.added b/http_check/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/http_check/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/http_check/hatch.toml b/http_check/hatch.toml index 91dfdd58ecd27..1bd2b7ae9753c 100644 --- a/http_check/hatch.toml +++ b/http_check/hatch.toml @@ -6,4 +6,4 @@ dependencies = [ ] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] diff --git a/http_check/pyproject.toml b/http_check/pyproject.toml index 414969c8f7962..078d51899ecaf 100644 --- a/http_check/pyproject.toml +++ b/http_check/pyproject.toml @@ -24,7 +24,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/hudi/changelog.d/18207.added b/hudi/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/hudi/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/hudi/hatch.toml b/hudi/hatch.toml index 74960053bd5e8..ff5633dcc7f65 100644 --- a/hudi/hatch.toml +++ b/hudi/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] [envs.default.env-vars] SPARK_VERSION = "3.2.0" diff --git a/hudi/pyproject.toml b/hudi/pyproject.toml index 24f95771d0f47..83e80c2170139 100644 --- a/hudi/pyproject.toml +++ b/hudi/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/hyperv/changelog.d/18207.added b/hyperv/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/hyperv/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/hyperv/hatch.toml b/hyperv/hatch.toml index dab04656200a9..0a3f9b470993b 100644 --- a/hyperv/hatch.toml +++ b/hyperv/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] [envs.default] platforms = [ diff --git a/hyperv/pyproject.toml b/hyperv/pyproject.toml index 572102aee2449..82f3a96acb8ac 100644 --- a/hyperv/pyproject.toml +++ b/hyperv/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/ibm_ace/changelog.d/18207.added b/ibm_ace/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/ibm_ace/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/ibm_ace/hatch.toml b/ibm_ace/hatch.toml index 9cd5c8e7d957b..6ad867c8725d4 100644 --- a/ibm_ace/hatch.toml +++ b/ibm_ace/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] version = ["12"] [envs.default.overrides] diff --git a/ibm_ace/pyproject.toml b/ibm_ace/pyproject.toml index 1b27a99a0bdde..b601bac23083d 100644 --- a/ibm_ace/pyproject.toml +++ b/ibm_ace/pyproject.toml @@ -23,7 +23,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/ibm_db2/changelog.d/18207.added b/ibm_db2/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/ibm_db2/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/ibm_db2/hatch.toml b/ibm_db2/hatch.toml index d7f555ee23243..e57d89a6eff93 100644 --- a/ibm_db2/hatch.toml +++ b/ibm_db2/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] version = ["11.1"] [envs.default.env-vars] diff --git a/ibm_db2/pyproject.toml b/ibm_db2/pyproject.toml index ed4b3a696092b..2078c82f10c23 100644 --- a/ibm_db2/pyproject.toml +++ b/ibm_db2/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/ibm_i/changelog.d/18207.added b/ibm_i/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/ibm_i/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/ibm_i/hatch.toml b/ibm_i/hatch.toml index 001e43ce25414..c85c5f07a7df2 100644 --- a/ibm_i/hatch.toml +++ b/ibm_i/hatch.toml @@ -1,4 +1,4 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] diff --git a/ibm_i/pyproject.toml b/ibm_i/pyproject.toml index fe8b3fe6d7229..104f0f2b53af7 100644 --- a/ibm_i/pyproject.toml +++ b/ibm_i/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/ibm_mq/changelog.d/18207.added b/ibm_mq/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/ibm_mq/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/ibm_mq/hatch.toml b/ibm_mq/hatch.toml index 4c4f57357d588..84e7759604957 100644 --- a/ibm_mq/hatch.toml +++ b/ibm_mq/hatch.toml @@ -6,11 +6,11 @@ mypy-deps = [ ] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] version = ["9"] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] version = ["9"] setup = ["cluster"] diff --git a/ibm_mq/pyproject.toml b/ibm_mq/pyproject.toml index c5d42f6be30c3..3cd5ce9f36be8 100644 --- a/ibm_mq/pyproject.toml +++ b/ibm_mq/pyproject.toml @@ -22,7 +22,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/ibm_was/changelog.d/18207.added b/ibm_was/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/ibm_was/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/ibm_was/hatch.toml b/ibm_was/hatch.toml index 845a7f93657c4..87d4bcf1402c2 100644 --- a/ibm_was/hatch.toml +++ b/ibm_was/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] # We do not test IBM WAS v8. IBM themselves no longer build a docker image for this # version and we don't see the need to build and host an image ourselves. For reference: # https://github.com/WASdev/ci.docker.websphere-traditional/issues/198 diff --git a/ibm_was/pyproject.toml b/ibm_was/pyproject.toml index d56eac6de342a..59dfa8bd0dfd6 100644 --- a/ibm_was/pyproject.toml +++ b/ibm_was/pyproject.toml @@ -24,7 +24,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/ignite/changelog.d/18207.added b/ignite/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/ignite/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/ignite/hatch.toml b/ignite/hatch.toml index 759749b520cbd..29430b5f3ae3a 100644 --- a/ignite/hatch.toml +++ b/ignite/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] version = ["2.8", "2.14"] [envs.default.overrides] diff --git a/ignite/pyproject.toml b/ignite/pyproject.toml index e42014a2a7e6e..c60bab739962f 100644 --- a/ignite/pyproject.toml +++ b/ignite/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/iis/changelog.d/18207.added b/iis/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/iis/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/iis/hatch.toml b/iis/hatch.toml index 7e3cb4fcef5db..53691c07eaa4c 100644 --- a/iis/hatch.toml +++ b/iis/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] [envs.default] dependencies = [ diff --git a/iis/pyproject.toml b/iis/pyproject.toml index b367c7ed75362..2606590ffc384 100644 --- a/iis/pyproject.toml +++ b/iis/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/impala/changelog.d/18207.added b/impala/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/impala/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/impala/hatch.toml b/impala/hatch.toml index 866c962647def..a402ec87b3ca2 100644 --- a/impala/hatch.toml +++ b/impala/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] version = ["4.0.0"] [envs.default.overrides] diff --git a/impala/pyproject.toml b/impala/pyproject.toml index 1676f6ad0f0f4..cfbea39777d50 100644 --- a/impala/pyproject.toml +++ b/impala/pyproject.toml @@ -26,7 +26,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/istio/changelog.d/18207.added b/istio/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/istio/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/istio/hatch.toml b/istio/hatch.toml index 3d89bb385df58..5305eac462c46 100644 --- a/istio/hatch.toml +++ b/istio/hatch.toml @@ -6,7 +6,7 @@ dependencies = [ ] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] version = ["1.13"] [envs.default.overrides] diff --git a/istio/pyproject.toml b/istio/pyproject.toml index 7ca39cfb40e54..7bbf30c2ef909 100644 --- a/istio/pyproject.toml +++ b/istio/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/jboss_wildfly/changelog.d/18207.added b/jboss_wildfly/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/jboss_wildfly/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/jboss_wildfly/hatch.toml b/jboss_wildfly/hatch.toml index 001e43ce25414..c85c5f07a7df2 100644 --- a/jboss_wildfly/hatch.toml +++ b/jboss_wildfly/hatch.toml @@ -1,4 +1,4 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] diff --git a/jboss_wildfly/pyproject.toml b/jboss_wildfly/pyproject.toml index fba55b673c0b7..52c633037058f 100644 --- a/jboss_wildfly/pyproject.toml +++ b/jboss_wildfly/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/journald/changelog.d/18207.added b/journald/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/journald/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/journald/pyproject.toml b/journald/pyproject.toml index 010340badfe1b..968b34e9a8aff 100644 --- a/journald/pyproject.toml +++ b/journald/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/kafka/changelog.d/18207.added b/kafka/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/kafka/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/kafka/hatch.toml b/kafka/hatch.toml index c368c6753b5a1..2029b40f6bec6 100644 --- a/kafka/hatch.toml +++ b/kafka/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] version = ["2.8", "3.3"] [envs.default.overrides] diff --git a/kafka/pyproject.toml b/kafka/pyproject.toml index 19a35c76e8c35..e3551f1dbcdcd 100644 --- a/kafka/pyproject.toml +++ b/kafka/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/kafka_consumer/changelog.d/18207.added b/kafka_consumer/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/kafka_consumer/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/kafka_consumer/hatch.toml b/kafka_consumer/hatch.toml index c2ae46bf79694..2c54fc0f8f402 100644 --- a/kafka_consumer/hatch.toml +++ b/kafka_consumer/hatch.toml @@ -13,11 +13,11 @@ ZK_VERSION = "3.6.4" AUTHENTICATION = "noauth" [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] version = ["2.6", "3.3"] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] version = ["3.3"] auth = ["ssl", "kerberos"] diff --git a/kafka_consumer/pyproject.toml b/kafka_consumer/pyproject.toml index 3e3eb29708129..49cf77e4a3aea 100644 --- a/kafka_consumer/pyproject.toml +++ b/kafka_consumer/pyproject.toml @@ -22,7 +22,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/karpenter/changelog.d/18207.added b/karpenter/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/karpenter/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/karpenter/hatch.toml b/karpenter/hatch.toml index 001e43ce25414..c85c5f07a7df2 100644 --- a/karpenter/hatch.toml +++ b/karpenter/hatch.toml @@ -1,4 +1,4 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] diff --git a/karpenter/pyproject.toml b/karpenter/pyproject.toml index c2c65ee312910..3b9ed9d4acad5 100644 --- a/karpenter/pyproject.toml +++ b/karpenter/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", "Private :: Do Not Upload", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", ] dependencies = [ diff --git a/kong/changelog.d/18207.added b/kong/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/kong/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/kong/hatch.toml b/kong/hatch.toml index 8f714cf1fe6de..12f000163835e 100644 --- a/kong/hatch.toml +++ b/kong/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] version = ["1.5.0", "3.0.0"] [envs.default.overrides] diff --git a/kong/pyproject.toml b/kong/pyproject.toml index 5e637f640fd1c..a3e1d8f6a6b8c 100644 --- a/kong/pyproject.toml +++ b/kong/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/kube_apiserver_metrics/changelog.d/18207.added b/kube_apiserver_metrics/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/kube_apiserver_metrics/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/kube_apiserver_metrics/hatch.toml b/kube_apiserver_metrics/hatch.toml index 821e6ed8e37ca..f62423b83c762 100644 --- a/kube_apiserver_metrics/hatch.toml +++ b/kube_apiserver_metrics/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] [envs.default.env-vars] DDEV_SKIP_GENERIC_TAGS_CHECK = "true" diff --git a/kube_apiserver_metrics/pyproject.toml b/kube_apiserver_metrics/pyproject.toml index b13c476bffdb4..42aa982f44cdb 100644 --- a/kube_apiserver_metrics/pyproject.toml +++ b/kube_apiserver_metrics/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/kube_controller_manager/changelog.d/18207.added b/kube_controller_manager/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/kube_controller_manager/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/kube_controller_manager/hatch.toml b/kube_controller_manager/hatch.toml index e3a9dc3faa4d7..9184e295226c9 100644 --- a/kube_controller_manager/hatch.toml +++ b/kube_controller_manager/hatch.toml @@ -11,7 +11,7 @@ dependencies = [ ] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] [envs.default.env-vars] DDEV_SKIP_GENERIC_TAGS_CHECK = "true" diff --git a/kube_controller_manager/pyproject.toml b/kube_controller_manager/pyproject.toml index 17c7abbf86546..571bae1836952 100644 --- a/kube_controller_manager/pyproject.toml +++ b/kube_controller_manager/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/kube_dns/changelog.d/18207.added b/kube_dns/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/kube_dns/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/kube_dns/hatch.toml b/kube_dns/hatch.toml index 001e43ce25414..c85c5f07a7df2 100644 --- a/kube_dns/hatch.toml +++ b/kube_dns/hatch.toml @@ -1,4 +1,4 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] diff --git a/kube_dns/pyproject.toml b/kube_dns/pyproject.toml index 67ea433f081d0..6c8c2b793f826 100644 --- a/kube_dns/pyproject.toml +++ b/kube_dns/pyproject.toml @@ -24,7 +24,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/kube_metrics_server/changelog.d/18207.added b/kube_metrics_server/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/kube_metrics_server/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/kube_metrics_server/hatch.toml b/kube_metrics_server/hatch.toml index 001e43ce25414..c85c5f07a7df2 100644 --- a/kube_metrics_server/hatch.toml +++ b/kube_metrics_server/hatch.toml @@ -1,4 +1,4 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] diff --git a/kube_metrics_server/pyproject.toml b/kube_metrics_server/pyproject.toml index ed80bd29a6270..e169008a48904 100644 --- a/kube_metrics_server/pyproject.toml +++ b/kube_metrics_server/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/kube_proxy/changelog.d/18207.added b/kube_proxy/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/kube_proxy/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/kube_proxy/hatch.toml b/kube_proxy/hatch.toml index 7514917802f38..b9882b7af0527 100644 --- a/kube_proxy/hatch.toml +++ b/kube_proxy/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] [envs.default] platforms = [ diff --git a/kube_proxy/pyproject.toml b/kube_proxy/pyproject.toml index 6a4c1a1cdafc7..59b2b46a4ba28 100644 --- a/kube_proxy/pyproject.toml +++ b/kube_proxy/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/kube_scheduler/changelog.d/18207.added b/kube_scheduler/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/kube_scheduler/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/kube_scheduler/hatch.toml b/kube_scheduler/hatch.toml index cc70f94c080ac..1e5a9a43783c7 100644 --- a/kube_scheduler/hatch.toml +++ b/kube_scheduler/hatch.toml @@ -5,7 +5,7 @@ base-package-features = [ ] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] [envs.default] dependencies = [ diff --git a/kube_scheduler/pyproject.toml b/kube_scheduler/pyproject.toml index 620f541b83b70..871ea3e106fcc 100644 --- a/kube_scheduler/pyproject.toml +++ b/kube_scheduler/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/kubeflow/pyproject.toml b/kubeflow/pyproject.toml index a7a3f0c00afda..507e230c0b3f3 100644 --- a/kubeflow/pyproject.toml +++ b/kubeflow/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", "Private :: Do Not Upload", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", ] dependencies = [ diff --git a/kubelet/changelog.d/18207.added b/kubelet/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/kubelet/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/kubelet/hatch.toml b/kubelet/hatch.toml index 7797d96f22bc1..2c4aa4d9760ec 100644 --- a/kubelet/hatch.toml +++ b/kubelet/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] [envs.default] dependencies = [ diff --git a/kubelet/pyproject.toml b/kubelet/pyproject.toml index 7fd6a3cc53cc4..075bd5b2e3e23 100644 --- a/kubelet/pyproject.toml +++ b/kubelet/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/kubernetes_cluster_autoscaler/changelog.d/18207.added b/kubernetes_cluster_autoscaler/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/kubernetes_cluster_autoscaler/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/kubernetes_cluster_autoscaler/hatch.toml b/kubernetes_cluster_autoscaler/hatch.toml index 001e43ce25414..c85c5f07a7df2 100644 --- a/kubernetes_cluster_autoscaler/hatch.toml +++ b/kubernetes_cluster_autoscaler/hatch.toml @@ -1,4 +1,4 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] diff --git a/kubernetes_cluster_autoscaler/pyproject.toml b/kubernetes_cluster_autoscaler/pyproject.toml index 0b704a8614b3c..b3d94f3a08466 100644 --- a/kubernetes_cluster_autoscaler/pyproject.toml +++ b/kubernetes_cluster_autoscaler/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", "Private :: Do Not Upload", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", ] dependencies = [ diff --git a/kubernetes_state/changelog.d/18207.added b/kubernetes_state/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/kubernetes_state/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/kubernetes_state/hatch.toml b/kubernetes_state/hatch.toml index 821e6ed8e37ca..f62423b83c762 100644 --- a/kubernetes_state/hatch.toml +++ b/kubernetes_state/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] [envs.default.env-vars] DDEV_SKIP_GENERIC_TAGS_CHECK = "true" diff --git a/kubernetes_state/pyproject.toml b/kubernetes_state/pyproject.toml index 1d5704d3b20c1..ece60796d07fb 100644 --- a/kubernetes_state/pyproject.toml +++ b/kubernetes_state/pyproject.toml @@ -24,7 +24,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/kubevirt_api/hatch.toml b/kubevirt_api/hatch.toml index 001e43ce25414..c85c5f07a7df2 100644 --- a/kubevirt_api/hatch.toml +++ b/kubevirt_api/hatch.toml @@ -1,4 +1,4 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] diff --git a/kubevirt_api/pyproject.toml b/kubevirt_api/pyproject.toml index 0b70603c9b141..5b09028db4589 100644 --- a/kubevirt_api/pyproject.toml +++ b/kubevirt_api/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", "Private :: Do Not Upload", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", ] dependencies = [ diff --git a/kubevirt_controller/hatch.toml b/kubevirt_controller/hatch.toml index 001e43ce25414..c85c5f07a7df2 100644 --- a/kubevirt_controller/hatch.toml +++ b/kubevirt_controller/hatch.toml @@ -1,4 +1,4 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] diff --git a/kubevirt_controller/pyproject.toml b/kubevirt_controller/pyproject.toml index 7ef8a76365552..dfd3797339a24 100644 --- a/kubevirt_controller/pyproject.toml +++ b/kubevirt_controller/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", "Private :: Do Not Upload", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", ] dependencies = [ diff --git a/kyototycoon/changelog.d/18207.added b/kyototycoon/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/kyototycoon/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/kyototycoon/hatch.toml b/kyototycoon/hatch.toml index 001e43ce25414..c85c5f07a7df2 100644 --- a/kyototycoon/hatch.toml +++ b/kyototycoon/hatch.toml @@ -1,4 +1,4 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] diff --git a/kyototycoon/pyproject.toml b/kyototycoon/pyproject.toml index aac5a3f24775a..7fa95a604f50a 100644 --- a/kyototycoon/pyproject.toml +++ b/kyototycoon/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/kyverno/changelog.d/18207.added b/kyverno/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/kyverno/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/kyverno/hatch.toml b/kyverno/hatch.toml index 001e43ce25414..c85c5f07a7df2 100644 --- a/kyverno/hatch.toml +++ b/kyverno/hatch.toml @@ -1,4 +1,4 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] diff --git a/kyverno/pyproject.toml b/kyverno/pyproject.toml index 8769aace3c165..51079fea365e8 100644 --- a/kyverno/pyproject.toml +++ b/kyverno/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", "Private :: Do Not Upload", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", ] dependencies = [ diff --git a/lighttpd/changelog.d/18207.added b/lighttpd/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/lighttpd/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/lighttpd/hatch.toml b/lighttpd/hatch.toml index cafdb3788cad1..261c4802bdf29 100644 --- a/lighttpd/hatch.toml +++ b/lighttpd/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] flavor = ["auth", "noauth"] [envs.default.overrides] diff --git a/lighttpd/pyproject.toml b/lighttpd/pyproject.toml index 28eb6b254122e..ac8ff3cac5eb6 100644 --- a/lighttpd/pyproject.toml +++ b/lighttpd/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/linkerd/changelog.d/18207.added b/linkerd/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/linkerd/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/linkerd/hatch.toml b/linkerd/hatch.toml index c4f5c489e98a5..46952cc77c421 100644 --- a/linkerd/hatch.toml +++ b/linkerd/hatch.toml @@ -6,7 +6,7 @@ dependencies = [ ] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] [envs.default.env-vars] DDEV_SKIP_GENERIC_TAGS_CHECK = "true" diff --git a/linkerd/pyproject.toml b/linkerd/pyproject.toml index 45f685a892bf4..458a3e1fb10ac 100644 --- a/linkerd/pyproject.toml +++ b/linkerd/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/linux_proc_extras/changelog.d/18207.added b/linux_proc_extras/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/linux_proc_extras/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/linux_proc_extras/hatch.toml b/linux_proc_extras/hatch.toml index 001e43ce25414..c85c5f07a7df2 100644 --- a/linux_proc_extras/hatch.toml +++ b/linux_proc_extras/hatch.toml @@ -1,4 +1,4 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] diff --git a/linux_proc_extras/pyproject.toml b/linux_proc_extras/pyproject.toml index b92af0be3195b..9d77452d0b67f 100644 --- a/linux_proc_extras/pyproject.toml +++ b/linux_proc_extras/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/mapr/changelog.d/18207.added b/mapr/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/mapr/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/mapr/hatch.toml b/mapr/hatch.toml index 0089c2a95f7eb..efa3452fe613c 100644 --- a/mapr/hatch.toml +++ b/mapr/hatch.toml @@ -7,4 +7,4 @@ dependencies = [ ] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] diff --git a/mapr/pyproject.toml b/mapr/pyproject.toml index ad39020b8805b..dd3c0e309f8f1 100644 --- a/mapr/pyproject.toml +++ b/mapr/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/mapreduce/changelog.d/18207.added b/mapreduce/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/mapreduce/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/mapreduce/hatch.toml b/mapreduce/hatch.toml index 85c9830af9b9e..11adac3e3342c 100644 --- a/mapreduce/hatch.toml +++ b/mapreduce/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] [envs.default] dependencies = [ diff --git a/mapreduce/pyproject.toml b/mapreduce/pyproject.toml index 1f77e1e9e45e8..cf57dab669bf7 100644 --- a/mapreduce/pyproject.toml +++ b/mapreduce/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/marathon/changelog.d/18207.added b/marathon/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/marathon/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/marathon/hatch.toml b/marathon/hatch.toml index 821e6ed8e37ca..f62423b83c762 100644 --- a/marathon/hatch.toml +++ b/marathon/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] [envs.default.env-vars] DDEV_SKIP_GENERIC_TAGS_CHECK = "true" diff --git a/marathon/pyproject.toml b/marathon/pyproject.toml index be2ec12a08957..0417f459c148c 100644 --- a/marathon/pyproject.toml +++ b/marathon/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/marklogic/changelog.d/18207.added b/marklogic/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/marklogic/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/marklogic/hatch.toml b/marklogic/hatch.toml index 4883100aa7231..b7093e50a81fc 100644 --- a/marklogic/hatch.toml +++ b/marklogic/hatch.toml @@ -6,7 +6,7 @@ mypy-deps = [ ] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] version = ["9.0", "10.0", "11.0"] [envs.default.overrides] diff --git a/marklogic/pyproject.toml b/marklogic/pyproject.toml index c36da31616366..32a75ba8be794 100644 --- a/marklogic/pyproject.toml +++ b/marklogic/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/mcache/changelog.d/18207.added b/mcache/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/mcache/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/mcache/hatch.toml b/mcache/hatch.toml index 821e6ed8e37ca..f62423b83c762 100644 --- a/mcache/hatch.toml +++ b/mcache/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] [envs.default.env-vars] DDEV_SKIP_GENERIC_TAGS_CHECK = "true" diff --git a/mcache/pyproject.toml b/mcache/pyproject.toml index 535b4bb5a02ab..3416ceb26b7ab 100644 --- a/mcache/pyproject.toml +++ b/mcache/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/mesos_master/changelog.d/18207.added b/mesos_master/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/mesos_master/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/mesos_master/hatch.toml b/mesos_master/hatch.toml index b34d77b810e1f..9fd016769c5bf 100644 --- a/mesos_master/hatch.toml +++ b/mesos_master/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] version = ["1.7"] [envs.default.overrides] diff --git a/mesos_master/pyproject.toml b/mesos_master/pyproject.toml index e08054638645c..ceed88f8f8ba7 100644 --- a/mesos_master/pyproject.toml +++ b/mesos_master/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/mesos_slave/changelog.d/18207.added b/mesos_slave/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/mesos_slave/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/mesos_slave/hatch.toml b/mesos_slave/hatch.toml index 6c838971ec02e..6b47bf4bf3e3e 100644 --- a/mesos_slave/hatch.toml +++ b/mesos_slave/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] version = ["1.7"] [envs.default.overrides] diff --git a/mesos_slave/pyproject.toml b/mesos_slave/pyproject.toml index ef842e6069708..caa42505979eb 100644 --- a/mesos_slave/pyproject.toml +++ b/mesos_slave/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/mongo/changelog.d/18207.added b/mongo/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/mongo/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/mongo/hatch.toml b/mongo/hatch.toml index 57204ea122ef2..30c0e594ed026 100644 --- a/mongo/hatch.toml +++ b/mongo/hatch.toml @@ -7,7 +7,7 @@ dependencies = [ ] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] version = ["4.4", "5.0", "6.0", "7.0"] flavor = ["standalone", "shard", "auth", "tls"] diff --git a/mongo/pyproject.toml b/mongo/pyproject.toml index ae31a970f390d..de71c7d49ce1a 100644 --- a/mongo/pyproject.toml +++ b/mongo/pyproject.toml @@ -24,7 +24,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/mysql/changelog.d/18207.added b/mysql/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/mysql/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/mysql/hatch.toml b/mysql/hatch.toml index 103bee10d87a2..2f6784e9ce02d 100644 --- a/mysql/hatch.toml +++ b/mysql/hatch.toml @@ -14,7 +14,7 @@ mypy-deps = [ ] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] version = [ "5.7", # EOL October 21, 2023 "8.0.36", # EOL April, 2026 @@ -22,12 +22,12 @@ version = [ ] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] version = ["8.0"] replication = ["group"] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] flavor = ["mariadb"] version = [ "10.2", # EOL 23 May 2022 (ended) diff --git a/mysql/pyproject.toml b/mysql/pyproject.toml index 01d6c9a97c5ba..097ffe0d70c8c 100644 --- a/mysql/pyproject.toml +++ b/mysql/pyproject.toml @@ -23,7 +23,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/nagios/changelog.d/18207.added b/nagios/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/nagios/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/nagios/hatch.toml b/nagios/hatch.toml index 8ca57231866cf..1002bb8b79cb3 100644 --- a/nagios/hatch.toml +++ b/nagios/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] version = ["4.4"] [envs.default.overrides] diff --git a/nagios/pyproject.toml b/nagios/pyproject.toml index 67de6ad22912d..559e5828704ce 100644 --- a/nagios/pyproject.toml +++ b/nagios/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/network/changelog.d/18207.added b/network/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/network/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/network/hatch.toml b/network/hatch.toml index 265647e85c0b8..89a24834a388b 100644 --- a/network/hatch.toml +++ b/network/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] [envs.default.overrides] platform.windows.e2e-env = { value = false } diff --git a/network/pyproject.toml b/network/pyproject.toml index f038e2d5eeb73..dc9cc589fcbac 100644 --- a/network/pyproject.toml +++ b/network/pyproject.toml @@ -24,7 +24,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/nfsstat/changelog.d/18207.added b/nfsstat/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/nfsstat/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/nfsstat/hatch.toml b/nfsstat/hatch.toml index 001e43ce25414..c85c5f07a7df2 100644 --- a/nfsstat/hatch.toml +++ b/nfsstat/hatch.toml @@ -1,4 +1,4 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] diff --git a/nfsstat/pyproject.toml b/nfsstat/pyproject.toml index 78f478a8bf029..63ba1349c8801 100644 --- a/nfsstat/pyproject.toml +++ b/nfsstat/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/nginx/changelog.d/18207.added b/nginx/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/nginx/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/nginx/hatch.toml b/nginx/hatch.toml index 63446a3942773..0a0a7e7475720 100644 --- a/nginx/hatch.toml +++ b/nginx/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] version = ["1.12", "1.13", "vts"] [envs.default.env-vars] diff --git a/nginx/pyproject.toml b/nginx/pyproject.toml index 0c6c5a37a3f5a..739baaff89c9c 100644 --- a/nginx/pyproject.toml +++ b/nginx/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/nginx_ingress_controller/changelog.d/18207.added b/nginx_ingress_controller/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/nginx_ingress_controller/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/nginx_ingress_controller/hatch.toml b/nginx_ingress_controller/hatch.toml index 821e6ed8e37ca..f62423b83c762 100644 --- a/nginx_ingress_controller/hatch.toml +++ b/nginx_ingress_controller/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] [envs.default.env-vars] DDEV_SKIP_GENERIC_TAGS_CHECK = "true" diff --git a/nginx_ingress_controller/pyproject.toml b/nginx_ingress_controller/pyproject.toml index c6ade0257b65b..8fe0a7fdaea79 100644 --- a/nginx_ingress_controller/pyproject.toml +++ b/nginx_ingress_controller/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/nvidia_triton/changelog.d/18207.added b/nvidia_triton/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/nvidia_triton/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/nvidia_triton/hatch.toml b/nvidia_triton/hatch.toml index 001e43ce25414..c85c5f07a7df2 100644 --- a/nvidia_triton/hatch.toml +++ b/nvidia_triton/hatch.toml @@ -1,4 +1,4 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] diff --git a/nvidia_triton/pyproject.toml b/nvidia_triton/pyproject.toml index 17e9c35e0f92a..2803d702469d5 100644 --- a/nvidia_triton/pyproject.toml +++ b/nvidia_triton/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", "Private :: Do Not Upload", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", ] dependencies = [ diff --git a/openldap/changelog.d/18207.added b/openldap/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/openldap/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/openldap/hatch.toml b/openldap/hatch.toml index 912bbc34f2b46..3ffcb288c9702 100644 --- a/openldap/hatch.toml +++ b/openldap/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] version = ["2.4", "2.6"] [envs.default.overrides] diff --git a/openldap/pyproject.toml b/openldap/pyproject.toml index cdbbc1723d277..49683d15481fb 100644 --- a/openldap/pyproject.toml +++ b/openldap/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/openmetrics/changelog.d/18207.added b/openmetrics/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/openmetrics/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/openmetrics/hatch.toml b/openmetrics/hatch.toml index 001e43ce25414..c85c5f07a7df2 100644 --- a/openmetrics/hatch.toml +++ b/openmetrics/hatch.toml @@ -1,4 +1,4 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] diff --git a/openmetrics/pyproject.toml b/openmetrics/pyproject.toml index 999a17e36146f..3175dc1edc5a3 100644 --- a/openmetrics/pyproject.toml +++ b/openmetrics/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/openstack/changelog.d/18207.added b/openstack/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/openstack/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/openstack/hatch.toml b/openstack/hatch.toml index 001e43ce25414..c85c5f07a7df2 100644 --- a/openstack/hatch.toml +++ b/openstack/hatch.toml @@ -1,4 +1,4 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] diff --git a/openstack/pyproject.toml b/openstack/pyproject.toml index 096272e19f7de..7a8e5c4b24995 100644 --- a/openstack/pyproject.toml +++ b/openstack/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/openstack_controller/changelog.d/18207.added b/openstack_controller/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/openstack_controller/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/openstack_controller/hatch.toml b/openstack_controller/hatch.toml index 8d6ec546035fd..a7511eb2248ce 100644 --- a/openstack_controller/hatch.toml +++ b/openstack_controller/hatch.toml @@ -1,19 +1,19 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] setup = ["legacy"] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] setup = ["gcp"] [envs.default.overrides] -name."^py3.11$".e2e-env = { value = true } -name."^py3.11-legacy$".e2e-env = { value = true } +name."^py3.12$".e2e-env = { value = true } +name."^py3.12-legacy$".e2e-env = { value = true } matrix.setup.e2e-env = { value = true, if = ["gcp"], env = ["TF_VAR_credentials_file", "TF_VAR_instance_name", "TF_VAR_desired_status", "TF_VAR_nat_ip", "TF_VAR_network_ip", "TF_VAR_user"] } matrix.setup.env-vars = [ { key = "USE_OPENSTACK_GCP", value = "true", if = ["gcp"] }, diff --git a/openstack_controller/pyproject.toml b/openstack_controller/pyproject.toml index 92241eb9e8f30..44382e19555e2 100644 --- a/openstack_controller/pyproject.toml +++ b/openstack_controller/pyproject.toml @@ -23,7 +23,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/oracle/changelog.d/18207.added b/oracle/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/oracle/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/oracle/pyproject.toml b/oracle/pyproject.toml index ffdc20014a96f..13d54f400810e 100644 --- a/oracle/pyproject.toml +++ b/oracle/pyproject.toml @@ -23,7 +23,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/ossec_security/pyproject.toml b/ossec_security/pyproject.toml index 29062b7091fdf..13e4a768ca451 100644 --- a/ossec_security/pyproject.toml +++ b/ossec_security/pyproject.toml @@ -24,7 +24,7 @@ classifiers = [ "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", "Private :: Do Not Upload", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", ] dependencies = [ diff --git a/pan_firewall/changelog.d/18207.added b/pan_firewall/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/pan_firewall/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/pan_firewall/pyproject.toml b/pan_firewall/pyproject.toml index 4fba326e78205..1a856ed8e9e84 100644 --- a/pan_firewall/pyproject.toml +++ b/pan_firewall/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/pdh_check/changelog.d/18207.added b/pdh_check/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/pdh_check/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/pdh_check/hatch.toml b/pdh_check/hatch.toml index 413e6c51fe94c..0f9f06cf58486 100644 --- a/pdh_check/hatch.toml +++ b/pdh_check/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] [envs.default] dependencies = [ diff --git a/pdh_check/pyproject.toml b/pdh_check/pyproject.toml index 92fb3eebe16b4..998815602818c 100644 --- a/pdh_check/pyproject.toml +++ b/pdh_check/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/pgbouncer/changelog.d/18207.added b/pgbouncer/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/pgbouncer/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/pgbouncer/hatch.toml b/pgbouncer/hatch.toml index 9e4ba1af299e8..b951fdc61f075 100644 --- a/pgbouncer/hatch.toml +++ b/pgbouncer/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] version = ["1.7", "1.8", "1.12"] [envs.default.env-vars] diff --git a/pgbouncer/pyproject.toml b/pgbouncer/pyproject.toml index 884c207d131e9..a79a5e653d120 100644 --- a/pgbouncer/pyproject.toml +++ b/pgbouncer/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/php_fpm/changelog.d/18207.added b/php_fpm/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/php_fpm/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/php_fpm/hatch.toml b/php_fpm/hatch.toml index 001e43ce25414..c85c5f07a7df2 100644 --- a/php_fpm/hatch.toml +++ b/php_fpm/hatch.toml @@ -1,4 +1,4 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] diff --git a/php_fpm/pyproject.toml b/php_fpm/pyproject.toml index dee5c967d0b09..959b853eab6bb 100644 --- a/php_fpm/pyproject.toml +++ b/php_fpm/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/ping_federate/pyproject.toml b/ping_federate/pyproject.toml index 8f38a7050384b..2940f7c233d93 100644 --- a/ping_federate/pyproject.toml +++ b/ping_federate/pyproject.toml @@ -24,7 +24,7 @@ classifiers = [ "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", "Private :: Do Not Upload", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", ] dependencies = [ diff --git a/postfix/changelog.d/18207.added b/postfix/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/postfix/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/postfix/hatch.toml b/postfix/hatch.toml index 001e43ce25414..c85c5f07a7df2 100644 --- a/postfix/hatch.toml +++ b/postfix/hatch.toml @@ -1,4 +1,4 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] diff --git a/postfix/pyproject.toml b/postfix/pyproject.toml index 004b7abc40e11..c9ff6a26814ab 100644 --- a/postfix/pyproject.toml +++ b/postfix/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/postgres/changelog.d/18207.added b/postgres/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/postgres/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/postgres/hatch.toml b/postgres/hatch.toml index 97823d4e37957..1f920fae0f74b 100644 --- a/postgres/hatch.toml +++ b/postgres/hatch.toml @@ -11,7 +11,7 @@ mypy-deps = [ ] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] version = ["9.6", "10.0", "11.0", "12.17", "13.0", "14.0", "15.0", "16.0"] [envs.default.overrides] diff --git a/postgres/pyproject.toml b/postgres/pyproject.toml index 8a9846261bd37..deed82366ec5d 100644 --- a/postgres/pyproject.toml +++ b/postgres/pyproject.toml @@ -23,7 +23,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/powerdns_recursor/changelog.d/18207.added b/powerdns_recursor/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/powerdns_recursor/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/powerdns_recursor/hatch.toml b/powerdns_recursor/hatch.toml index 2aa20377ea5df..457d6dc0e2ba0 100644 --- a/powerdns_recursor/hatch.toml +++ b/powerdns_recursor/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] version = ["3.7", "4.0"] [envs.default.overrides] diff --git a/powerdns_recursor/pyproject.toml b/powerdns_recursor/pyproject.toml index b1a671cb02c8c..86d4a8e7397ca 100644 --- a/powerdns_recursor/pyproject.toml +++ b/powerdns_recursor/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/presto/changelog.d/18207.added b/presto/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/presto/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/presto/hatch.toml b/presto/hatch.toml index d434b8b3fe36b..a57d0922250db 100644 --- a/presto/hatch.toml +++ b/presto/hatch.toml @@ -6,7 +6,7 @@ dependencies = [ ] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] version = ["346"] [envs.default.overrides] diff --git a/presto/pyproject.toml b/presto/pyproject.toml index d26d8a9ddb241..6e385d08d26e4 100644 --- a/presto/pyproject.toml +++ b/presto/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/process/changelog.d/18207.added b/process/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/process/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/process/hatch.toml b/process/hatch.toml index 3d61446b5d5b6..48af19a1f76fa 100644 --- a/process/hatch.toml +++ b/process/hatch.toml @@ -1,6 +1,6 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] [envs.bench] \ No newline at end of file diff --git a/process/pyproject.toml b/process/pyproject.toml index 8978b66f0fb6b..c55c26760190c 100644 --- a/process/pyproject.toml +++ b/process/pyproject.toml @@ -24,7 +24,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/prometheus/changelog.d/18207.added b/prometheus/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/prometheus/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/prometheus/hatch.toml b/prometheus/hatch.toml index 2e077452e744b..ec24537869bac 100644 --- a/prometheus/hatch.toml +++ b/prometheus/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] [envs.default] platforms = [ diff --git a/prometheus/pyproject.toml b/prometheus/pyproject.toml index 51a4fe0408275..410862fb39e9c 100644 --- a/prometheus/pyproject.toml +++ b/prometheus/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/proxysql/changelog.d/18207.added b/proxysql/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/proxysql/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/proxysql/hatch.toml b/proxysql/hatch.toml index b5cd0a46d8dbd..1640d02d24d12 100644 --- a/proxysql/hatch.toml +++ b/proxysql/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] version = ["1.4", "2.0"] [envs.default.overrides] diff --git a/proxysql/pyproject.toml b/proxysql/pyproject.toml index 09647491222ab..b01ff51923b82 100644 --- a/proxysql/pyproject.toml +++ b/proxysql/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/pulsar/changelog.d/18207.added b/pulsar/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/pulsar/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/pulsar/hatch.toml b/pulsar/hatch.toml index 3de6c23c1a3b5..0355df95f829c 100644 --- a/pulsar/hatch.toml +++ b/pulsar/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] version = ["2.9"] [envs.default.overrides] diff --git a/pulsar/pyproject.toml b/pulsar/pyproject.toml index e6b54194e4b4b..51820fbd38004 100644 --- a/pulsar/pyproject.toml +++ b/pulsar/pyproject.toml @@ -24,7 +24,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/rabbitmq/changelog.d/18207.added b/rabbitmq/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/rabbitmq/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/rabbitmq/hatch.toml b/rabbitmq/hatch.toml index 72ec427026312..1e6d2c4ba85a6 100644 --- a/rabbitmq/hatch.toml +++ b/rabbitmq/hatch.toml @@ -9,7 +9,7 @@ dependencies = [ # Rabbitmq versions 3.8+ introduce the Prometheus plugin. This is the preferred way to collect metrics. # We still support metrics from management plugin as a legacy option. [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] version = ["3.7", "3.11"] flavor = ["mgmt", "openmetrics"] diff --git a/rabbitmq/pyproject.toml b/rabbitmq/pyproject.toml index fd4bd248d71ee..a88015878d478 100644 --- a/rabbitmq/pyproject.toml +++ b/rabbitmq/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/ray/changelog.d/18207.added b/ray/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/ray/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/ray/hatch.toml b/ray/hatch.toml index 4f348c29ba42d..ce67c426c90cd 100644 --- a/ray/hatch.toml +++ b/ray/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] version = ["2.8"] [envs.default.overrides] diff --git a/ray/pyproject.toml b/ray/pyproject.toml index d523a530380f0..2fb0787081869 100644 --- a/ray/pyproject.toml +++ b/ray/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", "Private :: Do Not Upload", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", ] dependencies = [ diff --git a/redisdb/changelog.d/18207.added b/redisdb/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/redisdb/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/redisdb/hatch.toml b/redisdb/hatch.toml index ecf7d726206f8..b4f02e3d707e8 100644 --- a/redisdb/hatch.toml +++ b/redisdb/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] version = ["5.0", "6.0", "7.0", "cloud"] [envs.default.overrides] diff --git a/redisdb/pyproject.toml b/redisdb/pyproject.toml index 8bf1515cfd702..357cff34f951f 100644 --- a/redisdb/pyproject.toml +++ b/redisdb/pyproject.toml @@ -24,7 +24,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/rethinkdb/changelog.d/18207.added b/rethinkdb/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/rethinkdb/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/rethinkdb/hatch.toml b/rethinkdb/hatch.toml index 53e68f253fad0..f379e04150daf 100644 --- a/rethinkdb/hatch.toml +++ b/rethinkdb/hatch.toml @@ -6,7 +6,7 @@ mypy-deps = [ ] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] # Can't support lower 2.3 patch versions due to: https://github.com/rethinkdb/rethinkdb/issues/6108 version = ["2.3", "2.4"] diff --git a/rethinkdb/pyproject.toml b/rethinkdb/pyproject.toml index 88caf7d283c61..722a36b217430 100644 --- a/rethinkdb/pyproject.toml +++ b/rethinkdb/pyproject.toml @@ -24,7 +24,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/riak/changelog.d/18207.added b/riak/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/riak/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/riak/hatch.toml b/riak/hatch.toml index 001e43ce25414..c85c5f07a7df2 100644 --- a/riak/hatch.toml +++ b/riak/hatch.toml @@ -1,4 +1,4 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] diff --git a/riak/pyproject.toml b/riak/pyproject.toml index f956b2e9eda5b..5aea83a4b044c 100644 --- a/riak/pyproject.toml +++ b/riak/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/riakcs/changelog.d/18207.added b/riakcs/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/riakcs/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/riakcs/hatch.toml b/riakcs/hatch.toml index bd72781786c80..769d79d915cc1 100644 --- a/riakcs/hatch.toml +++ b/riakcs/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] [envs.default] e2e-env = false \ No newline at end of file diff --git a/riakcs/pyproject.toml b/riakcs/pyproject.toml index 3f8df8b6a4c56..2f90c1aa18457 100644 --- a/riakcs/pyproject.toml +++ b/riakcs/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/sap_hana/changelog.d/18207.added b/sap_hana/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/sap_hana/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/sap_hana/hatch.toml b/sap_hana/hatch.toml index 7dcab11f35177..ada46df010993 100644 --- a/sap_hana/hatch.toml +++ b/sap_hana/hatch.toml @@ -6,7 +6,7 @@ dependencies = [ ] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] version = ["2.0"] [envs.default.overrides] diff --git a/sap_hana/pyproject.toml b/sap_hana/pyproject.toml index 826b4166ae66e..04d415ff21bb9 100644 --- a/sap_hana/pyproject.toml +++ b/sap_hana/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/scylla/changelog.d/18207.added b/scylla/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/scylla/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/scylla/hatch.toml b/scylla/hatch.toml index 1b37eeb75e5cd..7f0416adaee51 100644 --- a/scylla/hatch.toml +++ b/scylla/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] version = ["3.1", "3.3", "5.2"] [envs.default.overrides] diff --git a/scylla/pyproject.toml b/scylla/pyproject.toml index 5e4355b0f5383..18be1cee6fab3 100644 --- a/scylla/pyproject.toml +++ b/scylla/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/sidekiq/changelog.d/18207.added b/sidekiq/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/sidekiq/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/sidekiq/pyproject.toml b/sidekiq/pyproject.toml index 9cc7236bfd39c..c2ca2e73f32c9 100644 --- a/sidekiq/pyproject.toml +++ b/sidekiq/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/silk/changelog.d/18207.added b/silk/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/silk/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/silk/hatch.toml b/silk/hatch.toml index b271cca6d3e45..0169207d8cffb 100644 --- a/silk/hatch.toml +++ b/silk/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] [envs.default] dependencies = [ diff --git a/silk/pyproject.toml b/silk/pyproject.toml index c3fc8baac47f1..02fbc039bf98c 100644 --- a/silk/pyproject.toml +++ b/silk/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/singlestore/changelog.d/18207.added b/singlestore/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/singlestore/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/singlestore/hatch.toml b/singlestore/hatch.toml index c671e67928494..142ef97f7acb9 100644 --- a/singlestore/hatch.toml +++ b/singlestore/hatch.toml @@ -6,4 +6,4 @@ mypy-deps = [ ] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] diff --git a/singlestore/pyproject.toml b/singlestore/pyproject.toml index 86e5e9297b603..1016e3715cf42 100644 --- a/singlestore/pyproject.toml +++ b/singlestore/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/snmp/changelog.d/18207.added b/snmp/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/snmp/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/snmp/hatch.toml b/snmp/hatch.toml index dddf61c8ad498..10e870f34fb6d 100644 --- a/snmp/hatch.toml +++ b/snmp/hatch.toml @@ -2,7 +2,7 @@ check-types = false [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] snmplistener = ["false", "true"] [envs.default.overrides] diff --git a/snmp/pyproject.toml b/snmp/pyproject.toml index 9e4b5c3e439ae..6ca2c91126604 100644 --- a/snmp/pyproject.toml +++ b/snmp/pyproject.toml @@ -24,7 +24,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/snowflake/changelog.d/18207.added b/snowflake/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/snowflake/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/snowflake/hatch.toml b/snowflake/hatch.toml index 8dd3b0acc729e..0165ed0f9b0d0 100644 --- a/snowflake/hatch.toml +++ b/snowflake/hatch.toml @@ -6,4 +6,4 @@ mypy-deps = [ ] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] diff --git a/snowflake/pyproject.toml b/snowflake/pyproject.toml index ee5113b4c0b6f..735f4c2543306 100644 --- a/snowflake/pyproject.toml +++ b/snowflake/pyproject.toml @@ -23,7 +23,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/solr/changelog.d/18207.added b/solr/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/solr/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/solr/hatch.toml b/solr/hatch.toml index ac11d1443411d..d71e1005cecdf 100644 --- a/solr/hatch.toml +++ b/solr/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] version = ["8.11", "9.3"] [envs.default.overrides] diff --git a/solr/pyproject.toml b/solr/pyproject.toml index 849370d716b6d..7eb592a6226e8 100644 --- a/solr/pyproject.toml +++ b/solr/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/sonarqube/changelog.d/18207.added b/sonarqube/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/sonarqube/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/sonarqube/hatch.toml b/sonarqube/hatch.toml index 543b1094534fa..7f7f353c14ddb 100644 --- a/sonarqube/hatch.toml +++ b/sonarqube/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] version = ["7.9", "8.5", "9.3"] # You can use M1 env if you want to run on M1 processor machine diff --git a/sonarqube/pyproject.toml b/sonarqube/pyproject.toml index b67b915b76d58..c1c3183c7a03b 100644 --- a/sonarqube/pyproject.toml +++ b/sonarqube/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/spark/changelog.d/18207.added b/spark/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/spark/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/spark/hatch.toml b/spark/hatch.toml index ef871733dabda..137a0e3269190 100644 --- a/spark/hatch.toml +++ b/spark/hatch.toml @@ -6,7 +6,7 @@ dependencies = [ ] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] version = ["2.4", "3.0"] [envs.default.overrides] diff --git a/spark/pyproject.toml b/spark/pyproject.toml index e81d5ee0df3d3..cf6722c7ceb09 100644 --- a/spark/pyproject.toml +++ b/spark/pyproject.toml @@ -24,7 +24,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/sqlserver/changelog.d/18207.added b/sqlserver/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/sqlserver/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/sqlserver/hatch.toml b/sqlserver/hatch.toml index a305f161e8fcf..843ea38127536 100644 --- a/sqlserver/hatch.toml +++ b/sqlserver/hatch.toml @@ -2,7 +2,7 @@ base-package-features = ["deps", "db", "json"] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] os = ["linux"] version = ["2017", "2019", "2022"] setup = ["single", "ha"] @@ -12,7 +12,7 @@ setup = ["single", "ha"] # time out. until we're able to modify and parallelize the work we'll limit the per-driver tests to only a single # sqlserver version [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] os = ["windows"] driver = ["SQLOLEDB", "SQLNCLI11", "MSOLEDBSQL", "odbc"] version = ["2019", "2022"] diff --git a/sqlserver/pyproject.toml b/sqlserver/pyproject.toml index 14015e55328f5..18c28b1749927 100644 --- a/sqlserver/pyproject.toml +++ b/sqlserver/pyproject.toml @@ -23,7 +23,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/squid/changelog.d/18207.added b/squid/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/squid/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/squid/hatch.toml b/squid/hatch.toml index d423b3a2de5a0..d5db0f0834835 100644 --- a/squid/hatch.toml +++ b/squid/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] version = ["4.13", "5.6"] [envs.default.overrides] diff --git a/squid/pyproject.toml b/squid/pyproject.toml index 4184551c8ec92..8f144cfa18e63 100644 --- a/squid/pyproject.toml +++ b/squid/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/ssh_check/changelog.d/18207.added b/ssh_check/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/ssh_check/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/ssh_check/hatch.toml b/ssh_check/hatch.toml index ac510b7a85bc6..4913f773c9f75 100644 --- a/ssh_check/hatch.toml +++ b/ssh_check/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] version = ["8.1", "9.1"] [envs.default.overrides] diff --git a/ssh_check/pyproject.toml b/ssh_check/pyproject.toml index d3d8d32bf2bb5..c0131118f6a21 100644 --- a/ssh_check/pyproject.toml +++ b/ssh_check/pyproject.toml @@ -24,7 +24,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/statsd/changelog.d/18207.added b/statsd/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/statsd/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/statsd/hatch.toml b/statsd/hatch.toml index e7b33c9f27801..ea46b02834dad 100644 --- a/statsd/hatch.toml +++ b/statsd/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] version = ["0.9"] [envs.default.overrides] diff --git a/statsd/pyproject.toml b/statsd/pyproject.toml index f5e3fd2f1033d..c2639c6b9b4ef 100644 --- a/statsd/pyproject.toml +++ b/statsd/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/strimzi/changelog.d/18207.added b/strimzi/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/strimzi/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/strimzi/hatch.toml b/strimzi/hatch.toml index ac8e44b6c7f05..bce6b232560ab 100644 --- a/strimzi/hatch.toml +++ b/strimzi/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] version = ["0.34"] [envs.default.overrides] diff --git a/strimzi/pyproject.toml b/strimzi/pyproject.toml index a9f27b18da786..c9a760e36dfa2 100644 --- a/strimzi/pyproject.toml +++ b/strimzi/pyproject.toml @@ -27,7 +27,7 @@ classifiers = [ "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", "Private :: Do Not Upload", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", ] dependencies = [ diff --git a/supervisord/changelog.d/18207.added b/supervisord/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/supervisord/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/supervisord/hatch.toml b/supervisord/hatch.toml index ace1fdff1a0c7..17fa18cb5164e 100644 --- a/supervisord/hatch.toml +++ b/supervisord/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] version = ["3.3"] [envs.default.overrides] diff --git a/supervisord/pyproject.toml b/supervisord/pyproject.toml index d2c516ce1668a..e7eab7fbad10a 100644 --- a/supervisord/pyproject.toml +++ b/supervisord/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/suricata/pyproject.toml b/suricata/pyproject.toml index e32ad6774087e..efc4f57207b0f 100644 --- a/suricata/pyproject.toml +++ b/suricata/pyproject.toml @@ -24,7 +24,7 @@ classifiers = [ "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", "Private :: Do Not Upload", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", ] dependencies = [ diff --git a/system_core/changelog.d/18207.added b/system_core/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/system_core/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/system_core/hatch.toml b/system_core/hatch.toml index 001e43ce25414..c85c5f07a7df2 100644 --- a/system_core/hatch.toml +++ b/system_core/hatch.toml @@ -1,4 +1,4 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] diff --git a/system_core/pyproject.toml b/system_core/pyproject.toml index 6c1b9d79c7b6f..0892c8bb8766c 100644 --- a/system_core/pyproject.toml +++ b/system_core/pyproject.toml @@ -24,7 +24,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/system_swap/changelog.d/18207.added b/system_swap/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/system_swap/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/system_swap/hatch.toml b/system_swap/hatch.toml index 001e43ce25414..c85c5f07a7df2 100644 --- a/system_swap/hatch.toml +++ b/system_swap/hatch.toml @@ -1,4 +1,4 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] diff --git a/system_swap/pyproject.toml b/system_swap/pyproject.toml index 338cde59201cc..6d867738a2022 100644 --- a/system_swap/pyproject.toml +++ b/system_swap/pyproject.toml @@ -24,7 +24,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/tcp_check/changelog.d/18207.added b/tcp_check/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/tcp_check/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/tcp_check/hatch.toml b/tcp_check/hatch.toml index 001e43ce25414..c85c5f07a7df2 100644 --- a/tcp_check/hatch.toml +++ b/tcp_check/hatch.toml @@ -1,4 +1,4 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] diff --git a/tcp_check/pyproject.toml b/tcp_check/pyproject.toml index b19ac5ae591c2..abbf435a04d73 100644 --- a/tcp_check/pyproject.toml +++ b/tcp_check/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/teamcity/changelog.d/18207.added b/teamcity/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/teamcity/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/teamcity/hatch.toml b/teamcity/hatch.toml index 747bf45e6b75a..aba75ffb7e27b 100644 --- a/teamcity/hatch.toml +++ b/teamcity/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] impl = ["legacy", "openmetrics"] [envs.default.overrides] diff --git a/teamcity/pyproject.toml b/teamcity/pyproject.toml index 6c2a2dd305174..e235daa243e39 100644 --- a/teamcity/pyproject.toml +++ b/teamcity/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/tekton/changelog.d/18207.added b/tekton/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/tekton/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/tekton/hatch.toml b/tekton/hatch.toml index 001e43ce25414..c85c5f07a7df2 100644 --- a/tekton/hatch.toml +++ b/tekton/hatch.toml @@ -1,4 +1,4 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] diff --git a/tekton/pyproject.toml b/tekton/pyproject.toml index cd0031341db8e..cb1bf1423018b 100644 --- a/tekton/pyproject.toml +++ b/tekton/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", "Private :: Do Not Upload", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", ] dependencies = [ diff --git a/teleport/changelog.d/18207.added b/teleport/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/teleport/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/teleport/hatch.toml b/teleport/hatch.toml index 001e43ce25414..c85c5f07a7df2 100644 --- a/teleport/hatch.toml +++ b/teleport/hatch.toml @@ -1,4 +1,4 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] diff --git a/teleport/pyproject.toml b/teleport/pyproject.toml index 38b162f52c627..6572facaf8a00 100644 --- a/teleport/pyproject.toml +++ b/teleport/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", "Private :: Do Not Upload", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", ] dependencies = [ diff --git a/temporal/changelog.d/18207.added b/temporal/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/temporal/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/temporal/hatch.toml b/temporal/hatch.toml index ef9dd47791214..29f12d02b8667 100644 --- a/temporal/hatch.toml +++ b/temporal/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] version = ["1.19"] [envs.default.overrides] diff --git a/temporal/pyproject.toml b/temporal/pyproject.toml index 56607a9dbe9f9..b9f0768aa6151 100644 --- a/temporal/pyproject.toml +++ b/temporal/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", "Private :: Do Not Upload", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", ] dependencies = [ diff --git a/tenable/changelog.d/18207.added b/tenable/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/tenable/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/tenable/pyproject.toml b/tenable/pyproject.toml index 8ab5356276b6d..fe274a81abac0 100644 --- a/tenable/pyproject.toml +++ b/tenable/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/teradata/changelog.d/18207.added b/teradata/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/teradata/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/teradata/hatch.toml b/teradata/hatch.toml index 847ebdf063c08..3c87a67bab2a4 100644 --- a/teradata/hatch.toml +++ b/teradata/hatch.toml @@ -1,14 +1,14 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] setup = ["sandbox"] [envs.default.overrides] -name."^py3.11$".e2e-env = { value = true } +name."^py3.12$".e2e-env = { value = true } matrix.setup.e2e-env = { value = true, if = ["sandbox"], env = ["TERADATA_SERVER"] } matrix.setup.env-vars = [ { key = "USE_TD_SANDBOX", value = "true", if = ["sandbox"] }, diff --git a/teradata/pyproject.toml b/teradata/pyproject.toml index d1199e10cce27..e102a5fd8a90b 100644 --- a/teradata/pyproject.toml +++ b/teradata/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/tibco_ems/hatch.toml b/tibco_ems/hatch.toml index bd72781786c80..769d79d915cc1 100644 --- a/tibco_ems/hatch.toml +++ b/tibco_ems/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] [envs.default] e2e-env = false \ No newline at end of file diff --git a/tibco_ems/pyproject.toml b/tibco_ems/pyproject.toml index c903069b76e4c..282743eef2730 100644 --- a/tibco_ems/pyproject.toml +++ b/tibco_ems/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", "Private :: Do Not Upload", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", ] dependencies = [ diff --git a/tls/changelog.d/18207.added b/tls/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/tls/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/tls/hatch.toml b/tls/hatch.toml index 91dfdd58ecd27..1bd2b7ae9753c 100644 --- a/tls/hatch.toml +++ b/tls/hatch.toml @@ -6,4 +6,4 @@ dependencies = [ ] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] diff --git a/tls/pyproject.toml b/tls/pyproject.toml index 4ae35db079046..42a568fe730bd 100644 --- a/tls/pyproject.toml +++ b/tls/pyproject.toml @@ -24,7 +24,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/tomcat/changelog.d/18207.added b/tomcat/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/tomcat/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/tomcat/hatch.toml b/tomcat/hatch.toml index 4b877dba0897c..94161a21b76ed 100644 --- a/tomcat/hatch.toml +++ b/tomcat/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] version = ["9.0", "10.0", "10.1"] flavor = ["standalone", "embedded"] diff --git a/tomcat/pyproject.toml b/tomcat/pyproject.toml index 0066e14daf1b0..9f0c160029382 100644 --- a/tomcat/pyproject.toml +++ b/tomcat/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/torchserve/changelog.d/18207.added b/torchserve/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/torchserve/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/torchserve/hatch.toml b/torchserve/hatch.toml index 3a72e9f93a15a..eff98ccee69bc 100644 --- a/torchserve/hatch.toml +++ b/torchserve/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] version = ["0.8"] [envs.default.overrides] diff --git a/torchserve/pyproject.toml b/torchserve/pyproject.toml index c759b902488f2..66bd379468471 100644 --- a/torchserve/pyproject.toml +++ b/torchserve/pyproject.toml @@ -26,7 +26,7 @@ classifiers = [ "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", "Private :: Do Not Upload", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", ] dependencies = [ diff --git a/traefik_mesh/changelog.d/18207.added b/traefik_mesh/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/traefik_mesh/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/traefik_mesh/hatch.toml b/traefik_mesh/hatch.toml index 001e43ce25414..c85c5f07a7df2 100644 --- a/traefik_mesh/hatch.toml +++ b/traefik_mesh/hatch.toml @@ -1,4 +1,4 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] diff --git a/traefik_mesh/pyproject.toml b/traefik_mesh/pyproject.toml index 2e6f16a26f5c8..69f2dfd774b3b 100644 --- a/traefik_mesh/pyproject.toml +++ b/traefik_mesh/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", "Private :: Do Not Upload", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", ] dependencies = [ diff --git a/traffic_server/changelog.d/18207.added b/traffic_server/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/traffic_server/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/traffic_server/hatch.toml b/traffic_server/hatch.toml index 6366d9ca2ae04..399e1b63b28b4 100644 --- a/traffic_server/hatch.toml +++ b/traffic_server/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] [envs.default.env-vars] TRAFFIC_SERVER_VERSION = "9.1.1" diff --git a/traffic_server/pyproject.toml b/traffic_server/pyproject.toml index 237a7c2169cd9..67931b928ec4f 100644 --- a/traffic_server/pyproject.toml +++ b/traffic_server/pyproject.toml @@ -24,7 +24,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/twemproxy/changelog.d/18207.added b/twemproxy/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/twemproxy/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/twemproxy/hatch.toml b/twemproxy/hatch.toml index ec2b228de127e..d81e8c9cf96b0 100644 --- a/twemproxy/hatch.toml +++ b/twemproxy/hatch.toml @@ -4,7 +4,7 @@ platforms = ["linux", "macos"] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] version = ["0.5.0"] [envs.default.env-vars] diff --git a/twemproxy/pyproject.toml b/twemproxy/pyproject.toml index 5dd08ee893c7d..08cbf57491673 100644 --- a/twemproxy/pyproject.toml +++ b/twemproxy/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/twistlock/changelog.d/18207.added b/twistlock/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/twistlock/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/twistlock/hatch.toml b/twistlock/hatch.toml index 001e43ce25414..c85c5f07a7df2 100644 --- a/twistlock/hatch.toml +++ b/twistlock/hatch.toml @@ -1,4 +1,4 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] diff --git a/twistlock/pyproject.toml b/twistlock/pyproject.toml index fd3d22e710ac1..0329130a09502 100644 --- a/twistlock/pyproject.toml +++ b/twistlock/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/varnish/changelog.d/18207.added b/varnish/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/varnish/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/varnish/hatch.toml b/varnish/hatch.toml index 315e875283099..e4feae8302fbe 100644 --- a/varnish/hatch.toml +++ b/varnish/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] version = ["5.2", "6.5"] [envs.default.overrides] diff --git a/varnish/pyproject.toml b/varnish/pyproject.toml index cecf1e2183a9f..4411d3db54f02 100644 --- a/varnish/pyproject.toml +++ b/varnish/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/vault/changelog.d/18207.added b/vault/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/vault/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/vault/hatch.toml b/vault/hatch.toml index 36bb2bc062149..13be1af587ad7 100644 --- a/vault/hatch.toml +++ b/vault/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] version = ["1.9.0"] auth = ["token-auth", "noauth"] diff --git a/vault/pyproject.toml b/vault/pyproject.toml index 4decac34151a4..730bb5f9aca6c 100644 --- a/vault/pyproject.toml +++ b/vault/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/vertica/changelog.d/18207.added b/vertica/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/vertica/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/vertica/hatch.toml b/vertica/hatch.toml index 129c75f0003c9..bdb9496b55c49 100644 --- a/vertica/hatch.toml +++ b/vertica/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] version = ["10.1", "11.1", "12.0", "23.3"] [envs.default.overrides] diff --git a/vertica/pyproject.toml b/vertica/pyproject.toml index c54b67d12edc5..bd239d161cf12 100644 --- a/vertica/pyproject.toml +++ b/vertica/pyproject.toml @@ -24,7 +24,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/vllm/changelog.d/18207.added b/vllm/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/vllm/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/vllm/hatch.toml b/vllm/hatch.toml index 001e43ce25414..c85c5f07a7df2 100644 --- a/vllm/hatch.toml +++ b/vllm/hatch.toml @@ -1,4 +1,4 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] diff --git a/vllm/pyproject.toml b/vllm/pyproject.toml index ada53013ef377..5df58ba7e3c8e 100644 --- a/vllm/pyproject.toml +++ b/vllm/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", "Private :: Do Not Upload", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", ] dependencies = [ diff --git a/voltdb/changelog.d/18207.added b/voltdb/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/voltdb/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/voltdb/hatch.toml b/voltdb/hatch.toml index 2353ada783248..fde300a90930a 100644 --- a/voltdb/hatch.toml +++ b/voltdb/hatch.toml @@ -9,7 +9,7 @@ mypy-deps = [ DDEV_SKIP_GENERIC_TAGS_CHECK = "true" [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] version = ["8.4", "10.0"] tls = ["false", "true"] diff --git a/voltdb/pyproject.toml b/voltdb/pyproject.toml index e14e6759a5607..1e81ac3c6f957 100644 --- a/voltdb/pyproject.toml +++ b/voltdb/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/vsphere/changelog.d/18207.added b/vsphere/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/vsphere/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/vsphere/hatch.toml b/vsphere/hatch.toml index a0f6624b570b8..4686ee4444838 100644 --- a/vsphere/hatch.toml +++ b/vsphere/hatch.toml @@ -3,7 +3,7 @@ check-types = false [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] version = ["6.7", "7.0"] [envs.default] diff --git a/vsphere/pyproject.toml b/vsphere/pyproject.toml index d60958d76573a..399ca557e460b 100644 --- a/vsphere/pyproject.toml +++ b/vsphere/pyproject.toml @@ -24,7 +24,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/weaviate/changelog.d/18207.added b/weaviate/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/weaviate/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/weaviate/hatch.toml b/weaviate/hatch.toml index 332a4ab893ca6..c27560d0695d7 100644 --- a/weaviate/hatch.toml +++ b/weaviate/hatch.toml @@ -4,7 +4,7 @@ DDEV_SKIP_GENERIC_TAGS_CHECK = "true" [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] version = ["1.20"] auth = ["no-auth", "auth"] diff --git a/weaviate/pyproject.toml b/weaviate/pyproject.toml index abcac32f8d2d9..3e5cf3924bb78 100644 --- a/weaviate/pyproject.toml +++ b/weaviate/pyproject.toml @@ -26,7 +26,7 @@ classifiers = [ "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", "Private :: Do Not Upload", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", ] dependencies = [ diff --git a/weblogic/changelog.d/18207.added b/weblogic/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/weblogic/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/weblogic/hatch.toml b/weblogic/hatch.toml index 9bfdb9f9df723..a117109c05929 100644 --- a/weblogic/hatch.toml +++ b/weblogic/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] version = ["12", "14"] [envs.default.overrides] diff --git a/weblogic/pyproject.toml b/weblogic/pyproject.toml index 0ecbb9d3d9d0e..d59c3481ab998 100644 --- a/weblogic/pyproject.toml +++ b/weblogic/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/win32_event_log/changelog.d/18207.added b/win32_event_log/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/win32_event_log/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/win32_event_log/hatch.toml b/win32_event_log/hatch.toml index 443ff4847ff5b..3400c2946f954 100644 --- a/win32_event_log/hatch.toml +++ b/win32_event_log/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] [envs.default] e2e-env = false diff --git a/win32_event_log/pyproject.toml b/win32_event_log/pyproject.toml index 6ddc69694ebfc..5405823ebfb06 100644 --- a/win32_event_log/pyproject.toml +++ b/win32_event_log/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/windows_performance_counters/changelog.d/18207.added b/windows_performance_counters/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/windows_performance_counters/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/windows_performance_counters/hatch.toml b/windows_performance_counters/hatch.toml index 6026000e5fb50..ec12a94baac91 100644 --- a/windows_performance_counters/hatch.toml +++ b/windows_performance_counters/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] [envs.default] platforms = [ diff --git a/windows_performance_counters/pyproject.toml b/windows_performance_counters/pyproject.toml index 98b5e81533591..851a85f31d463 100644 --- a/windows_performance_counters/pyproject.toml +++ b/windows_performance_counters/pyproject.toml @@ -26,7 +26,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/windows_service/changelog.d/18207.added b/windows_service/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/windows_service/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/windows_service/hatch.toml b/windows_service/hatch.toml index 71ba922f308fd..15860b143c506 100644 --- a/windows_service/hatch.toml +++ b/windows_service/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] [envs.default] platforms = [ diff --git a/windows_service/pyproject.toml b/windows_service/pyproject.toml index c7bcbe2d7e58e..b3d5f1be9560c 100644 --- a/windows_service/pyproject.toml +++ b/windows_service/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/wmi_check/changelog.d/18207.added b/wmi_check/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/wmi_check/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/wmi_check/hatch.toml b/wmi_check/hatch.toml index 44f3195b61fa0..353f4419cb7cf 100644 --- a/wmi_check/hatch.toml +++ b/wmi_check/hatch.toml @@ -3,7 +3,7 @@ check-types = false [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] [envs.default] e2e-env = false diff --git a/wmi_check/pyproject.toml b/wmi_check/pyproject.toml index b53c3a3091a41..9637eeb6797dd 100644 --- a/wmi_check/pyproject.toml +++ b/wmi_check/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/yarn/changelog.d/18207.added b/yarn/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/yarn/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/yarn/hatch.toml b/yarn/hatch.toml index b55a0ed85515f..2a7ba6534e9df 100644 --- a/yarn/hatch.toml +++ b/yarn/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] [envs.default.env-vars] DDEV_SKIP_GENERIC_TAGS_CHECK = "true" diff --git a/yarn/pyproject.toml b/yarn/pyproject.toml index e4a4d85c2365c..964786604a91e 100644 --- a/yarn/pyproject.toml +++ b/yarn/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] diff --git a/zk/changelog.d/18207.added b/zk/changelog.d/18207.added new file mode 100644 index 0000000000000..624cd9836c9fe --- /dev/null +++ b/zk/changelog.d/18207.added @@ -0,0 +1 @@ +Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/zk/hatch.toml b/zk/hatch.toml index bf778cb3b4d12..1f5458b389203 100644 --- a/zk/hatch.toml +++ b/zk/hatch.toml @@ -1,7 +1,7 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] version = ["3.5.8", "3.6.2"] ssl = ["true", "false"] diff --git a/zk/pyproject.toml b/zk/pyproject.toml index 1474dab090bf5..8cc1312dac750 100644 --- a/zk/pyproject.toml +++ b/zk/pyproject.toml @@ -25,7 +25,7 @@ classifiers = [ "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: BSD License", - "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring", "Private :: Do Not Upload", ] From 968a0fed63de32332a927ba63fc3b49157c4c041 Mon Sep 17 00:00:00 2001 From: savandalasaniya-crest Date: Tue, 17 Sep 2024 22:52:27 +0530 Subject: [PATCH 03/34] Add Freshservice integration with no assets (#18563) Co-authored-by: ravindrasojitra-crest --- .github/CODEOWNERS | 5 +++ .github/workflows/config/labeler.yml | 2 ++ freshservice/CHANGELOG.md | 7 ++++ freshservice/README.md | 43 +++++++++++++++++++++++++ freshservice/assets/service_checks.json | 1 + freshservice/manifest.json | 41 +++++++++++++++++++++++ 6 files changed, 99 insertions(+) create mode 100644 freshservice/CHANGELOG.md create mode 100644 freshservice/README.md create mode 100644 freshservice/assets/service_checks.json create mode 100644 freshservice/manifest.json diff --git a/.github/CODEOWNERS b/.github/CODEOWNERS index b4914e8ab5194..aa04297fa2021 100644 --- a/.github/CODEOWNERS +++ b/.github/CODEOWNERS @@ -260,6 +260,11 @@ datadog_checks_base/datadog_checks/base/checks/windows/ @DataDog/wi /contentful/manifest.json @DataDog/saas-integrations @DataDog/documentation /contentful/assets/logs/ @DataDog/saas-integrations @DataDog/documentation @DataDog/logs-backend +/freshservice/ @DataDog/saas-integrations +/freshservice/*.md @DataDog/saas-integrations @DataDog/documentation +/freshservice/manifest.json @DataDog/saas-integrations @DataDog/documentation +/freshservice/assets/logs/ @DataDog/saas-integrations @DataDog/documentation @DataDog/logs-backend + /godaddy/ @DataDog/saas-integrations /godaddy/*.md @DataDog/saas-integrations @DataDog/documentation /godaddy/manifest.json @DataDog/saas-integrations @DataDog/documentation diff --git a/.github/workflows/config/labeler.yml b/.github/workflows/config/labeler.yml index c9a6794590730..bebee0a7ae6f8 100644 --- a/.github/workflows/config/labeler.yml +++ b/.github/workflows/config/labeler.yml @@ -193,6 +193,8 @@ integration/fly_io: - fly_io/**/* integration/foundationdb: - foundationdb/**/* +integration/freshservice: +- freshservice/**/* integration/gearmand: - gearmand/**/* integration/gitlab: diff --git a/freshservice/CHANGELOG.md b/freshservice/CHANGELOG.md new file mode 100644 index 0000000000000..7a82ad4c0ef67 --- /dev/null +++ b/freshservice/CHANGELOG.md @@ -0,0 +1,7 @@ +# CHANGELOG - freshservice + +## 1.0.0 / 2024-09-11 + +***Added***: + +* Initial Release diff --git a/freshservice/README.md b/freshservice/README.md new file mode 100644 index 0000000000000..f01d2756c87a3 --- /dev/null +++ b/freshservice/README.md @@ -0,0 +1,43 @@ +# Agent Check: freshservice + +## Overview + +This check monitors [freshservice][1]. + +## Setup + +### Installation + +The Freshservice check is included in the [Datadog Agent][2] package. +No additional installation is needed on your server. + +### Configuration + +!!! Add list of steps to set up this integration !!! + +### Validation + +!!! Add steps to validate integration is functioning as expected !!! + +## Data Collected + +### Metrics + +Freshservice does not include any metrics. + +### Service Checks + +Freshservice does not include any service checks. + +### Events + +Freshservice does not include any events. + +## Troubleshooting + +Need help? Contact [Datadog support][3]. + +[1]: **LINK_TO_INTEGRATION_SITE** +[2]: https://app.datadoghq.com/account/settings/agent/latest +[3]: https://docs.datadoghq.com/help/ + diff --git a/freshservice/assets/service_checks.json b/freshservice/assets/service_checks.json new file mode 100644 index 0000000000000..fe51488c7066f --- /dev/null +++ b/freshservice/assets/service_checks.json @@ -0,0 +1 @@ +[] diff --git a/freshservice/manifest.json b/freshservice/manifest.json new file mode 100644 index 0000000000000..3682b14a2e436 --- /dev/null +++ b/freshservice/manifest.json @@ -0,0 +1,41 @@ +{ + "manifest_version": "2.0.0", + "app_uuid": "44b2812e-fea7-4d6a-b0c1-b5e836627090", + "app_id": "freshservice", + "display_on_public_website": false, + "tile": { + "overview": "README.md#Overview", + "configuration": "README.md#Setup", + "support": "README.md#Support", + "changelog": "CHANGELOG.md", + "description": "Gain insights into Freshservice logs", + "title": "Freshservice", + "media": [], + "classifier_tags": [ + "Category::Log Collection", + "Category::Security", + "Offering::Integration", + "Submitted Data Type::Logs" + ] + }, + "assets": { + "integration": { + "auto_install": false, + "source_type_id": 25506291, + "source_type_name": "Freshservice", + "events": { + "creates_events": false + }, + "service_checks": { + "metadata_path": "assets/service_checks.json" + } + } + }, + "author": { + "support_email": "help@datadoghq.com", + "name": "Datadog", + "homepage": "https://www.datadoghq.com", + "sales_email": "info@datadoghq.com" + }, + "oauth": {} +} \ No newline at end of file From 59b93a8558b45e701dccd237c4328cf98b98c577 Mon Sep 17 00:00:00 2001 From: Steven Yuen Date: Tue, 17 Sep 2024 15:42:53 -0400 Subject: [PATCH 04/34] [Release] Bumped sqlserver version to 17.5.3 (#18609) * [Release] Bumped sqlserver version to 17.5.3 * delete changelog file --- requirements-agent-release.txt | 2 +- sqlserver/CHANGELOG.md | 6 ++++++ sqlserver/changelog.d/18586.fixed | 1 - sqlserver/datadog_checks/sqlserver/__about__.py | 2 +- 4 files changed, 8 insertions(+), 3 deletions(-) delete mode 100644 sqlserver/changelog.d/18586.fixed diff --git a/requirements-agent-release.txt b/requirements-agent-release.txt index b102e0633871b..ca87469758a64 100644 --- a/requirements-agent-release.txt +++ b/requirements-agent-release.txt @@ -159,7 +159,7 @@ datadog-snowflake==5.9.0 datadog-solr==1.13.0 datadog-sonarqube==3.2.2 datadog-spark==4.3.1 -datadog-sqlserver==17.5.2 +datadog-sqlserver==17.5.3 datadog-squid==2.5.1 datadog-ssh-check==2.10.0 datadog-statsd==1.12.0 diff --git a/sqlserver/CHANGELOG.md b/sqlserver/CHANGELOG.md index e8696079bb49f..f47c2f59107a9 100644 --- a/sqlserver/CHANGELOG.md +++ b/sqlserver/CHANGELOG.md @@ -2,6 +2,12 @@ +## 17.5.3 / 2024-09-17 + +***Fixed***: + +* Fix ODBC config handling for Linux ([#18586](https://github.com/DataDog/integrations-core/pull/18586)) + ## 17.5.2 / 2024-08-28 / Agent 7.57.0 ***Fixed***: diff --git a/sqlserver/changelog.d/18586.fixed b/sqlserver/changelog.d/18586.fixed deleted file mode 100644 index e6177056b7c48..0000000000000 --- a/sqlserver/changelog.d/18586.fixed +++ /dev/null @@ -1 +0,0 @@ -[sqlserver] Fix ODBC config handling for Linux diff --git a/sqlserver/datadog_checks/sqlserver/__about__.py b/sqlserver/datadog_checks/sqlserver/__about__.py index 6a7af39f3ed9a..76ae2e29005a6 100644 --- a/sqlserver/datadog_checks/sqlserver/__about__.py +++ b/sqlserver/datadog_checks/sqlserver/__about__.py @@ -2,4 +2,4 @@ # All rights reserved # Licensed under a 3-clause BSD style license (see LICENSE) -__version__ = '17.5.2' +__version__ = '17.5.3' From 2a419f75fa05ababbe2a9a2fa29472f7321d9f90 Mon Sep 17 00:00:00 2001 From: "agent-platform-auto-pr[bot]" <153269286+agent-platform-auto-pr[bot]@users.noreply.github.com> Date: Tue, 17 Sep 2024 15:57:24 -0400 Subject: [PATCH 05/34] Update dependency resolution (#18605) Co-authored-by: Kyle-Neale --- .deps/image_digests.json | 6 +- .deps/resolved/linux-aarch64_3.12.txt | 128 ++++++++++++++++++++++++ .deps/resolved/linux-x86_64_3.12.txt | 129 +++++++++++++++++++++++++ .deps/resolved/macos-x86_64_3.12.txt | 128 ++++++++++++++++++++++++ .deps/resolved/windows-x86_64_3.12.txt | 126 ++++++++++++++++++++++++ 5 files changed, 514 insertions(+), 3 deletions(-) create mode 100644 .deps/resolved/linux-aarch64_3.12.txt create mode 100644 .deps/resolved/linux-x86_64_3.12.txt create mode 100644 .deps/resolved/macos-x86_64_3.12.txt create mode 100644 .deps/resolved/windows-x86_64_3.12.txt diff --git a/.deps/image_digests.json b/.deps/image_digests.json index 18afec2ebf7af..ad73cfcda588d 100644 --- a/.deps/image_digests.json +++ b/.deps/image_digests.json @@ -1,5 +1,5 @@ { - "linux-aarch64": "sha256:63f46c0aaad6c81830b99754b158625e6ff5dbf50cd74e62c098fa46ab6808ec", - "linux-x86_64": "sha256:e854e6d6f8258c2bb80c5f525a9c9f545ce95bd555583e9ae04bbeede9d5666a", - "windows-x86_64": "sha256:ce5ac8bd45f7816a8ea3f55f54d70c8856667ad2cf8d638b4c62728368a65652" + "linux-aarch64": "sha256:f3834f630ad669a0876dafb95a9ffd5b485171536cffbc1170ffbd3cbfab9c1d", + "linux-x86_64": "sha256:952783e733c9c0d08568394a0e925d1f8428d8f04507a2b10bcfd2a1486da2db", + "windows-x86_64": "sha256:2fd9816b42a300080a0bcc57753e577f0140bb8b0d06e14c54fc8f634d93a2db" } diff --git a/.deps/resolved/linux-aarch64_3.12.txt b/.deps/resolved/linux-aarch64_3.12.txt new file mode 100644 index 0000000000000..6503dd7bf1656 --- /dev/null +++ b/.deps/resolved/linux-aarch64_3.12.txt @@ -0,0 +1,128 @@ +aerospike @ https://agent-int-packages.datadoghq.com/built/aerospike/aerospike-7.1.1-20240917155304-cp312-cp312-manylinux2014_aarch64.manylinux_2_17_aarch64.whl#sha256=8573244c130f49445cf9af1c1423f13e0c2a2aef7395045e21a0e3cf760c2213 +annotated-types @ https://agent-int-packages.datadoghq.com/external/annotated-types/annotated_types-0.7.0-py3-none-any.whl#sha256=1f02e8b43a8fbbc3f3e0d4f0f4bfc8131bcb4eebe8849b8e5c773f3a1c582a53 +asn1crypto @ https://agent-int-packages.datadoghq.com/external/asn1crypto/asn1crypto-1.5.1-py2.py3-none-any.whl#sha256=db4e40728b728508912cbb3d44f19ce188f218e9eba635821bb4b68564f8fd67 +attrs @ https://agent-int-packages.datadoghq.com/external/attrs/attrs-24.2.0-py3-none-any.whl#sha256=81921eb96de3191c8258c199618104dd27ac608d9366f5e35d011eae1867ede2 +aws-requests-auth @ https://agent-int-packages.datadoghq.com/external/aws-requests-auth/aws_requests_auth-0.4.3-py2.py3-none-any.whl#sha256=646bc37d62140ea1c709d20148f5d43197e6bd2d63909eb36fa4bb2345759977 +azure-core @ https://agent-int-packages.datadoghq.com/external/azure-core/azure_core-1.31.0-py3-none-any.whl#sha256=22954de3777e0250029360ef31d80448ef1be13b80a459bff80ba7073379e2cd +azure-identity @ https://agent-int-packages.datadoghq.com/external/azure-identity/azure_identity-1.17.1-py3-none-any.whl#sha256=db8d59c183b680e763722bfe8ebc45930e6c57df510620985939f7f3191e0382 +bcrypt @ https://agent-int-packages.datadoghq.com/external/bcrypt/bcrypt-4.2.0-cp39-abi3-manylinux_2_17_aarch64.manylinux2014_aarch64.whl#sha256=3bbbfb2734f0e4f37c5136130405332640a1e46e6b23e000eeff2ba8d005da68 +beautifulsoup4 @ https://agent-int-packages.datadoghq.com/external/beautifulsoup4/beautifulsoup4-4.12.3-py3-none-any.whl#sha256=b80878c9f40111313e55da8ba20bdba06d8fa3969fc68304167741bbf9e082ed +binary @ https://agent-int-packages.datadoghq.com/external/binary/binary-1.0.0-py2.py3-none-any.whl#sha256=e1b61f3a5c002717d1a28e4d9d2dc8acbc9d6b12baf7b1e4ab25d743da97e323 +boto3 @ https://agent-int-packages.datadoghq.com/external/boto3/boto3-1.35.10-py3-none-any.whl#sha256=add26dd58e076dfd387013da4704716d5cff215cf14f6d4347c4b9b7fc1f0b8e +botocore @ https://agent-int-packages.datadoghq.com/external/botocore/botocore-1.35.10-py3-none-any.whl#sha256=0d96d023b9b0cea99a0a428a431d011329d3a958730aee6ed6a6fec5d9bfbc03 +bytecode @ https://agent-int-packages.datadoghq.com/external/bytecode/bytecode-0.15.1-py3-none-any.whl#sha256=0a1dc340cac823cff605609b8b214f7f9bf80418c6b9e0fc8c6db1793c27137d +cachetools @ https://agent-int-packages.datadoghq.com/external/cachetools/cachetools-5.5.0-py3-none-any.whl#sha256=02134e8439cdc2ffb62023ce1debca2944c3f289d66bb17ead3ab3dede74b292 +cattrs @ https://agent-int-packages.datadoghq.com/external/cattrs/cattrs-24.1.1-py3-none-any.whl#sha256=ec8ce8fdc725de9d07547cd616f968670687c6fa7a2e263b088370c46d834d97 +certifi @ https://agent-int-packages.datadoghq.com/external/certifi/certifi-2024.8.30-py3-none-any.whl#sha256=922820b53db7a7257ffbda3f597266d435245903d80737e34f8a45ff3e3230d8 +cffi @ https://agent-int-packages.datadoghq.com/external/cffi/cffi-1.17.1-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl#sha256=da95af8214998d77a98cc14e3a3bd00aa191526343078b530ceb0bd710fb48a5 +charset-normalizer @ https://agent-int-packages.datadoghq.com/external/charset-normalizer/charset_normalizer-3.3.2-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl#sha256=8f4a014bc36d3c57402e2977dada34f9c12300af536839dc38c0beab8878f38a +clickhouse-cityhash @ https://agent-int-packages.datadoghq.com/external/clickhouse-cityhash/clickhouse_cityhash-1.0.2.4-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl#sha256=bbfd83713e5a7a700c4a8200e921bc580fd7cba5f3b9d732172a5d82b12b3e20 +clickhouse-driver @ https://agent-int-packages.datadoghq.com/external/clickhouse-driver/clickhouse_driver-0.2.9-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl#sha256=4a8d8e2888a857d8db3d98765a5ad23ab561241feaef68bbffc5a0bd9c142342 +cm-client @ https://agent-int-packages.datadoghq.com/built/cm-client/cm_client-45.0.4-20240402155018-py3-none-manylinux2014_aarch64.whl#sha256=aba3c1683ef1b2099933e030464d29b3ad1c206784ebd15d8a7147ecd6ba24e1 +confluent-kafka @ https://agent-int-packages.datadoghq.com/built/confluent-kafka/confluent_kafka-2.5.0-20240917155305-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl#sha256=6dd31639ba4c15fcb59f65842e446ecda997e8c826dec9e9b1bec01748b5066d +cryptography @ https://agent-int-packages.datadoghq.com/external/cryptography/cryptography-43.0.0-cp39-abi3-manylinux_2_17_aarch64.manylinux2014_aarch64.whl#sha256=ac1955ce000cb29ab40def14fd1bbfa7af2017cca696ee696925615cafd0dce5 +ddsketch @ https://agent-int-packages.datadoghq.com/external/ddsketch/ddsketch-3.0.1-py3-none-any.whl#sha256=6d047b455fe2837c43d366ff1ae6ba0c3166e15499de8688437a75cea914224e +ddtrace @ https://agent-int-packages.datadoghq.com/external/ddtrace/ddtrace-2.10.6-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl#sha256=081bb12a54c46c9804e0645320d827deaff626b9035ba13ac97567149e07cdb5 +decorator @ https://agent-int-packages.datadoghq.com/external/decorator/decorator-5.1.1-py3-none-any.whl#sha256=b8c3f85900b9dc423225913c5aace94729fe1fa9763b38939a95226f02d37186 +deprecated @ https://agent-int-packages.datadoghq.com/external/deprecated/Deprecated-1.2.14-py2.py3-none-any.whl#sha256=6fac8b097794a90302bdbb17b9b815e732d3c4720583ff1b198499d78470466c +dnspython @ https://agent-int-packages.datadoghq.com/external/dnspython/dnspython-2.6.1-py3-none-any.whl#sha256=5ef3b9680161f6fa89daf8ad451b5f1a33b18ae8a1c6778cdf4b43f08c0a6e50 +dogpile-cache @ https://agent-int-packages.datadoghq.com/external/dogpile-cache/dogpile.cache-1.3.3-py3-none-any.whl#sha256=5e211c4902ebdf88c678d268e22454b41e68071632daa9402d8ee24e825ed8ca +envier @ https://agent-int-packages.datadoghq.com/external/envier/envier-0.5.2-py3-none-any.whl#sha256=65099cf3aa9b3b3b4b92db2f7d29e2910672e085b76f7e587d2167561a834add +filelock @ https://agent-int-packages.datadoghq.com/external/filelock/filelock-3.16.0-py3-none-any.whl#sha256=f6ed4c963184f4c84dd5557ce8fece759a3724b37b80c6c4f20a2f63a4dc6609 +foundationdb @ https://agent-int-packages.datadoghq.com/built/foundationdb/foundationdb-6.3.24-20240402155019-py3-none-manylinux2014_aarch64.whl#sha256=14259f824080062cc890965747597ff00a9d6c76a1eb926673fed68a45860ccd +google-auth @ https://agent-int-packages.datadoghq.com/external/google-auth/google_auth-2.34.0-py2.py3-none-any.whl#sha256=72fd4733b80b6d777dcde515628a9eb4a577339437012874ea286bca7261ee65 +gssapi @ https://agent-int-packages.datadoghq.com/built/gssapi/gssapi-1.8.3-20240917155305-cp312-cp312-manylinux2014_aarch64.manylinux_2_17_aarch64.whl#sha256=4043e268278e0e4430d92cfb51db3103bd238b5c658197b7a9c2acc35c789d67 +hazelcast-python-client @ https://agent-int-packages.datadoghq.com/external/hazelcast-python-client/hazelcast_python_client-5.4.0-py3-none-any.whl#sha256=16195cd58feb2dd3be1594d08d42527ae00797548a6a9d6a601aae2e8514ff5f +idna @ https://agent-int-packages.datadoghq.com/external/idna/idna-3.10-py3-none-any.whl#sha256=946d195a0d259cbba61165e88e65941f16e9b36ea6ddb97f00452bae8b1287d3 +importlib-metadata @ https://agent-int-packages.datadoghq.com/external/importlib-metadata/importlib_metadata-8.4.0-py3-none-any.whl#sha256=66f342cc6ac9818fc6ff340576acd24d65ba0b3efabb2b4ac08b598965a4a2f1 +in-toto @ https://agent-int-packages.datadoghq.com/external/in-toto/in_toto-2.0.0-py3-none-any.whl#sha256=eaf71733cdd5b2309b58b24169c13c24701da72baa43b7432d1dffc308e7bf32 +iso8601 @ https://agent-int-packages.datadoghq.com/external/iso8601/iso8601-2.1.0-py3-none-any.whl#sha256=aac4145c4dcb66ad8b648a02830f5e2ff6c24af20f4f482689be402db2429242 +jellyfish @ https://agent-int-packages.datadoghq.com/external/jellyfish/jellyfish-1.1.0-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl#sha256=87dc2a82c45b773a579fb695a5956a54106c1187f27c9ccee8508726d2e59cfc +jmespath @ https://agent-int-packages.datadoghq.com/external/jmespath/jmespath-1.0.1-py3-none-any.whl#sha256=02e2e4cc71b5bcab88332eebf907519190dd9e6e82107fa7f83b1003a6252980 +jsonpatch @ https://agent-int-packages.datadoghq.com/external/jsonpatch/jsonpatch-1.33-py2.py3-none-any.whl#sha256=0ae28c0cd062bbd8b8ecc26d7d164fbbea9652a1a3693f3b956c1eae5145dade +jsonpointer @ https://agent-int-packages.datadoghq.com/external/jsonpointer/jsonpointer-3.0.0-py2.py3-none-any.whl#sha256=13e088adc14fca8b6aa8177c044e12701e6ad4b28ff10e65f2267a90109c9942 +keystoneauth1 @ https://agent-int-packages.datadoghq.com/external/keystoneauth1/keystoneauth1-5.8.0-py3-none-any.whl#sha256=e69dff80c509ab64d4de4494658d914e81f26af720828dc584ceee74ecd666d9 +krb5 @ https://agent-int-packages.datadoghq.com/built/krb5/krb5-0.6.0-20240917155306-cp312-cp312-manylinux2014_aarch64.manylinux_2_17_aarch64.whl#sha256=47e12d5a37cd54c721688b99389cf6bdd555597ad08ab6787273ddc63a10a1e7 +kubernetes @ https://agent-int-packages.datadoghq.com/external/kubernetes/kubernetes-30.1.0-py2.py3-none-any.whl#sha256=e212e8b7579031dd2e512168b617373bc1e03888d41ac4e04039240a292d478d +ldap3 @ https://agent-int-packages.datadoghq.com/external/ldap3/ldap3-2.9.1-py2.py3-none-any.whl#sha256=5869596fc4948797020d3f03b7939da938778a0f9e2009f7a072ccf92b8e8d70 +lxml @ https://agent-int-packages.datadoghq.com/built/lxml/lxml-4.9.4-20240917155306-cp312-cp312-manylinux2014_aarch64.manylinux_2_17_aarch64.whl#sha256=06e8662fc23fce90e509842155152aacaca06be101edc42de77614f2e2c726db +lz4 @ https://agent-int-packages.datadoghq.com/external/lz4/lz4-4.3.3-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl#sha256=33c9a6fd20767ccaf70649982f8f3eeb0884035c150c0b818ea660152cf3c809 +mmh3 @ https://agent-int-packages.datadoghq.com/external/mmh3/mmh3-4.1.0-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl#sha256=ba245e94b8d54765e14c2d7b6214e832557e7856d5183bc522e17884cab2f45d +msal @ https://agent-int-packages.datadoghq.com/external/msal/msal-1.31.0-py3-none-any.whl#sha256=96bc37cff82ebe4b160d5fc0f1196f6ca8b50e274ecd0ec5bf69c438514086e7 +msal-extensions @ https://agent-int-packages.datadoghq.com/external/msal-extensions/msal_extensions-1.2.0-py3-none-any.whl#sha256=cf5ba83a2113fa6dc011a254a72f1c223c88d7dfad74cc30617c4679a417704d +netifaces @ https://agent-int-packages.datadoghq.com/built/netifaces/netifaces-0.11.0-20240917155307-cp312-cp312-manylinux2014_aarch64.manylinux_2_17_aarch64.whl#sha256=b18f919b60520879078083d19caaf9bb25c52381d9622ceae708245cf9826b25 +oauthlib @ https://agent-int-packages.datadoghq.com/external/oauthlib/oauthlib-3.2.2-py3-none-any.whl#sha256=8139f29aac13e25d502680e9e19963e83f16838d48a0d71c287fe40e7067fbca +openstacksdk @ https://agent-int-packages.datadoghq.com/external/openstacksdk/openstacksdk-3.3.0-py3-none-any.whl#sha256=e6d4121b87354984caf0e3c032e2ebf4d4440374f86c81c27ec52ca5df359157 +opentelemetry-api @ https://agent-int-packages.datadoghq.com/external/opentelemetry-api/opentelemetry_api-1.27.0-py3-none-any.whl#sha256=953d5871815e7c30c81b56d910c707588000fff7a3ca1c73e6531911d53065e7 +orjson @ https://agent-int-packages.datadoghq.com/external/orjson/orjson-3.10.7-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl#sha256=76ac14cd57df0572453543f8f2575e2d01ae9e790c21f57627803f5e79b0d3c3 +os-service-types @ https://agent-int-packages.datadoghq.com/external/os-service-types/os_service_types-1.7.0-py2.py3-none-any.whl#sha256=0505c72205690910077fb72b88f2a1f07533c8d39f2fe75b29583481764965d6 +packaging @ https://agent-int-packages.datadoghq.com/external/packaging/packaging-24.1-py3-none-any.whl#sha256=5b8f2217dbdbd2f7f384c41c628544e6d52f2d0f53c6d0c3ea61aa5d1d7ff124 +paramiko @ https://agent-int-packages.datadoghq.com/external/paramiko/paramiko-3.4.1-py3-none-any.whl#sha256=8e49fd2f82f84acf7ffd57c64311aa2b30e575370dc23bdb375b10262f7eac32 +pathspec @ https://agent-int-packages.datadoghq.com/external/pathspec/pathspec-0.12.1-py3-none-any.whl#sha256=a0d503e138a4c123b27490a4f7beda6a01c6f288df0e4a8b79c7eb0dc7b4cc08 +pbr @ https://agent-int-packages.datadoghq.com/external/pbr/pbr-6.1.0-py2.py3-none-any.whl#sha256=a776ae228892d8013649c0aeccbb3d5f99ee15e005a4cbb7e61d55a067b28a2a +platformdirs @ https://agent-int-packages.datadoghq.com/external/platformdirs/platformdirs-4.3.3-py3-none-any.whl#sha256=50a5450e2e84f44539718293cbb1da0a0885c9d14adf21b77bae4e66fc99d9b5 +ply @ https://agent-int-packages.datadoghq.com/external/ply/ply-3.11-py2.py3-none-any.whl#sha256=096f9b8350b65ebd2fd1346b12452efe5b9607f7482813ffca50c22722a807ce +portalocker @ https://agent-int-packages.datadoghq.com/external/portalocker/portalocker-2.10.1-py3-none-any.whl#sha256=53a5984ebc86a025552264b459b46a2086e269b21823cb572f8f28ee759e45bf +prometheus-client @ https://agent-int-packages.datadoghq.com/external/prometheus-client/prometheus_client-0.20.0-py3-none-any.whl#sha256=cde524a85bce83ca359cc837f28b8c0db5cac7aa653a588fd7e84ba061c329e7 +protobuf @ https://agent-int-packages.datadoghq.com/external/protobuf/protobuf-5.27.3-cp38-abi3-manylinux2014_aarch64.whl#sha256=b8a994fb3d1c11156e7d1e427186662b64694a62b55936b2b9348f0a7c6625ce +psutil @ https://agent-int-packages.datadoghq.com/built/psutil/psutil-5.9.6-20240917155307-cp311-abi3-manylinux2014_aarch64.manylinux_2_17_aarch64.whl#sha256=2ba6dce6cf517094b99ea70c0e82d657bdf4156a3eb5d775cf67bc7f44e02005 +psycopg2-binary @ https://agent-int-packages.datadoghq.com/external/psycopg2-binary/psycopg2_binary-2.9.9-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl#sha256=8f8544b092a29a6ddd72f3556a9fcf249ec412e10ad28be6a0c0d948924f2212 +pyasn1 @ https://agent-int-packages.datadoghq.com/external/pyasn1/pyasn1-0.4.8-py2.py3-none-any.whl#sha256=39c7e2ec30515947ff4e87fb6f456dfc6e84857d34be479c9d4a4ba4bf46aa5d +pyasn1-modules @ https://agent-int-packages.datadoghq.com/external/pyasn1-modules/pyasn1_modules-0.4.1-py3-none-any.whl#sha256=49bfa96b45a292b711e986f222502c1c9a5e1f4e568fc30e2574a6c7d07838fd +pyasyncore @ https://agent-int-packages.datadoghq.com/external/pyasyncore/pyasyncore-1.0.4-py3-none-any.whl#sha256=9e5f6dc9dc057c56370b7a5cdb4c4670fd4b0556de2913ed1f428cd6a5366895 +pycparser @ https://agent-int-packages.datadoghq.com/external/pycparser/pycparser-2.22-py3-none-any.whl#sha256=c3702b6d3dd8c7abc1afa565d7e63d53a1d0bd86cdc24edd75470f4de499cfcc +pycryptodomex @ https://agent-int-packages.datadoghq.com/external/pycryptodomex/pycryptodomex-3.20.0-cp35-abi3-manylinux_2_17_aarch64.manylinux2014_aarch64.whl#sha256=91852d4480a4537d169c29a9d104dda44094c78f1f5b67bca76c29a91042b623 +pydantic @ https://agent-int-packages.datadoghq.com/external/pydantic/pydantic-2.8.2-py3-none-any.whl#sha256=73ee9fddd406dc318b885c7a2eab8a6472b68b8fb5ba8150949fc3db939f23c8 +pydantic-core @ https://agent-int-packages.datadoghq.com/external/pydantic-core/pydantic_core-2.20.1-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl#sha256=f9aa05d09ecf4c75157197f27cdc9cfaeb7c5f15021c6373932bf3e124af029f +pyjwt @ https://agent-int-packages.datadoghq.com/external/pyjwt/PyJWT-2.9.0-py3-none-any.whl#sha256=3b02fb0f44517787776cf48f2ae25d8e14f300e6d7545a4315cee571a415e850 +pymongo @ https://agent-int-packages.datadoghq.com/external/pymongo/pymongo-4.8.0-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl#sha256=c6b804bb4f2d9dc389cc9e827d579fa327272cdb0629a99bfe5b83cb3e269ebf +pymysql @ https://agent-int-packages.datadoghq.com/external/pymysql/PyMySQL-1.1.1-py3-none-any.whl#sha256=4de15da4c61dc132f4fb9ab763063e693d521a80fd0e87943b9a453dd4c19d6c +pynacl @ https://agent-int-packages.datadoghq.com/external/pynacl/PyNaCl-1.5.0-cp36-abi3-manylinux_2_17_aarch64.manylinux2014_aarch64.whl#sha256=a36d4a9dda1f19ce6e03c9a784a2921a4b726b02e1c736600ca9c22029474394 +pyodbc @ https://agent-int-packages.datadoghq.com/built/pyodbc/pyodbc-5.1.0-20240917155308-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl#sha256=677380a4d52d3c02d426d11a77fcf205f19926da759f070eeed4007e546f6b4a +pyopenssl @ https://agent-int-packages.datadoghq.com/external/pyopenssl/pyOpenSSL-24.2.1-py3-none-any.whl#sha256=967d5719b12b243588573f39b0c677637145c7a1ffedcd495a487e58177fbb8d +pysmi @ https://agent-int-packages.datadoghq.com/external/pysmi/pysmi-1.2.1-py3-none-any.whl#sha256=d97c60de9f81d33ab2899124d95a94fa7fefacc86ab6e00cbfec543a073e6d33 +pysnmp @ https://agent-int-packages.datadoghq.com/external/pysnmp/pysnmp-5.1.0-py3-none-any.whl#sha256=375a8adfc6820faf24ace6761a6d20544e60580d714ff7266df272850c39b439 +pysnmp-mibs @ https://agent-int-packages.datadoghq.com/external/pysnmp-mibs/pysnmp_mibs-0.1.6-py2.py3-none-any.whl#sha256=5e153ebe8e767c07940cea435f866c623ff6b2376155c7da75085b08d3774d48 +pysnmpcrypto @ https://agent-int-packages.datadoghq.com/external/pysnmpcrypto/pysnmpcrypto-0.0.4-py2.py3-none-any.whl#sha256=5889733caa030f45d9e03ea9d6370fb06426a8cb7f839aabbcdde33c6f634679 +pysocks @ https://agent-int-packages.datadoghq.com/external/pysocks/PySocks-1.7.1-py3-none-any.whl#sha256=2725bd0a9925919b9b51739eea5f9e2bae91e83288108a9ad338b2e3a4435ee5 +pyspnego @ https://agent-int-packages.datadoghq.com/external/pyspnego/pyspnego-0.11.1-py3-none-any.whl#sha256=129a4294f2c4d681d5875240ef87accc6f1d921e8983737fb0b59642b397951e +python-binary-memcached @ https://agent-int-packages.datadoghq.com/external/python-binary-memcached/python_binary_memcached-0.31.2-py3-none-any.whl#sha256=e5b93d54429e835cab7d5b33988649f9748344aa49adaed8eed94b37e714d562 +python-dateutil @ https://agent-int-packages.datadoghq.com/external/python-dateutil/python_dateutil-2.9.0.post0-py2.py3-none-any.whl#sha256=a8b2bc7bffae282281c8140a97d3aa9c14da0b136dfe83f850eea9a5f7470427 +python3-gearman @ https://agent-int-packages.datadoghq.com/external/python3-gearman/python3_gearman-0.1.0-py3-none-any.whl#sha256=4a5808d3a0bfc6c243548ad57e7aab4bee62c9cba2b1c3a860fdd292d46a112d +pytz @ https://agent-int-packages.datadoghq.com/external/pytz/pytz-2024.2-py2.py3-none-any.whl#sha256=31c7c1817eb7fae7ca4b8c7ee50c72f93aa2dd863de768e1ef4245d426aa0725 +pyvmomi @ https://agent-int-packages.datadoghq.com/built/pyvmomi/pyvmomi-8.0.3.0.1-20240702172038-py2.py3-none-manylinux2014_aarch64.whl#sha256=e173daf28895975b57850fef301837f24fba59dd8ff1d931795732e9be281d57 +pyyaml @ https://agent-int-packages.datadoghq.com/external/pyyaml/PyYAML-6.0.2-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl#sha256=1f71ea527786de97d1a0cc0eacd1defc0985dcf6b3f17bb77dcfc8c34bec4dc5 +redis @ https://agent-int-packages.datadoghq.com/external/redis/redis-5.0.8-py3-none-any.whl#sha256=56134ee08ea909106090934adc36f65c9bcbbaecea5b21ba704ba6fb561f8eb4 +requests @ https://agent-int-packages.datadoghq.com/external/requests/requests-2.32.3-py3-none-any.whl#sha256=70761cfe03c773ceb22aa2f671b4757976145175cdfca038c02654d061d6dcc6 +requests-kerberos @ https://agent-int-packages.datadoghq.com/external/requests-kerberos/requests_kerberos-0.15.0-py2.py3-none-any.whl#sha256=ba9b0980b8489c93bfb13854fd118834e576d6700bfea3745cb2e62278cd16a6 +requests-ntlm @ https://agent-int-packages.datadoghq.com/external/requests-ntlm/requests_ntlm-1.3.0-py3-none-any.whl#sha256=4c7534a7d0e482bb0928531d621be4b2c74ace437e88c5a357ceb7452d25a510 +requests-oauthlib @ https://agent-int-packages.datadoghq.com/external/requests-oauthlib/requests_oauthlib-2.0.0-py2.py3-none-any.whl#sha256=7dd8a5c40426b779b0868c404bdef9768deccf22749cde15852df527e6269b36 +requests-toolbelt @ https://agent-int-packages.datadoghq.com/external/requests-toolbelt/requests_toolbelt-1.0.0-py2.py3-none-any.whl#sha256=cccfdd665f0a24fcf4726e690f65639d272bb0637b9b92dfd91a5568ccf6bd06 +requests-unixsocket2 @ https://agent-int-packages.datadoghq.com/external/requests-unixsocket2/requests_unixsocket2-0.4.2-py3-none-any.whl#sha256=701fcd49d74bc0f759bbe45c4dfda0045fd89652948c2b473b1a312214c3770b +requestsexceptions @ https://agent-int-packages.datadoghq.com/external/requestsexceptions/requestsexceptions-1.4.0-py2.py3-none-any.whl#sha256=3083d872b6e07dc5c323563ef37671d992214ad9a32b0ca4a3d7f5500bf38ce3 +rethinkdb @ https://agent-int-packages.datadoghq.com/external/rethinkdb/rethinkdb-2.4.9-py2.py3-none-any.whl#sha256=dd2455b3eab5266d8006eeefeb5eb970a3697dfe6aa18e1ab6e1d99233badfcb +rsa @ https://agent-int-packages.datadoghq.com/external/rsa/rsa-4.9-py3-none-any.whl#sha256=90260d9058e514786967344d0ef75fa8727eed8a7d2e43ce9f4bcf1b536174f7 +s3transfer @ https://agent-int-packages.datadoghq.com/external/s3transfer/s3transfer-0.10.2-py3-none-any.whl#sha256=eca1c20de70a39daee580aef4986996620f365c4e0fda6a86100231d62f1bf69 +securesystemslib @ https://agent-int-packages.datadoghq.com/external/securesystemslib/securesystemslib-0.28.0-py3-none-any.whl#sha256=9e6b9abe36a511d4f52c759069db8f6f650362ba82d6efc7bc7466a458b3f499 +semver @ https://agent-int-packages.datadoghq.com/external/semver/semver-3.0.2-py3-none-any.whl#sha256=b1ea4686fe70b981f85359eda33199d60c53964284e0cfb4977d243e37cf4bf4 +service-identity @ https://agent-int-packages.datadoghq.com/external/service-identity/service_identity-24.1.0-py3-none-any.whl#sha256=a28caf8130c8a5c1c7a6f5293faaf239bbfb7751e4862436920ee6f2616f568a +setuptools @ https://agent-int-packages.datadoghq.com/external/setuptools/setuptools-75.1.0-py3-none-any.whl#sha256=35ab7fd3bcd95e6b7fd704e4a1539513edad446c097797f2985e0e4b960772f2 +simplejson @ https://agent-int-packages.datadoghq.com/external/simplejson/simplejson-3.19.3-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl#sha256=d0b0efc7279d768db7c74d3d07f0b5c81280d16ae3fb14e9081dc903e8360771 +six @ https://agent-int-packages.datadoghq.com/external/six/six-1.16.0-py2.py3-none-any.whl#sha256=8abb2f1d86890a2dfb989f9a77cfcfd3e47c2a354b01111771326f8aa26e0254 +snowflake-connector-python @ https://agent-int-packages.datadoghq.com/external/snowflake-connector-python/snowflake_connector_python-3.12.1-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl#sha256=473642c0e628b8b9f264cbf31c7f4de44974373db43052b6542a66e751159caf +sortedcontainers @ https://agent-int-packages.datadoghq.com/external/sortedcontainers/sortedcontainers-2.4.0-py2.py3-none-any.whl#sha256=a163dcaede0f1c021485e957a39245190e74249897e2ae4b2aa38595db237ee0 +soupsieve @ https://agent-int-packages.datadoghq.com/external/soupsieve/soupsieve-2.6-py3-none-any.whl#sha256=e72c4ff06e4fb6e4b5a9f0f55fe6e81514581fca1515028625d0f299c602ccc9 +stevedore @ https://agent-int-packages.datadoghq.com/external/stevedore/stevedore-5.3.0-py3-none-any.whl#sha256=1efd34ca08f474dad08d9b19e934a22c68bb6fe416926479ba29e5013bcc8f78 +supervisor @ https://agent-int-packages.datadoghq.com/external/supervisor/supervisor-4.2.5-py2.py3-none-any.whl#sha256=2ecaede32fc25af814696374b79e42644ecaba5c09494c51016ffda9602d0f08 +tomlkit @ https://agent-int-packages.datadoghq.com/external/tomlkit/tomlkit-0.13.2-py3-none-any.whl#sha256=7a974427f6e119197f670fbbbeae7bef749a6c14e793db934baefc1b5f03efde +tuf @ https://agent-int-packages.datadoghq.com/external/tuf/tuf-4.0.0-py3-none-any.whl#sha256=a22ab5fa6daf910b3052929fdce42ccad8a300e5e85715daaff9592aed980f7a +typing-extensions @ https://agent-int-packages.datadoghq.com/external/typing-extensions/typing_extensions-4.12.2-py3-none-any.whl#sha256=04e5ca0351e0f3f85c6853954072df659d0d13fac324d0072316b67d7794700d +tzlocal @ https://agent-int-packages.datadoghq.com/external/tzlocal/tzlocal-5.2-py3-none-any.whl#sha256=49816ef2fe65ea8ac19d19aa7a1ae0551c834303d5014c6d5a62e4cbda8047b8 +uhashring @ https://agent-int-packages.datadoghq.com/external/uhashring/uhashring-2.3-py3-none-any.whl#sha256=7ee8a25ca495a97effad10bd563c83b4054a6d7606d9530757049a04edab9297 +uptime @ https://agent-int-packages.datadoghq.com/built/uptime/uptime-3.0.1-20240917155308-cp312-cp312-manylinux2014_aarch64.manylinux_2_17_aarch64.whl#sha256=6e4a3bb156fe2f24c0f00f17ee735a6caccb674aa6f307a164248c989baa8914 +urllib3 @ https://agent-int-packages.datadoghq.com/external/urllib3/urllib3-2.2.3-py3-none-any.whl#sha256=ca899ca043dcb1bafa3e262d73aa25c465bfb49e0bd9dd5d59f1d0acba2f8fac +vertica-python @ https://agent-int-packages.datadoghq.com/external/vertica-python/vertica_python-1.4.0-py3-none-any.whl#sha256=50fecd7687f4b0b9f6dee6e2b35c195af2a4f702ece01bd12e080b51756e000b +websocket-client @ https://agent-int-packages.datadoghq.com/external/websocket-client/websocket_client-1.8.0-py3-none-any.whl#sha256=17b44cc997f5c498e809b22cdf2d9c7a9e71c02c8cc2b6c56e7c2d1239bfa526 +wrapt @ https://agent-int-packages.datadoghq.com/external/wrapt/wrapt-1.16.0-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl#sha256=94265b00870aa407bd0cbcfd536f17ecde43b94fb8d228560a1e9d3041462d73 +xmltodict @ https://agent-int-packages.datadoghq.com/external/xmltodict/xmltodict-0.13.0-py2.py3-none-any.whl#sha256=aa89e8fd76320154a40d19a0df04a4695fb9dc5ba977cbb68ab3e4eb225e7852 +zipp @ https://agent-int-packages.datadoghq.com/external/zipp/zipp-3.20.2-py3-none-any.whl#sha256=a817ac80d6cf4b23bf7f2828b7cabf326f15a001bea8b1f9b49631780ba28350 diff --git a/.deps/resolved/linux-x86_64_3.12.txt b/.deps/resolved/linux-x86_64_3.12.txt new file mode 100644 index 0000000000000..018f55a9b844d --- /dev/null +++ b/.deps/resolved/linux-x86_64_3.12.txt @@ -0,0 +1,129 @@ +aerospike @ https://agent-int-packages.datadoghq.com/built/aerospike/aerospike-7.1.1-20240917155248-cp312-cp312-manylinux2014_x86_64.manylinux_2_17_x86_64.whl#sha256=50a14738d4102572c0dd7ccddcc735074f30902813190369f0765f2f4a087300 +annotated-types @ https://agent-int-packages.datadoghq.com/external/annotated-types/annotated_types-0.7.0-py3-none-any.whl#sha256=1f02e8b43a8fbbc3f3e0d4f0f4bfc8131bcb4eebe8849b8e5c773f3a1c582a53 +asn1crypto @ https://agent-int-packages.datadoghq.com/external/asn1crypto/asn1crypto-1.5.1-py2.py3-none-any.whl#sha256=db4e40728b728508912cbb3d44f19ce188f218e9eba635821bb4b68564f8fd67 +attrs @ https://agent-int-packages.datadoghq.com/external/attrs/attrs-24.2.0-py3-none-any.whl#sha256=81921eb96de3191c8258c199618104dd27ac608d9366f5e35d011eae1867ede2 +aws-requests-auth @ https://agent-int-packages.datadoghq.com/external/aws-requests-auth/aws_requests_auth-0.4.3-py2.py3-none-any.whl#sha256=646bc37d62140ea1c709d20148f5d43197e6bd2d63909eb36fa4bb2345759977 +azure-core @ https://agent-int-packages.datadoghq.com/external/azure-core/azure_core-1.31.0-py3-none-any.whl#sha256=22954de3777e0250029360ef31d80448ef1be13b80a459bff80ba7073379e2cd +azure-identity @ https://agent-int-packages.datadoghq.com/external/azure-identity/azure_identity-1.17.1-py3-none-any.whl#sha256=db8d59c183b680e763722bfe8ebc45930e6c57df510620985939f7f3191e0382 +bcrypt @ https://agent-int-packages.datadoghq.com/external/bcrypt/bcrypt-4.2.0-cp39-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl#sha256=3413bd60460f76097ee2e0a493ccebe4a7601918219c02f503984f0a7ee0aebe +beautifulsoup4 @ https://agent-int-packages.datadoghq.com/external/beautifulsoup4/beautifulsoup4-4.12.3-py3-none-any.whl#sha256=b80878c9f40111313e55da8ba20bdba06d8fa3969fc68304167741bbf9e082ed +binary @ https://agent-int-packages.datadoghq.com/external/binary/binary-1.0.0-py2.py3-none-any.whl#sha256=e1b61f3a5c002717d1a28e4d9d2dc8acbc9d6b12baf7b1e4ab25d743da97e323 +boto3 @ https://agent-int-packages.datadoghq.com/external/boto3/boto3-1.35.10-py3-none-any.whl#sha256=add26dd58e076dfd387013da4704716d5cff215cf14f6d4347c4b9b7fc1f0b8e +botocore @ https://agent-int-packages.datadoghq.com/external/botocore/botocore-1.35.10-py3-none-any.whl#sha256=0d96d023b9b0cea99a0a428a431d011329d3a958730aee6ed6a6fec5d9bfbc03 +bytecode @ https://agent-int-packages.datadoghq.com/external/bytecode/bytecode-0.15.1-py3-none-any.whl#sha256=0a1dc340cac823cff605609b8b214f7f9bf80418c6b9e0fc8c6db1793c27137d +cachetools @ https://agent-int-packages.datadoghq.com/external/cachetools/cachetools-5.5.0-py3-none-any.whl#sha256=02134e8439cdc2ffb62023ce1debca2944c3f289d66bb17ead3ab3dede74b292 +cattrs @ https://agent-int-packages.datadoghq.com/external/cattrs/cattrs-24.1.1-py3-none-any.whl#sha256=ec8ce8fdc725de9d07547cd616f968670687c6fa7a2e263b088370c46d834d97 +certifi @ https://agent-int-packages.datadoghq.com/external/certifi/certifi-2024.8.30-py3-none-any.whl#sha256=922820b53db7a7257ffbda3f597266d435245903d80737e34f8a45ff3e3230d8 +cffi @ https://agent-int-packages.datadoghq.com/external/cffi/cffi-1.17.1-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl#sha256=b62ce867176a75d03a665bad002af8e6d54644fad99a3c70905c543130e39d93 +charset-normalizer @ https://agent-int-packages.datadoghq.com/external/charset-normalizer/charset_normalizer-3.3.2-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl#sha256=90d558489962fd4918143277a773316e56c72da56ec7aa3dc3dbbe20fdfed15b +clickhouse-cityhash @ https://agent-int-packages.datadoghq.com/external/clickhouse-cityhash/clickhouse_cityhash-1.0.2.4-cp312-cp312-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl#sha256=f1f8fec4027cd648f72009ef59c9b76c5a27a33ca166b4e79e46542009429813 +clickhouse-driver @ https://agent-int-packages.datadoghq.com/external/clickhouse-driver/clickhouse_driver-0.2.9-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl#sha256=6dbcee870c60d9835e5dce1456ab6b9d807e6669246357f4b321ef747b90fa43 +cm-client @ https://agent-int-packages.datadoghq.com/built/cm-client/cm_client-45.0.4-20240402154838-py3-none-manylinux2014_x86_64.whl#sha256=aba3c1683ef1b2099933e030464d29b3ad1c206784ebd15d8a7147ecd6ba24e1 +confluent-kafka @ https://agent-int-packages.datadoghq.com/built/confluent-kafka/confluent_kafka-2.5.0-20240917155249-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl#sha256=aea2f7e617f82476c56eed42d0b5cf6b944a2769d93c8bf7baf0adef19d19bf7 +cryptography @ https://agent-int-packages.datadoghq.com/external/cryptography/cryptography-43.0.0-cp39-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl#sha256=299d3da8e00b7e2b54bb02ef58d73cd5f55fb31f33ebbf33bd00d9aa6807df7e +ddsketch @ https://agent-int-packages.datadoghq.com/external/ddsketch/ddsketch-3.0.1-py3-none-any.whl#sha256=6d047b455fe2837c43d366ff1ae6ba0c3166e15499de8688437a75cea914224e +ddtrace @ https://agent-int-packages.datadoghq.com/external/ddtrace/ddtrace-2.10.6-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl#sha256=5fc70ac472093093d9908965d95d977206372a3ddc8a2562acf9dfd57c6864d8 +decorator @ https://agent-int-packages.datadoghq.com/external/decorator/decorator-5.1.1-py3-none-any.whl#sha256=b8c3f85900b9dc423225913c5aace94729fe1fa9763b38939a95226f02d37186 +deprecated @ https://agent-int-packages.datadoghq.com/external/deprecated/Deprecated-1.2.14-py2.py3-none-any.whl#sha256=6fac8b097794a90302bdbb17b9b815e732d3c4720583ff1b198499d78470466c +dnspython @ https://agent-int-packages.datadoghq.com/external/dnspython/dnspython-2.6.1-py3-none-any.whl#sha256=5ef3b9680161f6fa89daf8ad451b5f1a33b18ae8a1c6778cdf4b43f08c0a6e50 +dogpile-cache @ https://agent-int-packages.datadoghq.com/external/dogpile-cache/dogpile.cache-1.3.3-py3-none-any.whl#sha256=5e211c4902ebdf88c678d268e22454b41e68071632daa9402d8ee24e825ed8ca +envier @ https://agent-int-packages.datadoghq.com/external/envier/envier-0.5.2-py3-none-any.whl#sha256=65099cf3aa9b3b3b4b92db2f7d29e2910672e085b76f7e587d2167561a834add +filelock @ https://agent-int-packages.datadoghq.com/external/filelock/filelock-3.16.0-py3-none-any.whl#sha256=f6ed4c963184f4c84dd5557ce8fece759a3724b37b80c6c4f20a2f63a4dc6609 +foundationdb @ https://agent-int-packages.datadoghq.com/built/foundationdb/foundationdb-6.3.24-20240402154840-py3-none-manylinux2014_x86_64.whl#sha256=14259f824080062cc890965747597ff00a9d6c76a1eb926673fed68a45860ccd +google-auth @ https://agent-int-packages.datadoghq.com/external/google-auth/google_auth-2.34.0-py2.py3-none-any.whl#sha256=72fd4733b80b6d777dcde515628a9eb4a577339437012874ea286bca7261ee65 +gssapi @ https://agent-int-packages.datadoghq.com/built/gssapi/gssapi-1.8.3-20240917155249-cp312-cp312-manylinux2014_x86_64.manylinux_2_17_x86_64.whl#sha256=c2bf203c7f213368bc8c2de664495f8e4699235ab2d58251a45cb30acbd92c08 +hazelcast-python-client @ https://agent-int-packages.datadoghq.com/external/hazelcast-python-client/hazelcast_python_client-5.4.0-py3-none-any.whl#sha256=16195cd58feb2dd3be1594d08d42527ae00797548a6a9d6a601aae2e8514ff5f +idna @ https://agent-int-packages.datadoghq.com/external/idna/idna-3.10-py3-none-any.whl#sha256=946d195a0d259cbba61165e88e65941f16e9b36ea6ddb97f00452bae8b1287d3 +importlib-metadata @ https://agent-int-packages.datadoghq.com/external/importlib-metadata/importlib_metadata-8.4.0-py3-none-any.whl#sha256=66f342cc6ac9818fc6ff340576acd24d65ba0b3efabb2b4ac08b598965a4a2f1 +in-toto @ https://agent-int-packages.datadoghq.com/external/in-toto/in_toto-2.0.0-py3-none-any.whl#sha256=eaf71733cdd5b2309b58b24169c13c24701da72baa43b7432d1dffc308e7bf32 +iso8601 @ https://agent-int-packages.datadoghq.com/external/iso8601/iso8601-2.1.0-py3-none-any.whl#sha256=aac4145c4dcb66ad8b648a02830f5e2ff6c24af20f4f482689be402db2429242 +jellyfish @ https://agent-int-packages.datadoghq.com/external/jellyfish/jellyfish-1.1.0-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl#sha256=e9d4002d01252f18eb26f28b66f6c9ce0696221804d8769553c5912b2f221a18 +jmespath @ https://agent-int-packages.datadoghq.com/external/jmespath/jmespath-1.0.1-py3-none-any.whl#sha256=02e2e4cc71b5bcab88332eebf907519190dd9e6e82107fa7f83b1003a6252980 +jsonpatch @ https://agent-int-packages.datadoghq.com/external/jsonpatch/jsonpatch-1.33-py2.py3-none-any.whl#sha256=0ae28c0cd062bbd8b8ecc26d7d164fbbea9652a1a3693f3b956c1eae5145dade +jsonpointer @ https://agent-int-packages.datadoghq.com/external/jsonpointer/jsonpointer-3.0.0-py2.py3-none-any.whl#sha256=13e088adc14fca8b6aa8177c044e12701e6ad4b28ff10e65f2267a90109c9942 +keystoneauth1 @ https://agent-int-packages.datadoghq.com/external/keystoneauth1/keystoneauth1-5.8.0-py3-none-any.whl#sha256=e69dff80c509ab64d4de4494658d914e81f26af720828dc584ceee74ecd666d9 +krb5 @ https://agent-int-packages.datadoghq.com/built/krb5/krb5-0.6.0-20240917155250-cp312-cp312-manylinux2014_x86_64.manylinux_2_17_x86_64.whl#sha256=f90b6b9bf1e964f081525253d1e20ebda0ca15f05df2194a16647d0c89c6293e +kubernetes @ https://agent-int-packages.datadoghq.com/external/kubernetes/kubernetes-30.1.0-py2.py3-none-any.whl#sha256=e212e8b7579031dd2e512168b617373bc1e03888d41ac4e04039240a292d478d +ldap3 @ https://agent-int-packages.datadoghq.com/external/ldap3/ldap3-2.9.1-py2.py3-none-any.whl#sha256=5869596fc4948797020d3f03b7939da938778a0f9e2009f7a072ccf92b8e8d70 +lxml @ https://agent-int-packages.datadoghq.com/built/lxml/lxml-4.9.4-20240917155250-cp312-cp312-manylinux2014_x86_64.manylinux_2_17_x86_64.whl#sha256=0c1d96942a08b30d1ae8192b59249d91254eccd4198361eabdcd1ef2e76158dc +lz4 @ https://agent-int-packages.datadoghq.com/external/lz4/lz4-4.3.3-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl#sha256=bca8fccc15e3add173da91be8f34121578dc777711ffd98d399be35487c934bf +mmh3 @ https://agent-int-packages.datadoghq.com/external/mmh3/mmh3-4.1.0-cp312-cp312-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl#sha256=3b02268be6e0a8eeb8a924d7db85f28e47344f35c438c1e149878bb1c47b1cd3 +msal @ https://agent-int-packages.datadoghq.com/external/msal/msal-1.31.0-py3-none-any.whl#sha256=96bc37cff82ebe4b160d5fc0f1196f6ca8b50e274ecd0ec5bf69c438514086e7 +msal-extensions @ https://agent-int-packages.datadoghq.com/external/msal-extensions/msal_extensions-1.2.0-py3-none-any.whl#sha256=cf5ba83a2113fa6dc011a254a72f1c223c88d7dfad74cc30617c4679a417704d +netifaces @ https://agent-int-packages.datadoghq.com/built/netifaces/netifaces-0.11.0-20240917155251-cp312-cp312-manylinux2014_x86_64.manylinux_2_17_x86_64.whl#sha256=9f873ab2793981115a8d0fdfba8869ae6b5d312c79045db1867d9294541dcb3d +oauthlib @ https://agent-int-packages.datadoghq.com/external/oauthlib/oauthlib-3.2.2-py3-none-any.whl#sha256=8139f29aac13e25d502680e9e19963e83f16838d48a0d71c287fe40e7067fbca +openstacksdk @ https://agent-int-packages.datadoghq.com/external/openstacksdk/openstacksdk-3.3.0-py3-none-any.whl#sha256=e6d4121b87354984caf0e3c032e2ebf4d4440374f86c81c27ec52ca5df359157 +opentelemetry-api @ https://agent-int-packages.datadoghq.com/external/opentelemetry-api/opentelemetry_api-1.27.0-py3-none-any.whl#sha256=953d5871815e7c30c81b56d910c707588000fff7a3ca1c73e6531911d53065e7 +orjson @ https://agent-int-packages.datadoghq.com/external/orjson/orjson-3.10.7-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl#sha256=a0c6a008e91d10a2564edbb6ee5069a9e66df3fbe11c9a005cb411f441fd2c09 +os-service-types @ https://agent-int-packages.datadoghq.com/external/os-service-types/os_service_types-1.7.0-py2.py3-none-any.whl#sha256=0505c72205690910077fb72b88f2a1f07533c8d39f2fe75b29583481764965d6 +packaging @ https://agent-int-packages.datadoghq.com/external/packaging/packaging-24.1-py3-none-any.whl#sha256=5b8f2217dbdbd2f7f384c41c628544e6d52f2d0f53c6d0c3ea61aa5d1d7ff124 +paramiko @ https://agent-int-packages.datadoghq.com/external/paramiko/paramiko-3.4.1-py3-none-any.whl#sha256=8e49fd2f82f84acf7ffd57c64311aa2b30e575370dc23bdb375b10262f7eac32 +pathspec @ https://agent-int-packages.datadoghq.com/external/pathspec/pathspec-0.12.1-py3-none-any.whl#sha256=a0d503e138a4c123b27490a4f7beda6a01c6f288df0e4a8b79c7eb0dc7b4cc08 +pbr @ https://agent-int-packages.datadoghq.com/external/pbr/pbr-6.1.0-py2.py3-none-any.whl#sha256=a776ae228892d8013649c0aeccbb3d5f99ee15e005a4cbb7e61d55a067b28a2a +platformdirs @ https://agent-int-packages.datadoghq.com/external/platformdirs/platformdirs-4.3.3-py3-none-any.whl#sha256=50a5450e2e84f44539718293cbb1da0a0885c9d14adf21b77bae4e66fc99d9b5 +ply @ https://agent-int-packages.datadoghq.com/external/ply/ply-3.11-py2.py3-none-any.whl#sha256=096f9b8350b65ebd2fd1346b12452efe5b9607f7482813ffca50c22722a807ce +portalocker @ https://agent-int-packages.datadoghq.com/external/portalocker/portalocker-2.10.1-py3-none-any.whl#sha256=53a5984ebc86a025552264b459b46a2086e269b21823cb572f8f28ee759e45bf +prometheus-client @ https://agent-int-packages.datadoghq.com/external/prometheus-client/prometheus_client-0.20.0-py3-none-any.whl#sha256=cde524a85bce83ca359cc837f28b8c0db5cac7aa653a588fd7e84ba061c329e7 +protobuf @ https://agent-int-packages.datadoghq.com/external/protobuf/protobuf-5.27.3-cp38-abi3-manylinux2014_x86_64.whl#sha256=a55c48f2a2092d8e213bd143474df33a6ae751b781dd1d1f4d953c128a415b25 +psutil @ https://agent-int-packages.datadoghq.com/external/psutil/psutil-5.9.6-cp36-abi3-manylinux_2_12_x86_64.manylinux2010_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl#sha256=748c9dd2583ed86347ed65d0035f45fa8c851e8d90354c122ab72319b5f366f4 +psycopg2-binary @ https://agent-int-packages.datadoghq.com/external/psycopg2-binary/psycopg2_binary-2.9.9-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl#sha256=6e6f98446430fdf41bd36d4faa6cb409f5140c1c2cf58ce0bbdaf16af7d3f119 +pyasn1 @ https://agent-int-packages.datadoghq.com/external/pyasn1/pyasn1-0.4.8-py2.py3-none-any.whl#sha256=39c7e2ec30515947ff4e87fb6f456dfc6e84857d34be479c9d4a4ba4bf46aa5d +pyasn1-modules @ https://agent-int-packages.datadoghq.com/external/pyasn1-modules/pyasn1_modules-0.4.1-py3-none-any.whl#sha256=49bfa96b45a292b711e986f222502c1c9a5e1f4e568fc30e2574a6c7d07838fd +pyasyncore @ https://agent-int-packages.datadoghq.com/external/pyasyncore/pyasyncore-1.0.4-py3-none-any.whl#sha256=9e5f6dc9dc057c56370b7a5cdb4c4670fd4b0556de2913ed1f428cd6a5366895 +pycparser @ https://agent-int-packages.datadoghq.com/external/pycparser/pycparser-2.22-py3-none-any.whl#sha256=c3702b6d3dd8c7abc1afa565d7e63d53a1d0bd86cdc24edd75470f4de499cfcc +pycryptodomex @ https://agent-int-packages.datadoghq.com/external/pycryptodomex/pycryptodomex-3.20.0-cp35-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl#sha256=bca649483d5ed251d06daf25957f802e44e6bb6df2e8f218ae71968ff8f8edc4 +pydantic @ https://agent-int-packages.datadoghq.com/external/pydantic/pydantic-2.8.2-py3-none-any.whl#sha256=73ee9fddd406dc318b885c7a2eab8a6472b68b8fb5ba8150949fc3db939f23c8 +pydantic-core @ https://agent-int-packages.datadoghq.com/external/pydantic-core/pydantic_core-2.20.1-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl#sha256=26dc97754b57d2fd00ac2b24dfa341abffc380b823211994c4efac7f13b9e90e +pyjwt @ https://agent-int-packages.datadoghq.com/external/pyjwt/PyJWT-2.9.0-py3-none-any.whl#sha256=3b02fb0f44517787776cf48f2ae25d8e14f300e6d7545a4315cee571a415e850 +pymongo @ https://agent-int-packages.datadoghq.com/external/pymongo/pymongo-4.8.0-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl#sha256=940d456774b17814bac5ea7fc28188c7a1338d4a233efbb6ba01de957bded2e8 +pymqi @ https://agent-int-packages.datadoghq.com/built/pymqi/pymqi-1.12.10-20240917155251-cp312-cp312-manylinux2014_x86_64.manylinux_2_17_x86_64.whl#sha256=f62a434af2e4da5b681bea1c2203f0b2d6a3cc4b27ee74b3dfb57bf0a558049c +pymysql @ https://agent-int-packages.datadoghq.com/external/pymysql/PyMySQL-1.1.1-py3-none-any.whl#sha256=4de15da4c61dc132f4fb9ab763063e693d521a80fd0e87943b9a453dd4c19d6c +pynacl @ https://agent-int-packages.datadoghq.com/external/pynacl/PyNaCl-1.5.0-cp36-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl#sha256=06b8f6fa7f5de8d5d2f7573fe8c863c051225a27b61e6860fd047b1775807858 +pyodbc @ https://agent-int-packages.datadoghq.com/built/pyodbc/pyodbc-5.1.0-20240917155252-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl#sha256=967887428c3564ee39edc65074e94d1f0a94fb3d1477cfcff080baa699525754 +pyopenssl @ https://agent-int-packages.datadoghq.com/external/pyopenssl/pyOpenSSL-24.2.1-py3-none-any.whl#sha256=967d5719b12b243588573f39b0c677637145c7a1ffedcd495a487e58177fbb8d +pysmi @ https://agent-int-packages.datadoghq.com/external/pysmi/pysmi-1.2.1-py3-none-any.whl#sha256=d97c60de9f81d33ab2899124d95a94fa7fefacc86ab6e00cbfec543a073e6d33 +pysnmp @ https://agent-int-packages.datadoghq.com/external/pysnmp/pysnmp-5.1.0-py3-none-any.whl#sha256=375a8adfc6820faf24ace6761a6d20544e60580d714ff7266df272850c39b439 +pysnmp-mibs @ https://agent-int-packages.datadoghq.com/external/pysnmp-mibs/pysnmp_mibs-0.1.6-py2.py3-none-any.whl#sha256=5e153ebe8e767c07940cea435f866c623ff6b2376155c7da75085b08d3774d48 +pysnmpcrypto @ https://agent-int-packages.datadoghq.com/external/pysnmpcrypto/pysnmpcrypto-0.0.4-py2.py3-none-any.whl#sha256=5889733caa030f45d9e03ea9d6370fb06426a8cb7f839aabbcdde33c6f634679 +pysocks @ https://agent-int-packages.datadoghq.com/external/pysocks/PySocks-1.7.1-py3-none-any.whl#sha256=2725bd0a9925919b9b51739eea5f9e2bae91e83288108a9ad338b2e3a4435ee5 +pyspnego @ https://agent-int-packages.datadoghq.com/external/pyspnego/pyspnego-0.11.1-py3-none-any.whl#sha256=129a4294f2c4d681d5875240ef87accc6f1d921e8983737fb0b59642b397951e +python-binary-memcached @ https://agent-int-packages.datadoghq.com/external/python-binary-memcached/python_binary_memcached-0.31.2-py3-none-any.whl#sha256=e5b93d54429e835cab7d5b33988649f9748344aa49adaed8eed94b37e714d562 +python-dateutil @ https://agent-int-packages.datadoghq.com/external/python-dateutil/python_dateutil-2.9.0.post0-py2.py3-none-any.whl#sha256=a8b2bc7bffae282281c8140a97d3aa9c14da0b136dfe83f850eea9a5f7470427 +python3-gearman @ https://agent-int-packages.datadoghq.com/external/python3-gearman/python3_gearman-0.1.0-py3-none-any.whl#sha256=4a5808d3a0bfc6c243548ad57e7aab4bee62c9cba2b1c3a860fdd292d46a112d +pytz @ https://agent-int-packages.datadoghq.com/external/pytz/pytz-2024.2-py2.py3-none-any.whl#sha256=31c7c1817eb7fae7ca4b8c7ee50c72f93aa2dd863de768e1ef4245d426aa0725 +pyvmomi @ https://agent-int-packages.datadoghq.com/built/pyvmomi/pyvmomi-8.0.3.0.1-20240702172050-py2.py3-none-manylinux2014_x86_64.whl#sha256=e173daf28895975b57850fef301837f24fba59dd8ff1d931795732e9be281d57 +pyyaml @ https://agent-int-packages.datadoghq.com/external/pyyaml/PyYAML-6.0.2-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl#sha256=80bab7bfc629882493af4aa31a4cfa43a4c57c83813253626916b8c7ada83476 +redis @ https://agent-int-packages.datadoghq.com/external/redis/redis-5.0.8-py3-none-any.whl#sha256=56134ee08ea909106090934adc36f65c9bcbbaecea5b21ba704ba6fb561f8eb4 +requests @ https://agent-int-packages.datadoghq.com/external/requests/requests-2.32.3-py3-none-any.whl#sha256=70761cfe03c773ceb22aa2f671b4757976145175cdfca038c02654d061d6dcc6 +requests-kerberos @ https://agent-int-packages.datadoghq.com/external/requests-kerberos/requests_kerberos-0.15.0-py2.py3-none-any.whl#sha256=ba9b0980b8489c93bfb13854fd118834e576d6700bfea3745cb2e62278cd16a6 +requests-ntlm @ https://agent-int-packages.datadoghq.com/external/requests-ntlm/requests_ntlm-1.3.0-py3-none-any.whl#sha256=4c7534a7d0e482bb0928531d621be4b2c74ace437e88c5a357ceb7452d25a510 +requests-oauthlib @ https://agent-int-packages.datadoghq.com/external/requests-oauthlib/requests_oauthlib-2.0.0-py2.py3-none-any.whl#sha256=7dd8a5c40426b779b0868c404bdef9768deccf22749cde15852df527e6269b36 +requests-toolbelt @ https://agent-int-packages.datadoghq.com/external/requests-toolbelt/requests_toolbelt-1.0.0-py2.py3-none-any.whl#sha256=cccfdd665f0a24fcf4726e690f65639d272bb0637b9b92dfd91a5568ccf6bd06 +requests-unixsocket2 @ https://agent-int-packages.datadoghq.com/external/requests-unixsocket2/requests_unixsocket2-0.4.2-py3-none-any.whl#sha256=701fcd49d74bc0f759bbe45c4dfda0045fd89652948c2b473b1a312214c3770b +requestsexceptions @ https://agent-int-packages.datadoghq.com/external/requestsexceptions/requestsexceptions-1.4.0-py2.py3-none-any.whl#sha256=3083d872b6e07dc5c323563ef37671d992214ad9a32b0ca4a3d7f5500bf38ce3 +rethinkdb @ https://agent-int-packages.datadoghq.com/external/rethinkdb/rethinkdb-2.4.9-py2.py3-none-any.whl#sha256=dd2455b3eab5266d8006eeefeb5eb970a3697dfe6aa18e1ab6e1d99233badfcb +rsa @ https://agent-int-packages.datadoghq.com/external/rsa/rsa-4.9-py3-none-any.whl#sha256=90260d9058e514786967344d0ef75fa8727eed8a7d2e43ce9f4bcf1b536174f7 +s3transfer @ https://agent-int-packages.datadoghq.com/external/s3transfer/s3transfer-0.10.2-py3-none-any.whl#sha256=eca1c20de70a39daee580aef4986996620f365c4e0fda6a86100231d62f1bf69 +securesystemslib @ https://agent-int-packages.datadoghq.com/external/securesystemslib/securesystemslib-0.28.0-py3-none-any.whl#sha256=9e6b9abe36a511d4f52c759069db8f6f650362ba82d6efc7bc7466a458b3f499 +semver @ https://agent-int-packages.datadoghq.com/external/semver/semver-3.0.2-py3-none-any.whl#sha256=b1ea4686fe70b981f85359eda33199d60c53964284e0cfb4977d243e37cf4bf4 +service-identity @ https://agent-int-packages.datadoghq.com/external/service-identity/service_identity-24.1.0-py3-none-any.whl#sha256=a28caf8130c8a5c1c7a6f5293faaf239bbfb7751e4862436920ee6f2616f568a +setuptools @ https://agent-int-packages.datadoghq.com/external/setuptools/setuptools-75.1.0-py3-none-any.whl#sha256=35ab7fd3bcd95e6b7fd704e4a1539513edad446c097797f2985e0e4b960772f2 +simplejson @ https://agent-int-packages.datadoghq.com/external/simplejson/simplejson-3.19.3-cp312-cp312-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl#sha256=7017329ca8d4dca94ad5e59f496e5fc77630aecfc39df381ffc1d37fb6b25832 +six @ https://agent-int-packages.datadoghq.com/external/six/six-1.16.0-py2.py3-none-any.whl#sha256=8abb2f1d86890a2dfb989f9a77cfcfd3e47c2a354b01111771326f8aa26e0254 +snowflake-connector-python @ https://agent-int-packages.datadoghq.com/external/snowflake-connector-python/snowflake_connector_python-3.12.1-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl#sha256=bddc4cdcd991f9538726a7c293d2637bb5aed43db68246e06c92c49a6df2b692 +sortedcontainers @ https://agent-int-packages.datadoghq.com/external/sortedcontainers/sortedcontainers-2.4.0-py2.py3-none-any.whl#sha256=a163dcaede0f1c021485e957a39245190e74249897e2ae4b2aa38595db237ee0 +soupsieve @ https://agent-int-packages.datadoghq.com/external/soupsieve/soupsieve-2.6-py3-none-any.whl#sha256=e72c4ff06e4fb6e4b5a9f0f55fe6e81514581fca1515028625d0f299c602ccc9 +stevedore @ https://agent-int-packages.datadoghq.com/external/stevedore/stevedore-5.3.0-py3-none-any.whl#sha256=1efd34ca08f474dad08d9b19e934a22c68bb6fe416926479ba29e5013bcc8f78 +supervisor @ https://agent-int-packages.datadoghq.com/external/supervisor/supervisor-4.2.5-py2.py3-none-any.whl#sha256=2ecaede32fc25af814696374b79e42644ecaba5c09494c51016ffda9602d0f08 +tomlkit @ https://agent-int-packages.datadoghq.com/external/tomlkit/tomlkit-0.13.2-py3-none-any.whl#sha256=7a974427f6e119197f670fbbbeae7bef749a6c14e793db934baefc1b5f03efde +tuf @ https://agent-int-packages.datadoghq.com/external/tuf/tuf-4.0.0-py3-none-any.whl#sha256=a22ab5fa6daf910b3052929fdce42ccad8a300e5e85715daaff9592aed980f7a +typing-extensions @ https://agent-int-packages.datadoghq.com/external/typing-extensions/typing_extensions-4.12.2-py3-none-any.whl#sha256=04e5ca0351e0f3f85c6853954072df659d0d13fac324d0072316b67d7794700d +tzlocal @ https://agent-int-packages.datadoghq.com/external/tzlocal/tzlocal-5.2-py3-none-any.whl#sha256=49816ef2fe65ea8ac19d19aa7a1ae0551c834303d5014c6d5a62e4cbda8047b8 +uhashring @ https://agent-int-packages.datadoghq.com/external/uhashring/uhashring-2.3-py3-none-any.whl#sha256=7ee8a25ca495a97effad10bd563c83b4054a6d7606d9530757049a04edab9297 +uptime @ https://agent-int-packages.datadoghq.com/built/uptime/uptime-3.0.1-20240917155252-cp312-cp312-manylinux2014_x86_64.manylinux_2_17_x86_64.whl#sha256=d3ea8a4efb33d233338f185e0cc1fd46b04fdc0d11b69d0764a4f2260430a984 +urllib3 @ https://agent-int-packages.datadoghq.com/external/urllib3/urllib3-2.2.3-py3-none-any.whl#sha256=ca899ca043dcb1bafa3e262d73aa25c465bfb49e0bd9dd5d59f1d0acba2f8fac +vertica-python @ https://agent-int-packages.datadoghq.com/external/vertica-python/vertica_python-1.4.0-py3-none-any.whl#sha256=50fecd7687f4b0b9f6dee6e2b35c195af2a4f702ece01bd12e080b51756e000b +websocket-client @ https://agent-int-packages.datadoghq.com/external/websocket-client/websocket_client-1.8.0-py3-none-any.whl#sha256=17b44cc997f5c498e809b22cdf2d9c7a9e71c02c8cc2b6c56e7c2d1239bfa526 +wrapt @ https://agent-int-packages.datadoghq.com/external/wrapt/wrapt-1.16.0-cp312-cp312-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl#sha256=98b5e1f498a8ca1858a1cdbffb023bfd954da4e3fa2c0cb5853d40014557248b +xmltodict @ https://agent-int-packages.datadoghq.com/external/xmltodict/xmltodict-0.13.0-py2.py3-none-any.whl#sha256=aa89e8fd76320154a40d19a0df04a4695fb9dc5ba977cbb68ab3e4eb225e7852 +zipp @ https://agent-int-packages.datadoghq.com/external/zipp/zipp-3.20.2-py3-none-any.whl#sha256=a817ac80d6cf4b23bf7f2828b7cabf326f15a001bea8b1f9b49631780ba28350 diff --git a/.deps/resolved/macos-x86_64_3.12.txt b/.deps/resolved/macos-x86_64_3.12.txt new file mode 100644 index 0000000000000..7e02cd7a4bece --- /dev/null +++ b/.deps/resolved/macos-x86_64_3.12.txt @@ -0,0 +1,128 @@ +annotated-types @ https://agent-int-packages.datadoghq.com/external/annotated-types/annotated_types-0.7.0-py3-none-any.whl#sha256=1f02e8b43a8fbbc3f3e0d4f0f4bfc8131bcb4eebe8849b8e5c773f3a1c582a53 +asn1crypto @ https://agent-int-packages.datadoghq.com/external/asn1crypto/asn1crypto-1.5.1-py2.py3-none-any.whl#sha256=db4e40728b728508912cbb3d44f19ce188f218e9eba635821bb4b68564f8fd67 +attrs @ https://agent-int-packages.datadoghq.com/external/attrs/attrs-24.2.0-py3-none-any.whl#sha256=81921eb96de3191c8258c199618104dd27ac608d9366f5e35d011eae1867ede2 +aws-requests-auth @ https://agent-int-packages.datadoghq.com/external/aws-requests-auth/aws_requests_auth-0.4.3-py2.py3-none-any.whl#sha256=646bc37d62140ea1c709d20148f5d43197e6bd2d63909eb36fa4bb2345759977 +azure-core @ https://agent-int-packages.datadoghq.com/external/azure-core/azure_core-1.31.0-py3-none-any.whl#sha256=22954de3777e0250029360ef31d80448ef1be13b80a459bff80ba7073379e2cd +azure-identity @ https://agent-int-packages.datadoghq.com/external/azure-identity/azure_identity-1.17.1-py3-none-any.whl#sha256=db8d59c183b680e763722bfe8ebc45930e6c57df510620985939f7f3191e0382 +bcrypt @ https://agent-int-packages.datadoghq.com/external/bcrypt/bcrypt-4.2.0-cp39-abi3-macosx_10_12_universal2.whl#sha256=c52aac18ea1f4a4f65963ea4f9530c306b56ccd0c6f8c8da0c06976e34a6e841 +beautifulsoup4 @ https://agent-int-packages.datadoghq.com/external/beautifulsoup4/beautifulsoup4-4.12.3-py3-none-any.whl#sha256=b80878c9f40111313e55da8ba20bdba06d8fa3969fc68304167741bbf9e082ed +binary @ https://agent-int-packages.datadoghq.com/external/binary/binary-1.0.0-py2.py3-none-any.whl#sha256=e1b61f3a5c002717d1a28e4d9d2dc8acbc9d6b12baf7b1e4ab25d743da97e323 +boto3 @ https://agent-int-packages.datadoghq.com/external/boto3/boto3-1.35.10-py3-none-any.whl#sha256=add26dd58e076dfd387013da4704716d5cff215cf14f6d4347c4b9b7fc1f0b8e +botocore @ https://agent-int-packages.datadoghq.com/external/botocore/botocore-1.35.10-py3-none-any.whl#sha256=0d96d023b9b0cea99a0a428a431d011329d3a958730aee6ed6a6fec5d9bfbc03 +bytecode @ https://agent-int-packages.datadoghq.com/external/bytecode/bytecode-0.15.1-py3-none-any.whl#sha256=0a1dc340cac823cff605609b8b214f7f9bf80418c6b9e0fc8c6db1793c27137d +cachetools @ https://agent-int-packages.datadoghq.com/external/cachetools/cachetools-5.5.0-py3-none-any.whl#sha256=02134e8439cdc2ffb62023ce1debca2944c3f289d66bb17ead3ab3dede74b292 +cattrs @ https://agent-int-packages.datadoghq.com/external/cattrs/cattrs-24.1.1-py3-none-any.whl#sha256=ec8ce8fdc725de9d07547cd616f968670687c6fa7a2e263b088370c46d834d97 +certifi @ https://agent-int-packages.datadoghq.com/external/certifi/certifi-2024.8.30-py3-none-any.whl#sha256=922820b53db7a7257ffbda3f597266d435245903d80737e34f8a45ff3e3230d8 +cffi @ https://agent-int-packages.datadoghq.com/external/cffi/cffi-1.17.1-cp312-cp312-macosx_10_9_x86_64.whl#sha256=805b4371bf7197c329fcb3ead37e710d1bca9da5d583f5073b799d5c5bd1eee4 +charset-normalizer @ https://agent-int-packages.datadoghq.com/external/charset-normalizer/charset_normalizer-3.3.2-cp312-cp312-macosx_10_9_x86_64.whl#sha256=ddbb2551d7e0102e7252db79ba445cdab71b26640817ab1e3e3648dad515003b +clickhouse-cityhash @ https://agent-int-packages.datadoghq.com/external/clickhouse-cityhash/clickhouse_cityhash-1.0.2.4-cp312-cp312-macosx_10_9_x86_64.whl#sha256=261fc1b0bf349de66b2d9e3d367879a561b516ca8e54e85e0c27b7c1a4f639b4 +clickhouse-driver @ https://agent-int-packages.datadoghq.com/external/clickhouse-driver/clickhouse_driver-0.2.9-cp312-cp312-macosx_10_9_x86_64.whl#sha256=fcb2fd00e58650ae206a6d5dbc83117240e622471aa5124733fbf2805eb8bda0 +cm-client @ https://agent-int-packages.datadoghq.com/built/cm-client/cm_client-45.0.4-20240402154932-py3-none-macosx_10_12_universal2.whl#sha256=aba3c1683ef1b2099933e030464d29b3ad1c206784ebd15d8a7147ecd6ba24e1 +confluent-kafka @ https://agent-int-packages.datadoghq.com/built/confluent-kafka/confluent_kafka-2.5.0-20240917155258-cp312-cp312-macosx_10_13_universal2.whl#sha256=b8d2785849784cb3ac8a33e0067bc0c6e4850d5dc4fe8bdf47973dcea5c94316 +cryptography @ https://agent-int-packages.datadoghq.com/external/cryptography/cryptography-43.0.0-cp39-abi3-macosx_10_9_universal2.whl#sha256=7b3f5fe74a5ca32d4d0f302ffe6680fcc5c28f8ef0dc0ae8f40c0f3a1b4fca66 +ddsketch @ https://agent-int-packages.datadoghq.com/external/ddsketch/ddsketch-3.0.1-py3-none-any.whl#sha256=6d047b455fe2837c43d366ff1ae6ba0c3166e15499de8688437a75cea914224e +ddtrace @ https://agent-int-packages.datadoghq.com/external/ddtrace/ddtrace-2.10.6-cp312-cp312-macosx_12_0_x86_64.whl#sha256=401f77b0564c3f990b58b9f21055331ca9efcdfa06dfa6ccff13cf21f8329ba5 +decorator @ https://agent-int-packages.datadoghq.com/external/decorator/decorator-5.1.1-py3-none-any.whl#sha256=b8c3f85900b9dc423225913c5aace94729fe1fa9763b38939a95226f02d37186 +deprecated @ https://agent-int-packages.datadoghq.com/external/deprecated/Deprecated-1.2.14-py2.py3-none-any.whl#sha256=6fac8b097794a90302bdbb17b9b815e732d3c4720583ff1b198499d78470466c +dnspython @ https://agent-int-packages.datadoghq.com/external/dnspython/dnspython-2.6.1-py3-none-any.whl#sha256=5ef3b9680161f6fa89daf8ad451b5f1a33b18ae8a1c6778cdf4b43f08c0a6e50 +dogpile-cache @ https://agent-int-packages.datadoghq.com/external/dogpile-cache/dogpile.cache-1.3.3-py3-none-any.whl#sha256=5e211c4902ebdf88c678d268e22454b41e68071632daa9402d8ee24e825ed8ca +envier @ https://agent-int-packages.datadoghq.com/external/envier/envier-0.5.2-py3-none-any.whl#sha256=65099cf3aa9b3b3b4b92db2f7d29e2910672e085b76f7e587d2167561a834add +filelock @ https://agent-int-packages.datadoghq.com/external/filelock/filelock-3.16.0-py3-none-any.whl#sha256=f6ed4c963184f4c84dd5557ce8fece759a3724b37b80c6c4f20a2f63a4dc6609 +foundationdb @ https://agent-int-packages.datadoghq.com/built/foundationdb/foundationdb-6.3.24-20240402154934-py3-none-macosx_10_12_universal2.whl#sha256=14259f824080062cc890965747597ff00a9d6c76a1eb926673fed68a45860ccd +google-auth @ https://agent-int-packages.datadoghq.com/external/google-auth/google_auth-2.34.0-py2.py3-none-any.whl#sha256=72fd4733b80b6d777dcde515628a9eb4a577339437012874ea286bca7261ee65 +gssapi @ https://agent-int-packages.datadoghq.com/external/gssapi/gssapi-1.8.3-cp312-cp312-macosx_10_9_x86_64.whl#sha256=19c373b3ba63ce19cd3163aa1495635e3d01b0de6cc4ff1126095eded1df6e01 +hazelcast-python-client @ https://agent-int-packages.datadoghq.com/external/hazelcast-python-client/hazelcast_python_client-5.4.0-py3-none-any.whl#sha256=16195cd58feb2dd3be1594d08d42527ae00797548a6a9d6a601aae2e8514ff5f +idna @ https://agent-int-packages.datadoghq.com/external/idna/idna-3.10-py3-none-any.whl#sha256=946d195a0d259cbba61165e88e65941f16e9b36ea6ddb97f00452bae8b1287d3 +importlib-metadata @ https://agent-int-packages.datadoghq.com/external/importlib-metadata/importlib_metadata-8.4.0-py3-none-any.whl#sha256=66f342cc6ac9818fc6ff340576acd24d65ba0b3efabb2b4ac08b598965a4a2f1 +in-toto @ https://agent-int-packages.datadoghq.com/external/in-toto/in_toto-2.0.0-py3-none-any.whl#sha256=eaf71733cdd5b2309b58b24169c13c24701da72baa43b7432d1dffc308e7bf32 +iso8601 @ https://agent-int-packages.datadoghq.com/external/iso8601/iso8601-2.1.0-py3-none-any.whl#sha256=aac4145c4dcb66ad8b648a02830f5e2ff6c24af20f4f482689be402db2429242 +jellyfish @ https://agent-int-packages.datadoghq.com/external/jellyfish/jellyfish-1.1.0-cp312-cp312-macosx_10_12_x86_64.whl#sha256=828a7000d369cbd4d812b88510c01fdab20b73dc54c63cdbe03bdff67ab362d0 +jmespath @ https://agent-int-packages.datadoghq.com/external/jmespath/jmespath-1.0.1-py3-none-any.whl#sha256=02e2e4cc71b5bcab88332eebf907519190dd9e6e82107fa7f83b1003a6252980 +jsonpatch @ https://agent-int-packages.datadoghq.com/external/jsonpatch/jsonpatch-1.33-py2.py3-none-any.whl#sha256=0ae28c0cd062bbd8b8ecc26d7d164fbbea9652a1a3693f3b956c1eae5145dade +jsonpointer @ https://agent-int-packages.datadoghq.com/external/jsonpointer/jsonpointer-3.0.0-py2.py3-none-any.whl#sha256=13e088adc14fca8b6aa8177c044e12701e6ad4b28ff10e65f2267a90109c9942 +keystoneauth1 @ https://agent-int-packages.datadoghq.com/external/keystoneauth1/keystoneauth1-5.8.0-py3-none-any.whl#sha256=e69dff80c509ab64d4de4494658d914e81f26af720828dc584ceee74ecd666d9 +krb5 @ https://agent-int-packages.datadoghq.com/external/krb5/krb5-0.6.0-cp312-cp312-macosx_10_9_x86_64.whl#sha256=8f1419ad4c4052f687aafd8ed0497de3e10905d7c4bfd0645d02fd63e24ab4b8 +kubernetes @ https://agent-int-packages.datadoghq.com/external/kubernetes/kubernetes-30.1.0-py2.py3-none-any.whl#sha256=e212e8b7579031dd2e512168b617373bc1e03888d41ac4e04039240a292d478d +ldap3 @ https://agent-int-packages.datadoghq.com/external/ldap3/ldap3-2.9.1-py2.py3-none-any.whl#sha256=5869596fc4948797020d3f03b7939da938778a0f9e2009f7a072ccf92b8e8d70 +lxml @ https://agent-int-packages.datadoghq.com/external/lxml/lxml-4.9.4-cp312-cp312-macosx_11_0_universal2.whl#sha256=dbcb2dc07308453db428a95a4d03259bd8caea97d7f0776842299f2d00c72fc8 +lz4 @ https://agent-int-packages.datadoghq.com/external/lz4/lz4-4.3.3-cp312-cp312-macosx_10_9_x86_64.whl#sha256=e36cd7b9d4d920d3bfc2369840da506fa68258f7bb176b8743189793c055e43d +mmh3 @ https://agent-int-packages.datadoghq.com/external/mmh3/mmh3-4.1.0-cp312-cp312-macosx_10_9_x86_64.whl#sha256=d6af3e2287644b2b08b5924ed3a88c97b87b44ad08e79ca9f93d3470a54a41c5 +msal @ https://agent-int-packages.datadoghq.com/external/msal/msal-1.31.0-py3-none-any.whl#sha256=96bc37cff82ebe4b160d5fc0f1196f6ca8b50e274ecd0ec5bf69c438514086e7 +msal-extensions @ https://agent-int-packages.datadoghq.com/external/msal-extensions/msal_extensions-1.2.0-py3-none-any.whl#sha256=cf5ba83a2113fa6dc011a254a72f1c223c88d7dfad74cc30617c4679a417704d +netifaces @ https://agent-int-packages.datadoghq.com/built/netifaces/netifaces-0.11.0-20240917155259-cp312-cp312-macosx_10_13_universal2.whl#sha256=e36f29cbdbd18847c6ee2d512ef10114d086f32e56e1654fd0230552dde5c24e +oauthlib @ https://agent-int-packages.datadoghq.com/external/oauthlib/oauthlib-3.2.2-py3-none-any.whl#sha256=8139f29aac13e25d502680e9e19963e83f16838d48a0d71c287fe40e7067fbca +openstacksdk @ https://agent-int-packages.datadoghq.com/external/openstacksdk/openstacksdk-3.3.0-py3-none-any.whl#sha256=e6d4121b87354984caf0e3c032e2ebf4d4440374f86c81c27ec52ca5df359157 +opentelemetry-api @ https://agent-int-packages.datadoghq.com/external/opentelemetry-api/opentelemetry_api-1.27.0-py3-none-any.whl#sha256=953d5871815e7c30c81b56d910c707588000fff7a3ca1c73e6531911d53065e7 +orjson @ https://agent-int-packages.datadoghq.com/external/orjson/orjson-3.10.7-cp312-cp312-macosx_10_15_x86_64.macosx_11_0_arm64.macosx_10_15_universal2.whl#sha256=44a96f2d4c3af51bfac6bc4ef7b182aa33f2f054fd7f34cc0ee9a320d051d41f +os-service-types @ https://agent-int-packages.datadoghq.com/external/os-service-types/os_service_types-1.7.0-py2.py3-none-any.whl#sha256=0505c72205690910077fb72b88f2a1f07533c8d39f2fe75b29583481764965d6 +packaging @ https://agent-int-packages.datadoghq.com/external/packaging/packaging-24.1-py3-none-any.whl#sha256=5b8f2217dbdbd2f7f384c41c628544e6d52f2d0f53c6d0c3ea61aa5d1d7ff124 +paramiko @ https://agent-int-packages.datadoghq.com/external/paramiko/paramiko-3.4.1-py3-none-any.whl#sha256=8e49fd2f82f84acf7ffd57c64311aa2b30e575370dc23bdb375b10262f7eac32 +pathspec @ https://agent-int-packages.datadoghq.com/external/pathspec/pathspec-0.12.1-py3-none-any.whl#sha256=a0d503e138a4c123b27490a4f7beda6a01c6f288df0e4a8b79c7eb0dc7b4cc08 +pbr @ https://agent-int-packages.datadoghq.com/external/pbr/pbr-6.1.0-py2.py3-none-any.whl#sha256=a776ae228892d8013649c0aeccbb3d5f99ee15e005a4cbb7e61d55a067b28a2a +platformdirs @ https://agent-int-packages.datadoghq.com/external/platformdirs/platformdirs-4.3.3-py3-none-any.whl#sha256=50a5450e2e84f44539718293cbb1da0a0885c9d14adf21b77bae4e66fc99d9b5 +ply @ https://agent-int-packages.datadoghq.com/external/ply/ply-3.11-py2.py3-none-any.whl#sha256=096f9b8350b65ebd2fd1346b12452efe5b9607f7482813ffca50c22722a807ce +portalocker @ https://agent-int-packages.datadoghq.com/external/portalocker/portalocker-2.10.1-py3-none-any.whl#sha256=53a5984ebc86a025552264b459b46a2086e269b21823cb572f8f28ee759e45bf +prometheus-client @ https://agent-int-packages.datadoghq.com/external/prometheus-client/prometheus_client-0.20.0-py3-none-any.whl#sha256=cde524a85bce83ca359cc837f28b8c0db5cac7aa653a588fd7e84ba061c329e7 +protobuf @ https://agent-int-packages.datadoghq.com/external/protobuf/protobuf-5.27.3-cp38-abi3-macosx_10_9_universal2.whl#sha256=68248c60d53f6168f565a8c76dc58ba4fa2ade31c2d1ebdae6d80f969cdc2d4f +psutil @ https://agent-int-packages.datadoghq.com/external/psutil/psutil-5.9.6-cp36-abi3-macosx_10_9_x86_64.whl#sha256=c69596f9fc2f8acd574a12d5f8b7b1ba3765a641ea5d60fb4736bf3c08a8214a +psycopg2-binary @ https://agent-int-packages.datadoghq.com/external/psycopg2-binary/psycopg2_binary-2.9.9-cp312-cp312-macosx_10_9_x86_64.whl#sha256=8532fd6e6e2dc57bcb3bc90b079c60de896d2128c5d9d6f24a63875a95a088cf +pyasn1 @ https://agent-int-packages.datadoghq.com/external/pyasn1/pyasn1-0.4.8-py2.py3-none-any.whl#sha256=39c7e2ec30515947ff4e87fb6f456dfc6e84857d34be479c9d4a4ba4bf46aa5d +pyasn1-modules @ https://agent-int-packages.datadoghq.com/external/pyasn1-modules/pyasn1_modules-0.4.1-py3-none-any.whl#sha256=49bfa96b45a292b711e986f222502c1c9a5e1f4e568fc30e2574a6c7d07838fd +pyasyncore @ https://agent-int-packages.datadoghq.com/external/pyasyncore/pyasyncore-1.0.4-py3-none-any.whl#sha256=9e5f6dc9dc057c56370b7a5cdb4c4670fd4b0556de2913ed1f428cd6a5366895 +pycparser @ https://agent-int-packages.datadoghq.com/external/pycparser/pycparser-2.22-py3-none-any.whl#sha256=c3702b6d3dd8c7abc1afa565d7e63d53a1d0bd86cdc24edd75470f4de499cfcc +pycryptodomex @ https://agent-int-packages.datadoghq.com/external/pycryptodomex/pycryptodomex-3.20.0-cp35-abi3-macosx_10_9_x86_64.whl#sha256=82ee7696ed8eb9a82c7037f32ba9b7c59e51dda6f105b39f043b6ef293989cb3 +pydantic @ https://agent-int-packages.datadoghq.com/external/pydantic/pydantic-2.8.2-py3-none-any.whl#sha256=73ee9fddd406dc318b885c7a2eab8a6472b68b8fb5ba8150949fc3db939f23c8 +pydantic-core @ https://agent-int-packages.datadoghq.com/external/pydantic-core/pydantic_core-2.20.1-cp312-cp312-macosx_10_12_x86_64.whl#sha256=595ba5be69b35777474fa07f80fc260ea71255656191adb22a8c53aba4479231 +pyjwt @ https://agent-int-packages.datadoghq.com/external/pyjwt/PyJWT-2.9.0-py3-none-any.whl#sha256=3b02fb0f44517787776cf48f2ae25d8e14f300e6d7545a4315cee571a415e850 +pymongo @ https://agent-int-packages.datadoghq.com/external/pymongo/pymongo-4.8.0-cp312-cp312-macosx_10_9_x86_64.whl#sha256=e6a720a3d22b54183352dc65f08cd1547204d263e0651b213a0a2e577e838526 +pymqi @ https://agent-int-packages.datadoghq.com/built/pymqi/pymqi-1.12.10-20240917155259-cp312-cp312-macosx_10_13_universal2.whl#sha256=63625dcc0c82f16ad6f8b5a2b60e4c5a3b51caa541a0a4e9805ef7bafbce5d95 +pymysql @ https://agent-int-packages.datadoghq.com/external/pymysql/PyMySQL-1.1.1-py3-none-any.whl#sha256=4de15da4c61dc132f4fb9ab763063e693d521a80fd0e87943b9a453dd4c19d6c +pynacl @ https://agent-int-packages.datadoghq.com/external/pynacl/PyNaCl-1.5.0-cp36-abi3-macosx_10_10_universal2.whl#sha256=401002a4aaa07c9414132aaed7f6836ff98f59277a234704ff66878c2ee4a0d1 +pyodbc @ https://agent-int-packages.datadoghq.com/external/pyodbc/pyodbc-5.1.0-cp312-cp312-macosx_10_9_x86_64.whl#sha256=d3d9cc4af703c4817b6e604315910b0cf5dcb68056d52b25ca072dd59c52dcbc +pyopenssl @ https://agent-int-packages.datadoghq.com/external/pyopenssl/pyOpenSSL-24.2.1-py3-none-any.whl#sha256=967d5719b12b243588573f39b0c677637145c7a1ffedcd495a487e58177fbb8d +pysmi @ https://agent-int-packages.datadoghq.com/external/pysmi/pysmi-1.2.1-py3-none-any.whl#sha256=d97c60de9f81d33ab2899124d95a94fa7fefacc86ab6e00cbfec543a073e6d33 +pysnmp @ https://agent-int-packages.datadoghq.com/external/pysnmp/pysnmp-5.1.0-py3-none-any.whl#sha256=375a8adfc6820faf24ace6761a6d20544e60580d714ff7266df272850c39b439 +pysnmp-mibs @ https://agent-int-packages.datadoghq.com/external/pysnmp-mibs/pysnmp_mibs-0.1.6-py2.py3-none-any.whl#sha256=5e153ebe8e767c07940cea435f866c623ff6b2376155c7da75085b08d3774d48 +pysnmpcrypto @ https://agent-int-packages.datadoghq.com/external/pysnmpcrypto/pysnmpcrypto-0.0.4-py2.py3-none-any.whl#sha256=5889733caa030f45d9e03ea9d6370fb06426a8cb7f839aabbcdde33c6f634679 +pysocks @ https://agent-int-packages.datadoghq.com/external/pysocks/PySocks-1.7.1-py3-none-any.whl#sha256=2725bd0a9925919b9b51739eea5f9e2bae91e83288108a9ad338b2e3a4435ee5 +pyspnego @ https://agent-int-packages.datadoghq.com/external/pyspnego/pyspnego-0.11.1-py3-none-any.whl#sha256=129a4294f2c4d681d5875240ef87accc6f1d921e8983737fb0b59642b397951e +python-binary-memcached @ https://agent-int-packages.datadoghq.com/external/python-binary-memcached/python_binary_memcached-0.31.2-py3-none-any.whl#sha256=e5b93d54429e835cab7d5b33988649f9748344aa49adaed8eed94b37e714d562 +python-dateutil @ https://agent-int-packages.datadoghq.com/external/python-dateutil/python_dateutil-2.9.0.post0-py2.py3-none-any.whl#sha256=a8b2bc7bffae282281c8140a97d3aa9c14da0b136dfe83f850eea9a5f7470427 +python3-gearman @ https://agent-int-packages.datadoghq.com/external/python3-gearman/python3_gearman-0.1.0-py3-none-any.whl#sha256=4a5808d3a0bfc6c243548ad57e7aab4bee62c9cba2b1c3a860fdd292d46a112d +pytz @ https://agent-int-packages.datadoghq.com/external/pytz/pytz-2024.2-py2.py3-none-any.whl#sha256=31c7c1817eb7fae7ca4b8c7ee50c72f93aa2dd863de768e1ef4245d426aa0725 +pyvmomi @ https://agent-int-packages.datadoghq.com/built/pyvmomi/pyvmomi-8.0.3.0.1-20240702172113-py2.py3-none-macosx_10_12_universal2.whl#sha256=e173daf28895975b57850fef301837f24fba59dd8ff1d931795732e9be281d57 +pyyaml @ https://agent-int-packages.datadoghq.com/external/pyyaml/PyYAML-6.0.2-cp312-cp312-macosx_10_9_x86_64.whl#sha256=c70c95198c015b85feafc136515252a261a84561b7b1d51e3384e0655ddf25ab +redis @ https://agent-int-packages.datadoghq.com/external/redis/redis-5.0.8-py3-none-any.whl#sha256=56134ee08ea909106090934adc36f65c9bcbbaecea5b21ba704ba6fb561f8eb4 +requests @ https://agent-int-packages.datadoghq.com/external/requests/requests-2.32.3-py3-none-any.whl#sha256=70761cfe03c773ceb22aa2f671b4757976145175cdfca038c02654d061d6dcc6 +requests-kerberos @ https://agent-int-packages.datadoghq.com/external/requests-kerberos/requests_kerberos-0.15.0-py2.py3-none-any.whl#sha256=ba9b0980b8489c93bfb13854fd118834e576d6700bfea3745cb2e62278cd16a6 +requests-ntlm @ https://agent-int-packages.datadoghq.com/external/requests-ntlm/requests_ntlm-1.3.0-py3-none-any.whl#sha256=4c7534a7d0e482bb0928531d621be4b2c74ace437e88c5a357ceb7452d25a510 +requests-oauthlib @ https://agent-int-packages.datadoghq.com/external/requests-oauthlib/requests_oauthlib-2.0.0-py2.py3-none-any.whl#sha256=7dd8a5c40426b779b0868c404bdef9768deccf22749cde15852df527e6269b36 +requests-toolbelt @ https://agent-int-packages.datadoghq.com/external/requests-toolbelt/requests_toolbelt-1.0.0-py2.py3-none-any.whl#sha256=cccfdd665f0a24fcf4726e690f65639d272bb0637b9b92dfd91a5568ccf6bd06 +requests-unixsocket2 @ https://agent-int-packages.datadoghq.com/external/requests-unixsocket2/requests_unixsocket2-0.4.2-py3-none-any.whl#sha256=701fcd49d74bc0f759bbe45c4dfda0045fd89652948c2b473b1a312214c3770b +requestsexceptions @ https://agent-int-packages.datadoghq.com/external/requestsexceptions/requestsexceptions-1.4.0-py2.py3-none-any.whl#sha256=3083d872b6e07dc5c323563ef37671d992214ad9a32b0ca4a3d7f5500bf38ce3 +rethinkdb @ https://agent-int-packages.datadoghq.com/external/rethinkdb/rethinkdb-2.4.9-py2.py3-none-any.whl#sha256=dd2455b3eab5266d8006eeefeb5eb970a3697dfe6aa18e1ab6e1d99233badfcb +rsa @ https://agent-int-packages.datadoghq.com/external/rsa/rsa-4.9-py3-none-any.whl#sha256=90260d9058e514786967344d0ef75fa8727eed8a7d2e43ce9f4bcf1b536174f7 +s3transfer @ https://agent-int-packages.datadoghq.com/external/s3transfer/s3transfer-0.10.2-py3-none-any.whl#sha256=eca1c20de70a39daee580aef4986996620f365c4e0fda6a86100231d62f1bf69 +securesystemslib @ https://agent-int-packages.datadoghq.com/external/securesystemslib/securesystemslib-0.28.0-py3-none-any.whl#sha256=9e6b9abe36a511d4f52c759069db8f6f650362ba82d6efc7bc7466a458b3f499 +semver @ https://agent-int-packages.datadoghq.com/external/semver/semver-3.0.2-py3-none-any.whl#sha256=b1ea4686fe70b981f85359eda33199d60c53964284e0cfb4977d243e37cf4bf4 +service-identity @ https://agent-int-packages.datadoghq.com/external/service-identity/service_identity-24.1.0-py3-none-any.whl#sha256=a28caf8130c8a5c1c7a6f5293faaf239bbfb7751e4862436920ee6f2616f568a +setuptools @ https://agent-int-packages.datadoghq.com/external/setuptools/setuptools-75.1.0-py3-none-any.whl#sha256=35ab7fd3bcd95e6b7fd704e4a1539513edad446c097797f2985e0e4b960772f2 +simplejson @ https://agent-int-packages.datadoghq.com/external/simplejson/simplejson-3.19.3-cp312-cp312-macosx_10_9_x86_64.whl#sha256=6ef9383c5e05f445be60f1735c1816163c874c0b1ede8bb4390aff2ced34f333 +six @ https://agent-int-packages.datadoghq.com/external/six/six-1.16.0-py2.py3-none-any.whl#sha256=8abb2f1d86890a2dfb989f9a77cfcfd3e47c2a354b01111771326f8aa26e0254 +snowflake-connector-python @ https://agent-int-packages.datadoghq.com/external/snowflake-connector-python/snowflake_connector_python-3.12.1-cp312-cp312-macosx_11_0_x86_64.whl#sha256=dadd262196cce0132ca7e766f055e00c00497a88fdf83fd48143eb4a469a4527 +sortedcontainers @ https://agent-int-packages.datadoghq.com/external/sortedcontainers/sortedcontainers-2.4.0-py2.py3-none-any.whl#sha256=a163dcaede0f1c021485e957a39245190e74249897e2ae4b2aa38595db237ee0 +soupsieve @ https://agent-int-packages.datadoghq.com/external/soupsieve/soupsieve-2.6-py3-none-any.whl#sha256=e72c4ff06e4fb6e4b5a9f0f55fe6e81514581fca1515028625d0f299c602ccc9 +stevedore @ https://agent-int-packages.datadoghq.com/external/stevedore/stevedore-5.3.0-py3-none-any.whl#sha256=1efd34ca08f474dad08d9b19e934a22c68bb6fe416926479ba29e5013bcc8f78 +supervisor @ https://agent-int-packages.datadoghq.com/external/supervisor/supervisor-4.2.5-py2.py3-none-any.whl#sha256=2ecaede32fc25af814696374b79e42644ecaba5c09494c51016ffda9602d0f08 +tomlkit @ https://agent-int-packages.datadoghq.com/external/tomlkit/tomlkit-0.13.2-py3-none-any.whl#sha256=7a974427f6e119197f670fbbbeae7bef749a6c14e793db934baefc1b5f03efde +tuf @ https://agent-int-packages.datadoghq.com/external/tuf/tuf-4.0.0-py3-none-any.whl#sha256=a22ab5fa6daf910b3052929fdce42ccad8a300e5e85715daaff9592aed980f7a +typing-extensions @ https://agent-int-packages.datadoghq.com/external/typing-extensions/typing_extensions-4.12.2-py3-none-any.whl#sha256=04e5ca0351e0f3f85c6853954072df659d0d13fac324d0072316b67d7794700d +tzlocal @ https://agent-int-packages.datadoghq.com/external/tzlocal/tzlocal-5.2-py3-none-any.whl#sha256=49816ef2fe65ea8ac19d19aa7a1ae0551c834303d5014c6d5a62e4cbda8047b8 +uhashring @ https://agent-int-packages.datadoghq.com/external/uhashring/uhashring-2.3-py3-none-any.whl#sha256=7ee8a25ca495a97effad10bd563c83b4054a6d7606d9530757049a04edab9297 +uptime @ https://agent-int-packages.datadoghq.com/built/uptime/uptime-3.0.1-20240917155300-cp312-cp312-macosx_10_13_universal2.whl#sha256=40c99fc8803f14b345a3bbfa38fe3ffd81cf4204491b6c31171b73cfec1431e9 +urllib3 @ https://agent-int-packages.datadoghq.com/external/urllib3/urllib3-2.2.3-py3-none-any.whl#sha256=ca899ca043dcb1bafa3e262d73aa25c465bfb49e0bd9dd5d59f1d0acba2f8fac +vertica-python @ https://agent-int-packages.datadoghq.com/external/vertica-python/vertica_python-1.4.0-py3-none-any.whl#sha256=50fecd7687f4b0b9f6dee6e2b35c195af2a4f702ece01bd12e080b51756e000b +websocket-client @ https://agent-int-packages.datadoghq.com/external/websocket-client/websocket_client-1.8.0-py3-none-any.whl#sha256=17b44cc997f5c498e809b22cdf2d9c7a9e71c02c8cc2b6c56e7c2d1239bfa526 +wrapt @ https://agent-int-packages.datadoghq.com/external/wrapt/wrapt-1.16.0-cp312-cp312-macosx_10_9_x86_64.whl#sha256=5eb404d89131ec9b4f748fa5cfb5346802e5ee8836f57d516576e61f304f3b7b +xmltodict @ https://agent-int-packages.datadoghq.com/external/xmltodict/xmltodict-0.13.0-py2.py3-none-any.whl#sha256=aa89e8fd76320154a40d19a0df04a4695fb9dc5ba977cbb68ab3e4eb225e7852 +zipp @ https://agent-int-packages.datadoghq.com/external/zipp/zipp-3.20.2-py3-none-any.whl#sha256=a817ac80d6cf4b23bf7f2828b7cabf326f15a001bea8b1f9b49631780ba28350 diff --git a/.deps/resolved/windows-x86_64_3.12.txt b/.deps/resolved/windows-x86_64_3.12.txt new file mode 100644 index 0000000000000..134d120f0b18f --- /dev/null +++ b/.deps/resolved/windows-x86_64_3.12.txt @@ -0,0 +1,126 @@ +annotated-types @ https://agent-int-packages.datadoghq.com/external/annotated-types/annotated_types-0.7.0-py3-none-any.whl#sha256=1f02e8b43a8fbbc3f3e0d4f0f4bfc8131bcb4eebe8849b8e5c773f3a1c582a53 +asn1crypto @ https://agent-int-packages.datadoghq.com/external/asn1crypto/asn1crypto-1.5.1-py2.py3-none-any.whl#sha256=db4e40728b728508912cbb3d44f19ce188f218e9eba635821bb4b68564f8fd67 +attrs @ https://agent-int-packages.datadoghq.com/external/attrs/attrs-24.2.0-py3-none-any.whl#sha256=81921eb96de3191c8258c199618104dd27ac608d9366f5e35d011eae1867ede2 +aws-requests-auth @ https://agent-int-packages.datadoghq.com/external/aws-requests-auth/aws_requests_auth-0.4.3-py2.py3-none-any.whl#sha256=646bc37d62140ea1c709d20148f5d43197e6bd2d63909eb36fa4bb2345759977 +azure-core @ https://agent-int-packages.datadoghq.com/external/azure-core/azure_core-1.31.0-py3-none-any.whl#sha256=22954de3777e0250029360ef31d80448ef1be13b80a459bff80ba7073379e2cd +azure-identity @ https://agent-int-packages.datadoghq.com/external/azure-identity/azure_identity-1.17.1-py3-none-any.whl#sha256=db8d59c183b680e763722bfe8ebc45930e6c57df510620985939f7f3191e0382 +bcrypt @ https://agent-int-packages.datadoghq.com/external/bcrypt/bcrypt-4.2.0-cp39-abi3-win_amd64.whl#sha256=61ed14326ee023917ecd093ee6ef422a72f3aec6f07e21ea5f10622b735538a9 +beautifulsoup4 @ https://agent-int-packages.datadoghq.com/external/beautifulsoup4/beautifulsoup4-4.12.3-py3-none-any.whl#sha256=b80878c9f40111313e55da8ba20bdba06d8fa3969fc68304167741bbf9e082ed +binary @ https://agent-int-packages.datadoghq.com/external/binary/binary-1.0.0-py2.py3-none-any.whl#sha256=e1b61f3a5c002717d1a28e4d9d2dc8acbc9d6b12baf7b1e4ab25d743da97e323 +boto3 @ https://agent-int-packages.datadoghq.com/external/boto3/boto3-1.35.10-py3-none-any.whl#sha256=add26dd58e076dfd387013da4704716d5cff215cf14f6d4347c4b9b7fc1f0b8e +botocore @ https://agent-int-packages.datadoghq.com/external/botocore/botocore-1.35.10-py3-none-any.whl#sha256=0d96d023b9b0cea99a0a428a431d011329d3a958730aee6ed6a6fec5d9bfbc03 +bytecode @ https://agent-int-packages.datadoghq.com/external/bytecode/bytecode-0.15.1-py3-none-any.whl#sha256=0a1dc340cac823cff605609b8b214f7f9bf80418c6b9e0fc8c6db1793c27137d +cachetools @ https://agent-int-packages.datadoghq.com/external/cachetools/cachetools-5.5.0-py3-none-any.whl#sha256=02134e8439cdc2ffb62023ce1debca2944c3f289d66bb17ead3ab3dede74b292 +cattrs @ https://agent-int-packages.datadoghq.com/external/cattrs/cattrs-24.1.1-py3-none-any.whl#sha256=ec8ce8fdc725de9d07547cd616f968670687c6fa7a2e263b088370c46d834d97 +certifi @ https://agent-int-packages.datadoghq.com/external/certifi/certifi-2024.8.30-py3-none-any.whl#sha256=922820b53db7a7257ffbda3f597266d435245903d80737e34f8a45ff3e3230d8 +cffi @ https://agent-int-packages.datadoghq.com/external/cffi/cffi-1.17.1-cp312-cp312-win_amd64.whl#sha256=51392eae71afec0d0c8fb1a53b204dbb3bcabcb3c9b807eedf3e1e6ccf2de903 +charset-normalizer @ https://agent-int-packages.datadoghq.com/external/charset-normalizer/charset_normalizer-3.3.2-cp312-cp312-win_amd64.whl#sha256=96b02a3dc4381e5494fad39be677abcb5e6634bf7b4fa83a6dd3112607547001 +clickhouse-cityhash @ https://agent-int-packages.datadoghq.com/external/clickhouse-cityhash/clickhouse_cityhash-1.0.2.4-cp312-cp312-win_amd64.whl#sha256=0409917be29f5ad80a6772712fce954b5e81450555636e8523290ee9740a2dbb +clickhouse-driver @ https://agent-int-packages.datadoghq.com/external/clickhouse-driver/clickhouse_driver-0.2.9-cp312-cp312-win_amd64.whl#sha256=de6624e28eeffd01668803d28ae89e3d4e359b1bff8b60e4933e1cb3c6f86f18 +cm-client @ https://agent-int-packages.datadoghq.com/built/cm-client/cm_client-45.0.4-20240402154627-py3-none-win_amd64.whl#sha256=1743b32a221d2a0804b4e425ffd53468e8f1754da217fe1e7bd9ff7800fd90f8 +confluent-kafka @ https://agent-int-packages.datadoghq.com/external/confluent-kafka/confluent_kafka-2.5.0-cp312-cp312-win_amd64.whl#sha256=d668b5c426af595271bf6fce2917a6c3a15453656077a59db85f440958b5ccc2 +cryptography @ https://agent-int-packages.datadoghq.com/external/cryptography/cryptography-43.0.0-cp39-abi3-win_amd64.whl#sha256=0663585d02f76929792470451a5ba64424acc3cd5227b03921dab0e2f27b1709 +ddsketch @ https://agent-int-packages.datadoghq.com/external/ddsketch/ddsketch-3.0.1-py3-none-any.whl#sha256=6d047b455fe2837c43d366ff1ae6ba0c3166e15499de8688437a75cea914224e +ddtrace @ https://agent-int-packages.datadoghq.com/external/ddtrace/ddtrace-2.10.6-cp312-cp312-win_amd64.whl#sha256=bb183a535e5b24828a45901babd9fd15a1350c9d5096de5ba463287d0c8c64d1 +decorator @ https://agent-int-packages.datadoghq.com/external/decorator/decorator-5.1.1-py3-none-any.whl#sha256=b8c3f85900b9dc423225913c5aace94729fe1fa9763b38939a95226f02d37186 +deprecated @ https://agent-int-packages.datadoghq.com/external/deprecated/Deprecated-1.2.14-py2.py3-none-any.whl#sha256=6fac8b097794a90302bdbb17b9b815e732d3c4720583ff1b198499d78470466c +dnspython @ https://agent-int-packages.datadoghq.com/external/dnspython/dnspython-2.6.1-py3-none-any.whl#sha256=5ef3b9680161f6fa89daf8ad451b5f1a33b18ae8a1c6778cdf4b43f08c0a6e50 +dogpile-cache @ https://agent-int-packages.datadoghq.com/external/dogpile-cache/dogpile.cache-1.3.3-py3-none-any.whl#sha256=5e211c4902ebdf88c678d268e22454b41e68071632daa9402d8ee24e825ed8ca +envier @ https://agent-int-packages.datadoghq.com/external/envier/envier-0.5.2-py3-none-any.whl#sha256=65099cf3aa9b3b3b4b92db2f7d29e2910672e085b76f7e587d2167561a834add +filelock @ https://agent-int-packages.datadoghq.com/external/filelock/filelock-3.16.0-py3-none-any.whl#sha256=f6ed4c963184f4c84dd5557ce8fece759a3724b37b80c6c4f20a2f63a4dc6609 +foundationdb @ https://agent-int-packages.datadoghq.com/built/foundationdb/foundationdb-6.3.24-20240402154628-py3-none-win_amd64.whl#sha256=07e8e97e51dc9248d58d60d33076b82380135c31ab3727a33b885cea17e34bc7 +google-auth @ https://agent-int-packages.datadoghq.com/external/google-auth/google_auth-2.34.0-py2.py3-none-any.whl#sha256=72fd4733b80b6d777dcde515628a9eb4a577339437012874ea286bca7261ee65 +hazelcast-python-client @ https://agent-int-packages.datadoghq.com/external/hazelcast-python-client/hazelcast_python_client-5.4.0-py3-none-any.whl#sha256=16195cd58feb2dd3be1594d08d42527ae00797548a6a9d6a601aae2e8514ff5f +idna @ https://agent-int-packages.datadoghq.com/external/idna/idna-3.10-py3-none-any.whl#sha256=946d195a0d259cbba61165e88e65941f16e9b36ea6ddb97f00452bae8b1287d3 +importlib-metadata @ https://agent-int-packages.datadoghq.com/external/importlib-metadata/importlib_metadata-8.4.0-py3-none-any.whl#sha256=66f342cc6ac9818fc6ff340576acd24d65ba0b3efabb2b4ac08b598965a4a2f1 +in-toto @ https://agent-int-packages.datadoghq.com/external/in-toto/in_toto-2.0.0-py3-none-any.whl#sha256=eaf71733cdd5b2309b58b24169c13c24701da72baa43b7432d1dffc308e7bf32 +iso8601 @ https://agent-int-packages.datadoghq.com/external/iso8601/iso8601-2.1.0-py3-none-any.whl#sha256=aac4145c4dcb66ad8b648a02830f5e2ff6c24af20f4f482689be402db2429242 +jellyfish @ https://agent-int-packages.datadoghq.com/external/jellyfish/jellyfish-1.1.0-cp312-none-win_amd64.whl#sha256=2b928bad2887c662783a4d9b5828ed1fa0e943f680589f7fc002c456fc02e184 +jmespath @ https://agent-int-packages.datadoghq.com/external/jmespath/jmespath-1.0.1-py3-none-any.whl#sha256=02e2e4cc71b5bcab88332eebf907519190dd9e6e82107fa7f83b1003a6252980 +jsonpatch @ https://agent-int-packages.datadoghq.com/external/jsonpatch/jsonpatch-1.33-py2.py3-none-any.whl#sha256=0ae28c0cd062bbd8b8ecc26d7d164fbbea9652a1a3693f3b956c1eae5145dade +jsonpointer @ https://agent-int-packages.datadoghq.com/external/jsonpointer/jsonpointer-3.0.0-py2.py3-none-any.whl#sha256=13e088adc14fca8b6aa8177c044e12701e6ad4b28ff10e65f2267a90109c9942 +keystoneauth1 @ https://agent-int-packages.datadoghq.com/external/keystoneauth1/keystoneauth1-5.8.0-py3-none-any.whl#sha256=e69dff80c509ab64d4de4494658d914e81f26af720828dc584ceee74ecd666d9 +kubernetes @ https://agent-int-packages.datadoghq.com/external/kubernetes/kubernetes-30.1.0-py2.py3-none-any.whl#sha256=e212e8b7579031dd2e512168b617373bc1e03888d41ac4e04039240a292d478d +ldap3 @ https://agent-int-packages.datadoghq.com/external/ldap3/ldap3-2.9.1-py2.py3-none-any.whl#sha256=5869596fc4948797020d3f03b7939da938778a0f9e2009f7a072ccf92b8e8d70 +lxml @ https://agent-int-packages.datadoghq.com/external/lxml/lxml-4.9.4-cp312-cp312-win_amd64.whl#sha256=f1faee2a831fe249e1bae9cbc68d3cd8a30f7e37851deee4d7962b17c410dd56 +lz4 @ https://agent-int-packages.datadoghq.com/external/lz4/lz4-4.3.3-cp312-cp312-win_amd64.whl#sha256=5d35533bf2cee56f38ced91f766cd0038b6abf46f438a80d50c52750088be93f +mmh3 @ https://agent-int-packages.datadoghq.com/external/mmh3/mmh3-4.1.0-cp312-cp312-win_amd64.whl#sha256=bebc3ecb6ba18292e3d40c8712482b4477abd6981c2ebf0e60869bd90f8ac3a9 +msal @ https://agent-int-packages.datadoghq.com/external/msal/msal-1.31.0-py3-none-any.whl#sha256=96bc37cff82ebe4b160d5fc0f1196f6ca8b50e274ecd0ec5bf69c438514086e7 +msal-extensions @ https://agent-int-packages.datadoghq.com/external/msal-extensions/msal_extensions-1.2.0-py3-none-any.whl#sha256=cf5ba83a2113fa6dc011a254a72f1c223c88d7dfad74cc30617c4679a417704d +netifaces @ https://agent-int-packages.datadoghq.com/built/netifaces/netifaces-0.11.0-20240830145553-cp312-cp312-win_amd64.whl#sha256=a1ba522e63fb6b220e7fe668767f334662afa9c56eca18b361bd1f88863ab59a +oauthlib @ https://agent-int-packages.datadoghq.com/external/oauthlib/oauthlib-3.2.2-py3-none-any.whl#sha256=8139f29aac13e25d502680e9e19963e83f16838d48a0d71c287fe40e7067fbca +openstacksdk @ https://agent-int-packages.datadoghq.com/external/openstacksdk/openstacksdk-3.3.0-py3-none-any.whl#sha256=e6d4121b87354984caf0e3c032e2ebf4d4440374f86c81c27ec52ca5df359157 +opentelemetry-api @ https://agent-int-packages.datadoghq.com/external/opentelemetry-api/opentelemetry_api-1.27.0-py3-none-any.whl#sha256=953d5871815e7c30c81b56d910c707588000fff7a3ca1c73e6531911d53065e7 +orjson @ https://agent-int-packages.datadoghq.com/external/orjson/orjson-3.10.7-cp312-none-win_amd64.whl#sha256=1d9c0e733e02ada3ed6098a10a8ee0052dd55774de3d9110d29868d24b17faa1 +os-service-types @ https://agent-int-packages.datadoghq.com/external/os-service-types/os_service_types-1.7.0-py2.py3-none-any.whl#sha256=0505c72205690910077fb72b88f2a1f07533c8d39f2fe75b29583481764965d6 +packaging @ https://agent-int-packages.datadoghq.com/external/packaging/packaging-24.1-py3-none-any.whl#sha256=5b8f2217dbdbd2f7f384c41c628544e6d52f2d0f53c6d0c3ea61aa5d1d7ff124 +paramiko @ https://agent-int-packages.datadoghq.com/external/paramiko/paramiko-3.4.1-py3-none-any.whl#sha256=8e49fd2f82f84acf7ffd57c64311aa2b30e575370dc23bdb375b10262f7eac32 +pathspec @ https://agent-int-packages.datadoghq.com/external/pathspec/pathspec-0.12.1-py3-none-any.whl#sha256=a0d503e138a4c123b27490a4f7beda6a01c6f288df0e4a8b79c7eb0dc7b4cc08 +pbr @ https://agent-int-packages.datadoghq.com/external/pbr/pbr-6.1.0-py2.py3-none-any.whl#sha256=a776ae228892d8013649c0aeccbb3d5f99ee15e005a4cbb7e61d55a067b28a2a +platformdirs @ https://agent-int-packages.datadoghq.com/external/platformdirs/platformdirs-4.3.3-py3-none-any.whl#sha256=50a5450e2e84f44539718293cbb1da0a0885c9d14adf21b77bae4e66fc99d9b5 +ply @ https://agent-int-packages.datadoghq.com/external/ply/ply-3.11-py2.py3-none-any.whl#sha256=096f9b8350b65ebd2fd1346b12452efe5b9607f7482813ffca50c22722a807ce +portalocker @ https://agent-int-packages.datadoghq.com/external/portalocker/portalocker-2.10.1-py3-none-any.whl#sha256=53a5984ebc86a025552264b459b46a2086e269b21823cb572f8f28ee759e45bf +prometheus-client @ https://agent-int-packages.datadoghq.com/external/prometheus-client/prometheus_client-0.20.0-py3-none-any.whl#sha256=cde524a85bce83ca359cc837f28b8c0db5cac7aa653a588fd7e84ba061c329e7 +protobuf @ https://agent-int-packages.datadoghq.com/external/protobuf/protobuf-5.27.3-cp310-abi3-win_amd64.whl#sha256=16ddf3f8c6c41e1e803da7abea17b1793a97ef079a912e42351eabb19b2cffe7 +psutil @ https://agent-int-packages.datadoghq.com/external/psutil/psutil-5.9.6-cp37-abi3-win_amd64.whl#sha256=6e5fb8dc711a514da83098bc5234264e551ad980cec5f85dabf4d38ed6f15e9a +psycopg2-binary @ https://agent-int-packages.datadoghq.com/external/psycopg2-binary/psycopg2_binary-2.9.9-cp312-cp312-win_amd64.whl#sha256=81ff62668af011f9a48787564ab7eded4e9fb17a4a6a74af5ffa6a457400d2ab +pyasn1 @ https://agent-int-packages.datadoghq.com/external/pyasn1/pyasn1-0.4.8-py2.py3-none-any.whl#sha256=39c7e2ec30515947ff4e87fb6f456dfc6e84857d34be479c9d4a4ba4bf46aa5d +pyasn1-modules @ https://agent-int-packages.datadoghq.com/external/pyasn1-modules/pyasn1_modules-0.4.1-py3-none-any.whl#sha256=49bfa96b45a292b711e986f222502c1c9a5e1f4e568fc30e2574a6c7d07838fd +pyasyncore @ https://agent-int-packages.datadoghq.com/external/pyasyncore/pyasyncore-1.0.4-py3-none-any.whl#sha256=9e5f6dc9dc057c56370b7a5cdb4c4670fd4b0556de2913ed1f428cd6a5366895 +pycparser @ https://agent-int-packages.datadoghq.com/external/pycparser/pycparser-2.22-py3-none-any.whl#sha256=c3702b6d3dd8c7abc1afa565d7e63d53a1d0bd86cdc24edd75470f4de499cfcc +pycryptodomex @ https://agent-int-packages.datadoghq.com/external/pycryptodomex/pycryptodomex-3.20.0-cp35-abi3-win_amd64.whl#sha256=2a47bcc478741b71273b917232f521fd5704ab4b25d301669879e7273d3586cc +pydantic @ https://agent-int-packages.datadoghq.com/external/pydantic/pydantic-2.8.2-py3-none-any.whl#sha256=73ee9fddd406dc318b885c7a2eab8a6472b68b8fb5ba8150949fc3db939f23c8 +pydantic-core @ https://agent-int-packages.datadoghq.com/external/pydantic-core/pydantic_core-2.20.1-cp312-none-win_amd64.whl#sha256=035ede2e16da7281041f0e626459bcae33ed998cca6a0a007a5ebb73414ac72d +pyjwt @ https://agent-int-packages.datadoghq.com/external/pyjwt/PyJWT-2.9.0-py3-none-any.whl#sha256=3b02fb0f44517787776cf48f2ae25d8e14f300e6d7545a4315cee571a415e850 +pymongo @ https://agent-int-packages.datadoghq.com/external/pymongo/pymongo-4.8.0-cp312-cp312-win_amd64.whl#sha256=e84bc7707492f06fbc37a9f215374d2977d21b72e10a67f1b31893ec5a140ad8 +pymqi @ https://agent-int-packages.datadoghq.com/built/pymqi/pymqi-1.12.10-20240830145554-cp312-cp312-win_amd64.whl#sha256=d9749dd3ca4d1705f364cab34f679cf889d363f64602ed2f7c106a78876bf4ab +pymysql @ https://agent-int-packages.datadoghq.com/external/pymysql/PyMySQL-1.1.1-py3-none-any.whl#sha256=4de15da4c61dc132f4fb9ab763063e693d521a80fd0e87943b9a453dd4c19d6c +pynacl @ https://agent-int-packages.datadoghq.com/external/pynacl/PyNaCl-1.5.0-cp36-abi3-win_amd64.whl#sha256=20f42270d27e1b6a29f54032090b972d97f0a1b0948cc52392041ef7831fee93 +pyodbc @ https://agent-int-packages.datadoghq.com/external/pyodbc/pyodbc-5.1.0-cp312-cp312-win_amd64.whl#sha256=33f4984af38872e7bdec78007a34e4d43ae72bf9d0bae3344e79d9d0db157c0e +pyopenssl @ https://agent-int-packages.datadoghq.com/external/pyopenssl/pyOpenSSL-24.2.1-py3-none-any.whl#sha256=967d5719b12b243588573f39b0c677637145c7a1ffedcd495a487e58177fbb8d +pysmi @ https://agent-int-packages.datadoghq.com/external/pysmi/pysmi-1.2.1-py3-none-any.whl#sha256=d97c60de9f81d33ab2899124d95a94fa7fefacc86ab6e00cbfec543a073e6d33 +pysnmp @ https://agent-int-packages.datadoghq.com/external/pysnmp/pysnmp-5.1.0-py3-none-any.whl#sha256=375a8adfc6820faf24ace6761a6d20544e60580d714ff7266df272850c39b439 +pysnmp-mibs @ https://agent-int-packages.datadoghq.com/external/pysnmp-mibs/pysnmp_mibs-0.1.6-py2.py3-none-any.whl#sha256=5e153ebe8e767c07940cea435f866c623ff6b2376155c7da75085b08d3774d48 +pysnmpcrypto @ https://agent-int-packages.datadoghq.com/external/pysnmpcrypto/pysnmpcrypto-0.0.4-py2.py3-none-any.whl#sha256=5889733caa030f45d9e03ea9d6370fb06426a8cb7f839aabbcdde33c6f634679 +pysocks @ https://agent-int-packages.datadoghq.com/external/pysocks/PySocks-1.7.1-py3-none-any.whl#sha256=2725bd0a9925919b9b51739eea5f9e2bae91e83288108a9ad338b2e3a4435ee5 +pyspnego @ https://agent-int-packages.datadoghq.com/external/pyspnego/pyspnego-0.11.1-py3-none-any.whl#sha256=129a4294f2c4d681d5875240ef87accc6f1d921e8983737fb0b59642b397951e +python-dateutil @ https://agent-int-packages.datadoghq.com/external/python-dateutil/python_dateutil-2.9.0.post0-py2.py3-none-any.whl#sha256=a8b2bc7bffae282281c8140a97d3aa9c14da0b136dfe83f850eea9a5f7470427 +pytz @ https://agent-int-packages.datadoghq.com/external/pytz/pytz-2024.2-py2.py3-none-any.whl#sha256=31c7c1817eb7fae7ca4b8c7ee50c72f93aa2dd863de768e1ef4245d426aa0725 +pyvmomi @ https://agent-int-packages.datadoghq.com/built/pyvmomi/pyvmomi-8.0.3.0.1-20240702172100-py2.py3-none-win_amd64.whl#sha256=19446fe48dbdd8b64097eff5648cc4b5a19165ede40826507f5e1398e1032e12 +pywin32 @ https://agent-int-packages.datadoghq.com/external/pywin32/pywin32-306-cp312-cp312-win_amd64.whl#sha256=37257794c1ad39ee9be652da0462dc2e394c8159dfd913a8a4e8eb6fd346da0e +pyyaml @ https://agent-int-packages.datadoghq.com/external/pyyaml/PyYAML-6.0.2-cp312-cp312-win_amd64.whl#sha256=7e7401d0de89a9a855c839bc697c079a4af81cf878373abd7dc625847d25cbd8 +redis @ https://agent-int-packages.datadoghq.com/external/redis/redis-5.0.8-py3-none-any.whl#sha256=56134ee08ea909106090934adc36f65c9bcbbaecea5b21ba704ba6fb561f8eb4 +requests @ https://agent-int-packages.datadoghq.com/external/requests/requests-2.32.3-py3-none-any.whl#sha256=70761cfe03c773ceb22aa2f671b4757976145175cdfca038c02654d061d6dcc6 +requests-kerberos @ https://agent-int-packages.datadoghq.com/external/requests-kerberos/requests_kerberos-0.15.0-py2.py3-none-any.whl#sha256=ba9b0980b8489c93bfb13854fd118834e576d6700bfea3745cb2e62278cd16a6 +requests-ntlm @ https://agent-int-packages.datadoghq.com/external/requests-ntlm/requests_ntlm-1.3.0-py3-none-any.whl#sha256=4c7534a7d0e482bb0928531d621be4b2c74ace437e88c5a357ceb7452d25a510 +requests-oauthlib @ https://agent-int-packages.datadoghq.com/external/requests-oauthlib/requests_oauthlib-2.0.0-py2.py3-none-any.whl#sha256=7dd8a5c40426b779b0868c404bdef9768deccf22749cde15852df527e6269b36 +requests-toolbelt @ https://agent-int-packages.datadoghq.com/external/requests-toolbelt/requests_toolbelt-1.0.0-py2.py3-none-any.whl#sha256=cccfdd665f0a24fcf4726e690f65639d272bb0637b9b92dfd91a5568ccf6bd06 +requests-unixsocket2 @ https://agent-int-packages.datadoghq.com/external/requests-unixsocket2/requests_unixsocket2-0.4.2-py3-none-any.whl#sha256=701fcd49d74bc0f759bbe45c4dfda0045fd89652948c2b473b1a312214c3770b +requestsexceptions @ https://agent-int-packages.datadoghq.com/external/requestsexceptions/requestsexceptions-1.4.0-py2.py3-none-any.whl#sha256=3083d872b6e07dc5c323563ef37671d992214ad9a32b0ca4a3d7f5500bf38ce3 +rethinkdb @ https://agent-int-packages.datadoghq.com/external/rethinkdb/rethinkdb-2.4.9-py2.py3-none-any.whl#sha256=dd2455b3eab5266d8006eeefeb5eb970a3697dfe6aa18e1ab6e1d99233badfcb +rsa @ https://agent-int-packages.datadoghq.com/external/rsa/rsa-4.9-py3-none-any.whl#sha256=90260d9058e514786967344d0ef75fa8727eed8a7d2e43ce9f4bcf1b536174f7 +s3transfer @ https://agent-int-packages.datadoghq.com/external/s3transfer/s3transfer-0.10.2-py3-none-any.whl#sha256=eca1c20de70a39daee580aef4986996620f365c4e0fda6a86100231d62f1bf69 +securesystemslib @ https://agent-int-packages.datadoghq.com/external/securesystemslib/securesystemslib-0.28.0-py3-none-any.whl#sha256=9e6b9abe36a511d4f52c759069db8f6f650362ba82d6efc7bc7466a458b3f499 +semver @ https://agent-int-packages.datadoghq.com/external/semver/semver-3.0.2-py3-none-any.whl#sha256=b1ea4686fe70b981f85359eda33199d60c53964284e0cfb4977d243e37cf4bf4 +service-identity @ https://agent-int-packages.datadoghq.com/external/service-identity/service_identity-24.1.0-py3-none-any.whl#sha256=a28caf8130c8a5c1c7a6f5293faaf239bbfb7751e4862436920ee6f2616f568a +setuptools @ https://agent-int-packages.datadoghq.com/external/setuptools/setuptools-75.1.0-py3-none-any.whl#sha256=35ab7fd3bcd95e6b7fd704e4a1539513edad446c097797f2985e0e4b960772f2 +simplejson @ https://agent-int-packages.datadoghq.com/external/simplejson/simplejson-3.19.3-cp312-cp312-win_amd64.whl#sha256=1e662336db50ad665777e6548b5076329a94a0c3d4a0472971c588b3ef27de3a +six @ https://agent-int-packages.datadoghq.com/external/six/six-1.16.0-py2.py3-none-any.whl#sha256=8abb2f1d86890a2dfb989f9a77cfcfd3e47c2a354b01111771326f8aa26e0254 +snowflake-connector-python @ https://agent-int-packages.datadoghq.com/external/snowflake-connector-python/snowflake_connector_python-3.12.1-cp312-cp312-win_amd64.whl#sha256=b06c63ec0381df1f4da6c4326330a1a40c8fc21fd3dcc2f58df4de395d676893 +sortedcontainers @ https://agent-int-packages.datadoghq.com/external/sortedcontainers/sortedcontainers-2.4.0-py2.py3-none-any.whl#sha256=a163dcaede0f1c021485e957a39245190e74249897e2ae4b2aa38595db237ee0 +soupsieve @ https://agent-int-packages.datadoghq.com/external/soupsieve/soupsieve-2.6-py3-none-any.whl#sha256=e72c4ff06e4fb6e4b5a9f0f55fe6e81514581fca1515028625d0f299c602ccc9 +sspilib @ https://agent-int-packages.datadoghq.com/external/sspilib/sspilib-0.1.0-cp312-cp312-win_amd64.whl#sha256=b83825a2c43ff84ddff72d09b098057efaabf3841d3c42888078e154cf8e9595 +stevedore @ https://agent-int-packages.datadoghq.com/external/stevedore/stevedore-5.3.0-py3-none-any.whl#sha256=1efd34ca08f474dad08d9b19e934a22c68bb6fe416926479ba29e5013bcc8f78 +supervisor @ https://agent-int-packages.datadoghq.com/external/supervisor/supervisor-4.2.5-py2.py3-none-any.whl#sha256=2ecaede32fc25af814696374b79e42644ecaba5c09494c51016ffda9602d0f08 +tomlkit @ https://agent-int-packages.datadoghq.com/external/tomlkit/tomlkit-0.13.2-py3-none-any.whl#sha256=7a974427f6e119197f670fbbbeae7bef749a6c14e793db934baefc1b5f03efde +tuf @ https://agent-int-packages.datadoghq.com/external/tuf/tuf-4.0.0-py3-none-any.whl#sha256=a22ab5fa6daf910b3052929fdce42ccad8a300e5e85715daaff9592aed980f7a +typing-extensions @ https://agent-int-packages.datadoghq.com/external/typing-extensions/typing_extensions-4.12.2-py3-none-any.whl#sha256=04e5ca0351e0f3f85c6853954072df659d0d13fac324d0072316b67d7794700d +tzdata @ https://agent-int-packages.datadoghq.com/external/tzdata/tzdata-2024.1-py2.py3-none-any.whl#sha256=9068bc196136463f5245e51efda838afa15aaeca9903f49050dfa2679db4d252 +tzlocal @ https://agent-int-packages.datadoghq.com/external/tzlocal/tzlocal-5.2-py3-none-any.whl#sha256=49816ef2fe65ea8ac19d19aa7a1ae0551c834303d5014c6d5a62e4cbda8047b8 +uptime @ https://agent-int-packages.datadoghq.com/built/uptime/uptime-3.0.1-20240830145554-cp312-cp312-win_amd64.whl#sha256=3428f0d06d749e2b07f4b783429de36dca4bd437fe0b4df558074fa898cae5c0 +urllib3 @ https://agent-int-packages.datadoghq.com/external/urllib3/urllib3-2.2.3-py3-none-any.whl#sha256=ca899ca043dcb1bafa3e262d73aa25c465bfb49e0bd9dd5d59f1d0acba2f8fac +vertica-python @ https://agent-int-packages.datadoghq.com/external/vertica-python/vertica_python-1.4.0-py3-none-any.whl#sha256=50fecd7687f4b0b9f6dee6e2b35c195af2a4f702ece01bd12e080b51756e000b +websocket-client @ https://agent-int-packages.datadoghq.com/external/websocket-client/websocket_client-1.8.0-py3-none-any.whl#sha256=17b44cc997f5c498e809b22cdf2d9c7a9e71c02c8cc2b6c56e7c2d1239bfa526 +wrapt @ https://agent-int-packages.datadoghq.com/external/wrapt/wrapt-1.16.0-cp312-cp312-win_amd64.whl#sha256=dcdba5c86e368442528f7060039eda390cc4091bfd1dca41e8046af7c910dda8 +xmltodict @ https://agent-int-packages.datadoghq.com/external/xmltodict/xmltodict-0.13.0-py2.py3-none-any.whl#sha256=aa89e8fd76320154a40d19a0df04a4695fb9dc5ba977cbb68ab3e4eb225e7852 +zipp @ https://agent-int-packages.datadoghq.com/external/zipp/zipp-3.20.2-py3-none-any.whl#sha256=a817ac80d6cf4b23bf7f2828b7cabf326f15a001bea8b1f9b49631780ba28350 From d5250d3b50e30ba7331b32b3db2f02ab1e1e199b Mon Sep 17 00:00:00 2001 From: Ilia Kurenkov Date: Tue, 17 Sep 2024 22:36:04 +0200 Subject: [PATCH 06/34] Improve assertions in test for python upgrade script (#18606) * Improve assertions in test for python upgrade script * bump python version in conftest * lint --------- Co-authored-by: Kyle-Neale --- ddev/tests/cli/meta/scripts/conftest.py | 11 +++++------ ddev/tests/cli/meta/scripts/test_upgrade_python.py | 4 ++-- 2 files changed, 7 insertions(+), 8 deletions(-) diff --git a/ddev/tests/cli/meta/scripts/conftest.py b/ddev/tests/cli/meta/scripts/conftest.py index d92dc4481cfcc..3eef49fef96ff 100644 --- a/ddev/tests/cli/meta/scripts/conftest.py +++ b/ddev/tests/cli/meta/scripts/conftest.py @@ -5,8 +5,10 @@ from ddev.repo.core import Repository -OLD_PYTHON_VERSION = "3.11" -NEW_PYTHON_VERSION = "3.12" +# Whenenever we bump python version, we also need to bump the python +# version in the conftest.py. +OLD_PYTHON_VERSION = "3.12" +NEW_PYTHON_VERSION = "3.13" @pytest.fixture @@ -43,7 +45,7 @@ def fake_repo(tmp_path_factory, config_file, ddev): f"""[env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["2.7", "{OLD_PYTHON_VERSION}"] +python = ["{OLD_PYTHON_VERSION}"] """, ) @@ -68,7 +70,6 @@ def fake_repo(tmp_path_factory, config_file, ddev): "License :: OSI Approved :: BSD License", "Natural Language :: English", "Operating System :: OS Independent", - "Programming Language :: Python :: 2.7", "Programming Language :: Python :: {OLD_PYTHON_VERSION}", ] """, @@ -86,7 +87,6 @@ def fake_repo(tmp_path_factory, config_file, ddev): "License :: OSI Approved :: BSD License", "Natural Language :: English", "Operating System :: OS Independent", - "Programming Language :: Python :: 2.7", "Programming Language :: Python :: {OLD_PYTHON_VERSION}", ] """, @@ -134,7 +134,6 @@ def fake_repo(tmp_path_factory, config_file, ddev): "License :: OSI Approved :: BSD License", "Natural Language :: English", "Operating System :: OS Independent", - "Programming Language :: Python :: 2.7", "Programming Language :: Python :: {OLD_PYTHON_VERSION}", ] """, diff --git a/ddev/tests/cli/meta/scripts/test_upgrade_python.py b/ddev/tests/cli/meta/scripts/test_upgrade_python.py index 8302f47ec89ec..0f72910563c37 100644 --- a/ddev/tests/cli/meta/scripts/test_upgrade_python.py +++ b/ddev/tests/cli/meta/scripts/test_upgrade_python.py @@ -27,8 +27,8 @@ def test_upgrade_python(fake_repo, ddev): hatch_file = fake_repo.path / 'dummy' / 'hatch.toml' contents = hatch_file.read_text() - assert f'python = ["2.7", "{OLD_PYTHON_VERSION}"]' not in contents - assert f'python = ["2.7", "{NEW_PYTHON_VERSION}"]' in contents + assert f'python = ["{OLD_PYTHON_VERSION}"]' not in contents + assert f'python = ["{NEW_PYTHON_VERSION}"]' in contents for integration in ('dummy', 'datadog_checks_dependency_provider', 'logs_only'): pyproject_file = fake_repo.path / integration / 'pyproject.toml' From 9c901c4d8621d912c9e77a7ece95b3bcca28ffa2 Mon Sep 17 00:00:00 2001 From: Kyle Neale Date: Tue, 17 Sep 2024 17:40:01 -0400 Subject: [PATCH 07/34] update python env to 3.12.6 (#18612) --- kubeflow/hatch.toml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/kubeflow/hatch.toml b/kubeflow/hatch.toml index 001e43ce25414..c85c5f07a7df2 100644 --- a/kubeflow/hatch.toml +++ b/kubeflow/hatch.toml @@ -1,4 +1,4 @@ [env.collectors.datadog-checks] [[envs.default.matrix]] -python = ["3.11"] +python = ["3.12"] From 1ab760be48802333c9a3a5c3ba850667c31d368e Mon Sep 17 00:00:00 2001 From: "datadog-agent-integrations-bot[bot]" <159767151+datadog-agent-integrations-bot[bot]@users.noreply.github.com> Date: Wed, 18 Sep 2024 09:20:55 -0400 Subject: [PATCH 08/34] Finalize Agent release 7.57.1 (#18615) Co-authored-by: Kyle-Neale --- AGENT_CHANGELOG.md | 4 + AGENT_INTEGRATIONS.md | 201 ++++++++++++++++++++++++++++++++++++++++++ 2 files changed, 205 insertions(+) diff --git a/AGENT_CHANGELOG.md b/AGENT_CHANGELOG.md index 62986be17e7cc..3fc171d8a39f6 100644 --- a/AGENT_CHANGELOG.md +++ b/AGENT_CHANGELOG.md @@ -1,3 +1,7 @@ +## Datadog Agent version [7.57.1](https://github.com/DataDog/datadog-agent/blob/master/CHANGELOG.rst#7571) + +* There were no integration updates for this version of the Agent. + ## Datadog Agent version [7.57.0](https://github.com/DataDog/datadog-agent/blob/master/CHANGELOG.rst#7570) * Amazon Kafka [4.10.0](https://github.com/DataDog/integrations-core/blob/master/amazon_msk/CHANGELOG.md) diff --git a/AGENT_INTEGRATIONS.md b/AGENT_INTEGRATIONS.md index 5b810e1a04d87..3dfab715b2827 100644 --- a/AGENT_INTEGRATIONS.md +++ b/AGENT_INTEGRATIONS.md @@ -1,3 +1,204 @@ +## Datadog Agent version 7.57.1 + +* datadog-active-directory: 2.1.1 +* datadog-activemq-xml: 3.2.1 +* datadog-activemq: 3.1.0 +* datadog-aerospike: 2.2.2 +* datadog-airflow: 5.0.1 +* datadog-amazon-msk: 4.10.0 +* datadog-ambari: 4.2.1 +* datadog-apache: 4.5.1 +* datadog-arangodb: 2.2.2 +* datadog-argo-rollouts: 1.0.2 +* datadog-argo-workflows: 1.0.2 +* datadog-argocd: 2.4.3 +* datadog-aspdotnet: 2.1.1 +* datadog-avi-vantage: 4.2.2 +* datadog-aws-neuron: 1.0.0 +* datadog-azure-iot-edge: 4.2.1 +* datadog-boundary: 2.2.3 +* datadog-btrfs: 2.3.0 +* datadog-cacti: 2.1.1 +* datadog-calico: 2.2.2 +* datadog-cassandra-nodetool: 1.13.0 +* datadog-cassandra: 1.18.0 +* datadog-ceph: 2.10.0 +* datadog-cert-manager: 4.1.2 +* datadog-checkpoint-quantum-firewall: 1.0.0 +* datadog-checks-base: 36.13.0 +* datadog-checks-dependency-provider: 1.4.0 +* datadog-checks-downloader: 4.7.0 +* datadog-cilium: 3.5.1 +* datadog-cisco-aci: 2.10.2 +* datadog-cisco-secure-firewall: 1.0.0 +* datadog-citrix-hypervisor: 3.2.1 +* datadog-clickhouse: 3.6.0 +* datadog-cloud-foundry-api: 3.3.1 +* datadog-cloudera: 2.2.0 +* datadog-cockroachdb: 3.3.2 +* datadog-confluent-platform: 1.10.2 +* datadog-consul: 2.6.1 +* datadog-coredns: 3.2.3 +* datadog-couch: 6.2.1 +* datadog-couchbase: 3.2.1 +* datadog-crio: 2.6.1 +* datadog-datadog-cluster-agent: 3.2.0 +* datadog-dcgm: 2.3.2 +* datadog-directory: 2.1.1 +* datadog-disk: 5.3.0 +* datadog-dns-check: 3.3.0 +* datadog-dotnetclr: 2.1.1 +* datadog-druid: 2.5.1 +* datadog-ecs-fargate: 4.3.1 +* datadog-eks-fargate: 4.2.1 +* datadog-elastic: 6.3.1 +* datadog-envoy: 3.5.1 +* datadog-esxi: 1.2.0 +* datadog-etcd: 6.2.2 +* datadog-exchange-server: 2.1.1 +* datadog-external-dns: 3.2.1 +* datadog-flink: 1.5.0 +* datadog-fluentd: 3.2.1 +* datadog-fluxcd: 1.2.2 +* datadog-fly-io: 1.0.0 +* datadog-foundationdb: 1.4.0 +* datadog-gearmand: 3.1.0 +* datadog-gitlab-runner: 4.2.1 +* datadog-gitlab: 7.3.2 +* datadog-glusterfs: 1.7.0 +* datadog-go-expvar: 2.5.1 +* datadog-gunicorn: 2.7.1 +* datadog-haproxy: 5.2.2 +* datadog-harbor: 3.2.2 +* datadog-hazelcast: 3.2.1 +* datadog-hdfs-datanode: 4.2.1 +* datadog-hdfs-namenode: 4.2.1 +* datadog-hive: 1.10.0 +* datadog-hivemq: 1.8.0 +* datadog-http-check: 9.8.0 +* datadog-hudi: 2.3.0 +* datadog-hyperv: 1.11.1 +* datadog-ibm-ace: 2.2.2 +* datadog-ibm-db2: 2.2.0 +* datadog-ibm-i: 2.2.0 +* datadog-ibm-mq: 6.3.0 +* datadog-ibm-was: 3.3.2 +* datadog-ignite: 2.4.0 +* datadog-iis: 3.1.1 +* datadog-impala: 2.2.2 +* datadog-istio: 6.1.2 +* datadog-jboss-wildfly: 2.2.0 +* datadog-journald: 1.2.0 +* datadog-kafka-consumer: 4.6.0 +* datadog-kafka: 2.16.0 +* datadog-karpenter: 1.5.0 +* datadog-kong: 3.2.2 +* datadog-kube-apiserver-metrics: 4.3.1 +* datadog-kube-controller-manager: 5.1.1 +* datadog-kube-dns: 4.4.1 +* datadog-kube-metrics-server: 3.3.1 +* datadog-kube-proxy: 6.3.1 +* datadog-kube-scheduler: 4.10.1 +* datadog-kubelet: 7.13.2 +* datadog-kubernetes-cluster-autoscaler: 1.0.1 +* datadog-kubernetes-state: 8.1.1 +* datadog-kyototycoon: 2.5.1 +* datadog-kyverno: 1.0.2 +* datadog-lighttpd: 3.5.1 +* datadog-linkerd: 4.2.2 +* datadog-linux-proc-extras: 2.5.0 +* datadog-mapr: 1.11.0 +* datadog-mapreduce: 4.2.1 +* datadog-marathon: 2.3.1 +* datadog-marklogic: 4.2.1 +* datadog-mcache: 4.1.0 +* datadog-mesos-master: 3.3.1 +* datadog-mesos-slave: 3.3.1 +* datadog-mongo: 6.9.0 +* datadog-mysql: 12.7.0 +* datadog-nagios: 1.13.0 +* datadog-network: 3.3.0 +* datadog-nfsstat: 1.13.0 +* datadog-nginx-ingress-controller: 2.6.1 +* datadog-nginx: 6.3.1 +* datadog-nvidia-triton: 1.2.2 +* datadog-openldap: 1.12.0 +* datadog-openmetrics: 4.2.2 +* datadog-openstack-controller: 6.9.0 +* datadog-openstack: 2.0.0 +* datadog-oracle: 5.2.0 +* datadog-ossec-security: 1.0.0 +* datadog-palo-alto-panorama: 1.0.0 +* datadog-pan-firewall: 1.2.0 +* datadog-pdh-check: 2.1.0 +* datadog-pgbouncer: 6.2.0 +* datadog-php-fpm: 3.3.1 +* datadog-ping-federate: 1.0.0 +* datadog-postfix: 1.14.0 +* datadog-postgres: 19.1.0 +* datadog-powerdns-recursor: 2.5.1 +* datadog-presto: 2.8.0 +* datadog-process: 3.4.0 +* datadog-prometheus: 3.6.0 +* datadog-proxysql: 5.1.1 +* datadog-pulsar: 2.2.2 +* datadog-rabbitmq: 5.3.2 +* datadog-ray: 1.2.2 +* datadog-redisdb: 5.7.0 +* datadog-rethinkdb: 3.1.0 +* datadog-riak: 3.5.1 +* datadog-riakcs: 2.11.0 +* datadog-sap-hana: 3.3.0 +* datadog-scylla: 2.7.2 +* datadog-sidekiq: 1.4.0 +* datadog-silk: 2.2.1 +* datadog-singlestore: 2.2.1 +* datadog-snmp: 7.4.0 +* datadog-snowflake: 5.8.0 +* datadog-solr: 1.13.0 +* datadog-sonarqube: 3.2.2 +* datadog-spark: 4.3.1 +* datadog-sqlserver: 17.5.2 +* datadog-squid: 2.5.1 +* datadog-ssh-check: 2.9.0 +* datadog-statsd: 1.12.0 +* datadog-strimzi: 2.2.2 +* datadog-supervisord: 2.6.0 +* datadog-suricata: 1.0.0 +* datadog-system-core: 2.4.0 +* datadog-system-swap: 1.18.0 +* datadog-tcp-check: 4.9.0 +* datadog-teamcity: 4.3.1 +* datadog-tekton: 1.0.2 +* datadog-teleport: 1.0.0 +* datadog-temporal: 2.3.0 +* datadog-tenable: 1.5.0 +* datadog-teradata: 2.2.1 +* datadog-tibco-ems: 1.0.0 +* datadog-tls: 2.19.0 +* datadog-tokumx: 3.2.0 +* datadog-tomcat: 2.0.0 +* datadog-torchserve: 2.2.2 +* datadog-traefik-mesh: 1.0.1 +* datadog-traffic-server: 2.2.1 +* datadog-twemproxy: 1.15.0 +* datadog-twistlock: 3.6.1 +* datadog-varnish: 2.1.0 +* datadog-vault: 4.2.1 +* datadog-vertica: 4.6.0 +* datadog-vllm: 1.0.0 +* datadog-voltdb: 3.2.1 +* datadog-vsphere: 7.6.0 +* datadog-weaviate: 2.3.3 +* datadog-weblogic: 1.3.0 +* datadog-win32-event-log: 3.3.0 +* datadog-windows-performance-counters: 2.1.1 +* datadog-windows-service: 4.9.1 +* datadog-wmi-check: 1.18.0 +* datadog-yarn: 5.3.1 +* datadog-zeek: 1.0.0 +* datadog-zk: 4.5.0 + ## Datadog Agent version 7.57.0 * datadog-active-directory: 2.1.1 From d74438853551ceb4c20bc55517ea3142a152788b Mon Sep 17 00:00:00 2001 From: Zhengda Lu Date: Wed, 18 Sep 2024 10:04:57 -0400 Subject: [PATCH 09/34] add comments to metrics to indicate which collector it comes from (#18611) --- mongo/datadog_checks/mongo/metrics.py | 16 +++++++++++++++- 1 file changed, 15 insertions(+), 1 deletion(-) diff --git a/mongo/datadog_checks/mongo/metrics.py b/mongo/datadog_checks/mongo/metrics.py index 146cd07263387..f60a532481141 100644 --- a/mongo/datadog_checks/mongo/metrics.py +++ b/mongo/datadog_checks/mongo/metrics.py @@ -10,6 +10,7 @@ Core metrics collected by default. """ BASE_METRICS = { + # server_status collector "asserts.msg": RATE, "asserts.regular": RATE, "asserts.rollovers": RATE, @@ -125,14 +126,17 @@ "opcountersRepl.insert": RATE, "opcountersRepl.query": RATE, "opcountersRepl.update": RATE, + # replication_info collector "oplog.logSizeMB": GAUGE, "oplog.usedSizeMB": GAUGE, "oplog.timeDiff": GAUGE, + # replica collector "replSet.health": GAUGE, "replSet.replicationLag": GAUGE, "replSet.state": GAUGE, "replSet.votes": GAUGE, "replSet.voteFraction": GAUGE, + # db_stats collector "stats.avgObjSize": GAUGE, "stats.collections": GAUGE, "stats.dataSize": GAUGE, @@ -150,8 +154,10 @@ "stats.fsUsedSize": GAUGE, "stats.fsTotalSize": GAUGE, "stats.views": GAUGE, + # session_stats collector "sessions.count": GAUGE, "uptime": GAUGE, + # host_info collector "system.memSizeMB": (GAUGE, "system.mem.total"), # total amount of system memory "system.memLimitMB": (GAUGE, "system.mem.limit"), # memory usage limit "system.numCores": (GAUGE, "system.cpu.cores"), # number of CPU cores @@ -186,7 +192,8 @@ https://docs.mongodb.org/manual/reference/command/serverStatus/#serverStatus.metrics.commands """ COMMANDS_METRICS = { - # Required version > + # Required version > 3.0.0 + # server_status collector "metrics.commands.count.failed": RATE, "metrics.commands.count.total": GAUGE, "metrics.commands.createIndexes.failed": RATE, @@ -210,6 +217,7 @@ https://docs.mongodb.org/manual/reference/command/serverStatus/#server-status-locks """ LOCKS_METRICS = { + # server_status collector "locks.Collection.acquireCount.R": RATE, "locks.Collection.acquireCount.r": RATE, "locks.Collection.acquireCount.W": RATE, @@ -262,6 +270,7 @@ TCMalloc memory allocator report. """ TCMALLOC_METRICS = { + # server_status collector "tcmalloc.generic.current_allocated_bytes": GAUGE, "tcmalloc.generic.heap_size": GAUGE, "tcmalloc.tcmalloc.aggressive_memory_decommit": GAUGE, @@ -279,6 +288,7 @@ WiredTiger storage engine. """ WIREDTIGER_METRICS = { + # server_status collector "wiredTiger.cache.bytes currently in the cache": (GAUGE, "wiredTiger.cache.bytes_currently_in_cache"), "wiredTiger.cache.bytes read into cache": GAUGE, "wiredTiger.cache.bytes written from cache": GAUGE, @@ -315,6 +325,7 @@ https://docs.mongodb.org/v3.0/reference/command/top/ """ TOP_METRICS = { + # top collector "commands.count": RATE, "commands.time": GAUGE, "getmore.count": RATE, @@ -337,6 +348,7 @@ COLLECTION_METRICS = { # collection storage stats + # coll_stats collector 'collection.size': GAUGE, 'collection.avgObjSize': GAUGE, 'collection.count': GAUGE, @@ -365,6 +377,7 @@ } SHARDED_DATA_DISTRIBUTION_METRICS = { + # sharded_data_distribution collector 'numOrphanedDocs': (GAUGE, 'sharded_data_distribution.num_orphaned_docs'), 'numOwnedDocuments': (GAUGE, 'sharded_data_distribution.num_owned_documents'), 'ownedSizeBytes': (GAUGE, 'sharded_data_distribution.owned_size_bytes'), @@ -372,6 +385,7 @@ } INDEX_METRICS = { + # index stats collector 'indexes.accesses.ops': RATE, } From 3765e710e0d83088b3a3167037b9b0c9bdb71fc2 Mon Sep 17 00:00:00 2001 From: Ilia Kurenkov Date: Wed, 18 Sep 2024 16:56:13 +0200 Subject: [PATCH 10/34] Remove six library from some integrations (#18604) * Remove six library from some integrations * fix mapreduce --- datadog_checks_dev/datadog_checks/dev/_env.py | 8 +- disk/datadog_checks/disk/disk.py | 9 +- .../datadog_checks/ecs_fargate/ecs_fargate.py | 13 +- .../kube_apiserver_metrics.py | 4 +- .../datadog_checks/mapreduce/mapreduce.py | 16 +- mcache/datadog_checks/mcache/mcache.py | 7 +- .../mesos_master/mesos_master.py | 10 +- mongo/datadog_checks/mongo/collectors/base.py | 13 +- .../mongo/collectors/coll_stats.py | 3 +- mongo/datadog_checks/mongo/collectors/top.py | 5 +- .../datadog_checks/mysql/collection_utils.py | 6 +- mysql/datadog_checks/mysql/innodb_metrics.py | 178 +++++++++--------- mysql/datadog_checks/mysql/mysql.py | 27 ++- nginx/datadog_checks/nginx/nginx.py | 20 +- redisdb/datadog_checks/redisdb/redisdb.py | 16 +- riakcs/datadog_checks/riakcs/riakcs.py | 7 +- sap_hana/datadog_checks/sap_hana/sap_hana.py | 4 +- spark/datadog_checks/spark/constants.py | 8 +- squid/datadog_checks/squid/squid.py | 4 +- .../datadog_checks/system_core/system_core.py | 5 +- .../datadog_checks/twemproxy/twemproxy.py | 5 +- .../datadog_checks/twistlock/twistlock.py | 3 +- varnish/datadog_checks/varnish/varnish.py | 22 +-- vsphere/datadog_checks/vsphere/api_rest.py | 3 +- vsphere/datadog_checks/vsphere/config.py | 9 +- .../vsphere/legacy/mor_cache.py | 8 +- vsphere/datadog_checks/vsphere/utils.py | 3 +- vsphere/datadog_checks/vsphere/vsphere.py | 9 +- vsphere/tests/legacy/utils.py | 3 +- yarn/datadog_checks/yarn/yarn.py | 12 +- zk/datadog_checks/zk/zk.py | 21 +-- 31 files changed, 198 insertions(+), 263 deletions(-) diff --git a/datadog_checks_dev/datadog_checks/dev/_env.py b/datadog_checks_dev/datadog_checks/dev/_env.py index 35b761a785d05..84f96bcab491a 100644 --- a/datadog_checks_dev/datadog_checks/dev/_env.py +++ b/datadog_checks_dev/datadog_checks/dev/_env.py @@ -5,8 +5,6 @@ import os from base64 import urlsafe_b64decode, urlsafe_b64encode -from six import iteritems - DDTRACE_OPTIONS_LIST = [ 'DD_TAGS', 'DD_TRACE*', @@ -56,7 +54,7 @@ def e2e_testing(): def set_env_vars(env_vars): - for key, value in iteritems(env_vars): + for key, value in env_vars.items(): key = '{}{}'.format(E2E_ENV_VAR_PREFIX, key) os.environ[key] = value @@ -68,11 +66,11 @@ def remove_env_vars(env_vars): def get_env_vars(raw=False): if raw: - return {key: value for key, value in iteritems(os.environ) if key.startswith(E2E_ENV_VAR_PREFIX)} + return {key: value for key, value in os.environ.items() if key.startswith(E2E_ENV_VAR_PREFIX)} else: env_vars = {} - for key, value in iteritems(os.environ): + for key, value in os.environ.items(): _, found, ev = key.partition(E2E_ENV_VAR_PREFIX) if found: # Normalize casing for Windows diff --git a/disk/datadog_checks/disk/disk.py b/disk/datadog_checks/disk/disk.py index 5999e32c9dc33..e4574f89b248e 100644 --- a/disk/datadog_checks/disk/disk.py +++ b/disk/datadog_checks/disk/disk.py @@ -9,7 +9,6 @@ from xml.etree import ElementTree as ET import psutil -from six import iteritems, string_types from datadog_checks.base import AgentCheck, ConfigurationError, is_affirmative from datadog_checks.base.utils.platform import Platform @@ -151,7 +150,7 @@ def check(self, _): self.log.debug('Passed: %s', part.device) tags = self._get_tags(part) - for metric_name, metric_value in iteritems(self._collect_part_metrics(part, disk_usage)): + for metric_name, metric_value in self._collect_part_metrics(part, disk_usage).items(): self.gauge(metric_name, metric_value, tags=tags) # Add in a disk read write or read only check @@ -324,7 +323,7 @@ def _collect_inodes_metrics(self, mountpoint): return metrics def collect_latency_metrics(self): - for disk_name, disk in iteritems(psutil.disk_io_counters(perdisk=True)): + for disk_name, disk in psutil.disk_io_counters(perdisk=True).items(): self.log.debug('IO Counters: %s -> %s', disk_name, disk) try: metric_tags = [] if self._custom_tags is None else self._custom_tags[:] @@ -389,7 +388,7 @@ def _compile_pattern_filters(self, instance): def _compile_valid_patterns(self, patterns, casing=IGNORE_CASE, extra_patterns=None): valid_patterns = [] - if isinstance(patterns, string_types): + if isinstance(patterns, str): patterns = [patterns] else: patterns = list(patterns) @@ -419,7 +418,7 @@ def _compile_tag_re(self): Compile regex strings from device_tag_re option and return list of compiled regex/tag pairs """ device_tag_list = [] - for regex_str, tags in iteritems(self._device_tag_re): + for regex_str, tags in self._device_tag_re.items(): try: device_tag_list.append([re.compile(regex_str, IGNORE_CASE), [t.strip() for t in tags.split(',')]]) except TypeError: diff --git a/ecs_fargate/datadog_checks/ecs_fargate/ecs_fargate.py b/ecs_fargate/datadog_checks/ecs_fargate/ecs_fargate.py index 9d55916a926cc..4ae362cc879b8 100644 --- a/ecs_fargate/datadog_checks/ecs_fargate/ecs_fargate.py +++ b/ecs_fargate/datadog_checks/ecs_fargate/ecs_fargate.py @@ -7,7 +7,6 @@ import requests from dateutil import parser -from six import iteritems from datadog_checks.base import AgentCheck from datadog_checks.base.utils.common import round_value @@ -192,7 +191,7 @@ def check(self, _): ## Ephemeral Storage Metrics if 'EphemeralStorageMetrics' in metadata: es_metrics = metadata['EphemeralStorageMetrics'] - for field_name, metric_value in iteritems(es_metrics): + for field_name, metric_value in es_metrics.items(): metric_name = EPHEMERAL_STORAGE_GAUGE_METRICS.get(field_name) self.gauge(metric_name, metric_value, task_tags) @@ -229,7 +228,7 @@ def check(self, _): self.service_check('fargate_check', AgentCheck.WARNING, message=msg, tags=custom_tags) self.log.warning(msg, exc_info=True) - for container_id, container_stats in iteritems(stats): + for container_id, container_stats in stats.items(): if container_id not in exlcuded_cid: self.submit_perf_metrics(container_tags, container_id, container_stats) @@ -337,7 +336,7 @@ def submit_perf_metrics(self, container_tags, container_id, container_stats): self.gauge('ecs.fargate.mem.limit', value, tags) # I/O metrics - for blkio_cat, metric_name in iteritems(IO_METRICS): + for blkio_cat, metric_name in IO_METRICS.items(): read_counter = write_counter = 0 blkio_stats = container_stats.get("blkio_stats", {}).get(blkio_cat) @@ -363,13 +362,13 @@ def submit_perf_metrics(self, container_tags, container_id, container_stats): # Network metrics networks = container_stats.get('networks', {}) - for network_interface, network_stats in iteritems(networks): + for network_interface, network_stats in networks.items(): network_tags = tags + ["interface:{}".format(network_interface)] - for field_name, metric_name in iteritems(NETWORK_GAUGE_METRICS): + for field_name, metric_name in NETWORK_GAUGE_METRICS.items(): metric_value = network_stats.get(field_name) if metric_value is not None: self.gauge(metric_name, metric_value, network_tags) - for field_name, metric_name in iteritems(NETWORK_RATE_METRICS): + for field_name, metric_name in NETWORK_RATE_METRICS.items(): metric_value = network_stats.get(field_name) if metric_value is not None: self.rate(metric_name, metric_value, network_tags) diff --git a/kube_apiserver_metrics/datadog_checks/kube_apiserver_metrics/kube_apiserver_metrics.py b/kube_apiserver_metrics/datadog_checks/kube_apiserver_metrics/kube_apiserver_metrics.py index e8287b258a261..d07d8b168d411 100644 --- a/kube_apiserver_metrics/datadog_checks/kube_apiserver_metrics/kube_apiserver_metrics.py +++ b/kube_apiserver_metrics/datadog_checks/kube_apiserver_metrics/kube_apiserver_metrics.py @@ -4,8 +4,6 @@ from copy import deepcopy from re import match, search, sub -from six import iteritems - from datadog_checks.base.checks.openmetrics import OpenMetricsBaseCheck from datadog_checks.base.errors import CheckException @@ -200,7 +198,7 @@ def submit_metric(self, metric_suffix, metric, scraper_config, gauge=True, monot # Explicit shallow copy of the instance tags _tags = list(scraper_config['custom_tags']) - for label_name, label_value in iteritems(sample[self.SAMPLE_LABELS]): + for label_name, label_value in sample[self.SAMPLE_LABELS].items(): _tags.append('{}:{}'.format(label_name, label_value)) if gauge: # submit raw metric diff --git a/mapreduce/datadog_checks/mapreduce/mapreduce.py b/mapreduce/datadog_checks/mapreduce/mapreduce.py index 893a6f3c63d48..31b850abaad47 100644 --- a/mapreduce/datadog_checks/mapreduce/mapreduce.py +++ b/mapreduce/datadog_checks/mapreduce/mapreduce.py @@ -1,12 +1,10 @@ # (C) Datadog, Inc. 2010-present # All rights reserved # Licensed under Simplified BSD License (see LICENSE) - +from urllib.parse import urljoin, urlsplit, urlunsplit from requests.exceptions import ConnectionError, HTTPError, InvalidURL, Timeout from simplejson import JSONDecodeError -from six import iteritems, itervalues -from six.moves.urllib.parse import urljoin, urlsplit, urlunsplit from datadog_checks.base import AgentCheck, ConfigurationError, is_affirmative from datadog_checks.mapreduce.metrics import ( @@ -93,7 +91,7 @@ def check(self, instance): # Report success after gathering all metrics from Application Master if running_jobs: - job_id, metrics = next(iteritems(running_jobs)) + job_id, metrics = next(iter(running_jobs.items())) am_address = self._get_url_base(metrics['tracking_url']) self.service_check( @@ -249,7 +247,7 @@ def _mapreduce_job_metrics(self, running_apps, addl_tags): """ running_jobs = {} - for app_name, tracking_url in itervalues(running_apps): + for app_name, tracking_url in running_apps.values(): metrics_json = self._rest_request_to_json( tracking_url, self.MAPREDUCE_JOBS_PATH, self.MAPREDUCE_SERVICE_CHECK @@ -289,7 +287,7 @@ def _mapreduce_job_counters_metrics(self, running_jobs, addl_tags): """ Get custom metrics specified for each counter """ - for job_metrics in itervalues(running_jobs): + for job_metrics in running_jobs.values(): job_name = job_metrics['job_name'] # Check if the job_name exist in the custom metrics @@ -344,7 +342,7 @@ def _mapreduce_task_metrics(self, running_jobs, addl_tags): Get metrics for each MapReduce task Return a dictionary of {task_id: 'tracking_url'} for each MapReduce task """ - for job_stats in itervalues(running_jobs): + for job_stats in running_jobs.values(): metrics_json = self._rest_request_to_json( job_stats['tracking_url'], 'tasks', self.MAPREDUCE_SERVICE_CHECK, tags=addl_tags @@ -376,7 +374,7 @@ def _set_metrics_from_json(self, metrics_json, metrics, tags): """ Parse the JSON response and set the metrics """ - for status, (metric_name, metric_type) in iteritems(metrics): + for status, (metric_name, metric_type) in metrics.items(): metric_status = metrics_json.get(status) if metric_status is not None: @@ -415,7 +413,7 @@ def _rest_request_to_json(self, address, object_path, service_name=None, tags=No # Add kwargs as arguments if kwargs: - query = '&'.join(['{}={}'.format(key, value) for key, value in iteritems(kwargs)]) + query = '&'.join(['{}={}'.format(key, value) for key, value in kwargs.items()]) url = urljoin(url, '?' + query) try: diff --git a/mcache/datadog_checks/mcache/mcache.py b/mcache/datadog_checks/mcache/mcache.py index 92d42af29c083..07e945c7ea161 100644 --- a/mcache/datadog_checks/mcache/mcache.py +++ b/mcache/datadog_checks/mcache/mcache.py @@ -4,7 +4,6 @@ from __future__ import division import bmemcached -from six import iteritems, itervalues from datadog_checks.base import AgentCheck, ConfigurationError @@ -111,7 +110,7 @@ def _process_response(self, response): if len(response) != 1: raise BadResponseError("Malformed response: {}".format(response)) - stats = list(itervalues(response))[0] + stats = list(response.values())[0] if not len(stats): raise BadResponseError("Malformed response for host: {}".format(stats)) @@ -177,7 +176,7 @@ def _get_metrics(self, client, tags, service_check_tags=None): raise def _get_optional_metrics(self, client, tags, options=None): - for arg, metrics_args in iteritems(self.OPTIONAL_STATS): + for arg, metrics_args in self.OPTIONAL_STATS.items(): if not options or options.get(arg, False): try: optional_rates = metrics_args[0] @@ -187,7 +186,7 @@ def _get_optional_metrics(self, client, tags, options=None): stats = self._process_response(client.stats(arg)) prefix = "memcache.{}".format(arg) - for metric, val in iteritems(stats): + for metric, val in stats.items(): # Check if metric is a gauge or rate metric_tags = [] if optional_fn: diff --git a/mesos_master/datadog_checks/mesos_master/mesos_master.py b/mesos_master/datadog_checks/mesos_master/mesos_master.py index e82cc2fb33c95..aaa2adc6fa489 100644 --- a/mesos_master/datadog_checks/mesos_master/mesos_master.py +++ b/mesos_master/datadog_checks/mesos_master/mesos_master.py @@ -6,9 +6,9 @@ Collects metrics from mesos master node, only the leader is sending metrics. """ +from urllib.parse import urlparse + import requests -from six import iteritems -from six.moves.urllib.parse import urlparse from datadog_checks.base import AgentCheck from datadog_checks.base.errors import CheckException @@ -305,7 +305,7 @@ def check(self, instance): framework_tags = ['framework_name:' + framework['name']] + tags self.GAUGE('mesos.framework.total_tasks', len(framework['tasks']), tags=framework_tags) resources = framework['used_resources'] - for key_name, (metric_name, metric_func) in iteritems(self.FRAMEWORK_METRICS): + for key_name, (metric_name, metric_func) in self.FRAMEWORK_METRICS.items(): metric_func(self, metric_name, resources[key_name], tags=framework_tags) role_metrics = self._get_master_roles(url, instance_tags) @@ -314,7 +314,7 @@ def check(self, instance): role_tags = ['mesos_role:' + role['name']] + tags self.GAUGE('mesos.role.frameworks.count', len(role['frameworks']), tags=role_tags) self.GAUGE('mesos.role.weight', role['weight'], tags=role_tags) - for key_name, (metric_name, metric_func) in iteritems(self.ROLE_RESOURCES_METRICS): + for key_name, (metric_name, metric_func) in self.ROLE_RESOURCES_METRICS.items(): try: metric_func(self, metric_name, role['resources'][key_name], tags=role_tags) except KeyError: @@ -335,7 +335,7 @@ def check(self, instance): self.STATS_METRICS, ] for m in metrics: - for key_name, (metric_name, metric_func) in iteritems(m): + for key_name, (metric_name, metric_func) in m.items(): if key_name in stats_metrics: metric_func(self, metric_name, stats_metrics[key_name], tags=tags) diff --git a/mongo/datadog_checks/mongo/collectors/base.py b/mongo/datadog_checks/mongo/collectors/base.py index 68d884ea1a81d..be88dfcf179af 100644 --- a/mongo/datadog_checks/mongo/collectors/base.py +++ b/mongo/datadog_checks/mongo/collectors/base.py @@ -4,14 +4,9 @@ import re -from six import PY3, iteritems - from datadog_checks.base import AgentCheck from datadog_checks.mongo.metrics import CASE_SENSITIVE_METRIC_NAME_SUFFIXES -if PY3: - long = int - class MongoCollector(object): """The base collector object, can be considered abstract. @@ -46,7 +41,7 @@ def _normalize(self, metric_name, submit_method, prefix=None): metric_suffix = "ps" if submit_method == AgentCheck.rate else "" # Replace case-sensitive metric name characters - for pattern, repl in iteritems(CASE_SENSITIVE_METRIC_NAME_SUFFIXES): + for pattern, repl in CASE_SENSITIVE_METRIC_NAME_SUFFIXES.items(): metric_name = re.compile(pattern).sub(repl, metric_name) # Normalize, and wrap @@ -93,11 +88,9 @@ def _submit_payload(self, payload, additional_tags=None, metrics_to_collect=None continue # value is now status[x][y][z] - if not isinstance(value, (int, long, float)): + if not isinstance(value, (int, float)): raise TypeError( - u"{0} value is a {1}, it should be an int, a float or a long instead.".format( - metric_name, type(value) - ) + u"{0} value is a {1}, it should be an int, or a float instead.".format(metric_name, type(value)) ) # Submit the metric diff --git a/mongo/datadog_checks/mongo/collectors/coll_stats.py b/mongo/datadog_checks/mongo/collectors/coll_stats.py index 68778a5546e16..02c8809641a24 100644 --- a/mongo/datadog_checks/mongo/collectors/coll_stats.py +++ b/mongo/datadog_checks/mongo/collectors/coll_stats.py @@ -3,7 +3,6 @@ # Licensed under a 3-clause BSD style license (see LICENSE) from pymongo.errors import OperationFailure -from six import iteritems from datadog_checks.base import AgentCheck from datadog_checks.mongo.collectors.base import MongoCollector @@ -96,7 +95,7 @@ def collect(self, api): # Submit the indexSizes metrics manually if index_sizes: metric_name_alias = self._normalize("collection.indexSizes", AgentCheck.gauge) - for idx, val in iteritems(index_sizes): + for idx, val in index_sizes.items(): # we tag the index idx_tags = self.base_tags + additional_tags + ["index:%s" % idx] self.gauge(metric_name_alias, val, tags=idx_tags) diff --git a/mongo/datadog_checks/mongo/collectors/top.py b/mongo/datadog_checks/mongo/collectors/top.py index fe51fd92ae8ad..9c7b186661c7d 100644 --- a/mongo/datadog_checks/mongo/collectors/top.py +++ b/mongo/datadog_checks/mongo/collectors/top.py @@ -1,9 +1,6 @@ # (C) Datadog, Inc. 2020-present # All rights reserved # Licensed under a 3-clause BSD style license (see LICENSE) - -from six import iteritems - from datadog_checks.mongo.collectors.base import MongoCollector from datadog_checks.mongo.common import MongosDeployment, ReplicaSetDeployment from datadog_checks.mongo.metrics import TOP_METRICS @@ -26,7 +23,7 @@ def compatible_with(self, deployment): def collect(self, api): dbtop = api["admin"].command('top') - for ns, ns_metrics in iteritems(dbtop['totals']): + for ns, ns_metrics in dbtop['totals'].items(): if "." not in ns: continue diff --git a/mysql/datadog_checks/mysql/collection_utils.py b/mysql/datadog_checks/mysql/collection_utils.py index f87c952227385..0846401c70038 100644 --- a/mysql/datadog_checks/mysql/collection_utils.py +++ b/mysql/datadog_checks/mysql/collection_utils.py @@ -3,8 +3,6 @@ # Licensed under Simplified BSD License (see LICENSE) import logging -from six import iteritems, text_type - log = logging.getLogger(__name__) @@ -12,7 +10,7 @@ def collect_all_scalars(key, dictionary): if key not in dictionary or dictionary[key] is None: yield None, None elif isinstance(dictionary[key], dict): - for tag, _ in iteritems(dictionary[key]): + for tag, _ in dictionary[key].items(): yield tag, collect_type(tag, dictionary[key], float) else: yield None, collect_type(key, dictionary, float) @@ -23,7 +21,7 @@ def collect_scalar(key, mapping): def collect_string(key, mapping): - return collect_type(key, mapping, text_type) + return collect_type(key, mapping, str) def collect_type(key, mapping, the_type): diff --git a/mysql/datadog_checks/mysql/innodb_metrics.py b/mysql/datadog_checks/mysql/innodb_metrics.py index cd4debe4d8448..b33bcc3868ae1 100644 --- a/mysql/datadog_checks/mysql/innodb_metrics.py +++ b/mysql/datadog_checks/mysql/innodb_metrics.py @@ -6,7 +6,6 @@ from contextlib import closing import pymysql -from six import PY3, iteritems from datadog_checks.base import is_affirmative from datadog_checks.base.log import get_check_logger @@ -15,9 +14,6 @@ from .collection_utils import collect_scalar from .const import OPTIONAL_INNODB_VARS -if PY3: - long = int - def _are_values_numeric(array): return all(v.isdigit() for v in array) @@ -79,38 +75,38 @@ def get_stats_from_innodb_status(self, db): row = [item.strip(']') for item in row] if line.startswith('---BUFFER POOL'): - buffer_id = long(row[2]) + buffer_id = int(row[2]) # SEMAPHORES if line.find('Mutex spin waits') == 0: # Mutex spin waits 79626940, rounds 157459864, OS waits 698719 # Mutex spin waits 0, rounds 247280272495, OS waits 316513438 - results['Innodb_mutex_spin_waits'] = long(row[3]) - results['Innodb_mutex_spin_rounds'] = long(row[5]) - results['Innodb_mutex_os_waits'] = long(row[8]) + results['Innodb_mutex_spin_waits'] = int(row[3]) + results['Innodb_mutex_spin_rounds'] = int(row[5]) + results['Innodb_mutex_os_waits'] = int(row[8]) elif line.find('RW-shared spins') == 0 and line.find(';') > 0: # RW-shared spins 3859028, OS waits 2100750; RW-excl spins # 4641946, OS waits 1530310 - results['Innodb_s_lock_spin_waits'] = long(row[2]) - results['Innodb_x_lock_spin_waits'] = long(row[8]) - results['Innodb_s_lock_os_waits'] = long(row[5]) - results['Innodb_x_lock_os_waits'] = long(row[11]) + results['Innodb_s_lock_spin_waits'] = int(row[2]) + results['Innodb_x_lock_spin_waits'] = int(row[8]) + results['Innodb_s_lock_os_waits'] = int(row[5]) + results['Innodb_x_lock_os_waits'] = int(row[11]) elif line.find('RW-shared spins') == 0 and line.find('; RW-excl spins') == -1: # Post 5.5.17 SHOW ENGINE INNODB STATUS syntax # RW-shared spins 604733, rounds 8107431, OS waits 241268 - results['Innodb_s_lock_spin_waits'] = long(row[2]) - results['Innodb_s_lock_spin_rounds'] = long(row[4]) - results['Innodb_s_lock_os_waits'] = long(row[7]) + results['Innodb_s_lock_spin_waits'] = int(row[2]) + results['Innodb_s_lock_spin_rounds'] = int(row[4]) + results['Innodb_s_lock_os_waits'] = int(row[7]) elif line.find('RW-excl spins') == 0: # Post 5.5.17 SHOW ENGINE INNODB STATUS syntax # RW-excl spins 604733, rounds 8107431, OS waits 241268 - results['Innodb_x_lock_spin_waits'] = long(row[2]) - results['Innodb_x_lock_spin_rounds'] = long(row[4]) - results['Innodb_x_lock_os_waits'] = long(row[7]) + results['Innodb_x_lock_spin_waits'] = int(row[2]) + results['Innodb_x_lock_spin_rounds'] = int(row[4]) + results['Innodb_x_lock_os_waits'] = int(row[7]) elif line.find('seconds the semaphore:') > 0: # --Thread 907205 has waited at handler/ha_innodb.cc line 7156 for 1.00 seconds the semaphore: results['Innodb_semaphore_waits'] += 1 - results['Innodb_semaphore_wait_time'] += long(float(row[9])) * 1000 + results['Innodb_semaphore_wait_time'] += int(float(row[9])) * 1000 # TRANSACTIONS elif line.find('Trx id counter') == 0: @@ -121,7 +117,7 @@ def get_stats_from_innodb_status(self, db): txn_seen = True elif line.find('History list length') == 0: # History list length 132 - results['Innodb_history_list_length'] = long(row[3]) + results['Innodb_history_list_length'] = int(row[3]) elif txn_seen and line.find('---TRANSACTION') == 0: # ---TRANSACTION 0, not started, process no 13510, OS thread id 1170446656 results['Innodb_current_transactions'] += 1 @@ -129,76 +125,76 @@ def get_stats_from_innodb_status(self, db): results['Innodb_active_transactions'] += 1 elif line.find('read views open inside InnoDB') > 0: # 1 read views open inside InnoDB - results['Innodb_read_views'] = long(row[0]) + results['Innodb_read_views'] = int(row[0]) elif line.find('mysql tables in use') == 0: # mysql tables in use 2, locked 2 - results['Innodb_tables_in_use'] += long(row[4]) - results['Innodb_locked_tables'] += long(row[6]) + results['Innodb_tables_in_use'] += int(row[4]) + results['Innodb_locked_tables'] += int(row[6]) elif txn_seen and line.find('lock struct(s)') > 0: # 23 lock struct(s), heap size 3024, undo log entries 27 # LOCK WAIT 12 lock struct(s), heap size 3024, undo log entries 5 # LOCK WAIT 2 lock struct(s), heap size 368 if line.find('LOCK WAIT') == 0: - results['Innodb_lock_structs'] += long(row[2]) + results['Innodb_lock_structs'] += int(row[2]) results['Innodb_locked_transactions'] += 1 elif line.find('ROLLING BACK') == 0: # ROLLING BACK 127539 lock struct(s), heap size 15201832, # 4411492 row lock(s), undo log entries 1042488 - results['Innodb_lock_structs'] += long(row[2]) + results['Innodb_lock_structs'] += int(row[2]) else: - results['Innodb_lock_structs'] += long(row[0]) + results['Innodb_lock_structs'] += int(row[0]) # FILE I/O elif line.find(' OS file reads, ') > 0: # 8782182 OS file reads, 15635445 OS file writes, 947800 OS # fsyncs - results['Innodb_os_file_reads'] = long(row[0]) - results['Innodb_os_file_writes'] = long(row[4]) - results['Innodb_os_file_fsyncs'] = long(row[8]) + results['Innodb_os_file_reads'] = int(row[0]) + results['Innodb_os_file_writes'] = int(row[4]) + results['Innodb_os_file_fsyncs'] = int(row[8]) elif line.find('Pending normal aio reads:') == 0: try: if len(row) == 8: # (len(row) == 8) Pending normal aio reads: 0, aio writes: 0, - results['Innodb_pending_normal_aio_reads'] = long(row[4]) - results['Innodb_pending_normal_aio_writes'] = long(row[7]) + results['Innodb_pending_normal_aio_reads'] = int(row[4]) + results['Innodb_pending_normal_aio_writes'] = int(row[7]) elif len(row) == 14: # (len(row) == 14) Pending normal aio reads: 0 [0, 0] , aio writes: 0 [0, 0] , - results['Innodb_pending_normal_aio_reads'] = long(row[4]) - results['Innodb_pending_normal_aio_writes'] = long(row[10]) + results['Innodb_pending_normal_aio_reads'] = int(row[4]) + results['Innodb_pending_normal_aio_writes'] = int(row[10]) elif len(row) == 16: # (len(row) == 16) Pending normal aio reads: [0, 0, 0, 0] , aio writes: [0, 0, 0, 0] , if _are_values_numeric(row[4:8]) and _are_values_numeric(row[11:15]): results['Innodb_pending_normal_aio_reads'] = ( - long(row[4]) + long(row[5]) + long(row[6]) + long(row[7]) + int(row[4]) + int(row[5]) + int(row[6]) + int(row[7]) ) results['Innodb_pending_normal_aio_writes'] = ( - long(row[11]) + long(row[12]) + long(row[13]) + long(row[14]) + int(row[11]) + int(row[12]) + int(row[13]) + int(row[14]) ) # (len(row) == 16) Pending normal aio reads: 0 [0, 0, 0, 0] , aio writes: 0 [0, 0] , elif _are_values_numeric(row[4:9]) and _are_values_numeric(row[12:15]): - results['Innodb_pending_normal_aio_reads'] = long(row[4]) - results['Innodb_pending_normal_aio_writes'] = long(row[12]) + results['Innodb_pending_normal_aio_reads'] = int(row[4]) + results['Innodb_pending_normal_aio_writes'] = int(row[12]) else: self.log.warning("Can't parse result line %s", line) elif len(row) == 18: # (len(row) == 18) Pending normal aio reads: 0 [0, 0, 0, 0] , aio writes: 0 [0, 0, 0, 0] , - results['Innodb_pending_normal_aio_reads'] = long(row[4]) - results['Innodb_pending_normal_aio_writes'] = long(row[12]) + results['Innodb_pending_normal_aio_reads'] = int(row[4]) + results['Innodb_pending_normal_aio_writes'] = int(row[12]) elif len(row) == 22: # (len(row) == 22) # Pending normal aio reads: 0 [0, 0, 0, 0, 0, 0, 0, 0] , aio writes: 0 [0, 0, 0, 0] , - results['Innodb_pending_normal_aio_reads'] = long(row[4]) - results['Innodb_pending_normal_aio_writes'] = long(row[16]) + results['Innodb_pending_normal_aio_reads'] = int(row[4]) + results['Innodb_pending_normal_aio_writes'] = int(row[16]) except ValueError as e: self.log.warning("Can't parse result line %s: %s", line, e) elif line.find('ibuf aio reads') == 0: # ibuf aio reads: 0, log i/o's: 0, sync i/o's: 0 # or ibuf aio reads:, log i/o's:, sync i/o's: if len(row) == 10: - results['Innodb_pending_ibuf_aio_reads'] = long(row[3]) - results['Innodb_pending_aio_log_ios'] = long(row[6]) - results['Innodb_pending_aio_sync_ios'] = long(row[9]) + results['Innodb_pending_ibuf_aio_reads'] = int(row[3]) + results['Innodb_pending_aio_log_ios'] = int(row[6]) + results['Innodb_pending_aio_sync_ios'] = int(row[9]) elif len(row) == 7: results['Innodb_pending_ibuf_aio_reads'] = 0 results['Innodb_pending_aio_log_ios'] = 0 @@ -206,11 +202,11 @@ def get_stats_from_innodb_status(self, db): elif line.find('Pending flushes (fsync)') == 0: if len(row) == 4: # Pending flushes (fsync): 0 - results['Innodb_pending_buffer_pool_flushes'] = long(row[3]) + results['Innodb_pending_buffer_pool_flushes'] = int(row[3]) else: # Pending flushes (fsync) log: 0; buffer pool: 0 - results['Innodb_pending_log_flushes'] = long(row[4]) - results['Innodb_pending_buffer_pool_flushes'] = long(row[7]) + results['Innodb_pending_log_flushes'] = int(row[4]) + results['Innodb_pending_buffer_pool_flushes'] = int(row[7]) # INSERT BUFFER AND ADAPTIVE HASH INDEX elif line.find('Ibuf for space 0: size ') == 0: @@ -218,24 +214,24 @@ def get_stats_from_innodb_status(self, db): # had two lines in the output. Newer has just one line, see below. # Ibuf for space 0: size 1, free list len 887, seg size 889, is not empty # Ibuf for space 0: size 1, free list len 887, seg size 889, - results['Innodb_ibuf_size'] = long(row[5]) - results['Innodb_ibuf_free_list'] = long(row[9]) - results['Innodb_ibuf_segment_size'] = long(row[12]) + results['Innodb_ibuf_size'] = int(row[5]) + results['Innodb_ibuf_free_list'] = int(row[9]) + results['Innodb_ibuf_segment_size'] = int(row[12]) elif line.find('Ibuf: size ') == 0: # Ibuf: size 1, free list len 4634, seg size 4636, - results['Innodb_ibuf_size'] = long(row[2]) - results['Innodb_ibuf_free_list'] = long(row[6]) - results['Innodb_ibuf_segment_size'] = long(row[9]) + results['Innodb_ibuf_size'] = int(row[2]) + results['Innodb_ibuf_free_list'] = int(row[6]) + results['Innodb_ibuf_segment_size'] = int(row[9]) if line.find('merges') > -1: - results['Innodb_ibuf_merges'] = long(row[10]) + results['Innodb_ibuf_merges'] = int(row[10]) elif line.find(', delete mark ') > 0 and prev_line.find('merged operations:') == 0: # Output of show engine innodb status has changed in 5.5 # merged operations: # insert 593983, delete mark 387006, delete 73092 - results['Innodb_ibuf_merged_inserts'] = long(row[1]) - results['Innodb_ibuf_merged_delete_marks'] = long(row[4]) - results['Innodb_ibuf_merged_deletes'] = long(row[6]) + results['Innodb_ibuf_merged_inserts'] = int(row[1]) + results['Innodb_ibuf_merged_delete_marks'] = int(row[4]) + results['Innodb_ibuf_merged_deletes'] = int(row[6]) results['Innodb_ibuf_merged'] = ( results['Innodb_ibuf_merged_inserts'] + results['Innodb_ibuf_merged_delete_marks'] @@ -243,85 +239,85 @@ def get_stats_from_innodb_status(self, db): ) elif line.find(' merged recs, ') > 0: # 19817685 inserts, 19817684 merged recs, 3552620 merges - results['Innodb_ibuf_merged_inserts'] = long(row[0]) - results['Innodb_ibuf_merged'] = long(row[2]) - results['Innodb_ibuf_merges'] = long(row[5]) + results['Innodb_ibuf_merged_inserts'] = int(row[0]) + results['Innodb_ibuf_merged'] = int(row[2]) + results['Innodb_ibuf_merges'] = int(row[5]) elif line.find('Hash table size ') == 0: # In some versions of InnoDB, the used cells is omitted. # Hash table size 4425293, used cells 4229064, .... # Hash table size 57374437, node heap has 72964 buffer(s) <-- # no used cells - results['Innodb_hash_index_cells_total'] = long(row[3]) - results['Innodb_hash_index_cells_used'] = long(row[6]) if line.find('used cells') > 0 else 0 + results['Innodb_hash_index_cells_total'] = int(row[3]) + results['Innodb_hash_index_cells_used'] = int(row[6]) if line.find('used cells') > 0 else 0 # LOG elif line.find(" log i/o's done, ") > 0: # 3430041 log i/o's done, 17.44 log i/o's/second # 520835887 log i/o's done, 17.28 log i/o's/second, 518724686 # syncs, 2980893 checkpoints - results['Innodb_log_writes'] = long(row[0]) + results['Innodb_log_writes'] = int(row[0]) elif line.find(" pending log writes, ") > 0: # 0 pending log writes, 0 pending chkp writes - results['Innodb_pending_log_writes'] = long(row[0]) - results['Innodb_pending_checkpoint_writes'] = long(row[4]) + results['Innodb_pending_log_writes'] = int(row[0]) + results['Innodb_pending_checkpoint_writes'] = int(row[4]) elif line.find("Log sequence number") == 0: # This number is NOT printed in hex in InnoDB plugin. # Log sequence number 272588624 - results['Innodb_lsn_current'] = long(row[3]) + results['Innodb_lsn_current'] = int(row[3]) elif line.find("Log flushed up to") == 0: # This number is NOT printed in hex in InnoDB plugin. # Log flushed up to 272588624 - results['Innodb_lsn_flushed'] = long(row[4]) + results['Innodb_lsn_flushed'] = int(row[4]) elif line.find("Last checkpoint at") == 0: # Last checkpoint at 272588624 - results['Innodb_lsn_last_checkpoint'] = long(row[3]) + results['Innodb_lsn_last_checkpoint'] = int(row[3]) # BUFFER POOL AND MEMORY elif line.find("Total memory allocated") == 0 and line.find("in additional pool allocated") > 0: # Total memory allocated 29642194944; in additional pool allocated 0 # Total memory allocated by read views 96 - results['Innodb_mem_total'] = long(row[3]) - results['Innodb_mem_additional_pool'] = long(row[8]) + results['Innodb_mem_total'] = int(row[3]) + results['Innodb_mem_additional_pool'] = int(row[8]) elif line.find('Adaptive hash index ') == 0: # Adaptive hash index 1538240664 (186998824 + 1351241840) - results['Innodb_mem_adaptive_hash'] = long(row[3]) + results['Innodb_mem_adaptive_hash'] = int(row[3]) elif line.find('Page hash ') == 0: # Page hash 11688584 - results['Innodb_mem_page_hash'] = long(row[2]) + results['Innodb_mem_page_hash'] = int(row[2]) elif line.find('Dictionary cache ') == 0: # Dictionary cache 145525560 (140250984 + 5274576) - results['Innodb_mem_dictionary'] = long(row[2]) + results['Innodb_mem_dictionary'] = int(row[2]) elif line.find('File system ') == 0: # File system 313848 (82672 + 231176) - results['Innodb_mem_file_system'] = long(row[2]) + results['Innodb_mem_file_system'] = int(row[2]) elif line.find('Lock system ') == 0: # Lock system 29232616 (29219368 + 13248) - results['Innodb_mem_lock_system'] = long(row[2]) + results['Innodb_mem_lock_system'] = int(row[2]) elif line.find('Recovery system ') == 0: # Recovery system 0 (0 + 0) - results['Innodb_mem_recovery_system'] = long(row[2]) + results['Innodb_mem_recovery_system'] = int(row[2]) elif line.find('Threads ') == 0: # Threads 409336 (406936 + 2400) - results['Innodb_mem_thread_hash'] = long(row[1]) + results['Innodb_mem_thread_hash'] = int(row[1]) elif line.find("Buffer pool size ") == 0: # The " " after size is necessary to avoid matching the wrong line: # Buffer pool size 1769471 # Buffer pool size, bytes 28991012864 if buffer_id == -1: - results['Innodb_buffer_pool_pages_total'] = long(row[3]) + results['Innodb_buffer_pool_pages_total'] = int(row[3]) elif line.find("Free buffers") == 0: # Free buffers 0 if buffer_id == -1: - results['Innodb_buffer_pool_pages_free'] = long(row[2]) + results['Innodb_buffer_pool_pages_free'] = int(row[2]) elif line.find("Database pages") == 0: # Database pages 1696503 if buffer_id == -1: - results['Innodb_buffer_pool_pages_data'] = long(row[2]) + results['Innodb_buffer_pool_pages_data'] = int(row[2]) elif line.find("Modified db pages") == 0: # Modified db pages 160602 if buffer_id == -1: - results['Innodb_buffer_pool_pages_dirty'] = long(row[3]) + results['Innodb_buffer_pool_pages_dirty'] = int(row[3]) elif line.find("Pages read ahead") == 0: # Must do this BEFORE the next test, otherwise it'll get fooled by this # line from the new plugin: @@ -330,22 +326,22 @@ def get_stats_from_innodb_status(self, db): elif line.find("Pages read") == 0: # Pages read 15240822, created 1770238, written 21705836 if buffer_id == -1: - results['Innodb_pages_read'] = long(row[2]) - results['Innodb_pages_created'] = long(row[4]) - results['Innodb_pages_written'] = long(row[6]) + results['Innodb_pages_read'] = int(row[2]) + results['Innodb_pages_created'] = int(row[4]) + results['Innodb_pages_written'] = int(row[6]) # ROW OPERATIONS elif line.find('Number of rows inserted') == 0: # Number of rows inserted 50678311, updated 66425915, deleted # 20605903, read 454561562 - results['Innodb_rows_inserted'] = long(row[4]) - results['Innodb_rows_updated'] = long(row[6]) - results['Innodb_rows_deleted'] = long(row[8]) - results['Innodb_rows_read'] = long(row[10]) + results['Innodb_rows_inserted'] = int(row[4]) + results['Innodb_rows_updated'] = int(row[6]) + results['Innodb_rows_deleted'] = int(row[8]) + results['Innodb_rows_read'] = int(row[10]) elif line.find(" queries inside InnoDB, ") > 0: # 0 queries inside InnoDB, 0 queries in queue - results['Innodb_queries_inside'] = long(row[0]) - results['Innodb_queries_queued'] = long(row[4]) + results['Innodb_queries_inside'] = int(row[0]) + results['Innodb_queries_queued'] = int(row[4]) prev_line = line @@ -357,7 +353,7 @@ def get_stats_from_innodb_status(self, db): # Finally we change back the metrics values to string to make the values # consistent with how they are reported by SHOW GLOBAL STATUS - for metric, value in list(iteritems(results)): + for metric, value in list(results.items()): results[metric] = str(value) return results diff --git a/mysql/datadog_checks/mysql/mysql.py b/mysql/datadog_checks/mysql/mysql.py index a4c2eeb6a4ec2..de45c43013417 100644 --- a/mysql/datadog_checks/mysql/mysql.py +++ b/mysql/datadog_checks/mysql/mysql.py @@ -13,7 +13,6 @@ import pymysql from cachetools import TTLCache -from six import PY3, iteritems, itervalues from datadog_checks.base import AgentCheck, is_affirmative from datadog_checks.base.utils.db import QueryExecutor, QueryManager @@ -96,10 +95,6 @@ from ..stubs import datadog_agent -if PY3: - long = int - - class MySql(AgentCheck): SERVICE_CHECK_NAME = 'mysql.can_connect' SLAVE_SERVICE_CHECK_NAME = 'mysql.replication.slave_running' @@ -726,9 +721,9 @@ def _check_replication_status(self, results): if replica_sql_running is None: replica_sql_running = collect_type('Replica_SQL_Running', results, dict) if replica_io_running: - replica_io_running = any(v.lower().strip() == 'yes' for v in itervalues(replica_io_running)) + replica_io_running = any(v.lower().strip() == 'yes' for v in replica_io_running.values()) if replica_sql_running: - replica_sql_running = any(v.lower().strip() == 'yes' for v in itervalues(replica_sql_running)) + replica_sql_running = any(v.lower().strip() == 'yes' for v in replica_sql_running.values()) binlog_running = results.get('Binlog_enabled', False) # replicas will only be collected if user has PROCESS privileges. @@ -813,7 +808,7 @@ def _is_group_replication_active(self, db): return False def _submit_metrics(self, variables, db_results, tags): - for variable, metric in iteritems(variables): + for variable, metric in variables.items(): if isinstance(metric, list): for m in metric: metric_name, metric_type = m @@ -856,7 +851,7 @@ def _collect_dict(self, metric_type, field_metric_map, query, db, tags): cursor.execute(query) result = cursor.fetchone() if result is not None: - for field, metric in list(iteritems(field_metric_map)): + for field, metric in field_metric_map.items(): # Find the column name in the cursor description to identify the column index # http://www.python.org/dev/peps/pep-0249/ # cursor.description is a tuple of (column_name, ..., ...) @@ -903,7 +898,7 @@ def _get_runtime_aurora_tags(self, db): def _collect_system_metrics(self, host, db, tags): pid = None # The server needs to run locally, accessed by TCP or socket - if host in ["localhost", "127.0.0.1", "0.0.0.0"] or db.port == long(0): + if host in ["localhost", "127.0.0.1", "0.0.0.0"] or db.port == int(0): pid = self._get_server_pid(db) if pid: @@ -1021,7 +1016,7 @@ def _get_binary_log_stats(self, db): master_logs = {result[0]: result[1] for result in cursor_results} binary_log_space = 0 - for value in itervalues(master_logs): + for value in master_logs.values(): binary_log_space += value return binary_log_space @@ -1059,7 +1054,7 @@ def _get_replica_stats(self, db, is_mariadb, replication_channel): # MySQL <5.7 does not have Channel_Name. # For MySQL >=5.7 'Channel_Name' is set to an empty string by default channel = replication_channel or replica_result.get('Channel_Name') or 'default' - for key, value in iteritems(replica_result): + for key, value in replica_result.items(): if value is not None: replica_results[key]['channel:{0}'.format(channel)] = value except (pymysql.err.InternalError, pymysql.err.OperationalError) as e: @@ -1161,7 +1156,7 @@ def _query_exec_time_per_schema(self, db): schema_query_avg_run_time = {} for row in cursor.fetchall(): schema_name = str(row[0]) - avg_us = long(row[1]) + avg_us = int(row[1]) # set the tag as the dictionary key schema_query_avg_run_time["schema:{0}".format(schema_name)] = avg_us @@ -1216,7 +1211,7 @@ def _query_size_per_schema(self, db): schema_size = {} for row in cursor.fetchall(): schema_name = str(row[0]) - size = long(row[1]) + size = int(row[1]) # set the tag as the dictionary key schema_size["schema:{0}".format(schema_name)] = size @@ -1241,8 +1236,8 @@ def _query_rows_stats_per_table(self, db): for row in cursor.fetchall(): table_schema = str(row[0]) table_name = str(row[1]) - rows_read_total = long(row[2]) - rows_changed_total = long(row[3]) + rows_read_total = int(row[2]) + rows_changed_total = int(row[3]) # set the tag as the dictionary key table_rows_read_total["schema:{},table:{}".format(table_schema, table_name)] = rows_read_total diff --git a/nginx/datadog_checks/nginx/nginx.py b/nginx/datadog_checks/nginx/nginx.py index 0f7006abfd52f..3dc525a9eea7c 100644 --- a/nginx/datadog_checks/nginx/nginx.py +++ b/nginx/datadog_checks/nginx/nginx.py @@ -4,10 +4,9 @@ import re from datetime import datetime from itertools import chain +from urllib.parse import urljoin, urlparse import simplejson as json -from six import PY3, iteritems, text_type -from six.moves.urllib.parse import urljoin, urlparse from datadog_checks.base import AgentCheck, ConfigurationError, to_native_string from datadog_checks.base.utils.time import get_timestamp @@ -15,9 +14,6 @@ from .const import PLUS_API_ENDPOINTS, PLUS_API_STREAM_ENDPOINTS, TAGGED_KEYS from .metrics import COUNT_METRICS, METRICS_SEND_AS_COUNT, METRICS_SEND_AS_HISTOGRAM, VTS_METRIC_MAP -if PY3: - long = int - if hasattr(datetime, 'fromisoformat'): fromisoformat = datetime.fromisoformat else: @@ -133,7 +129,7 @@ def _get_enabled_endpoints(self): supported_endpoints = self._supported_endpoints(available_endpoints) self.log.debug("Supported endpoints are %s", supported_endpoints) - return chain(iteritems(supported_endpoints)) + return chain(supported_endpoints.items()) except Exception as e: self.log.warning( "Could not determine available endpoints from the API, " @@ -250,13 +246,13 @@ def _get_plus_api_endpoints(self, use_stream=False): Returns all of either stream or default endpoints that the integration supports collecting metrics from based on the Plus API version """ - endpoints = iteritems({}) + endpoints = iter([]) available_plus_endpoints = PLUS_API_STREAM_ENDPOINTS if use_stream else PLUS_API_ENDPOINTS for earliest_version, new_endpoints in available_plus_endpoints.items(): if int(self.plus_api_version) >= int(earliest_version): - endpoints = chain(endpoints, iteritems(new_endpoints)) + endpoints = chain(endpoints, new_endpoints.items()) return endpoints def _get_all_plus_api_endpoints(self): @@ -370,10 +366,10 @@ def _flatten_json(cls, metric_base, val, tags): if tags is None: tags = [] tags = tags + [server] - for key, val2 in iteritems(val): + for key, val2 in val.items(): if key in TAGGED_KEYS: metric_name = '%s.%s' % (metric_base, TAGGED_KEYS[key]) - for tag_val, data in iteritems(val2): + for tag_val, data in val2.items(): tag = '%s:%s' % (TAGGED_KEYS[key], tag_val) output.extend(cls._flatten_json(metric_name, data, tags + [tag])) else: @@ -387,10 +383,10 @@ def _flatten_json(cls, metric_base, val, tags): elif isinstance(val, bool): output.append((metric_base, int(val), tags, 'gauge')) - elif isinstance(val, (int, float, long)): + elif isinstance(val, (int, float)): output.append((metric_base, val, tags, 'gauge')) - elif isinstance(val, (text_type, str)) and val[-1] == "Z": + elif isinstance(val, str) and val[-1] == "Z": try: # In the new Plus API, timestamps are now formatted # strings, some include microseconds, some don't... diff --git a/redisdb/datadog_checks/redisdb/redisdb.py b/redisdb/datadog_checks/redisdb/redisdb.py index 4155bcdb03e81..86298169a0b26 100644 --- a/redisdb/datadog_checks/redisdb/redisdb.py +++ b/redisdb/datadog_checks/redisdb/redisdb.py @@ -9,7 +9,6 @@ from copy import deepcopy import redis -from six import PY2, iteritems from datadog_checks.base import AgentCheck, ConfigurationError, ensure_unicode, is_affirmative from datadog_checks.base.utils.common import round_value @@ -266,7 +265,7 @@ def _check_db(self): elif info_name in self.RATE_KEYS: self.rate(self.RATE_KEYS[info_name], info[info_name], tags=tags) - for config_key, value in iteritems(config): + for config_key, value in config.items(): metric_name = self.CONFIG_GAUGE_KEYS.get(config_key) if metric_name is not None: self.gauge(metric_name, value, tags=tags) @@ -404,7 +403,7 @@ def _check_key_lengths(self, conn, tags): lengths[text_key]["key_type"] = key_type # Send the metrics for each db in the redis instance. - for key, total in iteritems(lengths): + for key, total in lengths.items(): # Only send non-zeros if tagged per db. if total["length"] > 0: self.gauge( @@ -416,7 +415,7 @@ def _check_key_lengths(self, conn, tags): # Warn if a key is missing from the entire redis instance. # Send 0 if the key is missing/empty from the entire redis instance. - for key, total in iteritems(lengths_overall): + for key, total in lengths_overall.items(): if total == 0: key_tags = ['key:{}'.format(key)] if instance_db: @@ -552,7 +551,7 @@ def _check_command_stats(self, conn, tags): self.warning('Could not retrieve command stats from Redis. INFO COMMANDSTATS only works with Redis >= 2.6.') return - for key, stats in iteritems(command_stats): + for key, stats in command_stats.items(): command = key.split('_', 1)[1] command_tags = tags + ['command:{}'.format(command)] @@ -572,11 +571,8 @@ def _collect_metadata(self, info): self.set_metadata('version', info['redis_version']) -_timer = time.time if PY2 else time.perf_counter - - def _call_and_time(func): - start_time = _timer() + start_time = time.perf_counter() rv = func() - end_time = _timer() + end_time = time.perf_counter() return rv, round_value((end_time - start_time) * 1000, 2) diff --git a/riakcs/datadog_checks/riakcs/riakcs.py b/riakcs/datadog_checks/riakcs/riakcs.py index 5f5f12ec8dffe..8185985da4140 100644 --- a/riakcs/datadog_checks/riakcs/riakcs.py +++ b/riakcs/datadog_checks/riakcs/riakcs.py @@ -8,7 +8,6 @@ import boto3 import simplejson as json from botocore.config import Config -from six import iteritems from datadog_checks.base import AgentCheck from datadog_checks.base.config import _is_affirmative @@ -22,7 +21,7 @@ def multidict(ordered_pairs): d[k].append(v) # unpack lists that have only 1 item dict_copy = deepcopy(d) - for k, v in iteritems(dict_copy): + for k, v in dict_copy.items(): if len(v) == 1: d[k] = v[0] return dict(d) @@ -62,7 +61,7 @@ def process_stats(self, stats, tags, metrics): metrics.update(V21_DEFAULT_METRICS) else: metrics = V21_DEFAULT_METRICS - for key, value in iteritems(stats): + for key, value in stats.items(): if key not in metrics: continue suffix = key.rsplit("_", 1)[-1] @@ -72,7 +71,7 @@ def process_stats(self, stats, tags, metrics): # pre 2.1 stats format legends = {len(k): k for k in stats["legend"]} del stats["legend"] - for key, values in iteritems(stats): + for key, values in stats.items(): legend = legends[len(values)] for i, value in enumerate(values): metric_name = "riakcs.{0}.{1}".format(key, legend[i]) diff --git a/sap_hana/datadog_checks/sap_hana/sap_hana.py b/sap_hana/datadog_checks/sap_hana/sap_hana.py index d0f173b6c316e..227b93a411a39 100644 --- a/sap_hana/datadog_checks/sap_hana/sap_hana.py +++ b/sap_hana/datadog_checks/sap_hana/sap_hana.py @@ -18,8 +18,6 @@ from hdbcli.dbapi import Connection as HanaConnection except ImportError: HanaConnection = None -from six import iteritems -from six.moves import zip from datadog_checks.base import AgentCheck, is_affirmative from datadog_checks.base.utils.common import total_time_to_temporal_percent @@ -265,7 +263,7 @@ def query_connection_overview(self): for conn in self.iter_rows(queries.GlobalSystemConnectionsStatus(schema=self._schema)): db_counts[(conn['db_name'], conn['host'], conn['port'])][conn['status'].lower()] += conn['total'] - for (db, host, port), counts in iteritems(db_counts): + for (db, host, port), counts in db_counts.items(): tags = ['db:{}'.format(db), 'hana_port:{}'.format(port)] tags.extend(self._tags) tags.append('hana_host:{}'.format(host)) diff --git a/spark/datadog_checks/spark/constants.py b/spark/datadog_checks/spark/constants.py index d66b3b410977d..4097fce403f4f 100644 --- a/spark/datadog_checks/spark/constants.py +++ b/spark/datadog_checks/spark/constants.py @@ -3,8 +3,6 @@ # Licensed under a 3-clause BSD style license (see LICENSE) import re -from six import iteritems - # Identifier for cluster master address in `spark.yaml` MASTER_ADDRESS = 'spark_url' DEPRECATED_MASTER_ADDRESS = 'resourcemanager_uri' @@ -137,15 +135,15 @@ } SPARK_DRIVER_METRICS = { - key: (value[0].format('driver'), value[1]) for key, value in iteritems(SPARK_EXECUTOR_TEMPLATE_METRICS) + key: (value[0].format('driver'), value[1]) for key, value in SPARK_EXECUTOR_TEMPLATE_METRICS.items() } SPARK_EXECUTOR_METRICS = { - key: (value[0].format('executor'), value[1]) for key, value in iteritems(SPARK_EXECUTOR_TEMPLATE_METRICS) + key: (value[0].format('executor'), value[1]) for key, value in SPARK_EXECUTOR_TEMPLATE_METRICS.items() } SPARK_EXECUTOR_LEVEL_METRICS = { - key: (value[0].format('executor.id'), value[1]) for key, value in iteritems(SPARK_EXECUTOR_TEMPLATE_METRICS) + key: (value[0].format('executor.id'), value[1]) for key, value in SPARK_EXECUTOR_TEMPLATE_METRICS.items() } SPARK_RDD_METRICS = { diff --git a/squid/datadog_checks/squid/squid.py b/squid/datadog_checks/squid/squid.py index 42b2ead2f8406..a11875e6a74bb 100644 --- a/squid/datadog_checks/squid/squid.py +++ b/squid/datadog_checks/squid/squid.py @@ -5,9 +5,7 @@ import re import requests -from six import iteritems -# project from datadog_checks.base import AgentCheck EVENT_TYPE = SOURCE_TYPE_NAME = 'squid' @@ -100,7 +98,7 @@ def check(self, instance): counters = self.get_counters(host, port, tags + custom_tags) # Send these values as rate - for counter, value in iteritems(counters): + for counter, value in counters.items(): self.rate(counter, value, tags=tags + custom_tags) def get_counters(self, host, port, tags): diff --git a/system_core/datadog_checks/system_core/system_core.py b/system_core/datadog_checks/system_core/system_core.py index 8ab0d0a452b9a..979836880a5aa 100644 --- a/system_core/datadog_checks/system_core/system_core.py +++ b/system_core/datadog_checks/system_core/system_core.py @@ -2,7 +2,6 @@ # All rights reserved # Licensed under a 3-clause BSD style license (see LICENSE) import psutil -from six import iteritems from datadog_checks.base import AgentCheck from datadog_checks.base.utils.platform import Platform @@ -22,11 +21,11 @@ def check(self, instance): for i, cpu in enumerate(cpu_times): tags = instance_tags + ['core:{0}'.format(i)] - for key, value in iteritems(cpu._asdict()): + for key, value in cpu._asdict().items(): self.rate('system.core.{0}'.format(key), 100.0 * value, tags=tags) total_cpu_times = psutil.cpu_times() - for key, value in iteritems(total_cpu_times._asdict()): + for key, value in total_cpu_times._asdict().items(): self.rate('system.core.{0}.total'.format(key), 100.0 * value / n_cpus, tags=instance_tags) # https://psutil.readthedocs.io/en/latest/#psutil.cpu_freq diff --git a/twemproxy/datadog_checks/twemproxy/twemproxy.py b/twemproxy/datadog_checks/twemproxy/twemproxy.py index 303357af28eb6..b6a2534a5eadc 100644 --- a/twemproxy/datadog_checks/twemproxy/twemproxy.py +++ b/twemproxy/datadog_checks/twemproxy/twemproxy.py @@ -4,7 +4,6 @@ import socket import simplejson as json -from six import iteritems from datadog_checks.base import AgentCheck, ensure_unicode @@ -158,11 +157,11 @@ def parse_json(cls, raw, tags=None): version = parsed.get('version', None) - for key, val in iteritems(parsed): + for key, val in parsed.items(): if isinstance(val, dict): # server pool pool_tags = tags + ['pool:%s' % key] - for server_key, server_val in iteritems(val): + for server_key, server_val in val.items(): if isinstance(server_val, dict): # server server_tags = pool_tags + ['server:%s' % server_key] diff --git a/twistlock/datadog_checks/twistlock/twistlock.py b/twistlock/datadog_checks/twistlock/twistlock.py index e8dd42f60fdb7..f543ee88d733f 100644 --- a/twistlock/datadog_checks/twistlock/twistlock.py +++ b/twistlock/datadog_checks/twistlock/twistlock.py @@ -7,7 +7,6 @@ from datetime import datetime, timedelta from dateutil import parser, tz -from six import iteritems from datadog_checks.base import AgentCheck from datadog_checks.base.utils.serialization import json @@ -284,7 +283,7 @@ def _report_vuln_info(self, namespace, data, tags): cve_tags += ["package:{}".format(cve['packageName'])] self.gauge('{}.cve.details'.format(namespace), float(1), cve_tags) # Send counts to avoid no-data on zeroes - for severity, count in iteritems(summary): + for severity, count in summary.items(): cve_tags = SEVERITY_TAGS.get(severity, []) + tags self.gauge('{}.cve.count'.format(namespace), float(count), cve_tags) diff --git a/varnish/datadog_checks/varnish/varnish.py b/varnish/datadog_checks/varnish/varnish.py index 57bd381452d11..f617cbf27650e 100644 --- a/varnish/datadog_checks/varnish/varnish.py +++ b/varnish/datadog_checks/varnish/varnish.py @@ -8,16 +8,12 @@ from os import geteuid from packaging.version import Version -from six import PY3, iteritems from six.moves import filter from datadog_checks.base import ConfigurationError from datadog_checks.base.checks import AgentCheck from datadog_checks.base.utils.subprocess_output import get_subprocess_output -if PY3: - long = int - class BackendStatus(object): HEALTHY = 'healthy' @@ -100,11 +96,11 @@ def _end_element(self, name): if name == "stat": m_name = self.normalize(self._current_metric) if self._current_type in ("a", "c"): - self.rate(m_name, long(self._current_value), tags=self.tags) + self.rate(m_name, int(self._current_value), tags=self.tags) elif self._current_type in ("i", "g"): - self.gauge(m_name, long(self._current_value), tags=self.tags) + self.gauge(m_name, int(self._current_value), tags=self.tags) if 'n_purges' in m_name: - self.rate('varnish.n_purgesps', long(self._current_value), tags=self.tags) + self.rate('varnish.n_purgesps', int(self._current_value), tags=self.tags) else: # Unsupported data type, ignore self._reset() @@ -120,7 +116,7 @@ def _char_data(self, data): data = data.strip() if len(data) > 0 and self._current_element != "": if self._current_element == "value": - self._current_value = long(data) + self._current_value = int(data) elif self._current_element == "flag": self._current_type = data else: @@ -243,7 +239,7 @@ def _parse_varnishstat(self, output, varnishstat_format): json_output = json.loads(output) if "counters" in json_output: json_output = json_output["counters"] - for name, metric in iteritems(json_output): + for name, metric in json_output.items(): if not isinstance(metric, dict): # skip 'timestamp' field continue @@ -254,11 +250,11 @@ def _parse_varnishstat(self, output, varnishstat_format): value = metric.get("value", 0) if metric.get("flag") in ("a", "c"): - self.rate(metric_name, long(value), tags=self.tags) + self.rate(metric_name, int(value), tags=self.tags) elif metric.get("flag") in ("g", "i"): - self.gauge(metric_name, long(value), tags=self.tags) + self.gauge(metric_name, int(value), tags=self.tags) if 'n_purges' in self.normalize(name, prefix="varnish"): - self.rate('varnish.n_purgesps', long(value), tags=self.tags) + self.rate('varnish.n_purgesps', int(value), tags=self.tags) elif 'flag' not in metric: self.log.warning("Could not determine the type of metric %s, skipping submission", metric_name) self.log.debug("Raw metric %s is missing the `flag` field", str(metric)) @@ -367,7 +363,7 @@ def _submit_backend_service_checks(self, backends_by_status): if backends_by_status is None: return - for status, backends in iteritems(backends_by_status): + for status, backends in backends_by_status.items(): check_status = BackendStatus.to_check_status(status) for backend, message in backends: service_checks_tags = ['backend:%s' % backend] + self.custom_tags diff --git a/vsphere/datadog_checks/vsphere/api_rest.py b/vsphere/datadog_checks/vsphere/api_rest.py index 5de5d17b78f8d..cdf6df833c0c4 100644 --- a/vsphere/datadog_checks/vsphere/api_rest.py +++ b/vsphere/datadog_checks/vsphere/api_rest.py @@ -6,7 +6,6 @@ from typing import Any, Dict, Iterator, List, Set # noqa: F401 from pyVmomi import vim -from six import iteritems from datadog_checks.base.log import CheckLoggingAdapter # noqa: F401 from datadog_checks.base.utils.http import RequestsWrapper @@ -24,7 +23,7 @@ 'ClusterComputeResource': vim.ClusterComputeResource, } -MOR_TYPE_MAPPING_TO_STRING = {v: k for k, v in iteritems(MOR_TYPE_MAPPING_FROM_STRING)} +MOR_TYPE_MAPPING_TO_STRING = {v: k for k, v in MOR_TYPE_MAPPING_FROM_STRING.items()} class VSphereRestAPI(object): diff --git a/vsphere/datadog_checks/vsphere/config.py b/vsphere/datadog_checks/vsphere/config.py index 3cda46f780e52..366b091f3117d 100644 --- a/vsphere/datadog_checks/vsphere/config.py +++ b/vsphere/datadog_checks/vsphere/config.py @@ -6,7 +6,6 @@ from typing import Any, Dict, List # noqa: F401 from pyVmomi import vim -from six import iteritems, string_types from datadog_checks.base import ConfigurationError, is_affirmative from datadog_checks.base.log import CheckLoggingAdapter # noqa: F401 @@ -208,9 +207,7 @@ def _parse_resource_filters(self, all_resource_filters): ) # Check required fields and their types - for field, field_type in iteritems( - {'resource': string_types, 'property': string_types, 'type': string_types, 'patterns': list} - ): + for field, field_type in {'resource': str, 'property': str, 'type': str, 'patterns': list}.items(): if field not in resource_filter: self.log.warning( "Ignoring filter %r because it doesn't contain a %s field.", resource_filter, field @@ -282,7 +279,7 @@ def _parse_metric_regex_filters(self, all_metric_filters): # type: (MetricFilterConfig) -> MetricFilters allowed_resource_types = [MOR_TYPE_AS_STRING[k] for k in self.collected_resource_types] metric_filters = {} - for resource_type, filters in iteritems(all_metric_filters): + for resource_type, filters in all_metric_filters.items(): if resource_type not in allowed_resource_types: self.log.warning( "Ignoring metric_filter for resource '%s'. When collection_type is '%s', it should be one of '%s'", @@ -293,7 +290,7 @@ def _parse_metric_regex_filters(self, all_metric_filters): continue metric_filters[resource_type] = filters - return {k: [re.compile(r) for r in v] for k, v in iteritems(metric_filters)} + return {k: [re.compile(r) for r in v] for k, v in metric_filters.items()} def _normalize_event_resource_filters(self, filters): return [filter.lower() for filter in filters] diff --git a/vsphere/datadog_checks/vsphere/legacy/mor_cache.py b/vsphere/datadog_checks/vsphere/legacy/mor_cache.py index 276fa653a609f..2edb5081cf18b 100644 --- a/vsphere/datadog_checks/vsphere/legacy/mor_cache.py +++ b/vsphere/datadog_checks/vsphere/legacy/mor_cache.py @@ -4,8 +4,6 @@ import threading import time -from six import iteritems - from datadog_checks.vsphere.legacy.common import REALTIME_RESOURCES @@ -88,7 +86,7 @@ def mors(self, key): Generator returning all the mors in the cache for the given instance key. """ with self._mor_lock: - for k, v in iteritems(self._mor.get(key, {})): + for k, v in self._mor.get(key, {}).items(): yield k, v def mors_batch(self, key, batch_size, max_historical_metrics=None): @@ -111,7 +109,7 @@ def mors_batch(self, key, batch_size, max_historical_metrics=None): batch = {} nb_hist_metrics = 0 - for mor_name, mor in iteritems(mors_dict): + for mor_name, mor in mors_dict.items(): if mor['mor_type'] not in REALTIME_RESOURCES and mor.get('metrics'): # Those metrics are historical, let's make sure we don't have too # many of them in the same batch. @@ -182,7 +180,7 @@ def purge(self, key, ttl): with self._mor_lock: # Don't change the dict during iteration! # First collect the names of the Mors to remove... - for name, mor in iteritems(self._mor[key]): + for name, mor in self._mor[key].items(): age = now - mor['creation_time'] if age > ttl: mors_to_purge.append(name) diff --git a/vsphere/datadog_checks/vsphere/utils.py b/vsphere/datadog_checks/vsphere/utils.py index 136b041ed304b..99298b575cced 100644 --- a/vsphere/datadog_checks/vsphere/utils.py +++ b/vsphere/datadog_checks/vsphere/utils.py @@ -4,7 +4,6 @@ from typing import Any, Dict, List, Optional, Type # noqa: F401 from pyVmomi import vim -from six import iteritems from datadog_checks.base import to_string from datadog_checks.vsphere.constants import ( @@ -238,7 +237,7 @@ def get_mapped_instance_tag(metric_name): tag cannot be guessed by looking at the api results and has to be inferred using documentation or experience. This method acts as a utility to map a metric_name to the meaning of its instance tag. """ - for prefix, tag_key in iteritems(METRIC_TO_INSTANCE_TAG_MAPPING): + for prefix, tag_key in METRIC_TO_INSTANCE_TAG_MAPPING.items(): if metric_name.startswith(prefix): return tag_key return 'instance' diff --git a/vsphere/datadog_checks/vsphere/vsphere.py b/vsphere/datadog_checks/vsphere/vsphere.py index 80acdcb990ad9..f50cfdf73743c 100644 --- a/vsphere/datadog_checks/vsphere/vsphere.py +++ b/vsphere/datadog_checks/vsphere/vsphere.py @@ -11,7 +11,6 @@ from typing import Any, Callable, Dict, Generator, Iterable, List, Optional, Set, Type, cast # noqa: F401 from pyVmomi import vim, vmodl -from six import iteritems from datadog_checks.base import AgentCheck, is_affirmative, to_string from datadog_checks.base.checks.libs.timer import Timer @@ -182,7 +181,7 @@ def collect_tags(self, infrastructure_data): resource_filters_without_tags = [f for f in self._config.resource_filters if not isinstance(f, TagFilter)] filtered_infra_data = { mor: props - for mor, props in iteritems(infrastructure_data) + for mor, props in infrastructure_data.items() if isinstance(mor, tuple(self._config.collected_resource_types)) and is_resource_collected_by_filters(mor, infrastructure_data, resource_filters_without_tags) } @@ -234,7 +233,7 @@ def refresh_infrastructure_cache(self): all_tags = self.collect_tags(infrastructure_data) self.infrastructure_cache.set_all_tags(all_tags) - for mor, properties in iteritems(infrastructure_data): + for mor, properties in infrastructure_data.items(): if not isinstance(mor, tuple(self._config.collected_resource_types)): # Do nothing for the resource types we do not collect continue @@ -487,7 +486,7 @@ def make_query_specs(self): counters = self.metrics_metadata_cache.get_metadata(resource_type) metric_ids = [] # type: List[vim.PerformanceManager.MetricId] is_historical_batch = metric_type == HISTORICAL - for counter_key, metric_name in iteritems(counters): + for counter_key, metric_name in counters.items(): # PerformanceManager.MetricId `instance` kwarg: # - An asterisk (*) to specify all instances of the metric for the specified counterId # - Double-quotes ("") to specify aggregated statistics @@ -504,7 +503,7 @@ def make_query_specs(self): for batch in self.make_batch(mors, metric_ids, resource_type, is_historical_batch=is_historical_batch): query_specs = [] - for mor, metrics in iteritems(batch): + for mor, metrics in batch.items(): query_spec = vim.PerformanceManager.QuerySpec() # type: vim.PerformanceManager.QuerySpec query_spec.entity = mor query_spec.metricId = metrics diff --git a/vsphere/tests/legacy/utils.py b/vsphere/tests/legacy/utils.py index bee31bf3dc758..5842f9817fded 100644 --- a/vsphere/tests/legacy/utils.py +++ b/vsphere/tests/legacy/utils.py @@ -7,7 +7,6 @@ from mock import MagicMock, Mock from pyVmomi import vim -from six import iteritems HERE = os.path.abspath(os.path.dirname(__file__)) @@ -90,7 +89,7 @@ def assertMOR(check, instance, name=None, spec=None, tags=None, count=None, subs instance_name = instance['name'] candidates = [] - mor_list = [mor for _, mors in iteritems(check.mor_objects_queue._objects_queue[instance_name]) for mor in mors] + mor_list = [mor for mors in check.mor_objects_queue._objects_queue[instance_name].values() for mor in mors] for mor in mor_list: if name is not None and name != mor['hostname']: diff --git a/yarn/datadog_checks/yarn/yarn.py b/yarn/datadog_checks/yarn/yarn.py index 461a02dd0f620..c6d94672f5469 100644 --- a/yarn/datadog_checks/yarn/yarn.py +++ b/yarn/datadog_checks/yarn/yarn.py @@ -1,9 +1,9 @@ # (C) Datadog, Inc. 2018-present # All rights reserved # Licensed under a 3-clause BSD style license (see LICENSE) +from urllib.parse import urljoin, urlsplit, urlunsplit + from requests.exceptions import ConnectionError, HTTPError, InvalidURL, SSLError, Timeout -from six import iteritems -from six.moves.urllib.parse import urljoin, urlsplit, urlunsplit from datadog_checks.base import AgentCheck, is_affirmative from datadog_checks.base.errors import ConfigurationError @@ -198,7 +198,7 @@ def check(self, instance): app_tags = {} filtered_app_tags = {} - for dd_prefix, yarn_key in iteritems(app_tags): + for dd_prefix, yarn_key in app_tags.items(): if yarn_key in self._ALLOWED_APPLICATION_TAGS: filtered_app_tags[dd_prefix] = yarn_key app_tags = filtered_app_tags @@ -292,7 +292,7 @@ def _yarn_app_metrics(self, rm_address, app_tags, addl_tags): def _get_app_tags(self, app_json, app_tags): split_app_tags = self.instance.get('split_yarn_application_tags', DEFAULT_SPLIT_YARN_APPLICATION_TAGS) tags = [] - for dd_tag, yarn_key in iteritems(app_tags): + for dd_tag, yarn_key in app_tags.items(): try: val = app_json[yarn_key] if val: @@ -416,7 +416,7 @@ def _set_yarn_metrics_from_json(self, tags, metrics_json, yarn_metrics): """ Parse the JSON response and set the metrics """ - for dict_path, metric in iteritems(yarn_metrics): + for dict_path, metric in yarn_metrics.items(): metric_name, metric_type = metric metric_value = self._get_value_from_json(dict_path, metrics_json) @@ -465,7 +465,7 @@ def _rest_request_to_json(self, url, object_path, tags, *args, **kwargs): # Add kwargs as arguments if kwargs: - query = '&'.join(['{}={}'.format(key, value) for key, value in iteritems(kwargs)]) + query = '&'.join(['{}={}'.format(key, value) for key, value in kwargs.items()]) url = urljoin(url, '?' + query) try: diff --git a/zk/datadog_checks/zk/zk.py b/zk/datadog_checks/zk/zk.py index 05d9493e1889b..dfa63e9415cb8 100644 --- a/zk/datadog_checks/zk/zk.py +++ b/zk/datadog_checks/zk/zk.py @@ -61,15 +61,12 @@ import struct from collections import defaultdict from contextlib import closing +from io import StringIO from packaging.version import Version -from six import PY3, StringIO, iteritems from datadog_checks.base import AgentCheck, ensure_bytes, ensure_unicode, is_affirmative -if PY3: - long = int - class ZKConnectionFailure(Exception): """Raised when we are unable to connect or get the output of a command.""" @@ -226,7 +223,7 @@ def report_instance_mode(self, mode): tags = self.base_tags + ['mode:%s' % mode] self.gauge('zookeeper.instances', 1, tags=tags) gauges[mode] = 1 - for k, v in iteritems(gauges): + for k, v in gauges.items(): gauge_name = 'zookeeper.instances.%s' % k self.gauge(gauge_name, v, tags=self.base_tags) @@ -309,15 +306,15 @@ def parse_stat(self, buf): _, value = buf.readline().split(':') # Fixme: This metric name is wrong. It should be removed in a major version of the agent # See https://github.com/DataDog/integrations-core/issues/816 - metrics.append(ZKMetric('zookeeper.bytes_received', long(value.strip()))) - metrics.append(ZKMetric('zookeeper.packets.received', long(value.strip()), "rate")) + metrics.append(ZKMetric('zookeeper.bytes_received', int(value.strip()))) + metrics.append(ZKMetric('zookeeper.packets.received', int(value.strip()), "rate")) # Sent: 1324 _, value = buf.readline().split(':') # Fixme: This metric name is wrong. It should be removed in a major version of the agent # See https://github.com/DataDog/integrations-core/issues/816 - metrics.append(ZKMetric('zookeeper.bytes_sent', long(value.strip()))) - metrics.append(ZKMetric('zookeeper.packets.sent', long(value.strip()), "rate")) + metrics.append(ZKMetric('zookeeper.bytes_sent', int(value.strip()))) + metrics.append(ZKMetric('zookeeper.packets.sent', int(value.strip()), "rate")) if has_connections_val: # Connections: 1 @@ -330,12 +327,12 @@ def parse_stat(self, buf): # Outstanding: 0 _, value = buf.readline().split(':') - metrics.append(ZKMetric('zookeeper.outstanding_requests', long(value.strip()))) + metrics.append(ZKMetric('zookeeper.outstanding_requests', int(value.strip()))) # Zxid: 0x1034799c7 _, value = buf.readline().split(':') # Parse as a 64 bit hex int - zxid = long(value.strip(), 16) + zxid = int(value.strip(), 16) # convert to bytes zxid_bytes = struct.pack('>q', zxid) # the higher order 4 bytes is the epoch @@ -353,7 +350,7 @@ def parse_stat(self, buf): # Node count: 487 _, value = buf.readline().split(':') - metrics.append(ZKMetric('zookeeper.nodes', long(value.strip()))) + metrics.append(ZKMetric('zookeeper.nodes', int(value.strip()))) return metrics, tags, mode, version From 969a2aefd8d085593c1a9ac0317161b741a3f8e9 Mon Sep 17 00:00:00 2001 From: Zhengda Lu Date: Wed, 18 Sep 2024 12:48:00 -0400 Subject: [PATCH 11/34] [mongo] bump base package version (#18583) * bump base package version * add changelog * use 36.11.0 * remove file --- mongo/changelog.d/18583.added | 1 + mongo/pyproject.toml | 2 +- 2 files changed, 2 insertions(+), 1 deletion(-) create mode 100644 mongo/changelog.d/18583.added diff --git a/mongo/changelog.d/18583.added b/mongo/changelog.d/18583.added new file mode 100644 index 0000000000000..73092246d5ada --- /dev/null +++ b/mongo/changelog.d/18583.added @@ -0,0 +1 @@ +Bump datadog-checks-base dependency diff --git a/mongo/pyproject.toml b/mongo/pyproject.toml index de71c7d49ce1a..87eb3c784e024 100644 --- a/mongo/pyproject.toml +++ b/mongo/pyproject.toml @@ -29,7 +29,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=36.7.0", + "datadog-checks-base>=36.11.0", ] dynamic = [ "version", From ed1085fc604b5fc99a528cacf5b1e55a07860758 Mon Sep 17 00:00:00 2001 From: rahulkaukuntla <144174402+rahulkaukuntla@users.noreply.github.com> Date: Wed, 18 Sep 2024 16:53:54 -0400 Subject: [PATCH 12/34] [PLINT-508] Fixed excluded host tags for property metrics in vSphere (#18601) * accounting for excluded host tags for property metrics * added a changelog * lint * applying suggestion * addressing Sarah's comments --- vsphere/changelog.d/18601.fixed | 1 + vsphere/datadog_checks/vsphere/vsphere.py | 7 ++- vsphere/tests/test_unit.py | 63 ++++++++++++++++++----- 3 files changed, 56 insertions(+), 15 deletions(-) create mode 100644 vsphere/changelog.d/18601.fixed diff --git a/vsphere/changelog.d/18601.fixed b/vsphere/changelog.d/18601.fixed new file mode 100644 index 0000000000000..62efc32805c0e --- /dev/null +++ b/vsphere/changelog.d/18601.fixed @@ -0,0 +1 @@ +Fixed excluded host tags for property metrics diff --git a/vsphere/datadog_checks/vsphere/vsphere.py b/vsphere/datadog_checks/vsphere/vsphere.py index f50cfdf73743c..6c137ce2ff1df 100644 --- a/vsphere/datadog_checks/vsphere/vsphere.py +++ b/vsphere/datadog_checks/vsphere/vsphere.py @@ -901,7 +901,12 @@ def submit_property_metrics( ) return - base_tags = self._config.base_tags + resource_tags + base_tags = [] + if self._config.excluded_host_tags: + base_tags.extend([t for t in resource_tags if t.split(":", 1)[0] in self._config.excluded_host_tags]) + else: + base_tags.extend(resource_tags) + base_tags.extend(self._config.base_tags) if resource_type == vim.VirtualMachine: object_properties = self._config.object_properties_to_collect_by_mor.get(resource_metric_suffix, []) diff --git a/vsphere/tests/test_unit.py b/vsphere/tests/test_unit.py index fb6ce7133eb8c..0f3654809764c 100644 --- a/vsphere/tests/test_unit.py +++ b/vsphere/tests/test_unit.py @@ -2358,7 +2358,7 @@ def test_vm_property_metrics(aggregator, realtime_instance, dd_run_check, caplog base_tags_vm1 = base_tags + ['vsphere_host:host1'] base_tags_vm3 = base_tags + ['vsphere_host:host2'] - realtime_instance['excluded_host_tags'] = ['vsphere_host'] + realtime_instance['excluded_host_tags'] = ['vsphere_host', 'vsphere_type', 'vsphere_folder'] check = VSphereCheck('vsphere', {}, [realtime_instance]) caplog.set_level(logging.DEBUG) dd_run_check(check) @@ -2373,8 +2373,8 @@ def test_vm_property_metrics(aggregator, realtime_instance, dd_run_check, caplog ) assert ( "Could not submit property metric- no metric data: name=`vm.guest.guestFullName`, " - "value=`None`, hostname=`vm1`, base tags=`['vcenter_server:FAKE', 'vsphere_host:host1', " - "'vsphere_folder:unknown', 'vsphere_type:vm']` additional tags=`{}`" + "value=`None`, hostname=`vm1`, base tags=`['vsphere_host:host1', 'vsphere_folder:unknown', " + "'vsphere_type:vm', 'vcenter_server:FAKE']` additional tags=`{}`" ) in caplog.text aggregator.assert_metric( @@ -2433,8 +2433,8 @@ def test_vm_property_metrics(aggregator, realtime_instance, dd_run_check, caplog ) assert ( "Could not submit property metric- no metric data: name=`vm.guest.toolsRunningStatus`, " - "value=`None`, hostname=`vm1`, base tags=`['vcenter_server:FAKE', 'vsphere_host:host1', " - "'vsphere_folder:unknown', 'vsphere_type:vm']` additional tags=`{}`" + "value=`None`, hostname=`vm1`, base tags=`['vsphere_host:host1', 'vsphere_folder:unknown', " + "'vsphere_type:vm', 'vcenter_server:FAKE']` additional tags=`{}`" ) in caplog.text aggregator.assert_metric( @@ -2514,8 +2514,8 @@ def test_vm_property_metrics(aggregator, realtime_instance, dd_run_check, caplog assert ( "Submit property metric: name=`vm.config.memoryAllocation.limit`, value=`-1.0`, " - "hostname=`vm1`, tags=`['vcenter_server:FAKE', 'vsphere_host:host1', " - "'vsphere_folder:unknown', 'vsphere_type:vm']`, count=`False`" + "hostname=`vm1`, tags=`['vsphere_host:host1', 'vsphere_folder:unknown', " + "'vsphere_type:vm', 'vcenter_server:FAKE']`, count=`False`" ) in caplog.text aggregator.assert_metric( @@ -2526,8 +2526,8 @@ def test_vm_property_metrics(aggregator, realtime_instance, dd_run_check, caplog assert ( "Could not submit property metric- unexpected metric value: " "name=`vm.config.cpuAllocation.overheadLimit`, value=`None`, hostname=`vm1`, " - "base tags=`['vcenter_server:FAKE', 'vsphere_host:host1', 'vsphere_folder:unknown', " - "'vsphere_type:vm']` additional tags=`{}`" + "base tags=`['vsphere_host:host1', 'vsphere_folder:unknown', " + "'vsphere_type:vm', 'vcenter_server:FAKE']` additional tags=`{}`" ) in caplog.text aggregator.assert_metric( @@ -2538,8 +2538,8 @@ def test_vm_property_metrics(aggregator, realtime_instance, dd_run_check, caplog assert ( "Could not submit property metric- unexpected metric value: " "name=`vm.config.memoryAllocation.overheadLimit`, value=`None`, hostname=`vm1`, " - "base tags=`['vcenter_server:FAKE', 'vsphere_host:host1', 'vsphere_folder:unknown', " - "'vsphere_type:vm']` additional tags=`{}`" + "base tags=`['vsphere_host:host1', 'vsphere_folder:unknown', " + "'vsphere_type:vm', 'vcenter_server:FAKE']` additional tags=`{}`" ) in caplog.text # VM 3 @@ -2599,8 +2599,8 @@ def test_vm_property_metrics(aggregator, realtime_instance, dd_run_check, caplog ) assert ( "Could not submit property metric- unexpected metric value: name=`vm.summary.config.memorySizeMB`, " - "value=`None`, hostname=`vm3`, base tags=`['vcenter_server:FAKE', 'vsphere_host:host2', " - "'vsphere_folder:unknown', 'vsphere_type:vm']` additional tags=`{}`" + "value=`None`, hostname=`vm3`, base tags=`['vsphere_host:host2', " + "'vsphere_folder:unknown', 'vsphere_type:vm', 'vcenter_server:FAKE']` additional tags=`{}`" ) in caplog.text aggregator.assert_metric( @@ -2733,7 +2733,7 @@ def test_host_property_metrics(aggregator, realtime_instance, dd_run_check, capl assert ( "Could not submit property metric- no metric data: " "name=`host.hardware.cpuPowerManagementInfo.currentPolicy`, value=`None`, " - "hostname=`host2`, base tags=`['vcenter_server:FAKE', 'vsphere_type:host']` " + "hostname=`host2`, base tags=`['vsphere_type:host', 'vcenter_server:FAKE']` " "additional tags=`{}`" ) in caplog.text @@ -3181,6 +3181,41 @@ def test_property_metrics_invalid_ip_route_config_gateway( ) +def test_property_metrics_excluded_host_tags( + aggregator, realtime_instance, dd_run_check, service_instance, vm_properties_ex, datadog_agent +): + realtime_instance['collect_property_metrics'] = True + + service_instance.content.propertyCollector.RetrievePropertiesEx = vm_properties_ex + + realtime_instance['excluded_host_tags'] = ['vsphere_host', 'vsphere_folder'] + check = VSphereCheck('vsphere', {}, [realtime_instance]) + dd_run_check(check) + + aggregator.assert_metric( + 'vsphere.vm.summary.quickStats.uptimeSeconds', + count=1, + value=12184573.0, + tags=['vcenter_server:FAKE', 'vsphere_folder:unknown', 'vsphere_host:host1'], + hostname='vm1', + ) + datadog_agent.assert_external_tags( + 'vm1', + {'vsphere': ['vcenter_server:FAKE', 'vsphere_type:vm']}, + ) + aggregator.assert_metric( + 'vsphere.host.hardware.cpuPowerManagementInfo.currentPolicy', + count=1, + value=1, + tags=['currentPolicy:Balanced', 'vcenter_server:FAKE'], + hostname='host1', + ) + datadog_agent.assert_external_tags( + 'host1', + {'vsphere': ['vcenter_server:FAKE', 'vsphere_type:host']}, + ) + + @pytest.mark.parametrize( ('max_query_metrics', 'metrics_per_query', 'max_historical_metrics', 'expected_batch_num'), [ From 25c7fbea4e3bd01a1112553602c478bd04a4800a Mon Sep 17 00:00:00 2001 From: Ilia Kurenkov Date: Thu, 19 Sep 2024 06:42:12 +0200 Subject: [PATCH 13/34] Drop 'six' from some more checks. (#18616) * Drop 'six' from some more checks. Here we focus mostly on the different compatibility crutches in use by integrations * fix import * fix another import --- .../activemq_xml/activemq_xml.py | 3 +- ambari/datadog_checks/ambari/ambari.py | 13 +++--- btrfs/datadog_checks/btrfs/btrfs.py | 4 +- ceph/datadog_checks/ceph/ceph.py | 3 +- .../datadog_checks/cisco_aci/aci_metrics.py | 10 ++--- .../datadog_checks/cisco_aci/capacity.py | 10 ++--- cisco_aci/datadog_checks/cisco_aci/cisco.py | 6 +-- cisco_aci/datadog_checks/cisco_aci/fabric.py | 13 ++---- cisco_aci/datadog_checks/cisco_aci/ndm.py | 18 +++----- cisco_aci/datadog_checks/cisco_aci/tags.py | 6 +-- cisco_aci/datadog_checks/cisco_aci/tenant.py | 6 +-- .../datadog_checks/clickhouse/clickhouse.py | 6 +-- .../cloud_foundry_api/cloud_foundry_api.py | 2 +- .../datadog_checks/cloud_foundry_api/utils.py | 2 +- cockroachdb/tests/legacy/common.py | 4 +- consul/datadog_checks/consul/consul.py | 11 +++-- couch/datadog_checks/couch/couch.py | 29 ++++++------ .../datadog_checks/couchbase/couchbase.py | 5 +-- elastic/datadog_checks/elastic/config.py | 3 +- elastic/datadog_checks/elastic/elastic.py | 27 ++++++----- envoy/datadog_checks/envoy/check.py | 3 +- envoy/datadog_checks/envoy/envoy.py | 2 +- envoy/datadog_checks/envoy/parser.py | 2 - esxi/datadog_checks/esxi/check.py | 5 +-- esxi/datadog_checks/esxi/utils.py | 3 +- etcd/datadog_checks/etcd/etcd.py | 2 +- fluentd/datadog_checks/fluentd/fluentd.py | 3 +- fly_io/datadog_checks/fly_io/check.py | 2 +- .../gitlab_runner/gitlab_runner.py | 2 +- glusterfs/datadog_checks/glusterfs/check.py | 4 +- .../datadog_checks/go_expvar/go_expvar.py | 6 +-- .../hdfs_datanode/hdfs_datanode.py | 8 ++-- .../hdfs_namenode/hdfs_namenode.py | 8 ++-- .../datadog_checks/http_check/http_check.py | 9 +--- .../collectors/channel_metric_collector.py | 6 +-- .../collectors/queue_metric_collector.py | 8 ++-- ibm_mq/datadog_checks/ibm_mq/config.py | 6 +-- ibm_mq/datadog_checks/ibm_mq/ibm_mq.py | 4 +- ibm_was/datadog_checks/ibm_was/ibm_was.py | 5 +-- iis/datadog_checks/iis/iis.py | 10 ++--- kong/datadog_checks/kong/kong.py | 3 +- .../kube_controller_manager.py | 3 +- kube_dns/datadog_checks/kube_dns/kube_dns.py | 3 +- kubelet/datadog_checks/kubelet/cadvisor.py | 5 +-- kubelet/datadog_checks/kubelet/kubelet.py | 13 +++--- kubelet/datadog_checks/kubelet/prometheus.py | 12 +++-- .../kubernetes_state/kubernetes_state.py | 36 +++++++-------- lighttpd/datadog_checks/lighttpd/lighttpd.py | 3 +- .../linux_proc_extras/linux_proc_extras.py | 4 +- mapr/datadog_checks/mapr/mapr.py | 4 +- marathon/datadog_checks/marathon/marathon.py | 5 +-- .../marklogic/parsers/request.py | 4 +- .../marklogic/parsers/status.py | 6 +-- .../marklogic/parsers/storage.py | 6 +-- .../datadog_checks/mesos_slave/mesos_slave.py | 7 ++- .../mongo/collectors/replica.py | 3 +- mongo/datadog_checks/mongo/utils.py | 2 +- network/datadog_checks/network/check_bsd.py | 21 ++++----- network/datadog_checks/network/check_linux.py | 37 +++++++-------- .../datadog_checks/network/check_solaris.py | 9 +--- .../datadog_checks/network/check_windows.py | 7 +-- network/datadog_checks/network/network.py | 29 ++++-------- .../datadog_checks/openstack/openstack.py | 9 ++-- .../openstack_controller/legacy/api.py | 16 +++---- .../legacy/openstack_controller_legacy.py | 21 +++++---- .../datadog_checks/pgbouncer/pgbouncer.py | 2 +- php_fpm/datadog_checks/php_fpm/php_fpm.py | 45 +++++++++---------- postgres/datadog_checks/postgres/config.py | 17 +------ postgres/datadog_checks/postgres/postgres.py | 3 +- process/datadog_checks/process/process.py | 5 +-- rabbitmq/datadog_checks/rabbitmq/rabbitmq.py | 11 +++-- silk/datadog_checks/silk/check.py | 3 +- snmp/datadog_checks/snmp/snmp.py | 4 +- .../datadog_checks/sqlserver/connection.py | 4 +- statsd/datadog_checks/statsd/statsd.py | 3 +- tls/datadog_checks/tls/tls.py | 5 +-- voltdb/datadog_checks/voltdb/check.py | 3 +- voltdb/datadog_checks/voltdb/client.py | 2 +- voltdb/datadog_checks/voltdb/config.py | 3 +- vsphere/datadog_checks/vsphere/api.py | 3 +- vsphere/datadog_checks/vsphere/cache.py | 3 +- .../vsphere/legacy/vsphere_legacy.py | 4 +- vsphere/tests/legacy/test_mor_cache.py | 1 - .../windows_service/windows_service.py | 3 +- 84 files changed, 257 insertions(+), 414 deletions(-) diff --git a/activemq_xml/datadog_checks/activemq_xml/activemq_xml.py b/activemq_xml/datadog_checks/activemq_xml/activemq_xml.py index 6c274675dc844..18a6d2ccdd00b 100644 --- a/activemq_xml/datadog_checks/activemq_xml/activemq_xml.py +++ b/activemq_xml/datadog_checks/activemq_xml/activemq_xml.py @@ -4,7 +4,6 @@ from xml.etree import ElementTree import requests -from six import iteritems from datadog_checks.base import AgentCheck from datadog_checks.base.config import _is_affirmative @@ -95,7 +94,7 @@ def _process_data(self, data, el_type, tags, max_elements, detailed_elements): continue el_tags = tags + ["{0}:{1}".format(el_type, name)] - for attr_name, alias in iteritems(TOPIC_QUEUE_METRICS): + for attr_name, alias in TOPIC_QUEUE_METRICS.items(): metric_name = "activemq.{0}.{1}".format(el_type, alias) value = stats.get(attr_name, 0) self.gauge(metric_name, value, tags=el_tags) diff --git a/ambari/datadog_checks/ambari/ambari.py b/ambari/datadog_checks/ambari/ambari.py index e84ca963606c1..902bf894667d1 100644 --- a/ambari/datadog_checks/ambari/ambari.py +++ b/ambari/datadog_checks/ambari/ambari.py @@ -2,7 +2,6 @@ # All rights reserved # Licensed under a 3-clause BSD style license (see LICENSE) from requests.exceptions import ConnectionError, HTTPError, Timeout -from six import iteritems from datadog_checks.base import AgentCheck from datadog_checks.base.constants import ServiceCheck @@ -96,7 +95,7 @@ def get_host_metrics(self, clusters): continue metrics = self.flatten_host_metrics(host_metrics) - for metric_name, value in iteritems(metrics): + for metric_name, value in metrics.items(): metric_tags = self.base_tags + [cluster_tag] if isinstance(value, float): self._submit_gauge(metric_name, value, metric_tags, hostname) @@ -111,7 +110,7 @@ def get_service_status_and_metrics(self, clusters, collect_service_metrics, coll for cluster in clusters: tags = self.base_tags + [CLUSTER_TAG_TEMPLATE.format(cluster)] - for service, components in iteritems(self.included_services): + for service, components in self.included_services.items(): service_tags = tags + [SERVICE_TAG + service.lower()] if collect_service_metrics: @@ -144,7 +143,7 @@ def get_component_metrics(self, cluster, service, base_tags, component_included) component_metrics_endpoint = common.create_endpoint(self.base_url, cluster, service, COMPONENT_METRICS_QUERY) components_response = self._make_request(component_metrics_endpoint) - component_included = {k.upper(): v for k, v in iteritems(component_included)} + component_included = {k.upper(): v for k, v in component_included.items()} if components_response is None or 'items' not in components_response: self.log.warning("No components found for service %s.", service) @@ -171,7 +170,7 @@ def get_component_metrics(self, cluster, service, base_tags, component_included) metrics = self.flatten_service_metrics(component_metrics[header], header) component_tag = COMPONENT_TAG + component_name.lower() - for metric_name, value in iteritems(metrics): + for metric_name, value in metrics.items(): metric_tags = base_tags + [component_tag] if isinstance(value, float): self._submit_gauge(metric_name, value, metric_tags) @@ -206,7 +205,7 @@ def _submit_service_checks(self, name, value, tags, message=None): @classmethod def flatten_service_metrics(cls, metric_dict, prefix): flat_metrics = {} - for raw_metric_name, metric_value in iteritems(metric_dict): + for raw_metric_name, metric_value in metric_dict.items(): if isinstance(metric_value, dict): flat_metrics.update(cls.flatten_service_metrics(metric_value, prefix)) else: @@ -217,7 +216,7 @@ def flatten_service_metrics(cls, metric_dict, prefix): @classmethod def flatten_host_metrics(cls, metric_dict, prefix=""): flat_metrics = {} - for raw_metric_name, metric_value in iteritems(metric_dict): + for raw_metric_name, metric_value in metric_dict.items(): metric_name = '{}.{}'.format(prefix, raw_metric_name) if prefix else raw_metric_name if raw_metric_name == "boottime": flat_metrics["boottime"] = metric_value diff --git a/btrfs/datadog_checks/btrfs/btrfs.py b/btrfs/datadog_checks/btrfs/btrfs.py index c469b511dd5bf..3c5d0543ea23d 100644 --- a/btrfs/datadog_checks/btrfs/btrfs.py +++ b/btrfs/datadog_checks/btrfs/btrfs.py @@ -11,8 +11,6 @@ from collections import defaultdict import psutil -from six import iteritems -from six.moves import range from datadog_checks.base import AgentCheck @@ -175,7 +173,7 @@ def check(self, _): if len(btrfs_devices) == 0: raise Exception("No btrfs device found") - for device, mountpoint in iteritems(btrfs_devices): + for device, mountpoint in btrfs_devices.items(): for flags, total_bytes, used_bytes in self.get_usage(mountpoint): replication_type, usage_type = FLAGS_MAPPER[flags] tags = [ diff --git a/ceph/datadog_checks/ceph/ceph.py b/ceph/datadog_checks/ceph/ceph.py index 535bdb9e2ba09..e52d497db48f7 100644 --- a/ceph/datadog_checks/ceph/ceph.py +++ b/ceph/datadog_checks/ceph/ceph.py @@ -7,7 +7,6 @@ import re import simplejson as json -from six import iteritems from datadog_checks.base import AgentCheck from datadog_checks.base.config import _is_affirmative @@ -149,7 +148,7 @@ def _extract_metrics(self, raw, tags): # so we won't send the metric osd.pct_used if 'checks' in raw['health_detail']: checks = raw['health_detail']['checks'] - for check_name, check_detail in iteritems(checks): + for check_name, check_detail in checks.items(): if check_name == 'OSD_NEARFULL': health['num_near_full_osds'] = len(check_detail['detail']) if check_name == 'OSD_FULL': diff --git a/cisco_aci/datadog_checks/cisco_aci/aci_metrics.py b/cisco_aci/datadog_checks/cisco_aci/aci_metrics.py index 92ded035e1542..5a254eac9ab1f 100644 --- a/cisco_aci/datadog_checks/cisco_aci/aci_metrics.py +++ b/cisco_aci/datadog_checks/cisco_aci/aci_metrics.py @@ -2,8 +2,6 @@ # All rights reserved # Licensed under a 3-clause BSD style license (see LICENSE) -from six import iteritems - METRIC_PREFIX = "cisco_aci" FABRIC_PREFIX = METRIC_PREFIX + ".fabric" @@ -106,11 +104,11 @@ def make_tenant_metrics(): tenant_metrics = {"tenant": {}, "application": {}, "endpoint_group": {}} - for cisco_metric, metric_map in iteritems(metrics): + for cisco_metric, metric_map in metrics.items(): tenant_metrics["tenant"][cisco_metric] = {} tenant_metrics["application"][cisco_metric] = {} tenant_metrics["endpoint_group"][cisco_metric] = {} - for sub_metric, dd_metric in iteritems(metric_map): + for sub_metric, dd_metric in metric_map.items(): dd_tenant_metric = dd_metric.format(TENANT_PREFIX) tenant_metrics["tenant"][cisco_metric][sub_metric] = dd_tenant_metric dd_app_metric = dd_metric.format(APPLICATION_PREFIX) @@ -118,10 +116,10 @@ def make_tenant_metrics(): dd_epg_metric = dd_metric.format(ENDPOINT_GROUP_PREFIX) tenant_metrics["endpoint_group"][cisco_metric][sub_metric] = dd_epg_metric - for cisco_metric, metric_map in iteritems(endpoint_metrics): + for cisco_metric, metric_map in endpoint_metrics.items(): if not tenant_metrics.get("endpoint_group", {}).get(cisco_metric): tenant_metrics["endpoint_group"][cisco_metric] = {} - for sub_metric, dd_metric in iteritems(metric_map): + for sub_metric, dd_metric in metric_map.items(): dd_epg_metric = dd_metric.format(TENANT_PREFIX) tenant_metrics["endpoint_group"][cisco_metric][sub_metric] = dd_epg_metric diff --git a/cisco_aci/datadog_checks/cisco_aci/capacity.py b/cisco_aci/datadog_checks/cisco_aci/capacity.py index 1fe537388a4a5..692e6c77816c8 100644 --- a/cisco_aci/datadog_checks/cisco_aci/capacity.py +++ b/cisco_aci/datadog_checks/cisco_aci/capacity.py @@ -2,8 +2,6 @@ # All rights reserved # Licensed under a 3-clause BSD style license (see LICENSE) -from six import iteritems - from . import aci_metrics, exceptions, helpers @@ -49,7 +47,7 @@ def collect(self): self.log.info("finished collecting capacity data") def _get_eqpt_capacity(self): - for c, metric_dict in iteritems(aci_metrics.EQPT_CAPACITY_METRICS): + for c, metric_dict in aci_metrics.EQPT_CAPACITY_METRICS.items(): data = self.api.get_eqpt_capacity(c) for d in data: dn = d.get('attributes', {}).get('dn') @@ -63,14 +61,14 @@ def _get_eqpt_capacity(self): child_attrs = child.get(c, {}).get('attributes') if not child_attrs or type(child_attrs) is not dict: continue - for cisco_metric, dd_metric in iteritems(metric_dict): + for cisco_metric, dd_metric in metric_dict.items(): value = child_attrs.get(cisco_metric) if not value: continue self.gauge(dd_metric, value, tags=tags, hostname=hostname) def _get_contexts(self): - for c, metric_dict in iteritems(aci_metrics.CAPACITY_CONTEXT_METRICS): + for c, metric_dict in aci_metrics.CAPACITY_CONTEXT_METRICS.items(): dd_metric = metric_dict.get("metric_name") utilized_metric_name = dd_metric + ".utilized" # These Values are, for some reason, hardcoded in the UI @@ -105,7 +103,7 @@ def _get_apic_capacity_limits(self): def _get_apic_capacity_metrics(self): tags = self.user_tags + self.check_tags - for c, opts in iteritems(aci_metrics.APIC_CAPACITY_METRICS): + for c, opts in aci_metrics.APIC_CAPACITY_METRICS.items(): dd_metric = opts.get("metric_name") data = self.api.get_apic_capacity_metrics(c, query=opts.get("query_string")) if c == "fabricNode": diff --git a/cisco_aci/datadog_checks/cisco_aci/cisco.py b/cisco_aci/datadog_checks/cisco_aci/cisco.py index 4cfee34753db7..ee17248e34f11 100644 --- a/cisco_aci/datadog_checks/cisco_aci/cisco.py +++ b/cisco_aci/datadog_checks/cisco_aci/cisco.py @@ -1,8 +1,6 @@ # (C) Datadog, Inc. 2018-present # All rights reserved # Licensed under a 3-clause BSD style license (see LICENSE) -from six import iteritems - from datadog_checks.base import AgentCheck, ConfigurationError from datadog_checks.base.config import _is_affirmative from datadog_checks.base.utils.containers import hash_mutable @@ -146,7 +144,7 @@ def submit_metrics(self, metrics, tags, instance=None, obj_type="gauge", hostnam instance = {} user_tags = instance.get('tags', []) - for mname, mval in iteritems(metrics): + for mname, mval in metrics.items(): tags_to_send = [] if mval: if hostname: @@ -162,7 +160,7 @@ def submit_metrics(self, metrics, tags, instance=None, obj_type="gauge", hostnam def get_external_host_tags(self): external_host_tags = [] - for hostname, tags in iteritems(self.external_host_tags): + for hostname, tags in self.external_host_tags.items(): host_tags = tags + self.check_tags external_host_tags.append((hostname, {SOURCE_TYPE: host_tags})) return external_host_tags diff --git a/cisco_aci/datadog_checks/cisco_aci/fabric.py b/cisco_aci/datadog_checks/cisco_aci/fabric.py index c73baf0eb695a..1a99ebbb3425b 100644 --- a/cisco_aci/datadog_checks/cisco_aci/fabric.py +++ b/cisco_aci/datadog_checks/cisco_aci/fabric.py @@ -1,15 +1,10 @@ # (C) Datadog, Inc. 2018-present # All rights reserved # Licensed under a 3-clause BSD style license (see LICENSE) - - -from six import PY3, iteritems +import time from datadog_checks.base.utils.serialization import json -if PY3: - import time - from . import aci_metrics, exceptions, helpers, ndm VENDOR_CISCO = 'cisco' @@ -43,7 +38,7 @@ def __init__(self, check, api, instance, namespace): self.event_platform_event = check.event_platform_event def ndm_enabled(self): - return PY3 and self.send_ndm_metadata + return self.send_ndm_metadata def collect(self): fabric_pods = self.api.get_fabric_pods() @@ -170,10 +165,10 @@ def submit_fabric_metric(self, stats, tags, obj_type, hostname=None): continue metrics = {} - for n, ms in iteritems(aci_metrics.FABRIC_METRICS): + for n, ms in aci_metrics.FABRIC_METRICS.items(): if n not in name: continue - for cisco_metric, dd_metric in iteritems(ms): + for cisco_metric, dd_metric in ms.items(): mname = dd_metric.format(self.get_fabric_type(obj_type)) mval = s.get(name, {}).get("attributes", {}).get(cisco_metric) json_attrs = s.get(name, {}).get("attributes", {}) diff --git a/cisco_aci/datadog_checks/cisco_aci/ndm.py b/cisco_aci/datadog_checks/cisco_aci/ndm.py index 850a1c4ae6173..5681350d6b39f 100644 --- a/cisco_aci/datadog_checks/cisco_aci/ndm.py +++ b/cisco_aci/datadog_checks/cisco_aci/ndm.py @@ -2,17 +2,13 @@ # All rights reserved # Licensed under a 3-clause BSD style license (see LICENSE) - -from six import PY3 - -if PY3: - from datadog_checks.cisco_aci.models import ( - DeviceMetadata, - InterfaceMetadata, - NetworkDevicesMetadata, - Node, - PhysIf, - ) +from datadog_checks.cisco_aci.models import ( + DeviceMetadata, + InterfaceMetadata, + NetworkDevicesMetadata, + Node, + PhysIf, +) VENDOR_CISCO = 'cisco' PAYLOAD_METADATA_BATCH_SIZE = 100 diff --git a/cisco_aci/datadog_checks/cisco_aci/tags.py b/cisco_aci/datadog_checks/cisco_aci/tags.py index 2696c960f23ce..c9b1fab5e59de 100644 --- a/cisco_aci/datadog_checks/cisco_aci/tags.py +++ b/cisco_aci/datadog_checks/cisco_aci/tags.py @@ -4,8 +4,6 @@ import re -from six import iteritems - from datadog_checks.base.utils.containers import hash_mutable from . import exceptions, helpers @@ -103,7 +101,7 @@ def _tenant_mapper(self, edpt): application_meta = [] application_meta_map = self._edpt_tags_map(edpt) - for k, v in iteritems(application_meta_map): + for k, v in application_meta_map.items(): application_meta.append(k + ":" + v) tenant_name = application_meta_map.get("tenant") app_name = application_meta_map.get("application") @@ -112,7 +110,7 @@ def _tenant_mapper(self, edpt): # adding meta tags endpoint_meta = [] endpoint_meta_map = self._get_epg_meta_tags_map(tenant_name, app_name, epg_name) - for k, v in iteritems(endpoint_meta_map): + for k, v in endpoint_meta_map.items(): endpoint_meta.append(k + ":" + v) # adding application tags diff --git a/cisco_aci/datadog_checks/cisco_aci/tenant.py b/cisco_aci/datadog_checks/cisco_aci/tenant.py index f906dd4aa38ea..02b8b7dede03e 100644 --- a/cisco_aci/datadog_checks/cisco_aci/tenant.py +++ b/cisco_aci/datadog_checks/cisco_aci/tenant.py @@ -6,8 +6,6 @@ import re import time -from six import iteritems - from . import exceptions, helpers @@ -112,10 +110,10 @@ def submit_raw_obj(self, raw_stats, tags, obj_type): tenant_metrics = self.tenant_metrics.get(obj_type, {}) - for n, ms in iteritems(tenant_metrics): + for n, ms in tenant_metrics.items(): if n not in name: continue - for cisco_metric, dd_metric in iteritems(ms): + for cisco_metric, dd_metric in ms.items(): mval = s.get(name, {}).get("attributes", {}).get(cisco_metric) json_attrs = s.get(name, {}).get("attributes", {}) if mval and helpers.check_metric_can_be_zero(cisco_metric, mval, json_attrs): diff --git a/clickhouse/datadog_checks/clickhouse/clickhouse.py b/clickhouse/datadog_checks/clickhouse/clickhouse.py index 85e4af4d02266..e640169d9e749 100644 --- a/clickhouse/datadog_checks/clickhouse/clickhouse.py +++ b/clickhouse/datadog_checks/clickhouse/clickhouse.py @@ -2,7 +2,6 @@ # All rights reserved # Licensed under a 3-clause BSD style license (see LICENSE) import clickhouse_driver -from six import raise_from from datadog_checks.base import AgentCheck, ConfigurationError, is_affirmative from datadog_checks.base.utils.db import QueryManager @@ -119,10 +118,7 @@ def connect(self): self._error_sanitizer.clean(self._error_sanitizer.scrub(str(e))) ) self.service_check(self.SERVICE_CHECK_CONNECT, self.CRITICAL, message=error, tags=self._tags) - - # When an exception is raised in the context of another one, both will be printed. To avoid - # this we set the context to None. https://www.python.org/dev/peps/pep-0409/ - raise_from(type(e)(error), None) + raise type(e)(error) from None else: self.service_check(self.SERVICE_CHECK_CONNECT, self.OK, tags=self._tags) self._client = client diff --git a/cloud_foundry_api/datadog_checks/cloud_foundry_api/cloud_foundry_api.py b/cloud_foundry_api/datadog_checks/cloud_foundry_api/cloud_foundry_api.py index f7d8cadae870c..219f156a7ac85 100644 --- a/cloud_foundry_api/datadog_checks/cloud_foundry_api/cloud_foundry_api.py +++ b/cloud_foundry_api/datadog_checks/cloud_foundry_api/cloud_foundry_api.py @@ -5,10 +5,10 @@ import json import time from typing import Any, Dict, Generator, Tuple # noqa: F401 +from urllib.parse import urlparse from requests.exceptions import HTTPError, RequestException from semver import VersionInfo -from six.moves.urllib_parse import urlparse from datadog_checks.base import AgentCheck from datadog_checks.base.errors import CheckException, ConfigurationError diff --git a/cloud_foundry_api/datadog_checks/cloud_foundry_api/utils.py b/cloud_foundry_api/datadog_checks/cloud_foundry_api/utils.py index 193bb135baac0..6adc2ae883187 100644 --- a/cloud_foundry_api/datadog_checks/cloud_foundry_api/utils.py +++ b/cloud_foundry_api/datadog_checks/cloud_foundry_api/utils.py @@ -3,9 +3,9 @@ # Licensed under a 3-clause BSD style license (see LICENSE) from datetime import datetime from typing import Any, Dict # noqa: F401 +from urllib.parse import urljoin from dateutil import parser, tz -from six.moves.urllib_parse import urljoin def get_next_url(payload, version): diff --git a/cockroachdb/tests/legacy/common.py b/cockroachdb/tests/legacy/common.py index 81ad062f6812f..94bda6b0382f3 100644 --- a/cockroachdb/tests/legacy/common.py +++ b/cockroachdb/tests/legacy/common.py @@ -1,14 +1,12 @@ # (C) Datadog, Inc. 2024-present # All rights reserved # Licensed under a 3-clause BSD style license (see LICENSE) -from six import itervalues - from datadog_checks.cockroachdb.metrics import METRIC_MAP from datadog_checks.dev.utils import assert_service_checks def assert_check(aggregator): - for metric in itervalues(METRIC_MAP): + for metric in METRIC_MAP.values(): aggregator.assert_metric('cockroachdb.{}'.format(metric), at_least=0) assert aggregator.metrics_asserted_pct > 80, 'Missing metrics {}'.format(aggregator.not_asserted()) diff --git a/consul/datadog_checks/consul/consul.py b/consul/datadog_checks/consul/consul.py index cccd45fee7812..585197fb62b48 100644 --- a/consul/datadog_checks/consul/consul.py +++ b/consul/datadog_checks/consul/consul.py @@ -8,11 +8,10 @@ from itertools import islice from multiprocessing.pool import ThreadPool from time import time as timestamp +from urllib.parse import urljoin import requests from requests import HTTPError -from six import iteritems, iterkeys, itervalues -from six.moves.urllib.parse import urljoin from datadog_checks.base import ConfigurationError, OpenMetricsBaseCheck, is_affirmative from datadog_checks.base.utils.serialization import json @@ -309,7 +308,7 @@ def _cull_services_list(self, services): ) self.warning(log_line) - services = {s: services[s] for s in list(islice(iterkeys(allowed_services), 0, self.max_services))} + services = {s: services[s] for s in list(islice(allowed_services, 0, self.max_services))} return services @@ -381,7 +380,7 @@ def check(self, _): elif STATUS_SEVERITY[status] > STATUS_SEVERITY[sc[sc_id]['status']]: sc[sc_id]['status'] = status - for s in itervalues(sc): + for s in sc.values(): self.service_check(HEALTH_CHECK, s['status'], tags=main_tags + s['tags']) except Exception as e: @@ -437,7 +436,7 @@ def check(self, _): nodes_with_service[service] if self.thread_pool is None else nodes_with_service[service].get(), ) - for node, service_status in iteritems(nodes_to_service_status): + for node, service_status in nodes_to_service_status.items(): # For every node discovered for included services, gauge the following: # `consul.catalog.services_up` : Total services registered on node # `consul.catalog.services_passing` : Total passing services on node @@ -455,7 +454,7 @@ def check(self, _): tags=main_tags + node_tags, ) - for node_status, count in iteritems(nodes_per_service_tag_counts): + for node_status, count in nodes_per_service_tag_counts.items(): service_tags = [ 'consul_{}_service_tag:{}'.format(node_status.service_name, tag) for tag in node_status.service_tags_set diff --git a/couch/datadog_checks/couch/couch.py b/couch/datadog_checks/couch/couch.py index ad00f4105cbe3..37235cd84b2aa 100644 --- a/couch/datadog_checks/couch/couch.py +++ b/couch/datadog_checks/couch/couch.py @@ -5,10 +5,9 @@ from __future__ import division import math +from urllib.parse import quote, urljoin import requests -from six import iteritems -from six.moves.urllib.parse import quote, urljoin from datadog_checks.base import AgentCheck from datadog_checks.base.errors import CheckException, ConfigurationError @@ -115,14 +114,14 @@ def __init__(self, agent_check): def _create_metric(self, data, tags=None): overall_stats = data.get('stats', {}) - for key, stats in iteritems(overall_stats): - for metric, val in iteritems(stats): + for key, stats in overall_stats.items(): + for metric, val in stats.items(): if val['current'] is not None: metric_name = '.'.join(['couchdb', key, metric]) self.gauge(metric_name, val['current'], tags=tags) - for db_name, db_stats in iteritems(data.get('databases', {})): - for name, val in iteritems(db_stats): + for db_name, db_stats in data.get('databases', {}).items(): + for name, val in db_stats.items(): if name in ['doc_count', 'disk_size'] and val is not None: metric_name = '.'.join(['couchdb', 'by_db', name]) metric_tags = list(tags) @@ -204,10 +203,10 @@ def __init__(self, agent_check): self.instance = agent_check.instance def _build_metrics(self, data, tags, prefix='couchdb'): - for key, value in iteritems(data): + for key, value in data.items(): if "type" in value: if value["type"] == "histogram": - for metric, histo_value in iteritems(value["value"]): + for metric, histo_value in value["value"].items(): if metric == "histogram": continue elif metric == "percentile": @@ -221,7 +220,7 @@ def _build_metrics(self, data, tags, prefix='couchdb'): self._build_metrics(value, tags, "{0}.{1}".format(prefix, key)) def _build_db_metrics(self, data, tags): - for key, value in iteritems(data['sizes']): + for key, value in data['sizes'].items(): self.gauge("couchdb.by_db.{0}_size".format(key), value, tags) for key in ['doc_del_count', 'doc_count']: @@ -233,18 +232,18 @@ def _build_dd_metrics(self, info, tags): ddtags.append("design_document:{0}".format(info['name'])) ddtags.append("language:{0}".format(data['language'])) - for key, value in iteritems(data['sizes']): + for key, value in data['sizes'].items(): self.gauge("couchdb.by_ddoc.{0}_size".format(key), value, ddtags) - for key, value in iteritems(data['updates_pending']): + for key, value in data['updates_pending'].items(): self.gauge("couchdb.by_ddoc.{0}_updates_pending".format(key), value, ddtags) self.gauge("couchdb.by_ddoc.waiting_clients", data['waiting_clients'], ddtags) def _build_system_metrics(self, data, tags, prefix='couchdb.erlang'): - for key, value in iteritems(data): + for key, value in data.items(): if key == "message_queues": - for queue, val in iteritems(value): + for queue, val in value.items(): queue_tags = list(tags) queue_tags.append("queue:{0}".format(queue)) if isinstance(val, dict): @@ -257,7 +256,7 @@ def _build_system_metrics(self, data, tags, prefix='couchdb.erlang'): else: self.gauge("{0}.{1}.size".format(prefix, key), val, queue_tags) elif key == "distribution": - for node, metrics in iteritems(value): + for node, metrics in value.items(): dist_tags = list(tags) dist_tags.append("node:{0}".format(node)) self._build_system_metrics(metrics, dist_tags, "{0}.{1}".format(prefix, key)) @@ -305,7 +304,7 @@ def _build_active_tasks_metrics(self, data, tags, prefix='couchdb.active_tasks') if task.get(metric) is not None: self.gauge("{0}.view_compaction.{1}".format(prefix, metric), task[metric], rtags) - for metric, count in iteritems(counts): + for metric, count in counts.items(): if metric == "database_compaction": metric = "db_compaction" self.gauge("{0}.{1}.count".format(prefix, metric), count, tags) diff --git a/couchbase/datadog_checks/couchbase/couchbase.py b/couchbase/datadog_checks/couchbase/couchbase.py index 97281b8304eb0..b74703d838528 100644 --- a/couchbase/datadog_checks/couchbase/couchbase.py +++ b/couchbase/datadog_checks/couchbase/couchbase.py @@ -7,10 +7,9 @@ import re import time +from urllib.parse import urljoin import requests -from six import string_types -from six.moves.urllib.parse import urljoin from datadog_checks.base import AgentCheck, ConfigurationError from datadog_checks.couchbase.couchbase_consts import ( @@ -96,7 +95,7 @@ def _create_metrics(self, data): norm_metric_name = self.camel_case_to_joined_lower(metric_name) if norm_metric_name in QUERY_STATS: # for query times, the unit is part of the value, we need to extract it - if isinstance(val, string_types): + if isinstance(val, str): val = self.extract_seconds_value(val) full_metric_name = 'couchbase.query.{}'.format(self.camel_case_to_joined_lower(norm_metric_name)) diff --git a/elastic/datadog_checks/elastic/config.py b/elastic/datadog_checks/elastic/config.py index 1e79c236b4ee6..18b111307e1a9 100644 --- a/elastic/datadog_checks/elastic/config.py +++ b/elastic/datadog_checks/elastic/config.py @@ -2,8 +2,7 @@ # All rights reserved # Licensed under a 3-clause BSD style license (see LICENSE) from collections import namedtuple - -from six.moves.urllib.parse import urlparse +from urllib.parse import urlparse from datadog_checks.base import ConfigurationError, is_affirmative diff --git a/elastic/datadog_checks/elastic/elastic.py b/elastic/datadog_checks/elastic/elastic.py index aa41cf8eff95f..cb0b1cbd012de 100644 --- a/elastic/datadog_checks/elastic/elastic.py +++ b/elastic/datadog_checks/elastic/elastic.py @@ -6,10 +6,9 @@ from collections import defaultdict, namedtuple from copy import deepcopy from itertools import product +from urllib.parse import urljoin, urlparse import requests -from six import iteritems, itervalues -from six.moves.urllib.parse import urljoin, urlparse from datadog_checks.base import AgentCheck, is_affirmative, to_string @@ -262,13 +261,13 @@ def _get_index_metrics(self, admin_forwarder, version, base_tags): index_data['health_reverse'] = dd_health.reverse_status # Ensure that index_data does not contain None values - for key, value in list(iteritems(index_data)): + for key, value in index_data.items(): if value is None: del index_data[key] self.log.debug("The index %s has no metric data for %s", idx['index'], key) tags = base_tags + ['index_name:' + idx['index']] - for metric, desc in iteritems(index_stats_for_version(version)): + for metric, desc in index_stats_for_version(version).items(): self._process_metric(index_data, metric, *desc, tags=tags) self._get_index_search_stats(admin_forwarder, base_tags) @@ -282,7 +281,7 @@ def _get_template_metrics(self, admin_forwarder, base_tags): filtered_templates = [t for t in template_resp if not t['name'].startswith(TEMPLATE_EXCLUSION_LIST)] - for metric, desc in iteritems(TEMPLATE_METRICS): + for metric, desc in TEMPLATE_METRICS.items(): self._process_metric({'templates': filtered_templates}, metric, *desc, tags=base_tags) def _get_index_search_stats(self, admin_forwarder, base_tags): @@ -294,7 +293,7 @@ def _get_index_search_stats(self, admin_forwarder, base_tags): # The health we can get from /_cluster/health if we pass level=indices query param. Reference: # https://www.elastic.co/guide/en/elasticsearch/reference/current/cluster-health.html#cluster-health-api-query-params # noqa: E501 indices = self._get_data(self._join_url('/_stats/search', admin_forwarder))['indices'] - for (idx_name, data), (m_name, path) in product(iteritems(indices), INDEX_SEARCH_STATS): + for (idx_name, data), (m_name, path) in product(indices.items(), INDEX_SEARCH_STATS): tags = base_tags + ['index_name:' + idx_name] self._process_metric(data, m_name, 'gauge', path, tags=tags) @@ -362,7 +361,7 @@ def _process_pending_tasks_data(self, data, base_tags): p_tasks[task.get('priority')] += 1 average_time_in_queue += task.get('time_in_queue_millis', 0) - total = sum(itervalues(p_tasks)) + total = sum(p_tasks.values()) node_data = { 'pending_task_total': total, 'pending_tasks_priority_high': p_tasks['high'], @@ -377,7 +376,7 @@ def _process_pending_tasks_data(self, data, base_tags): self._process_metric(node_data, metric, *desc, tags=base_tags) def _process_stats_data(self, data, stats_metrics, base_tags): - for node_data in itervalues(data.get('nodes', {})): + for node_data in data.get('nodes', {}).values(): metric_hostname = None metrics_tags = list(base_tags) @@ -396,18 +395,18 @@ def _process_stats_data(self, data, stats_metrics, base_tags): metric_hostname = node_data[k] break - for metric, desc in iteritems(stats_metrics): + for metric, desc in stats_metrics.items(): self._process_metric(node_data, metric, *desc, tags=metrics_tags, hostname=metric_hostname) def _process_pshard_stats_data(self, data, pshard_stats_metrics, base_tags): - for metric, desc in iteritems(pshard_stats_metrics): + for metric, desc in pshard_stats_metrics.items(): pshard_tags = base_tags if desc[1].startswith('_all.'): pshard_tags = pshard_tags + ['index_name:_all'] self._process_metric(data, metric, *desc, tags=pshard_tags) # process index-level metrics if self._config.cluster_stats and self._config.detailed_index_stats: - for metric, desc in iteritems(pshard_stats_metrics): + for metric, desc in pshard_stats_metrics.items(): if desc[1].startswith('_all.'): for index in data['indices']: self.log.debug("Processing index %s", index) @@ -449,7 +448,7 @@ def _process_health_data(self, data, version, base_tags, service_check_tags): ): self.event(self._create_event(current_status, tags=base_tags)) - for metric, desc in iteritems(health_stats_for_version(version)): + for metric, desc in health_stats_for_version(version).items(): self._process_metric(data, metric, *desc, tags=base_tags) # Process the service check @@ -473,12 +472,12 @@ def _process_health_data(self, data, version, base_tags, service_check_tags): self.service_check(self.SERVICE_CHECK_CLUSTER_STATUS, dd_health.status, message=msg, tags=service_check_tags) def _process_policy_data(self, data, version, base_tags): - for policy, policy_data in iteritems(data): + for policy, policy_data in data.items(): repo = policy_data.get('policy', {}).get('repository', 'unknown') tags = base_tags + ['policy:{}'.format(policy), 'repository:{}'.format(repo)] slm_stats = slm_stats_for_version(version) - for metric, desc in iteritems(slm_stats): + for metric, desc in slm_stats.items(): self._process_metric(policy_data, metric, *desc, tags=tags) def _process_cat_allocation_data(self, admin_forwarder, version, base_tags): diff --git a/envoy/datadog_checks/envoy/check.py b/envoy/datadog_checks/envoy/check.py index dbc3576ca20b3..c825e5e48c11c 100644 --- a/envoy/datadog_checks/envoy/check.py +++ b/envoy/datadog_checks/envoy/check.py @@ -3,8 +3,7 @@ # Licensed under a 3-clause BSD style license (see LICENSE) import re from collections import defaultdict - -from six.moves.urllib.parse import urljoin, urlparse, urlunparse +from urllib.parse import urljoin, urlparse, urlunparse from datadog_checks.base import AgentCheck, OpenMetricsBaseCheckV2 diff --git a/envoy/datadog_checks/envoy/envoy.py b/envoy/datadog_checks/envoy/envoy.py index 747432d4a5950..b656fa34cfa2a 100644 --- a/envoy/datadog_checks/envoy/envoy.py +++ b/envoy/datadog_checks/envoy/envoy.py @@ -3,10 +3,10 @@ # Licensed under a 3-clause BSD style license (see LICENSE) import re from collections import defaultdict +from urllib.parse import urljoin import requests from six import PY2 -from six.moves.urllib.parse import urljoin from datadog_checks.base import AgentCheck, ConfigurationError, is_affirmative diff --git a/envoy/datadog_checks/envoy/parser.py b/envoy/datadog_checks/envoy/parser.py index 24e745e6f39ce..aa5be1f6dfd1b 100644 --- a/envoy/datadog_checks/envoy/parser.py +++ b/envoy/datadog_checks/envoy/parser.py @@ -6,8 +6,6 @@ from math import isnan from typing import Any, Dict, List, Tuple # noqa: F401 -from six.moves import range, zip - from .errors import UnknownMetric, UnknownTags from .metrics import LEGACY_TAG_OVERWRITE, METRIC_PREFIX, METRIC_TREE, MOD_METRICS diff --git a/esxi/datadog_checks/esxi/check.py b/esxi/datadog_checks/esxi/check.py index 6876c46976a6c..2a4d162a9bcb6 100644 --- a/esxi/datadog_checks/esxi/check.py +++ b/esxi/datadog_checks/esxi/check.py @@ -11,7 +11,6 @@ import socks from pyVim import connect from pyVmomi import vim, vmodl -from six import iteritems from datadog_checks.base import AgentCheck, is_affirmative from datadog_checks.base.utils.common import to_string @@ -132,7 +131,7 @@ def _validate_excluded_host_tags(self, excluded_host_tags): def _parse_metric_regex_filters(self, all_metric_filters): allowed_resource_types = RESOURCE_TYPE_TO_NAME.values() metric_filters = {} - for resource_type, filters in iteritems(all_metric_filters): + for resource_type, filters in all_metric_filters.items(): if resource_type not in allowed_resource_types: self.log.warning( "Ignoring metric_filter for resource '%s'. It should be one of '%s'", @@ -142,7 +141,7 @@ def _parse_metric_regex_filters(self, all_metric_filters): continue metric_filters[resource_type] = filters - return {k: [re.compile(r) for r in v] for k, v in iteritems(metric_filters)} + return {k: [re.compile(r) for r in v] for k, v in metric_filters.items()} def _parse_resource_filters(self, all_resource_filters): # Keep a list of resource filters ids (tuple of resource, property and type) that are already registered. diff --git a/esxi/datadog_checks/esxi/utils.py b/esxi/datadog_checks/esxi/utils.py index d4540881af139..bc2134e3c62ac 100644 --- a/esxi/datadog_checks/esxi/utils.py +++ b/esxi/datadog_checks/esxi/utils.py @@ -3,7 +3,6 @@ # Licensed under a 3-clause BSD style license (see LICENSE) from pyVmomi import vim -from six import iteritems from datadog_checks.base import to_string @@ -76,7 +75,7 @@ def get_mapped_instance_tag(metric_name): tag cannot be guessed by looking at the api results and has to be inferred using documentation or experience. This method acts as a utility to map a metric_name to the meaning of its instance tag. """ - for prefix, tag_key in iteritems(METRIC_TO_INSTANCE_TAG_MAPPING): + for prefix, tag_key in METRIC_TO_INSTANCE_TAG_MAPPING.items(): if metric_name.startswith(prefix): return tag_key return 'instance' diff --git a/etcd/datadog_checks/etcd/etcd.py b/etcd/datadog_checks/etcd/etcd.py index e219ff0575f29..a01ff4ea12373 100644 --- a/etcd/datadog_checks/etcd/etcd.py +++ b/etcd/datadog_checks/etcd/etcd.py @@ -1,7 +1,7 @@ # (C) Datadog, Inc. 2018-present # All rights reserved # Licensed under a 3-clause BSD style license (see LICENSE) -from six.moves.urllib.parse import urlparse +from urllib.parse import urlparse from datadog_checks.base import ConfigurationError, OpenMetricsBaseCheck, is_affirmative diff --git a/fluentd/datadog_checks/fluentd/fluentd.py b/fluentd/datadog_checks/fluentd/fluentd.py index f05e6cc2a116f..246cc2bfbdb84 100644 --- a/fluentd/datadog_checks/fluentd/fluentd.py +++ b/fluentd/datadog_checks/fluentd/fluentd.py @@ -4,8 +4,7 @@ # Licensed under Simplified BSD License (see LICENSE) import re - -from six.moves.urllib.parse import urlparse +from urllib.parse import urlparse from datadog_checks.base import AgentCheck, ConfigurationError from datadog_checks.base.utils.subprocess_output import get_subprocess_output diff --git a/fly_io/datadog_checks/fly_io/check.py b/fly_io/datadog_checks/fly_io/check.py index 840accdd57a02..7b44f5942e72d 100644 --- a/fly_io/datadog_checks/fly_io/check.py +++ b/fly_io/datadog_checks/fly_io/check.py @@ -2,7 +2,7 @@ # All rights reserved # Licensed under a 3-clause BSD style license (see LICENSE) -from six.moves.urllib.parse import quote_plus +from urllib.parse import quote_plus from datadog_checks.base import OpenMetricsBaseCheckV2 diff --git a/gitlab_runner/datadog_checks/gitlab_runner/gitlab_runner.py b/gitlab_runner/datadog_checks/gitlab_runner/gitlab_runner.py index 0e28fbaa39e54..54c625e03cc86 100644 --- a/gitlab_runner/datadog_checks/gitlab_runner/gitlab_runner.py +++ b/gitlab_runner/datadog_checks/gitlab_runner/gitlab_runner.py @@ -3,9 +3,9 @@ # Licensed under a 3-clause BSD style license (see LICENSE) from copy import deepcopy +from urllib.parse import urlparse import requests -from six.moves.urllib.parse import urlparse from datadog_checks.base.checks.openmetrics import OpenMetricsBaseCheck from datadog_checks.base.errors import CheckException diff --git a/glusterfs/datadog_checks/glusterfs/check.py b/glusterfs/datadog_checks/glusterfs/check.py index 71ff1cdcf6acb..64f84d5fe1dd7 100644 --- a/glusterfs/datadog_checks/glusterfs/check.py +++ b/glusterfs/datadog_checks/glusterfs/check.py @@ -12,8 +12,6 @@ import subprocess from typing import Dict, List # noqa: F401 -from six import iteritems - from datadog_checks.base import AgentCheck, ConfigurationError from datadog_checks.base.config import is_affirmative @@ -180,7 +178,7 @@ def submit_metrics(self, payload, prefix, metric_mapping, tags): Parse a payload with a given metric_mapping and submit metric for valid values. Some values contain measurements like `GiB` which should be removed and only submitted if consistent """ - for key, metric in iteritems(metric_mapping): + for key, metric in metric_mapping.items(): if key in payload: value = payload[key] diff --git a/go_expvar/datadog_checks/go_expvar/go_expvar.py b/go_expvar/datadog_checks/go_expvar/go_expvar.py index f26658bc38a3f..280d4adbf7783 100644 --- a/go_expvar/datadog_checks/go_expvar/go_expvar.py +++ b/go_expvar/datadog_checks/go_expvar/go_expvar.py @@ -5,9 +5,7 @@ import re from collections import defaultdict - -from six import iteritems -from six.moves.urllib.parse import urlparse +from urllib.parse import urlparse from datadog_checks.base import AgentCheck @@ -269,7 +267,7 @@ def items(self, object): for new_key, new_content in enumerate(object): yield str(new_key), new_content elif isinstance(object, dict): - for new_key, new_content in iteritems(object): + for new_key, new_content in object.items(): yield str(new_key), new_content else: self.log.warning("Could not parse this object, check the json served by the expvar") diff --git a/hdfs_datanode/datadog_checks/hdfs_datanode/hdfs_datanode.py b/hdfs_datanode/datadog_checks/hdfs_datanode/hdfs_datanode.py index 709b3f73540f3..f5fc7712adccd 100644 --- a/hdfs_datanode/datadog_checks/hdfs_datanode/hdfs_datanode.py +++ b/hdfs_datanode/datadog_checks/hdfs_datanode/hdfs_datanode.py @@ -1,10 +1,10 @@ # (C) Datadog, Inc. 2018-present # All rights reserved # Licensed under a 3-clause BSD style license (see LICENSE) +from urllib.parse import urljoin + from requests.exceptions import ConnectionError, HTTPError, InvalidURL, Timeout from simplejson import JSONDecodeError -from six import iteritems -from six.moves.urllib.parse import urljoin from datadog_checks.base import AgentCheck @@ -87,7 +87,7 @@ def _hdfs_datanode_metrics(self, beans, tags): self.log.debug("Bean name retrieved: %s", bean_name) - for metric, (metric_name, metric_type) in iteritems(self.HDFS_METRICS): + for metric, (metric_name, metric_type) in self.HDFS_METRICS.items(): metric_value = bean.get(metric) if metric_value is not None: self._set_metric(metric_name, metric_type, metric_value, tags) @@ -110,7 +110,7 @@ def _rest_request_to_json(self, url, object_path, query_params, tags): # Add query_params as arguments if query_params: - query = '&'.join(['{}={}'.format(key, value) for key, value in iteritems(query_params)]) + query = '&'.join(['{}={}'.format(key, value) for key, value in query_params.items()]) url = urljoin(url, '?' + query) self.log.debug('Attempting to connect to "%s"', url) diff --git a/hdfs_namenode/datadog_checks/hdfs_namenode/hdfs_namenode.py b/hdfs_namenode/datadog_checks/hdfs_namenode/hdfs_namenode.py index 9fd30a0c1dcc2..259703128e68c 100644 --- a/hdfs_namenode/datadog_checks/hdfs_namenode/hdfs_namenode.py +++ b/hdfs_namenode/datadog_checks/hdfs_namenode/hdfs_namenode.py @@ -3,10 +3,10 @@ # Licensed under a 3-clause BSD style license (see LICENSE) from __future__ import division +from urllib.parse import urljoin + from requests.exceptions import ConnectionError, HTTPError, InvalidURL, Timeout from simplejson import JSONDecodeError -from six import iteritems -from six.moves.urllib.parse import urljoin from datadog_checks.base import AgentCheck from datadog_checks.base.utils.common import compute_percent @@ -113,7 +113,7 @@ def _hdfs_namenode_metrics(self, beans, metrics, tags): if bean_name != bean_name: raise Exception("Unexpected bean name {}".format(bean_name)) - for metric, (metric_name, metric_type) in iteritems(metrics): + for metric, (metric_name, metric_type) in metrics.items(): metric_value = bean.get(metric) if metric_value is not None: @@ -146,7 +146,7 @@ def _rest_request_to_json(self, url, object_path, query_params, tags=None): # Add query_params as arguments if query_params: - query = '&'.join(['{}={}'.format(key, value) for key, value in iteritems(query_params)]) + query = '&'.join(['{}={}'.format(key, value) for key, value in query_params.items()]) url = urljoin(url, '?' + query) self.log.debug('Attempting to connect to "%s"', url) diff --git a/http_check/datadog_checks/http_check/http_check.py b/http_check/datadog_checks/http_check/http_check.py index 0bc665d290e9e..a3b93d3da6517 100644 --- a/http_check/datadog_checks/http_check/http_check.py +++ b/http_check/datadog_checks/http_check/http_check.py @@ -8,22 +8,17 @@ import socket import time from datetime import datetime +from urllib.parse import urlparse import requests from cryptography import x509 from requests import Response # noqa: F401 -from six import PY2, string_types -from six.moves.urllib.parse import urlparse from datadog_checks.base import AgentCheck, ensure_unicode, is_affirmative from .config import DEFAULT_EXPECTED_CODE, from_instance from .utils import get_ca_certs_path -# Apply thread-safety fix, see https://bugs.python.org/issue7980 -if PY2: - import _strptime # noqa - DEFAULT_EXPIRE_DAYS_WARNING = 14 DEFAULT_EXPIRE_DAYS_CRITICAL = 7 DEFAULT_EXPIRE_WARNING = DEFAULT_EXPIRE_DAYS_WARNING * 24 * 3600 @@ -132,7 +127,7 @@ def send_status_down(loginfo, down_msg): persist=True, stream=stream, json=data if method.upper() in DATA_METHODS and isinstance(data, dict) else None, - data=data if method.upper() in DATA_METHODS and isinstance(data, string_types) else None, + data=data if method.upper() in DATA_METHODS and isinstance(data, str) else None, ) except ( socket.timeout, diff --git a/ibm_mq/datadog_checks/ibm_mq/collectors/channel_metric_collector.py b/ibm_mq/datadog_checks/ibm_mq/collectors/channel_metric_collector.py index 61cf572d918ed..18145329f8ef2 100644 --- a/ibm_mq/datadog_checks/ibm_mq/collectors/channel_metric_collector.py +++ b/ibm_mq/datadog_checks/ibm_mq/collectors/channel_metric_collector.py @@ -3,8 +3,6 @@ # Licensed under a 3-clause BSD style license (see LICENSE) from typing import Callable, Dict, List # noqa: F401 -from six import iteritems - from datadog_checks.base import AgentCheck, to_string from datadog_checks.base.log import CheckLoggingAdapter # noqa: F401 @@ -168,7 +166,7 @@ def _submit_channel_status(self, queue_manager, search_channel_name, tags, chann def _submit_metrics_from_properties(self, channel_info, channel_name, metrics_map, tags): # type: (Dict, str, Dict[str, int], List[str] ) -> None - for metric_name, pymqi_type in iteritems(metrics_map): + for metric_name, pymqi_type in metrics_map.items(): metric_full_name = '{}.channel.{}'.format(metrics.METRIC_PREFIX, metric_name) if pymqi_type not in channel_info: self.log.debug("metric '%s' not found in channel: %s", metric_name, channel_name) @@ -181,7 +179,7 @@ def _submit_channel_count(self, channel_name, channel_status, channel_tags): self.log.warning("Status `%s` not found for channel `%s`", channel_status, channel_name) channel_status = STATUS_MQCHS_UNKNOWN - for status, status_label in iteritems(CHANNEL_STATUS_NAME_MAPPING): + for status, status_label in CHANNEL_STATUS_NAME_MAPPING.items(): status_active = int(status == channel_status) self.gauge( self.CHANNEL_COUNT_CHECK, diff --git a/ibm_mq/datadog_checks/ibm_mq/collectors/queue_metric_collector.py b/ibm_mq/datadog_checks/ibm_mq/collectors/queue_metric_collector.py index 2e7413fe3c68f..89ea372bcc564 100644 --- a/ibm_mq/datadog_checks/ibm_mq/collectors/queue_metric_collector.py +++ b/ibm_mq/datadog_checks/ibm_mq/collectors/queue_metric_collector.py @@ -4,8 +4,6 @@ import logging # noqa: F401 from typing import Any, Callable, Dict, List, Set # noqa: F401 -from six import iteritems - from datadog_checks.base import AgentCheck, to_string from datadog_checks.base.types import ServiceCheck # noqa: F401 from datadog_checks.ibm_mq.metrics import GAUGE @@ -150,7 +148,7 @@ def queue_manager_stats(self, queue_manager, tags): """ Get stats from the queue manager """ - for mname, pymqi_value in iteritems(metrics.queue_manager_metrics()): + for mname, pymqi_value in metrics.queue_manager_metrics().items(): try: m = queue_manager.inquire(pymqi_value) mname = '{}.queue_manager.{}'.format(metrics.METRIC_PREFIX, mname) @@ -198,7 +196,7 @@ def queue_stats(self, queue_manager, queue_name, tags): return enriched_tags def _submit_queue_stats(self, queue_info, queue_name, tags): - for metric_suffix, mq_attr in iteritems(metrics.queue_metrics()): + for metric_suffix, mq_attr in metrics.queue_metrics().items(): metric_name = '{}.queue.{}'.format(metrics.METRIC_PREFIX, metric_suffix) if callable(mq_attr): metric_value = mq_attr(queue_info) @@ -235,7 +233,7 @@ def get_pcf_queue_status_metrics(self, queue_manager, queue_name, tags): else: # Response is a list. It likely has only one member in it. for queue_info in response: - for mname, values in iteritems(metrics.pcf_metrics()): + for mname, values in metrics.pcf_metrics().items(): metric_name = '{}.queue.{}'.format(metrics.METRIC_PREFIX, mname) try: if callable(values): diff --git a/ibm_mq/datadog_checks/ibm_mq/config.py b/ibm_mq/datadog_checks/ibm_mq/config.py index 281ccab9c2dcd..b4a4385ce577a 100644 --- a/ibm_mq/datadog_checks/ibm_mq/config.py +++ b/ibm_mq/datadog_checks/ibm_mq/config.py @@ -6,7 +6,6 @@ import re from dateutil.tz import UTC -from six import PY2, iteritems from datadog_checks.base import AgentCheck, ConfigurationError, is_affirmative from datadog_checks.base.constants import ServiceCheck @@ -161,9 +160,6 @@ def __init__(self, instance, init_config): pattern = instance.get('queue_manager_process', init_config.get('queue_manager_process', '')) if pattern: - if PY2: - raise ConfigurationError('The `queue_manager_process` option is only supported on Agent 7') - pattern = pattern.replace('', re.escape(self.queue_manager_name)) self.queue_manager_process_pattern = re.compile(pattern) @@ -183,7 +179,7 @@ def _compile_tag_re(self): Compile regex strings from queue_tag_re option and return list of compiled regex/tag pairs """ queue_tag_list = [] - for regex_str, tags in iteritems(self._queue_tag_re): + for regex_str, tags in self._queue_tag_re.items(): try: queue_tag_list.append([re.compile(regex_str), [t.strip() for t in tags.split(',')]]) except TypeError: diff --git a/ibm_mq/datadog_checks/ibm_mq/ibm_mq.py b/ibm_mq/datadog_checks/ibm_mq/ibm_mq.py index 68e43342da3b4..9f72f4d110234 100644 --- a/ibm_mq/datadog_checks/ibm_mq/ibm_mq.py +++ b/ibm_mq/datadog_checks/ibm_mq/ibm_mq.py @@ -3,8 +3,6 @@ # Licensed under a 3-clause BSD style license (see LICENSE) import threading -from six import iteritems - from datadog_checks.base import AgentCheck from datadog_checks.ibm_mq.collectors.stats_collector import StatsCollector from datadog_checks.ibm_mq.metrics import COUNT, GAUGE @@ -121,7 +119,7 @@ def _collect_metadata(self, queue_manager): def send_metrics_from_properties(self, properties, metrics_map, prefix, tags): # type: (Dict, Dict, str, List[str]) -> None - for metric_name, (pymqi_type, metric_type) in iteritems(metrics_map): + for metric_name, (pymqi_type, metric_type) in metrics_map.items(): metric_full_name = '{}.{}'.format(prefix, metric_name) if pymqi_type not in properties: self.log.debug("MQ type `%s` not found in properties for metric `%s` and tags `%s`", metric_name, tags) diff --git a/ibm_was/datadog_checks/ibm_was/ibm_was.py b/ibm_was/datadog_checks/ibm_was/ibm_was.py index a5e47623beeb5..1ec764b3e9e47 100644 --- a/ibm_was/datadog_checks/ibm_was/ibm_was.py +++ b/ibm_was/datadog_checks/ibm_was/ibm_was.py @@ -5,7 +5,6 @@ import requests from lxml import etree -from six import iteritems from datadog_checks.base import AgentCheck, ConfigurationError, ensure_unicode, is_affirmative @@ -63,7 +62,7 @@ def check(self, _): server_tags = ['server:{}'.format(server.get('name'))] server_tags.extend(node_tags) - for category, prefix in iteritems(self.metric_categories): + for category, prefix in self.metric_categories.items(): if self.collect_stats.get(category): self.log.debug("Collecting %s stats", category) stats = self.get_node_from_name(server, category) @@ -151,7 +150,7 @@ def append_custom_queries(self): def setup_configured_stats(self): collect_stats = {} - for category, prefix in iteritems(metrics.METRIC_CATEGORIES): + for category, prefix in metrics.METRIC_CATEGORIES.items(): if is_affirmative(self.instance.get('collect_{}_stats'.format(prefix), True)): collect_stats[category] = True return collect_stats diff --git a/iis/datadog_checks/iis/iis.py b/iis/datadog_checks/iis/iis.py index 195ec391bf41e..97c6fd8295cfc 100644 --- a/iis/datadog_checks/iis/iis.py +++ b/iis/datadog_checks/iis/iis.py @@ -1,8 +1,6 @@ # (C) Datadog, Inc. 2010-present # All rights reserved # Licensed under Simplified BSD License (see LICENSE) -from six import PY3, iteritems - from datadog_checks.base import PDHBaseCheck, is_affirmative from .service_check import app_pool_service_check, site_service_check @@ -49,7 +47,7 @@ class IIS(PDHBaseCheck): APP_POOL = 'app_pool' def __new__(cls, name, init_config, instances): - if PY3 and not is_affirmative(instances[0].get('use_legacy_check_version', False)): + if not is_affirmative(instances[0].get('use_legacy_check_version', False)): from .check import IISCheckV2 return IISCheckV2(name, init_config, instances) @@ -84,7 +82,7 @@ def check(self, _): self.log.debug( "Unknown IIS counter: %s. Falling back to default submission.", counter.english_class_name ) - for instance_name, val in iteritems(counter_values): + for instance_name, val in counter_values.items(): tags = list(self._tags.get(self.instance_hash, [])) if not counter.is_single_instance(): @@ -102,7 +100,7 @@ def collect_sites(self, dd_name, metric_func, counter, counter_values): namespace = self.SITE remaining_sites = self._remaining_data[namespace] - for site_name, value in iteritems(counter_values): + for site_name, value in counter_values.items(): is_single_instance = counter.is_single_instance() if ( not is_single_instance @@ -139,7 +137,7 @@ def collect_app_pools(self, dd_name, metric_func, counter, counter_values): namespace = self.APP_POOL remaining_app_pools = self._remaining_data[namespace] - for app_pool_name, value in iteritems(counter_values): + for app_pool_name, value in counter_values.items(): is_single_instance = counter.is_single_instance() if ( not is_single_instance diff --git a/kong/datadog_checks/kong/kong.py b/kong/datadog_checks/kong/kong.py index 75c2ea5e345b2..ff5eb45768710 100644 --- a/kong/datadog_checks/kong/kong.py +++ b/kong/datadog_checks/kong/kong.py @@ -1,9 +1,10 @@ # (C) Datadog, Inc. 2010-present # All rights reserved # Licensed under Simplified BSD License (see LICENSE) +from urllib.parse import urlparse + import simplejson as json from six import PY2 -from six.moves.urllib.parse import urlparse from datadog_checks.base import AgentCheck, ConfigurationError diff --git a/kube_controller_manager/datadog_checks/kube_controller_manager/kube_controller_manager.py b/kube_controller_manager/datadog_checks/kube_controller_manager/kube_controller_manager.py index b8780f79a8398..70f372d78fdfd 100644 --- a/kube_controller_manager/datadog_checks/kube_controller_manager/kube_controller_manager.py +++ b/kube_controller_manager/datadog_checks/kube_controller_manager/kube_controller_manager.py @@ -5,7 +5,6 @@ import re import requests -from six import iteritems from datadog_checks.base import AgentCheck from datadog_checks.base.checks.kube_leader import KubeLeaderElectionMixin @@ -169,7 +168,7 @@ def check(self, instance): transformers[limiter + "_rate_limiter_use"] = self.rate_limiter_use queues = self.DEFAULT_QUEUES + instance.get("extra_queues", []) for queue in queues: - for metric, func in iteritems(self.QUEUE_METRICS_TRANSFORMERS): + for metric, func in self.QUEUE_METRICS_TRANSFORMERS.items(): transformers[queue + metric] = func # Support new metrics (introduced in v1.14.0) diff --git a/kube_dns/datadog_checks/kube_dns/kube_dns.py b/kube_dns/datadog_checks/kube_dns/kube_dns.py index 2efb8a8831144..799fd7586c50d 100644 --- a/kube_dns/datadog_checks/kube_dns/kube_dns.py +++ b/kube_dns/datadog_checks/kube_dns/kube_dns.py @@ -6,7 +6,6 @@ from copy import deepcopy import requests -from six import iteritems from datadog_checks.base import AgentCheck from datadog_checks.base.checks.openmetrics import OpenMetricsBaseCheck @@ -108,7 +107,7 @@ def submit_as_gauge_and_monotonic_count(self, metric_suffix, metric, scraper_con # Explicit shallow copy of the instance tags _tags = list(scraper_config['custom_tags']) - for label_name, label_value in iteritems(sample[self.SAMPLE_LABELS]): + for label_name, label_value in sample[self.SAMPLE_LABELS].items(): _tags.append('{}:{}'.format(label_name, label_value)) # submit raw metric self.gauge(metric_name, sample[self.SAMPLE_VALUE], _tags) diff --git a/kubelet/datadog_checks/kubelet/cadvisor.py b/kubelet/datadog_checks/kubelet/cadvisor.py index 274699b2145cf..18175f5c482c8 100644 --- a/kubelet/datadog_checks/kubelet/cadvisor.py +++ b/kubelet/datadog_checks/kubelet/cadvisor.py @@ -6,10 +6,9 @@ import logging import numbers from fnmatch import fnmatch +from urllib.parse import urlparse import requests -from six import iteritems -from six.moves.urllib.parse import urlparse from datadog_checks.base.utils.tagging import tagger @@ -118,7 +117,7 @@ def _publish_raw_metrics(self, metric, dat, tags, is_pod, depth=0): self.gauge(metric, float(dat), tags) elif isinstance(dat, dict): - for k, v in iteritems(dat): + for k, v in dat.items(): self._publish_raw_metrics(metric + '.%s' % k.lower(), v, tags, is_pod, depth + 1) elif isinstance(dat, list): diff --git a/kubelet/datadog_checks/kubelet/kubelet.py b/kubelet/datadog_checks/kubelet/kubelet.py index ac4d8f9826ef2..b028e067609c3 100644 --- a/kubelet/datadog_checks/kubelet/kubelet.py +++ b/kubelet/datadog_checks/kubelet/kubelet.py @@ -7,11 +7,10 @@ import sys from collections import defaultdict from copy import deepcopy +from urllib.parse import urlparse import requests from kubeutil import get_connection_info -from six import iteritems -from six.moves.urllib.parse import urlparse from datadog_checks.base import AgentCheck, OpenMetricsBaseCheck from datadog_checks.base.checks.kubelet_base.base import KubeletBase, KubeletCredentials, urljoin @@ -532,9 +531,9 @@ def _report_pods_running(self, pods, instance_tags): tags += instance_tags hash_tags = tuple(sorted(tags)) pods_tag_counter[hash_tags] += 1 - for tags, count in iteritems(pods_tag_counter): + for tags, count in pods_tag_counter.items(): self.gauge(self.NAMESPACE + '.pods.running', count, list(tags)) - for tags, count in iteritems(containers_tag_counter): + for tags, count in containers_tag_counter.items(): self.gauge(self.NAMESPACE + '.containers.running', count, list(tags)) def _report_container_spec_metrics(self, pod_list, instance_tags): @@ -578,14 +577,14 @@ def _report_container_spec_metrics(self, pod_list, instance_tags): tags += instance_tags try: - for resource, value_str in iteritems(ctr.get('resources', {}).get('requests', {})): + for resource, value_str in ctr.get('resources', {}).get('requests', {}).items(): value = self.parse_quantity(value_str) self.gauge('{}.{}.requests'.format(self.NAMESPACE, resource), value, tags) except (KeyError, AttributeError) as e: self.log.debug("Unable to retrieve container requests for %s: %s", c_name, e) try: - for resource, value_str in iteritems(ctr.get('resources', {}).get('limits', {})): + for resource, value_str in ctr.get('resources', {}).get('limits', {}).items(): value = self.parse_quantity(value_str) self.gauge('{}.{}.limits'.format(self.NAMESPACE, resource), value, tags) except (KeyError, AttributeError) as e: @@ -730,7 +729,7 @@ def append_pod_tags_to_volume_metrics(self, metric, scraper_config, hostname=Non # Determine the tags to send tags = self._metric_tags(metric.name, val, sample, scraper_config, hostname=custom_hostname) pvc_name, kube_ns = None, None - for label_name, label_value in iteritems(sample[self.SAMPLE_LABELS]): + for label_name, label_value in sample[self.SAMPLE_LABELS].items(): if label_name == "persistentvolumeclaim": pvc_name = label_value elif label_name == "namespace": diff --git a/kubelet/datadog_checks/kubelet/prometheus.py b/kubelet/datadog_checks/kubelet/prometheus.py index 7c5d53c0e7937..fa2537a598226 100644 --- a/kubelet/datadog_checks/kubelet/prometheus.py +++ b/kubelet/datadog_checks/kubelet/prometheus.py @@ -5,8 +5,6 @@ from copy import deepcopy -from six import iteritems - from datadog_checks.base.checks.kubelet_base.base import urljoin from datadog_checks.base.checks.openmetrics import OpenMetricsBaseCheck from datadog_checks.base.utils.tagging import tagger @@ -294,7 +292,7 @@ def _process_container_metric(self, type, metric_name, metric, scraper_config, l return samples = self._sum_values_by_context(metric, self._get_entity_id_if_container_metric) - for c_id, sample in iteritems(samples): + for c_id, sample in samples.items(): pod_uid = self._get_pod_uid(sample[self.SAMPLE_LABELS]) if self.pod_list_utils.is_excluded(c_id, pod_uid): continue @@ -340,7 +338,7 @@ def _process_pod_rate(self, metric_name, metric, scraper_config, labels=None): return samples = self._sum_values_by_context(metric, self._get_pod_uid_if_pod_metric) - for pod_uid, sample in iteritems(samples): + for pod_uid, sample in samples.items(): pod = get_pod_by_uid(pod_uid, self.pod_list) namespace = pod.get('metadata', {}).get('namespace', None) if self.pod_list_utils.is_namespace_excluded(namespace): @@ -372,7 +370,7 @@ def _process_usage_metric(self, m_name, metric, cache, scraper_config, labels=No seen_keys = {k: False for k in cache} samples = self._sum_values_by_context(metric, self._get_entity_id_if_container_metric) - for c_id, sample in iteritems(samples): + for c_id, sample in samples.items(): c_name = get_container_label(sample[self.SAMPLE_LABELS], 'name') if not c_name: continue @@ -407,7 +405,7 @@ def _process_usage_metric(self, m_name, metric, cache, scraper_config, labels=No self.gauge(m_name, val, tags) # purge the cache - for k, seen in iteritems(seen_keys): + for k, seen in seen_keys.items(): if not seen: del cache[k] @@ -418,7 +416,7 @@ def _process_limit_metric(self, m_name, metric, cache, scraper_config, pct_m_nam for each sample in the metric and reports the usage_pct """ samples = self._latest_value_by_context(metric, self._get_entity_id_if_container_metric) - for c_id, sample in iteritems(samples): + for c_id, sample in samples.items(): limit = sample[self.SAMPLE_VALUE] pod_uid = self._get_pod_uid(sample[self.SAMPLE_LABELS]) if self.pod_list_utils.is_excluded(c_id, pod_uid): diff --git a/kubernetes_state/datadog_checks/kubernetes_state/kubernetes_state.py b/kubernetes_state/datadog_checks/kubernetes_state/kubernetes_state.py index 28e4fd78d6a00..f74ff28cf8d77 100644 --- a/kubernetes_state/datadog_checks/kubernetes_state/kubernetes_state.py +++ b/kubernetes_state/datadog_checks/kubernetes_state/kubernetes_state.py @@ -7,8 +7,6 @@ from collections import Counter, defaultdict from copy import deepcopy -from six import iteritems - from datadog_checks.base.checks.openmetrics import OpenMetricsBaseCheck from datadog_checks.base.config import is_affirmative from datadog_checks.base.errors import CheckException @@ -175,18 +173,18 @@ def check(self, instance): self.process(scraper_config, metric_transformers=self.METRIC_TRANSFORMERS) # Logic for Cron Jobs - for job_tags, job in iteritems(self.failed_cron_job_counts): + for job_tags, job in self.failed_cron_job_counts.items(): self.monotonic_count(scraper_config['namespace'] + '.job.failed', job.count, list(job_tags)) job.set_previous_and_reset_current_ts() - for job_tags, job in iteritems(self.succeeded_cron_job_counts): + for job_tags, job in self.succeeded_cron_job_counts.items(): self.monotonic_count(scraper_config['namespace'] + '.job.succeeded', job.count, list(job_tags)) job.set_previous_and_reset_current_ts() # Logic for Jobs - for job_tags, job_count in iteritems(self.job_succeeded_count): + for job_tags, job_count in self.job_succeeded_count.items(): self.monotonic_count(scraper_config['namespace'] + '.job.succeeded', job_count, list(job_tags)) - for job_tags, job_count in iteritems(self.job_failed_count): + for job_tags, job_count in self.job_failed_count.items(): self.monotonic_count(scraper_config['namespace'] + '.job.failed', job_count, list(job_tags)) def _filter_metric(self, metric, scraper_config): @@ -508,7 +506,7 @@ def _condition_to_tag_check(self, sample, base_sc_name, mapping, scraper_config, metric_name = scraper_config['namespace'] + '.node.by_condition' metric_tags = [] - for label_name, label_value in iteritems(sample[self.SAMPLE_LABELS]): + for label_name, label_value in sample[self.SAMPLE_LABELS].items(): metric_tags += self._build_tags(label_name, label_value, scraper_config) self.gauge( metric_name, @@ -632,7 +630,7 @@ def kube_pod_status_phase(self, metric, scraper_config): ) status_phase_counter[tuple(sorted(tags))] += sample[self.SAMPLE_VALUE] - for tags, count in iteritems(status_phase_counter): + for tags, count in status_phase_counter.items(): self.gauge(metric_name, count, tags=list(tags)) def _submit_metric_kube_pod_container_status_reason( @@ -650,7 +648,7 @@ def _submit_metric_kube_pod_container_status_reason( else: continue - for label_name, label_value in iteritems(sample[self.SAMPLE_LABELS]): + for label_name, label_value in sample[self.SAMPLE_LABELS].items(): if label_name == "reason": continue @@ -687,7 +685,7 @@ def kube_cronjob_next_schedule_time(self, metric, scraper_config): for sample in metric.samples: on_schedule = int(sample[self.SAMPLE_VALUE]) - curr_time tags = [] - for label_name, label_value in iteritems(sample[self.SAMPLE_LABELS]): + for label_name, label_value in sample[self.SAMPLE_LABELS].items(): tags += self._build_tags(label_name, label_value, scraper_config) tags += scraper_config['custom_tags'] @@ -703,7 +701,7 @@ def kube_job_complete(self, metric, scraper_config): service_check_name = scraper_config['namespace'] + '.job.complete' for sample in metric.samples: tags = [] - for label_name, label_value in iteritems(sample[self.SAMPLE_LABELS]): + for label_name, label_value in sample[self.SAMPLE_LABELS].items(): if label_name == 'job' or label_name == 'job_name': tags += self._get_job_tags(label_name, label_value, scraper_config) else: @@ -714,7 +712,7 @@ def kube_job_failed(self, metric, scraper_config): service_check_name = scraper_config['namespace'] + '.job.complete' for sample in metric.samples: tags = [] - for label_name, label_value in iteritems(sample[self.SAMPLE_LABELS]): + for label_name, label_value in sample[self.SAMPLE_LABELS].items(): if label_name == 'job' or label_name == 'job_name': tags += self._get_job_tags(label_name, label_value, scraper_config) else: @@ -725,7 +723,7 @@ def kube_job_status_failed(self, metric, scraper_config): for sample in metric.samples: job_ts = None tags = [] + scraper_config['custom_tags'] - for label_name, label_value in iteritems(sample[self.SAMPLE_LABELS]): + for label_name, label_value in sample[self.SAMPLE_LABELS].items(): if label_name == 'job' or label_name == 'job_name': tags += self._get_job_tags(label_name, label_value, scraper_config) job_ts = self._extract_job_timestamp(label_value) @@ -742,7 +740,7 @@ def kube_job_status_succeeded(self, metric, scraper_config): for sample in metric.samples: job_ts = None tags = [] + scraper_config['custom_tags'] - for label_name, label_value in iteritems(sample[self.SAMPLE_LABELS]): + for label_name, label_value in sample[self.SAMPLE_LABELS].items(): if label_name == 'job' or label_name == 'job_name': tags += self._get_job_tags(label_name, label_value, scraper_config) job_ts = self._extract_job_timestamp(label_value) @@ -780,7 +778,7 @@ def kube_node_status_condition(self, metric, scraper_config): ) by_condition_counter[tuple(sorted(tags))] += sample[self.SAMPLE_VALUE] - for tags, count in iteritems(by_condition_counter): + for tags, count in by_condition_counter.items(): self.gauge(metric_name, count, tags=list(tags)) def kube_node_status_ready(self, metric, scraper_config): @@ -850,7 +848,7 @@ def kube_node_spec_unschedulable(self, metric, scraper_config): if metric.type in METRIC_TYPES: for sample in metric.samples: tags = [] - for label_name, label_value in iteritems(sample[self.SAMPLE_LABELS]): + for label_name, label_value in sample[self.SAMPLE_LABELS].items(): tags += self._build_tags(label_name, label_value, scraper_config) tags += scraper_config['custom_tags'] status = statuses[int(sample[self.SAMPLE_VALUE])] # value can be 0 or 1 @@ -920,7 +918,7 @@ def sum_values_by_tags(self, metric, scraper_config): tags = self._tags_for_count(sample, config, scraper_config) object_counter[tuple(sorted(tags))] += sample[self.SAMPLE_VALUE] - for tags, count in iteritems(object_counter): + for tags, count in object_counter.items(): self.gauge(metric_name, count, tags=list(tags)) def count_objects_by_tags(self, metric, scraper_config): @@ -933,7 +931,7 @@ def count_objects_by_tags(self, metric, scraper_config): tags = self._tags_for_count(sample, config, scraper_config) object_counter[tuple(sorted(tags))] += 1 - for tags, count in iteritems(object_counter): + for tags, count in object_counter.items(): self.gauge(metric_name, count, tags=list(tags)) def _tags_for_count(self, sample, count_config, scraper_config): @@ -980,7 +978,7 @@ def _metric_tags(self, metric_name, val, sample, scraper_config, hostname=None): custom_tags = scraper_config['custom_tags'] _tags = list(custom_tags) _tags += scraper_config['_metric_tags'] - for label_name, label_value in iteritems(sample[self.SAMPLE_LABELS]): + for label_name, label_value in sample[self.SAMPLE_LABELS].items(): if label_name not in scraper_config['exclude_labels']: _tags += self._build_tags(label_name, label_value, scraper_config) return self._finalize_tags_to_submit( diff --git a/lighttpd/datadog_checks/lighttpd/lighttpd.py b/lighttpd/datadog_checks/lighttpd/lighttpd.py index b3b05a8ac0cfb..0c9201891f20c 100644 --- a/lighttpd/datadog_checks/lighttpd/lighttpd.py +++ b/lighttpd/datadog_checks/lighttpd/lighttpd.py @@ -3,8 +3,7 @@ # Licensed under Simplified BSD License (see LICENSE) import re - -from six.moves.urllib.parse import urlparse +from urllib.parse import urlparse from datadog_checks.base import AgentCheck diff --git a/linux_proc_extras/datadog_checks/linux_proc_extras/linux_proc_extras.py b/linux_proc_extras/datadog_checks/linux_proc_extras/linux_proc_extras.py index d44a735bd01cf..18228bd74154f 100644 --- a/linux_proc_extras/datadog_checks/linux_proc_extras/linux_proc_extras.py +++ b/linux_proc_extras/datadog_checks/linux_proc_extras/linux_proc_extras.py @@ -5,8 +5,6 @@ from collections import defaultdict -from six import iteritems - from datadog_checks.base import AgentCheck from datadog_checks.base.utils.subprocess_output import get_subprocess_output @@ -53,7 +51,7 @@ def set_paths(self): "interrupts_info": "interrupts", } - for key, path in iteritems(self.proc_path_map): + for key, path in self.proc_path_map.items(): self.proc_path_map[key] = "{procfs}/{path}".format(procfs=proc_location, path=path) def get_inode_info(self): diff --git a/mapr/datadog_checks/mapr/mapr.py b/mapr/datadog_checks/mapr/mapr.py index 295b4e6364e53..994a17302a65f 100644 --- a/mapr/datadog_checks/mapr/mapr.py +++ b/mapr/datadog_checks/mapr/mapr.py @@ -6,8 +6,6 @@ import os import re -from six import iteritems - from datadog_checks.base import AgentCheck, ensure_unicode, is_affirmative from datadog_checks.base.errors import CheckException @@ -191,7 +189,7 @@ def should_collect_metric(self, metric_name): def submit_metric(self, metric): metric_name = metric['metric'] tags = copy.deepcopy(self.custom_tags) - for k, v in iteritems(metric['tags']): + for k, v in metric['tags'].items(): if k == 'clustername': tags.append("{}:{}".format('mapr_cluster', v)) if not self._disable_legacy_cluster_tag: diff --git a/marathon/datadog_checks/marathon/marathon.py b/marathon/datadog_checks/marathon/marathon.py index 5bacbf863f2c9..c5edf1d3bd393 100644 --- a/marathon/datadog_checks/marathon/marathon.py +++ b/marathon/datadog_checks/marathon/marathon.py @@ -2,10 +2,9 @@ # (C) graemej 2014 # All rights reserved # Licensed under a 3-clause BSD style license (see LICENSE) +from urllib.parse import urljoin import requests -from six import iteritems -from six.moves.urllib.parse import urljoin from datadog_checks.base import AgentCheck @@ -217,7 +216,7 @@ def process_queues(self, url, acs_url, tags=None, label_tags=None, group=None): queued.add(queue['app']['id']) - for m_type, sub_metric in iteritems(self.QUEUE_METRICS): + for m_type, sub_metric in self.QUEUE_METRICS.items(): if isinstance(sub_metric, list): for attr, name in sub_metric: try: diff --git a/marklogic/datadog_checks/marklogic/parsers/request.py b/marklogic/datadog_checks/marklogic/parsers/request.py index 8537926a95e77..7c54d3621d5b6 100644 --- a/marklogic/datadog_checks/marklogic/parsers/request.py +++ b/marklogic/datadog_checks/marklogic/parsers/request.py @@ -3,8 +3,6 @@ # Licensed under a 3-clause BSD style license (see LICENSE) from typing import Any, Dict, Generator, List, Tuple # noqa: F401 -from six import iteritems - from .common import build_metric_to_submit, is_metric @@ -22,7 +20,7 @@ def _parse_request_metrics(data, tags): # type: (Dict[str, Any], List[str]) -> Generator[Tuple, None, None] list_summary = data['request-default-list']['list-summary'] - for key, value in iteritems(list_summary): + for key, value in list_summary.items(): if is_metric(value): metric = build_metric_to_submit("requests.{}".format(key), value, tags) if metric is not None: diff --git a/marklogic/datadog_checks/marklogic/parsers/status.py b/marklogic/datadog_checks/marklogic/parsers/status.py index 50af6190ee4fa..20ca0459773fa 100644 --- a/marklogic/datadog_checks/marklogic/parsers/status.py +++ b/marklogic/datadog_checks/marklogic/parsers/status.py @@ -3,8 +3,6 @@ # Licensed under a 3-clause BSD style license (see LICENSE) from typing import Any, Dict, Generator, List, Tuple # noqa: F401 -from six import iteritems - from ..constants import RESOURCE_TYPES from .common import build_metric_to_submit, is_metric @@ -26,7 +24,7 @@ def parse_per_resource_status_metrics(resource_type, data, tags): def parse_summary_status_base_metrics(data, tags): # type: (Dict[str, Any], List[str]) -> Generator[Tuple, None, None] relations = data['local-cluster-status']['status-relations'] - for key, resource_data in iteritems(relations): + for key, resource_data in relations.items(): if not key.endswith('-status'): continue resource_type = resource_data['typeref'] @@ -40,7 +38,7 @@ def parse_summary_status_base_metrics(data, tags): def _parse_status_metrics(metric_prefix, metrics, tags): # type: (str, Dict[str, Any], List[str]) -> Generator[Tuple, None, None] - for key, data in iteritems(metrics): + for key, data in metrics.items(): if key in ['rate-properties', 'load-properties']: prop_type = key[: key.index('-properties')] total_key = 'total-' + prop_type diff --git a/marklogic/datadog_checks/marklogic/parsers/storage.py b/marklogic/datadog_checks/marklogic/parsers/storage.py index 2d131e371a305..aa5e1d51b84ba 100644 --- a/marklogic/datadog_checks/marklogic/parsers/storage.py +++ b/marklogic/datadog_checks/marklogic/parsers/storage.py @@ -3,8 +3,6 @@ # Licensed under a 3-clause BSD style license (see LICENSE) from typing import Any, Dict, Generator, List, Tuple # noqa: F401 -from six import iteritems - from .common import build_metric_to_submit, is_metric @@ -39,7 +37,7 @@ def _parse_storage_metrics(data, tags, include_location_forest): host_tags.append('marklogic_host_name:{}'.format(hosts_meta[host_id])) for location_data in host_data['locations']['location']: location_tags = host_tags + ['storage_path:{}'.format(location_data['path'])] - for host_key, host_value in iteritems(location_data): + for host_key, host_value in location_data.items(): if host_key == 'location-forests': location_value = host_value['location-forest'] for forest_data in location_value: @@ -48,7 +46,7 @@ def _parse_storage_metrics(data, tags, include_location_forest): "forest_name:{}".format(forest_data['nameref']), ] if include_location_forest: - for forest_key, forest_value in iteritems(forest_data): + for forest_key, forest_value in forest_data.items(): if forest_key == 'disk-size': metric = build_metric_to_submit( "forests.storage.{}".format(forest_key), forest_value, tags=forest_tags diff --git a/mesos_slave/datadog_checks/mesos_slave/mesos_slave.py b/mesos_slave/datadog_checks/mesos_slave/mesos_slave.py index 49f381e45ca84..4b53468797345 100644 --- a/mesos_slave/datadog_checks/mesos_slave/mesos_slave.py +++ b/mesos_slave/datadog_checks/mesos_slave/mesos_slave.py @@ -6,10 +6,9 @@ Collects metrics from mesos slave node. """ +from urllib.parse import urlparse from requests.exceptions import Timeout -from six import iteritems -from six.moves.urllib.parse import urlparse from datadog_checks.base import AgentCheck, ConfigurationError @@ -157,7 +156,7 @@ def _process_stats_info(self, url, tags): self.STATS_METRICS, ] for m in metrics: - for key_name, (metric_name, metric_func) in iteritems(m): + for key_name, (metric_name, metric_func) in m.items(): if key_name in stats_metrics: metric_func(self, metric_name, stats_metrics[key_name], tags=stats_tags) self.service_check(self.SERVICE_CHECK_NAME, AgentCheck.OK, tags=stats_tags) @@ -222,5 +221,5 @@ def _process_tasks(self, tasks, state_metrics, tags): task_tags = ['task_name:' + t['name']] task_tags.extend(tags) self.service_check(t['name'] + '.ok', self.TASK_STATUS[t['state']], tags=task_tags) - for key_name, (metric_name, metric_func) in iteritems(self.TASK_METRICS): + for key_name, (metric_name, metric_func) in self.TASK_METRICS.items(): metric_func(self, metric_name, t['resources'][key_name], tags=task_tags) diff --git a/mongo/datadog_checks/mongo/collectors/replica.py b/mongo/datadog_checks/mongo/collectors/replica.py index 95404d3dedfcd..bc0eec37afc89 100644 --- a/mongo/datadog_checks/mongo/collectors/replica.py +++ b/mongo/datadog_checks/mongo/collectors/replica.py @@ -3,8 +3,7 @@ # Licensed under a 3-clause BSD style license (see LICENSE) import time - -from six.moves.urllib.parse import urlsplit +from urllib.parse import urlsplit from datadog_checks.mongo.api import MongoApi from datadog_checks.mongo.collectors.base import MongoCollector diff --git a/mongo/datadog_checks/mongo/utils.py b/mongo/datadog_checks/mongo/utils.py index 24a84ca025c0f..f2c72e3303a94 100644 --- a/mongo/datadog_checks/mongo/utils.py +++ b/mongo/datadog_checks/mongo/utils.py @@ -2,9 +2,9 @@ # All rights reserved # Licensed under a 3-clause BSD style license (see LICENSE) import re +from urllib.parse import quote_plus, unquote_plus, urlencode, urlunparse import pymongo -from six.moves.urllib.parse import quote_plus, unquote_plus, urlencode, urlunparse def build_connection_string(hosts, scheme, username=None, password=None, database=None, options=None): diff --git a/network/datadog_checks/network/check_bsd.py b/network/datadog_checks/network/check_bsd.py index fd76bd9b447c5..545694c05ad4a 100644 --- a/network/datadog_checks/network/check_bsd.py +++ b/network/datadog_checks/network/check_bsd.py @@ -2,17 +2,12 @@ # All rights reserved # Licensed under Simplified BSD License (see LICENSE) -from six import PY3, iteritems - from datadog_checks.base.utils.platform import Platform from datadog_checks.base.utils.subprocess_output import SubprocessOutputEmptyError, get_subprocess_output from . import Network from .const import BSD_TCP_METRICS -if PY3: - long = int - class BSDNetwork(Network): def __init__(self, name, init_config, instances): @@ -80,15 +75,15 @@ def check(self, _): current = iface # Filter inactive interfaces - if self.parse_long(x[-5]) or self.parse_long(x[-2]): + if self.parse_int(x[-5]) or self.parse_int(x[-2]): iface = current metrics = { - 'bytes_rcvd': self.parse_long(x[-5]), - 'bytes_sent': self.parse_long(x[-2]), - 'packets_in.count': self.parse_long(x[-7]), - 'packets_in.error': self.parse_long(x[-6]), - 'packets_out.count': self.parse_long(x[-4]), - 'packets_out.error': self.parse_long(x[-3]), + 'bytes_rcvd': self.parse_int(x[-5]), + 'bytes_sent': self.parse_int(x[-2]), + 'packets_in.count': self.parse_int(x[-7]), + 'packets_in.error': self.parse_int(x[-6]), + 'packets_out.count': self.parse_int(x[-4]), + 'packets_out.error': self.parse_int(x[-3]), } self.submit_devicemetrics(iface, metrics, custom_tags) except SubprocessOutputEmptyError: @@ -139,7 +134,7 @@ def check(self, _): # udp6 0 0 :::41458 :::* metrics = self.parse_cx_state(lines[2:], self.tcp_states['netstat'], 5) - for metric, value in iteritems(metrics): + for metric, value in metrics.items(): self.gauge(metric, value, tags=custom_tags) except SubprocessOutputEmptyError: self.log.exception("Error collecting connection states.") diff --git a/network/datadog_checks/network/check_linux.py b/network/datadog_checks/network/check_linux.py index dc4d1683c213b..986003b763052 100644 --- a/network/datadog_checks/network/check_linux.py +++ b/network/datadog_checks/network/check_linux.py @@ -4,8 +4,6 @@ import os import socket -from six import PY3, iteritems - from datadog_checks.base import is_affirmative from datadog_checks.base.utils.common import pattern_filter from datadog_checks.base.utils.subprocess_output import SubprocessOutputEmptyError, get_subprocess_output @@ -19,9 +17,6 @@ except ImportError: from datadog_checks.base.stubs import datadog_agent -if PY3: - long = int - class LinuxNetwork(Network): def __init__(self, name, init_config, instances): @@ -89,7 +84,7 @@ def check(self, _): self.histogram('system.net.tcp.recv_q', recvq, custom_tags + ["state:" + state]) self.histogram('system.net.tcp.send_q', sendq, custom_tags + ["state:" + state]) - for metric, value in iteritems(metrics): + for metric, value in metrics.items(): self.gauge(metric, value, tags=custom_tags) except OSError as e: @@ -107,7 +102,7 @@ def check(self, _): # udp6 0 0 :::41458 :::* metrics = self.parse_cx_state(lines[2:], self.tcp_states['netstat'], 5) - for metric, value in iteritems(metrics): + for metric, value in metrics.items(): self.gauge(metric, value, tags=custom_tags) if self._collect_cx_queues: @@ -136,17 +131,17 @@ def check(self, _): cols = line.split(':', 1) x = cols[1].split() # Filter inactive interfaces - if self.parse_long(x[0]) or self.parse_long(x[8]): + if self.parse_int(x[0]) or self.parse_int(x[8]): iface = cols[0].strip() metrics = { - 'bytes_rcvd': self.parse_long(x[0]), - 'bytes_sent': self.parse_long(x[8]), - 'packets_in.count': self.parse_long(x[1]), - 'packets_in.drop': self.parse_long(x[3]), - 'packets_in.error': self.parse_long(x[2]) + self.parse_long(x[3]), - 'packets_out.count': self.parse_long(x[9]), - 'packets_out.drop': self.parse_long(x[11]), - 'packets_out.error': self.parse_long(x[10]) + self.parse_long(x[11]), + 'bytes_rcvd': self.parse_int(x[0]), + 'bytes_sent': self.parse_int(x[8]), + 'packets_in.count': self.parse_int(x[1]), + 'packets_in.drop': self.parse_int(x[3]), + 'packets_in.error': self.parse_int(x[2]) + self.parse_int(x[3]), + 'packets_out.count': self.parse_int(x[9]), + 'packets_out.drop': self.parse_int(x[11]), + 'packets_out.error': self.parse_int(x[10]) + self.parse_int(x[11]), } self.submit_devicemetrics(iface, metrics, custom_tags) self._handle_ethtool_stats(iface, custom_tags) @@ -251,14 +246,14 @@ def check(self, _): for met in nstat_metrics_names[k]: if met in netstat_data.get(k, {}): self.submit_netmetric( - nstat_metrics_names[k][met], self.parse_long(netstat_data[k][met]), tags=custom_tags + nstat_metrics_names[k][met], self.parse_int(netstat_data[k][met]), tags=custom_tags ) for k in nstat_metrics_gauge_names: for met in nstat_metrics_gauge_names[k]: if met in netstat_data.get(k, {}): self._submit_netmetric_gauge( - nstat_metrics_gauge_names[k][met], self.parse_long(netstat_data[k][met]), tags=custom_tags + nstat_metrics_gauge_names[k][met], self.parse_int(netstat_data[k][met]), tags=custom_tags ) # Get the conntrack -S information @@ -460,7 +455,7 @@ def _submit_ena_metrics(self, iface, vals_by_metric, tags): assert m in allowed count = 0 - for metric, val in iteritems(vals_by_metric): + for metric, val in vals_by_metric.items(): self.log.debug("Submitting system.net.%s", metric) self.gauge('system.net.%s' % metric, val, tags=metric_tags) count += 1 @@ -480,9 +475,9 @@ def _submit_ethtool_metrics(self, iface, ethtool_metrics, base_tags): base_tags_with_device.append('device:{}'.format(iface)) count = 0 - for ethtool_tag, metric_map in iteritems(ethtool_metrics): + for ethtool_tag, metric_map in ethtool_metrics.items(): tags = base_tags_with_device + [ethtool_tag] - for metric, val in iteritems(metric_map): + for metric, val in metric_map.items(): self.log.debug("Submitting system.net.%s", metric) self.monotonic_count('system.net.%s' % metric, val, tags=tags) count += 1 diff --git a/network/datadog_checks/network/check_solaris.py b/network/datadog_checks/network/check_solaris.py index 62be4697b88ce..da5868ec0f4e8 100644 --- a/network/datadog_checks/network/check_solaris.py +++ b/network/datadog_checks/network/check_solaris.py @@ -2,16 +2,11 @@ # All rights reserved # Licensed under Simplified BSD License (see LICENSE) -from six import PY3, iteritems - from datadog_checks.base.utils.subprocess_output import SubprocessOutputEmptyError, get_subprocess_output from . import Network from .const import SOLARIS_TCP_METRICS -if PY3: - long = int - class SolarisNetwork(Network): def __init__(self, name, init_config, instances): @@ -25,7 +20,7 @@ def check(self, instance): try: netstat, _, _ = get_subprocess_output(["kstat", "-p", "link:0:"], self.log) metrics_by_interface = self._parse_solaris_netstat(netstat) - for interface, metrics in iteritems(metrics_by_interface): + for interface, metrics in metrics_by_interface.items(): self.submit_devicemetrics(interface, metrics, custom_tags) except SubprocessOutputEmptyError: self.log.exception("Error collecting kstat stats.") @@ -136,7 +131,7 @@ def _parse_solaris_netstat(self, netstat_output): # Add it to this interface's list of metrics. metrics = metrics_by_interface.get(iface, {}) - metrics[ddname] = self.parse_long(cols[1]) + metrics[ddname] = self.parse_int(cols[1]) metrics_by_interface[iface] = metrics return metrics_by_interface diff --git a/network/datadog_checks/network/check_windows.py b/network/datadog_checks/network/check_windows.py index 0115a4588fb3a..fa56342dc77b7 100644 --- a/network/datadog_checks/network/check_windows.py +++ b/network/datadog_checks/network/check_windows.py @@ -7,13 +7,10 @@ from ctypes.wintypes import DWORD import psutil -from six import PY3, iteritems from . import Network Iphlpapi = windll.Iphlpapi -if PY3: - long = int class TCPSTATS(Structure): @@ -82,7 +79,7 @@ def _cx_state_psutil(self, tags=None): else: metrics[metric] += 1 - for metric, value in iteritems(metrics): + for metric, value in metrics.items(): self.gauge(metric, value, tags=tags) def _cx_counters_psutil(self, tags=None): @@ -90,7 +87,7 @@ def _cx_counters_psutil(self, tags=None): Collect metrics about interfaces counters using psutil """ tags = [] if tags is None else tags - for iface, counters in iteritems(psutil.net_io_counters(pernic=True)): + for iface, counters in psutil.net_io_counters(pernic=True).items(): metrics = { 'bytes_rcvd': counters.bytes_recv, 'bytes_sent': counters.bytes_sent, diff --git a/network/datadog_checks/network/network.py b/network/datadog_checks/network/network.py index ef20c46657bc1..b9f6d5cf97cac 100644 --- a/network/datadog_checks/network/network.py +++ b/network/datadog_checks/network/network.py @@ -7,35 +7,24 @@ """ import re +import shutil import socket import psutil -from six import PY3, iteritems, itervalues from datadog_checks.base import AgentCheck, ConfigurationError from datadog_checks.base.errors import CheckException from datadog_checks.base.utils.platform import Platform +# fcntl only available on Unix systems try: import fcntl except ImportError: fcntl = None -if PY3: - long = int - -# Use a different find_executable implementation depending on Python version, -# because we want to avoid depending on distutils. -if PY3: - import shutil - - def find_executable(name): - return shutil.which(name) - -else: - # Fallback to distutils for Python 2 as shutil.which was added on Python 3.3 - from distutils.spawn import find_executable +def find_executable(name): + return shutil.which(name) class Network(AgentCheck): @@ -257,7 +246,7 @@ def submit_devicemetrics(self, iface, vals_by_metric, tags): ) count = 0 - for metric, val in iteritems(vals_by_metric): + for metric, val in vals_by_metric.items(): self.rate('system.net.%s' % metric, val, tags=metric_tags) count += 1 self.log.debug("tracked %s network metrics for interface %s", count, iface) @@ -273,9 +262,9 @@ def get_expected_metrics(self): ] return expected_metrics - def parse_long(self, v): + def parse_int(self, v): try: - return long(v) + return int(v) except ValueError: return 0 @@ -291,7 +280,7 @@ def submit_regexed_values(self, output, regex_list, tags): for regex, metric in regex_list: value = re.match(regex, line) if value: - self.submit_netmetric(metric, self.parse_long(value.group(1)), tags=tags) + self.submit_netmetric(metric, self.parse_int(value.group(1)), tags=tags) def is_collect_cx_state_runnable(self, proc_location): """ @@ -325,7 +314,7 @@ def get_net_proc_base_location(proc_location): return net_proc_base_location def _get_metrics(self): - return {val: 0 for val in itervalues(self.cx_state_gauge)} + return {val: 0 for val in self.cx_state_gauge.values()} def parse_cx_state(self, lines, tcp_states, state_col, protocol=None, ip_version=None): """ diff --git a/openstack/datadog_checks/openstack/openstack.py b/openstack/datadog_checks/openstack/openstack.py index c8a3bfae19c44..e4f141197b98f 100644 --- a/openstack/datadog_checks/openstack/openstack.py +++ b/openstack/datadog_checks/openstack/openstack.py @@ -8,11 +8,10 @@ import re import time from datetime import datetime, timedelta +from urllib.parse import urljoin import requests import simplejson as json -from six import iteritems -from six.moves.urllib.parse import urljoin from datadog_checks.base import AgentCheck, is_affirmative @@ -852,7 +851,7 @@ def get_stats_for_single_hypervisor(self, hyp_id, instance, host_tags=None, cust else: self.service_check(self.HYPERVISOR_SC, AgentCheck.OK, tags=service_check_tags) - for label, val in iteritems(hyp): + for label, val in hyp.items(): if label in NOVA_HYPERVISOR_METRICS: metric_label = "openstack.nova.{0}".format(label) self.gauge(metric_label, val, tags=tags) @@ -1196,7 +1195,7 @@ def check(self, instance): # The scopes we iterate over should all be OpenStackProjectScope # instances projects = [] - for _, scope in iteritems(scope_map): + for _, scope in scope_map.items(): # Store the scope on the object so we don't have to keep passing it around self._current_scope = scope @@ -1408,7 +1407,7 @@ def get_external_host_tags(self): """ self.log.debug("Collecting external_host_tags now") external_host_tags = [] - for k, v in iteritems(self.external_host_tags): + for k, v in self.external_host_tags.items(): external_host_tags.append((k, {SOURCE_TYPE: v})) self.log.debug("Sending external_host_tags: %s", external_host_tags) diff --git a/openstack_controller/datadog_checks/openstack_controller/legacy/api.py b/openstack_controller/datadog_checks/openstack_controller/legacy/api.py index ce758ccb9c0eb..6d115ef6e0163 100644 --- a/openstack_controller/datadog_checks/openstack_controller/legacy/api.py +++ b/openstack_controller/datadog_checks/openstack_controller/legacy/api.py @@ -4,11 +4,10 @@ import copy from os import environ +from urllib.parse import urljoin import requests from openstack import connection -from six import PY3 -from six.moves.urllib.parse import urljoin from .exceptions import ( AuthenticationNeeded, @@ -215,15 +214,10 @@ def get_os_hypervisors_detail(self): return self.connection.list_hypervisors() def get_os_hypervisor_uptime(self, hypervisor): - if PY3: - if hypervisor.uptime is None: - self._check_authentication() - self.connection.compute.get_hypervisor_uptime(hypervisor) - return hypervisor.uptime - else: - # Hypervisor uptime is not available in openstacksdk 0.24.0. - self.logger.warning("Hypervisor uptime is not available with this version of openstacksdk") - raise NotImplementedError() + if hypervisor.uptime is None: + self._check_authentication() + self.connection.compute.get_hypervisor_uptime(hypervisor) + return hypervisor.uptime def get_os_aggregates(self): # Each aggregate is missing the 'uuid' attribute compared to what is returned by SimpleApi diff --git a/openstack_controller/datadog_checks/openstack_controller/legacy/openstack_controller_legacy.py b/openstack_controller/datadog_checks/openstack_controller/legacy/openstack_controller_legacy.py index 2f567008fe4aa..0be77873e84dc 100644 --- a/openstack_controller/datadog_checks/openstack_controller/legacy/openstack_controller_legacy.py +++ b/openstack_controller/datadog_checks/openstack_controller/legacy/openstack_controller_legacy.py @@ -10,7 +10,6 @@ import requests from openstack.config.loader import OpenStackConfig -from six import iteritems, itervalues from datadog_checks.base import AgentCheck, is_affirmative from datadog_checks.base.utils.common import pattern_filter @@ -213,7 +212,7 @@ def collect_hypervisors_metrics( """ # Create a dictionary with hypervisor hostname as key and the list of project names as value hyp_project_names = defaultdict(set) - for server in itervalues(servers): + for server in servers.values(): hypervisor_hostname = server.get('hypervisor_hostname') if not hypervisor_hostname: self.log.debug( @@ -276,7 +275,7 @@ def get_stats_for_single_hypervisor( if not collect_hypervisor_metrics: return - for label, val in iteritems(hyp): + for label, val in hyp.items(): if label in NOVA_HYPERVISOR_METRICS: metric_label = "openstack.nova.{}".format(label) self.gauge(metric_label, val, tags=tags) @@ -365,7 +364,7 @@ def populate_servers_cache(self, projects, exclude_server_id_rules): # "tenant_id can also be requested which is alias of project_id but that is not # recommended to use as that will be removed in future." tenant_to_name = {} - for name, p in iteritems(projects): + for name, p in projects.items(): tenant_to_name[p.get('id')] = name cached_servers = self.servers_cache.get('servers') @@ -379,7 +378,7 @@ def populate_servers_cache(self, projects, exclude_server_id_rules): # Filter out excluded servers servers = {} - for updated_server_id, updated_server in iteritems(updated_servers): + for updated_server_id, updated_server in updated_servers.items(): if not any(re.match(rule, updated_server_id) for rule in exclude_server_id_rules): servers[updated_server_id] = updated_server @@ -717,7 +716,7 @@ def check(self, instance): projects = self.get_projects(include_project_name_rules, exclude_project_name_rules) if collect_project_metrics: - for project in itervalues(projects): + for project in projects.values(): self.collect_project_limit(project, custom_tags) servers = self.populate_servers_cache(projects, exclude_server_id_rules) @@ -733,16 +732,16 @@ def check(self, instance): if collect_server_diagnostic_metrics or collect_server_flavor_metrics: if collect_server_diagnostic_metrics: self.log.debug("Fetching stats from %s server(s)", len(servers)) - for server in itervalues(servers): + for server in servers.values(): self.collect_server_diagnostic_metrics(server, tags=custom_tags, use_shortname=use_shortname) if collect_server_flavor_metrics: - if len(servers) >= 1 and 'flavor_id' in next(itervalues(servers)): + if len(servers) >= 1 and 'flavor_id' in next(iter(servers.values())): self.log.debug("Fetching server flavors") # If flavors are not part of servers detail (new in version 2.47) then we need to fetch them flavors = self.get_flavors() else: flavors = None - for server in itervalues(servers): + for server in servers.values(): self.collect_server_flavor_metrics( server, flavors, tags=custom_tags, use_shortname=use_shortname ) @@ -790,7 +789,7 @@ def get_external_host_tags(self): """ self.log.debug("Collecting external_host_tags") external_host_tags = [] - for k, v in iteritems(self.external_host_tags): + for k, v in self.external_host_tags.items(): external_host_tags.append((k, {SOURCE_TYPE: v})) self.log.debug("Sending external_host_tags: %s", external_host_tags) @@ -831,7 +830,7 @@ def get_projects(self, include_project_name_rules, exclude_project_name_rules): filtered_project_names = pattern_filter( list(project_by_name), whitelist=include_project_name_rules, blacklist=exclude_project_name_rules ) - result = {name: v for (name, v) in iteritems(project_by_name) if name in filtered_project_names} + result = {name: v for (name, v) in project_by_name.items() if name in filtered_project_names} return result # Neutron Proxy Methods diff --git a/pgbouncer/datadog_checks/pgbouncer/pgbouncer.py b/pgbouncer/datadog_checks/pgbouncer/pgbouncer.py index 8a95e0d5f07c0..8401d06a81a2c 100644 --- a/pgbouncer/datadog_checks/pgbouncer/pgbouncer.py +++ b/pgbouncer/datadog_checks/pgbouncer/pgbouncer.py @@ -3,10 +3,10 @@ # Licensed under Simplified BSD License (see LICENSE) import re import time +from urllib.parse import urlparse import psycopg2 as pg from psycopg2 import extras as pgextras -from six.moves.urllib.parse import urlparse from datadog_checks.base import AgentCheck, ConfigurationError, is_affirmative from datadog_checks.pgbouncer.metrics import ( diff --git a/php_fpm/datadog_checks/php_fpm/php_fpm.py b/php_fpm/datadog_checks/php_fpm/php_fpm.py index 5aab48ccab25c..44096e33a6eb3 100644 --- a/php_fpm/datadog_checks/php_fpm/php_fpm.py +++ b/php_fpm/datadog_checks/php_fpm/php_fpm.py @@ -5,36 +5,31 @@ import random import socket import time - -from six import PY3, StringIO, iteritems, string_types -from six.moves.urllib.parse import urlparse +from io import StringIO +from urllib.parse import urlparse from datadog_checks.base import AgentCheck, is_affirmative from datadog_checks.base.utils.time import get_precise_time -if PY3: - # Flup package does not exist anymore so what's needed is vendored - # flup.client.fcgi_app.FCGIApp flup-py3 version 1.0.3 - from .vendor.fcgi_app import FCGIApp - - def get_connection(self): - if self._connect is not None: - if isinstance(self._connect, string_types): - sock = socket.socket(socket.AF_UNIX, socket.SOCK_STREAM) - sock.connect(self._connect) - elif hasattr(socket, 'create_connection'): - sock = socket.create_connection(self._connect) - else: - sock = socket.socket(socket.AF_INET, socket.SOCK_STREAM) - sock.connect(self._connect) - return sock +# Flup package does not exist anymore so what's needed is vendored +# flup.client.fcgi_app.FCGIApp flup-py3 version 1.0.3 +from .vendor.fcgi_app import FCGIApp - FCGIApp._getConnection = get_connection -else: - # flup version 1.0.3.dev-20110405 - from .vendor.fcgi_app_py2 import FCGIApp + +def get_connection(self): + if self._connect is not None: + if isinstance(self._connect, str): + sock = socket.socket(socket.AF_UNIX, socket.SOCK_STREAM) + sock.connect(self._connect) + elif hasattr(socket, 'create_connection'): + sock = socket.create_connection(self._connect) + else: + sock = socket.socket(socket.AF_INET, socket.SOCK_STREAM) + sock.connect(self._connect) + return sock +FCGIApp._getConnection = get_connection # Relax param filtering FCGIApp._environPrefixes.extend(('DOCUMENT_', 'SCRIPT_')) DEFAULT_TIMEOUT = 10 @@ -142,13 +137,13 @@ def _process_status(self, status_url, tags, http_host, use_fastcgi): self.gauge(self.STATUS_DURATION_NAME, check_duration, tags=metric_tags) - for key, mname in iteritems(self.GAUGES): + for key, mname in self.GAUGES.items(): if key not in data: self.log.warning("Gauge metric %s is missing from FPM status", key) continue self.gauge(mname, int(data[key]), tags=metric_tags) - for key, mname in iteritems(self.MONOTONIC_COUNTS): + for key, mname in self.MONOTONIC_COUNTS.items(): if key not in data: self.log.warning("Counter metric %s is missing from FPM status", key) continue diff --git a/postgres/datadog_checks/postgres/config.py b/postgres/datadog_checks/postgres/config.py index e1af379df5b1e..77a9849fcb123 100644 --- a/postgres/datadog_checks/postgres/config.py +++ b/postgres/datadog_checks/postgres/config.py @@ -4,8 +4,6 @@ # https://www.postgresql.org/docs/current/libpq-connect.html#LIBPQ-PARAMKEYWORDS from typing import Optional -from six import PY2, PY3, iteritems - from datadog_checks.base import AgentCheck, ConfigurationError, is_affirmative from datadog_checks.base.utils.aws import rds_parse_tags_from_endpoint from datadog_checks.base.utils.db.utils import get_agent_host_tags @@ -51,7 +49,7 @@ def __init__(self, instance, init_config, check): ) self.application_name = instance.get('application_name', 'datadog-agent') - if not self.isascii(self.application_name): + if not self.application_name.isascii(): raise ConfigurationError("Application name can include only ASCII characters: %s", self.application_name) self.query_timeout = int(instance.get('query_timeout', 5000)) @@ -220,7 +218,7 @@ def _get_custom_metrics(custom_metrics): m['query'] = m['query'] % '{metrics_columns}' try: - for ref, (_, mtype) in iteritems(m['metrics']): + for ref, (_, mtype) in m['metrics'].items(): cap_mtype = mtype.upper() if cap_mtype not in ('RATE', 'GAUGE', 'MONOTONIC'): raise ConfigurationError( @@ -233,17 +231,6 @@ def _get_custom_metrics(custom_metrics): raise Exception('Error processing custom metric `{}`: {}'.format(m, e)) return custom_metrics - @staticmethod - def isascii(application_name): - if PY3: - return application_name.isascii() - elif PY2: - try: - application_name.encode('ascii') - return True - except UnicodeEncodeError: - return False - @staticmethod def _aws_managed_authentication(aws, password): if 'managed_authentication' not in aws: diff --git a/postgres/datadog_checks/postgres/postgres.py b/postgres/datadog_checks/postgres/postgres.py index 76d7e8bb2101e..03d77018b93ee 100644 --- a/postgres/datadog_checks/postgres/postgres.py +++ b/postgres/datadog_checks/postgres/postgres.py @@ -9,7 +9,6 @@ import psycopg2 from cachetools import TTLCache -from six import iteritems from datadog_checks.base import AgentCheck from datadog_checks.base.utils.db import QueryExecutor @@ -595,7 +594,7 @@ def _query_scope(self, cursor, scope, instance_tags, is_custom_metrics, dbname=N tags = copy.copy(instance_tags) # Add tags from descriptors. - tags += [("%s:%s" % (k, v)) for (k, v) in iteritems(desc_map)] + tags += [("%s:%s" % (k, v)) for (k, v) in desc_map.items()] # Submit metrics to the Agent. for column, value in zip(cols, column_values): diff --git a/process/datadog_checks/process/process.py b/process/datadog_checks/process/process.py index 467f40b5e9708..8df86cd8119b8 100644 --- a/process/datadog_checks/process/process.py +++ b/process/datadog_checks/process/process.py @@ -10,7 +10,6 @@ from collections import defaultdict import psutil -from six import iteritems from datadog_checks.base import AgentCheck, is_affirmative from datadog_checks.base.utils.platform import Platform @@ -464,7 +463,7 @@ def check(self, _): self.last_pid_cache_ts[self.name] = 0 self.process_list_cache.reset() - for attr, mname in iteritems(ATTR_TO_METRIC): + for attr, mname in ATTR_TO_METRIC.items(): vals = [x for x in proc_state[attr] if x is not None] # skip [] if vals: @@ -480,7 +479,7 @@ def check(self, _): if mname in ['ioread_bytes', 'iowrite_bytes']: self.monotonic_count('system.processes.{}_count'.format(mname), sum_vals, tags=tags) - for attr, mname in iteritems(ATTR_TO_METRIC_RATE): + for attr, mname in ATTR_TO_METRIC_RATE.items(): vals = [x for x in proc_state[attr] if x is not None] if vals: self.rate('system.processes.{}'.format(mname), sum(vals), tags=tags) diff --git a/rabbitmq/datadog_checks/rabbitmq/rabbitmq.py b/rabbitmq/datadog_checks/rabbitmq/rabbitmq.py index 20ed0e56517d9..f52d2d9c8d62f 100644 --- a/rabbitmq/datadog_checks/rabbitmq/rabbitmq.py +++ b/rabbitmq/datadog_checks/rabbitmq/rabbitmq.py @@ -5,10 +5,9 @@ import re import time from collections import defaultdict +from urllib.parse import quote_plus, urljoin, urlparse from requests.exceptions import RequestException -from six import iteritems -from six.moves.urllib.parse import quote_plus, urljoin, urlparse from datadog_checks.base import AgentCheck, is_affirmative, to_native_string @@ -88,8 +87,8 @@ def _get_config(self, instance): NODE_TYPE: {'explicit': instance.get('nodes', []), 'regexes': instance.get('nodes_regexes', [])}, } - for object_type, filters in iteritems(specified): - for _, filter_objects in iteritems(filters): + for object_type, filters in specified.items(): + for _, filter_objects in filters.items(): if type(filter_objects) != list: raise TypeError("{0} / {0}_regexes parameter must be a list".format(object_type)) @@ -499,10 +498,10 @@ def get_connections_stat(self, instance, base_url, object_type, vhosts, limit_vh # 'state' does not exist for direct type connections. connection_states[conn.get('state', 'direct')] += 1 - for vhost, nb_conn in iteritems(stats): + for vhost, nb_conn in stats.items(): self.gauge('rabbitmq.connections', nb_conn, tags=['{}_vhost:{}'.format(TAG_PREFIX, vhost)] + custom_tags) - for conn_state, nb_conn in iteritems(connection_states): + for conn_state, nb_conn in connection_states.items(): self.gauge( 'rabbitmq.connections.state', nb_conn, diff --git a/silk/datadog_checks/silk/check.py b/silk/datadog_checks/silk/check.py index 75e2bd68986d1..ba54f7203cc87 100644 --- a/silk/datadog_checks/silk/check.py +++ b/silk/datadog_checks/silk/check.py @@ -2,8 +2,7 @@ # All rights reserved # Licensed under a 3-clause BSD style license (see LICENSE) from copy import deepcopy - -from six.moves.urllib.parse import urljoin, urlparse +from urllib.parse import urljoin, urlparse from datadog_checks.base import AgentCheck, ConfigurationError from datadog_checks.base.utils.time import get_timestamp diff --git a/snmp/datadog_checks/snmp/snmp.py b/snmp/datadog_checks/snmp/snmp.py index 40dd1e2f8417b..99dd29279da8a 100644 --- a/snmp/datadog_checks/snmp/snmp.py +++ b/snmp/datadog_checks/snmp/snmp.py @@ -14,8 +14,6 @@ from concurrent import futures from typing import Any, DefaultDict, Dict, List, Optional, Pattern, Tuple # noqa: F401 -from six import iteritems - from datadog_checks.base import AgentCheck, ConfigurationError, is_affirmative from datadog_checks.base.errors import CheckException from datadog_checks.snmp.utils import extract_value @@ -515,7 +513,7 @@ def report_metrics( self.log.debug('Ignoring metric %s', name) continue if isinstance(metric, ParsedTableMetric): - for index, val in iteritems(results[name]): + for index, val in results[name].items(): metric_tags = tags + self.get_index_tags(index, results, metric.index_tags, metric.column_tags) self.submit_metric( name, val, metric.forced_type, metric_tags, metric.options, metric.extract_value_pattern diff --git a/sqlserver/datadog_checks/sqlserver/connection.py b/sqlserver/datadog_checks/sqlserver/connection.py index 32af256ab57e0..525b852a626b2 100644 --- a/sqlserver/datadog_checks/sqlserver/connection.py +++ b/sqlserver/datadog_checks/sqlserver/connection.py @@ -5,8 +5,6 @@ import socket from contextlib import closing, contextmanager -from six import raise_from - from datadog_checks.base import AgentCheck, ConfigurationError from datadog_checks.base.log import get_check_logger from datadog_checks.sqlserver.cursor import CommenterCursorWrapper @@ -316,7 +314,7 @@ def open_db_connections(self, db_key, db_name=None, is_default=True, key_prefix= if is_default: # the message that is raised here (along with the exception stack trace) # is what will be seen in the agent status output. - raise_from(SQLConnectionError(check_err_message), None) + raise SQLConnectionError(check_err_message) from None else: # if not the default db, we should still log this exception # to give the customer an opportunity to fix the issue diff --git a/statsd/datadog_checks/statsd/statsd.py b/statsd/datadog_checks/statsd/statsd.py index a8bf7530252b3..9a9a4eddc5f1b 100644 --- a/statsd/datadog_checks/statsd/statsd.py +++ b/statsd/datadog_checks/statsd/statsd.py @@ -4,8 +4,7 @@ import re import socket - -from six import BytesIO +from io import BytesIO from datadog_checks.base import AgentCheck, ensure_bytes, ensure_unicode diff --git a/tls/datadog_checks/tls/tls.py b/tls/datadog_checks/tls/tls.py index 872f0ffb073fd..0f167ff6b4950 100644 --- a/tls/datadog_checks/tls/tls.py +++ b/tls/datadog_checks/tls/tls.py @@ -4,10 +4,9 @@ import socket import ssl from datetime import datetime +from urllib.parse import urlparse import service_identity -from six import text_type -from six.moves.urllib.parse import urlparse from datadog_checks.base import AgentCheck, is_affirmative @@ -149,7 +148,7 @@ def validate_certificate(self, cert): validator, host_type = self.validation_data try: - validator(cert, text_type(self._server_hostname)) + validator(cert, str(self._server_hostname)) except service_identity.VerificationError: message = 'The {} on the certificate does not match the given host'.format(host_type) self.log.debug(message) diff --git a/voltdb/datadog_checks/voltdb/check.py b/voltdb/datadog_checks/voltdb/check.py index 1452a9dc0ddc2..c868d06967763 100644 --- a/voltdb/datadog_checks/voltdb/check.py +++ b/voltdb/datadog_checks/voltdb/check.py @@ -4,7 +4,6 @@ from typing import Any, List, Optional, cast # noqa: F401 import requests # noqa: F401 -from six import raise_from from datadog_checks.base import AgentCheck from datadog_checks.base.utils.db import QueryManager @@ -52,7 +51,7 @@ def _raise_for_status_with_details(self, response): pass else: message += ' (details: {})'.format(details) - raise_from(Exception(message), exc) + raise Exception(message) from exc def _fetch_version(self): # type: () -> Optional[str] diff --git a/voltdb/datadog_checks/voltdb/client.py b/voltdb/datadog_checks/voltdb/client.py index cc276ed602cc8..29efb4a241a04 100644 --- a/voltdb/datadog_checks/voltdb/client.py +++ b/voltdb/datadog_checks/voltdb/client.py @@ -3,9 +3,9 @@ # Licensed under a 3-clause BSD style license (see LICENSE) import json from typing import Callable, Union # noqa: F401 +from urllib.parse import urljoin import requests -from six.moves.urllib.parse import urljoin class Client(object): diff --git a/voltdb/datadog_checks/voltdb/config.py b/voltdb/datadog_checks/voltdb/config.py index d827d2866073a..483ef1408d296 100644 --- a/voltdb/datadog_checks/voltdb/config.py +++ b/voltdb/datadog_checks/voltdb/config.py @@ -2,8 +2,7 @@ # All rights reserved # Licensed under a 3-clause BSD style license (see LICENSE) from typing import Callable, List, Optional # noqa: F401 - -from six.moves.urllib.parse import urlparse +from urllib.parse import urlparse from datadog_checks.base import ConfigurationError, is_affirmative diff --git a/vsphere/datadog_checks/vsphere/api.py b/vsphere/datadog_checks/vsphere/api.py index d11995233b89f..4d5e0de2dcda1 100644 --- a/vsphere/datadog_checks/vsphere/api.py +++ b/vsphere/datadog_checks/vsphere/api.py @@ -8,7 +8,6 @@ from pyVim import connect from pyVmomi import vim, vmodl -from six import itervalues from datadog_checks.base.log import CheckLoggingAdapter # noqa: F401 from datadog_checks.vsphere.config import VSphereConfig # noqa: F401 @@ -286,7 +285,7 @@ def get_infrastructure(self): # at this point they are custom pyvmomi objects and the attribute keys are not resolved. attribute_keys = {x.key: x.name for x in self._fetch_all_attributes()} - for props in itervalues(infrastructure_data): + for props in infrastructure_data.values(): mor_attributes = [] if self.config.collect_property_metrics: all_properties = {} diff --git a/vsphere/datadog_checks/vsphere/cache.py b/vsphere/datadog_checks/vsphere/cache.py index 1e2e6ec2741a7..687b8825b5248 100644 --- a/vsphere/datadog_checks/vsphere/cache.py +++ b/vsphere/datadog_checks/vsphere/cache.py @@ -6,7 +6,6 @@ from typing import Any, Dict, Generator, Iterator, List, Type # noqa: F401 from pyVmomi import vim # noqa: F401 -from six import iterkeys from datadog_checks.vsphere.types import CounterId, MetricName, ResourceTags # noqa: F401 @@ -140,7 +139,7 @@ def get_mor_props(self, mor, default=None): def get_mors(self, resource_type): # type: (Type[vim.ManagedEntity]) -> Iterator[vim.ManagedEntity] - return iterkeys(self._mors.get(resource_type, {})) + return iter(self._mors.get(resource_type, {}).keys()) def set_mor_props(self, mor, mor_data): # type: (vim.ManagedEntity, Dict[str, Any]) -> None diff --git a/vsphere/datadog_checks/vsphere/legacy/vsphere_legacy.py b/vsphere/datadog_checks/vsphere/legacy/vsphere_legacy.py index 8eca20328eeb9..158a87723ad01 100644 --- a/vsphere/datadog_checks/vsphere/legacy/vsphere_legacy.py +++ b/vsphere/datadog_checks/vsphere/legacy/vsphere_legacy.py @@ -17,8 +17,6 @@ vim, # pylint: disable=E0611 vmodl, # pylint: disable=E0611 ) -from six import itervalues -from six.moves import range from datadog_checks.base import AgentCheck, ensure_unicode, to_string from datadog_checks.base.checks.libs.thread_pool import SENTINEL, Pool @@ -951,7 +949,7 @@ def collect_metrics(self, instance): batch_size = self.batch_morlist_size or n_mors for batch in mors_batch_method(i_key, batch_size, max_historical_metrics): query_specs = [] - for mor in itervalues(batch): + for mor in batch.values(): if mor['mor_type'] == 'vm': vm_count += 1 if mor['mor_type'] not in REALTIME_RESOURCES and ('metrics' not in mor or not mor['metrics']): diff --git a/vsphere/tests/legacy/test_mor_cache.py b/vsphere/tests/legacy/test_mor_cache.py index fb1c12c61b40d..76da1c3ca8dfd 100644 --- a/vsphere/tests/legacy/test_mor_cache.py +++ b/vsphere/tests/legacy/test_mor_cache.py @@ -5,7 +5,6 @@ import pytest from mock import MagicMock -from six.moves import range from datadog_checks.vsphere.legacy.mor_cache import MorCache, MorNotFoundError diff --git a/windows_service/datadog_checks/windows_service/windows_service.py b/windows_service/datadog_checks/windows_service/windows_service.py index fc387f761022b..3e2ed672fb1e2 100644 --- a/windows_service/datadog_checks/windows_service/windows_service.py +++ b/windows_service/datadog_checks/windows_service/windows_service.py @@ -7,7 +7,6 @@ import pywintypes import win32service import winerror -from six import raise_from from datadog_checks.base import AgentCheck @@ -42,7 +41,7 @@ def _init_patterns(self): pattern = self.name self._name_re = re.compile(pattern, SERVICE_PATTERN_FLAGS) except re.error as e: - raise_from(Exception("Regular expression syntax error in '{}': {}".format(pattern, str(e))), None) + raise Exception("Regular expression syntax error in '{}': {}".format(pattern, str(e))) from None def match(self, service_view): if self.name is not None: From 8c600ef77d5bc0086790e2941d8b19dbffe1ab58 Mon Sep 17 00:00:00 2001 From: Ilia Kurenkov Date: Thu, 19 Sep 2024 11:14:04 +0200 Subject: [PATCH 14/34] Remove some more basic Python 2 compatibility code (#18619) * Remove some more basic Python 2 compatibility code * fix spark --- apache/datadog_checks/apache/apache.py | 3 +-- gitlab/datadog_checks/gitlab/common.py | 2 +- scylla/datadog_checks/scylla/scylla.py | 3 ++- spark/datadog_checks/spark/spark.py | 27 +++++++++---------- spark/tests/test_spark.py | 4 +-- .../datadog_checks/supervisord/supervisord.py | 4 +-- supervisord/tests/conftest.py | 4 +-- supervisord/tests/test_supervisord_unit.py | 6 ++--- teamcity/datadog_checks/teamcity/common.py | 8 +----- tls/datadog_checks/tls/tls_remote.py | 15 ++++------- tls/datadog_checks/tls/utils.py | 16 +---------- tls/tests/utils.py | 5 ++-- varnish/datadog_checks/varnish/varnish.py | 1 - 13 files changed, 34 insertions(+), 64 deletions(-) diff --git a/apache/datadog_checks/apache/apache.py b/apache/datadog_checks/apache/apache.py index 64f87f2d92aee..9d24990073544 100644 --- a/apache/datadog_checks/apache/apache.py +++ b/apache/datadog_checks/apache/apache.py @@ -2,8 +2,7 @@ # All rights reserved # Licensed under Simplified BSD License (see LICENSE) import re - -from six.moves.urllib.parse import urlparse +from urllib.parse import urlparse from datadog_checks.base import AgentCheck, ConfigurationError from datadog_checks.base.errors import CheckException diff --git a/gitlab/datadog_checks/gitlab/common.py b/gitlab/datadog_checks/gitlab/common.py index 75df9027121e0..7995ec40780d5 100644 --- a/gitlab/datadog_checks/gitlab/common.py +++ b/gitlab/datadog_checks/gitlab/common.py @@ -1,7 +1,7 @@ # (C) Datadog, Inc. 2023-present # All rights reserved # Licensed under a 3-clause BSD style license (see LICENSE) -from six.moves.urllib.parse import urlparse +from urllib.parse import urlparse def get_gitlab_version(http, log, gitlab_url, api_token): diff --git a/scylla/datadog_checks/scylla/scylla.py b/scylla/datadog_checks/scylla/scylla.py index 349908068e5b0..f2354729967ee 100644 --- a/scylla/datadog_checks/scylla/scylla.py +++ b/scylla/datadog_checks/scylla/scylla.py @@ -1,8 +1,9 @@ # (C) Datadog, Inc. 2020-present # All rights reserved # Licensed under a 3-clause BSD style license (see LICENSE) +from urllib.parse import urlparse + from six import PY2 -from six.moves.urllib.parse import urlparse from datadog_checks.base import ConfigurationError, OpenMetricsBaseCheck diff --git a/spark/datadog_checks/spark/spark.py b/spark/datadog_checks/spark/spark.py index 221aee90c1897..cce92c87b2e41 100644 --- a/spark/datadog_checks/spark/spark.py +++ b/spark/datadog_checks/spark/spark.py @@ -1,12 +1,11 @@ # (C) Datadog, Inc. 2018-present # All rights reserved # Licensed under a 3-clause BSD style license (see LICENSE) +from urllib.parse import urljoin, urlparse, urlsplit, urlunsplit from bs4 import BeautifulSoup from requests.exceptions import ConnectionError, HTTPError, InvalidURL, Timeout from simplejson import JSONDecodeError -from six import iteritems, itervalues -from six.moves.urllib.parse import urljoin, urlparse, urlsplit, urlunsplit from datadog_checks.base import AgentCheck, ConfigurationError, is_affirmative @@ -114,7 +113,7 @@ def check(self, _): # Report success after gathering all metrics from the ApplicationMaster if spark_apps: - _, (_, tracking_url) = next(iteritems(spark_apps)) + _, (_, tracking_url) = next(iter(spark_apps.items())) base_url = self._get_request_url(tracking_url) am_address = self._get_url_base(base_url) @@ -372,7 +371,7 @@ def _get_spark_app_ids(self, running_apps, tags): """ spark_apps = {} version_set = False - for app_id, (app_name, tracking_url) in iteritems(running_apps): + for app_id, (app_name, tracking_url) in running_apps.items(): try: if not version_set: version_set = self._collect_version(tracking_url, tags) @@ -394,7 +393,7 @@ def _spark_job_metrics(self, running_apps, addl_tags): """ Get metrics for each Spark job. """ - for app_id, (app_name, tracking_url) in iteritems(running_apps): + for app_id, (app_name, tracking_url) in running_apps.items(): base_url = self._get_request_url(tracking_url) response = self._rest_request_to_json( @@ -423,7 +422,7 @@ def _spark_stage_metrics(self, running_apps, addl_tags): """ Get metrics for each Spark stage. """ - for app_id, (app_name, tracking_url) in iteritems(running_apps): + for app_id, (app_name, tracking_url) in running_apps.items(): base_url = self._get_request_url(tracking_url) response = self._rest_request_to_json( @@ -449,7 +448,7 @@ def _spark_executor_metrics(self, running_apps, addl_tags): """ Get metrics for each Spark executor. """ - for app_id, (app_name, tracking_url) in iteritems(running_apps): + for app_id, (app_name, tracking_url) in running_apps.items(): base_url = self._get_request_url(tracking_url) response = self._rest_request_to_json( @@ -479,7 +478,7 @@ def _spark_rdd_metrics(self, running_apps, addl_tags): """ Get metrics for each Spark RDD. """ - for app_id, (app_name, tracking_url) in iteritems(running_apps): + for app_id, (app_name, tracking_url) in running_apps.items(): base_url = self._get_request_url(tracking_url) response = self._rest_request_to_json( @@ -499,7 +498,7 @@ def _spark_streaming_statistics_metrics(self, running_apps, addl_tags): """ Get metrics for each application streaming statistics. """ - for app_id, (app_name, tracking_url) in iteritems(running_apps): + for app_id, (app_name, tracking_url) in running_apps.items(): try: base_url = self._get_request_url(tracking_url) response = self._rest_request_to_json( @@ -525,7 +524,7 @@ def _spark_structured_streams_metrics(self, running_apps, addl_tags): - `SET spark.sql.streaming.metricsEnabled=true` in the app """ - for app_name, tracking_url in itervalues(running_apps): + for app_name, tracking_url in running_apps.values(): try: base_url = self._get_request_url(tracking_url) response = self._rest_request_to_json( @@ -534,10 +533,10 @@ def _spark_structured_streams_metrics(self, running_apps, addl_tags): self.log.debug('Structured streaming metrics: %s', response) response = { metric_name: v['value'] - for metric_name, v in iteritems(response.get('gauges')) + for metric_name, v in response.get('gauges').items() if 'streaming' in metric_name and 'value' in v } - for gauge_name, value in iteritems(response): + for gauge_name, value in response.items(): match = STRUCTURED_STREAMS_METRICS_REGEX.match(gauge_name) if not match: self.log.debug("No regex match found for gauge: '%s'", str(gauge_name)) @@ -576,7 +575,7 @@ def _set_metrics_from_json(self, tags, metrics_json, metrics): if metrics_json is None: return - for status, (metric_name, metric_type) in iteritems(metrics): + for status, (metric_name, metric_type) in metrics.items(): # Metrics defined with a dot `.` are exposed in a nested dictionary. # {"foo": {"bar": "baz", "qux": "quux"}} # foo.bar -> baz @@ -625,7 +624,7 @@ def _rest_request(self, url, object_path, service_name, tags, *args, **kwargs): # Add kwargs as arguments if kwargs: - query = '&'.join(['{0}={1}'.format(key, value) for key, value in iteritems(kwargs)]) + query = '&'.join(['{0}={1}'.format(key, value) for key, value in kwargs.items()]) url = urljoin(url, '?' + query) try: diff --git a/spark/tests/test_spark.py b/spark/tests/test_spark.py index d9dab0e246274..27ea4d3a65adf 100644 --- a/spark/tests/test_spark.py +++ b/spark/tests/test_spark.py @@ -6,14 +6,14 @@ import ssl import threading import time +from http import server as BaseHTTPServer +from urllib.parse import parse_qsl, unquote_plus, urlencode, urljoin, urlparse, urlunparse import mock import pytest import urllib3 from requests import RequestException from six import iteritems -from six.moves import BaseHTTPServer -from six.moves.urllib.parse import parse_qsl, unquote_plus, urlencode, urljoin, urlparse, urlunparse from datadog_checks.dev.http import MockResponse from datadog_checks.dev.utils import get_metadata_metrics diff --git a/supervisord/datadog_checks/supervisord/supervisord.py b/supervisord/datadog_checks/supervisord/supervisord.py index c2f6d410c2499..63d57ca8868fd 100644 --- a/supervisord/datadog_checks/supervisord/supervisord.py +++ b/supervisord/datadog_checks/supervisord/supervisord.py @@ -5,10 +5,10 @@ import re import socket import time +import xmlrpc.client as xmlrpclib from collections import defaultdict import supervisor.xmlrpc -from six.moves import xmlrpc_client as xmlrpclib from datadog_checks.base import AgentCheck @@ -206,7 +206,7 @@ def _connect(instance): host = instance.get('host', DEFAULT_HOST) port = instance.get('port', DEFAULT_PORT) auth = '{}:{}@'.format(user, password) if user and password else '' - server = xmlrpclib.Server('http://{}{}:{}/RPC2'.format(auth, host, port)) + server = xmlrpclib.ServerProxy('http://{}{}:{}/RPC2'.format(auth, host, port)) return server.supervisor @staticmethod diff --git a/supervisord/tests/conftest.py b/supervisord/tests/conftest.py index f8b47842528cd..5170ffbc14366 100644 --- a/supervisord/tests/conftest.py +++ b/supervisord/tests/conftest.py @@ -4,9 +4,9 @@ import os from copy import deepcopy +from xmlrpc.client import ServerProxy import pytest -from six.moves import xmlrpc_client as xmlrpclib from datadog_checks.dev import docker_run from datadog_checks.supervisord.supervisord import SupervisordCheck @@ -32,6 +32,6 @@ def bad_instance(): @pytest.fixture(scope='session') def dd_environment(): with docker_run(compose_file=os.path.join(HERE, 'compose', 'supervisord.yaml'), endpoints=URL, mount_logs=True): - server = xmlrpclib.Server('{}/RPC2'.format(URL)) + server = ServerProxy('{}/RPC2'.format(URL)) server.supervisor.startAllProcesses() yield SUPERVISORD_CONFIG diff --git a/supervisord/tests/test_supervisord_unit.py b/supervisord/tests/test_supervisord_unit.py index c302bb09c0610..6e4ac05da968b 100644 --- a/supervisord/tests/test_supervisord_unit.py +++ b/supervisord/tests/test_supervisord_unit.py @@ -2,12 +2,12 @@ # All rights reserved # Licensed under a 3-clause BSD style license (see LICENSE) +import xmlrpc.client as xmlrpclib from socket import socket import mock import pytest from mock import patch -from six.moves import xmlrpc_client as xmlrpclib from datadog_checks.supervisord.supervisord import FORMAT_TIME # pylint: disable=import-error,no-name-in-module @@ -24,9 +24,7 @@ def mock_server(url, transport=None): def test_check(aggregator, check): """Integration test for supervisord check. Using a mocked supervisord.""" - with patch.object(xmlrpclib, 'Server', side_effect=mock_server), patch.object( - xmlrpclib, 'ServerProxy', side_effect=mock_server - ): + with patch.object(xmlrpclib, 'ServerProxy', side_effect=mock_server): for tc in TEST_CASES: for instance in tc['instances']: name = instance['name'] diff --git a/teamcity/datadog_checks/teamcity/common.py b/teamcity/datadog_checks/teamcity/common.py index 63105100341fa..fad03ec4a7ff4 100644 --- a/teamcity/datadog_checks/teamcity/common.py +++ b/teamcity/datadog_checks/teamcity/common.py @@ -5,13 +5,7 @@ import time from collections import OrderedDict from copy import deepcopy - -from six import PY2 - -if PY2: - from urlparse import urlparse -else: - from urllib.parse import urlparse +from urllib.parse import urlparse import requests diff --git a/tls/datadog_checks/tls/tls_remote.py b/tls/datadog_checks/tls/tls_remote.py index ae3cccf3d0e34..516b3a98b4cf1 100644 --- a/tls/datadog_checks/tls/tls_remote.py +++ b/tls/datadog_checks/tls/tls_remote.py @@ -14,7 +14,6 @@ from datadog_checks.base.utils.time import get_timestamp from .const import SERVICE_CHECK_CAN_CONNECT, SERVICE_CHECK_EXPIRATION, SERVICE_CHECK_VALIDATION -from .utils import closing class TLSRemoteCheck(object): @@ -58,13 +57,11 @@ def _get_cert_and_protocol_version(self, sock): self.log.debug("Could not validate certificate because there is no connection") return cert, protocol_version # Get the cert & TLS version from the connection - with closing(sock): + with sock: self.log.debug('Getting cert and TLS protocol version') try: - with closing( - self.agent_check.get_tls_context().wrap_socket( - sock, server_hostname=self.agent_check._server_hostname - ) + with self.agent_check.get_tls_context().wrap_socket( + sock, server_hostname=self.agent_check._server_hostname ) as secure_sock: protocol_version = secure_sock.version() der_cert = secure_sock.getpeercert(binary_form=True) @@ -180,14 +177,12 @@ def fetch_intermediate_certs(self): self.log.error('Error occurred while connecting to socket to discover intermediate certificates: %s', e) return - with closing(sock): + with sock: try: context = ssl.SSLContext(protocol=ssl.PROTOCOL_TLS) context.verify_mode = ssl.CERT_NONE - with closing( - context.wrap_socket(sock, server_hostname=self.agent_check._server_hostname) - ) as secure_sock: + with context.wrap_socket(sock, server_hostname=self.agent_check._server_hostname) as secure_sock: der_cert = secure_sock.getpeercert(binary_form=True) protocol_version = secure_sock.version() if protocol_version and protocol_version not in self.agent_check.allowed_versions: diff --git a/tls/datadog_checks/tls/utils.py b/tls/datadog_checks/tls/utils.py index 0fd72a62620f0..262cb50108328 100644 --- a/tls/datadog_checks/tls/utils.py +++ b/tls/datadog_checks/tls/utils.py @@ -5,8 +5,6 @@ from ipaddress import ip_address -from six import PY2, text_type - # https://github.com/python/cpython/blob/ef516d11c1a0f885dba0aba8cf5366502077cdd4/Lib/ssl.py#L158-L165 DEFAULT_PROTOCOL_VERSIONS = {'TLSv1.2', 'TLSv1.3'} SUPPORTED_PROTOCOL_VERSIONS = {'SSLv3', 'TLSv1', 'TLSv1.1', 'TLSv1.2', 'TLSv1.3'} @@ -46,7 +44,7 @@ def get_protocol_versions(versions): def is_ip_address(hostname): try: - ip_address(text_type(hostname)) + ip_address(str(hostname)) except ValueError: return False @@ -59,15 +57,3 @@ def days_to_seconds(days): def seconds_to_days(seconds): return seconds / 60 / 60 / 24 - - -if PY2: - from contextlib import closing as _closing - - def closing(sock): - return _closing(sock) - -else: - - def closing(sock): - return sock diff --git a/tls/tests/utils.py b/tls/tests/utils.py index 6aa4702e28112..74504d73bf28a 100644 --- a/tls/tests/utils.py +++ b/tls/tests/utils.py @@ -9,7 +9,6 @@ from urllib.parse import urlparse from datadog_checks.dev import TempDir -from datadog_checks.tls.utils import closing @contextmanager @@ -31,8 +30,8 @@ def download_cert(filepath, host, raw=False): for _ in range(20): try: - with closing(socket.create_connection((host, 443))) as sock: - with closing(context.wrap_socket(sock, server_hostname=host)) as secure_sock: + with socket.create_connection((host, 443)) as sock: + with context.wrap_socket(sock, server_hostname=host) as secure_sock: cert = secure_sock.getpeercert(binary_form=True) except Exception: # no cov time.sleep(3) diff --git a/varnish/datadog_checks/varnish/varnish.py b/varnish/datadog_checks/varnish/varnish.py index f617cbf27650e..4761d52caf774 100644 --- a/varnish/datadog_checks/varnish/varnish.py +++ b/varnish/datadog_checks/varnish/varnish.py @@ -8,7 +8,6 @@ from os import geteuid from packaging.version import Version -from six.moves import filter from datadog_checks.base import ConfigurationError from datadog_checks.base.checks import AgentCheck From 6faa14c7c02d29b8583c2cff4024a17e9c02c7a8 Mon Sep 17 00:00:00 2001 From: Ilia Kurenkov Date: Thu, 19 Sep 2024 13:54:47 +0200 Subject: [PATCH 15/34] Remove Python2 logic when loading checks (#18620) * Remove Python2 logic when loading checks * remove py2 test for argocd --- .../active_directory/active_directory.py | 7 ++----- .../datadog_checks/aerospike/aerospike.py | 19 +++++-------------- .../datadog_checks/amazon_msk/amazon_msk.py | 11 +---------- argocd/datadog_checks/argocd/check.py | 8 -------- argocd/tests/test_argocd.py | 9 --------- .../datadog_checks/aspdotnet/aspdotnet.py | 7 ++----- cilium/datadog_checks/cilium/cilium.py | 12 +----------- .../datadog_checks/cockroachdb/cockroachdb.py | 14 ++------------ coredns/datadog_checks/coredns/coredns.py | 12 +----------- .../datadog_checks/dotnetclr/dotnetclr.py | 7 ++----- envoy/datadog_checks/envoy/envoy.py | 11 +---------- .../exchange_server/exchange_server.py | 7 ++----- gitlab/datadog_checks/gitlab/gitlab.py | 13 ++----------- haproxy/datadog_checks/haproxy/check.py | 13 ++----------- hyperv/datadog_checks/hyperv/hyperv.py | 7 ++----- kong/datadog_checks/kong/kong.py | 14 +++----------- linkerd/datadog_checks/linkerd/linkerd.py | 13 +------------ .../openmetrics/config_models/validators.py | 8 +------- .../datadog_checks/openmetrics/openmetrics.py | 4 +--- rabbitmq/datadog_checks/rabbitmq/check.py | 11 ++--------- scylla/datadog_checks/scylla/scylla.py | 12 +----------- teamcity/datadog_checks/teamcity/check.py | 13 ++----------- .../datadog_checks/teamcity/teamcity_rest.py | 8 -------- vault/datadog_checks/vault/vault.py | 15 ++------------- .../datadog_checks/win32_event_log/check.py | 11 +---------- 25 files changed, 39 insertions(+), 227 deletions(-) diff --git a/active_directory/datadog_checks/active_directory/active_directory.py b/active_directory/datadog_checks/active_directory/active_directory.py index 98aa34b3a1e87..5965b545525d3 100644 --- a/active_directory/datadog_checks/active_directory/active_directory.py +++ b/active_directory/datadog_checks/active_directory/active_directory.py @@ -1,18 +1,15 @@ # (C) Datadog, Inc. 2013-present # All rights reserved # Licensed under Simplified BSD License (see LICENSE) -from six import PY3 - from datadog_checks.base import PDHBaseCheck, is_affirmative +from .check import ActiveDirectoryCheckV2 from .metrics import DEFAULT_COUNTERS class ActiveDirectoryCheck(PDHBaseCheck): def __new__(cls, name, init_config, instances): - if PY3 and not is_affirmative(instances[0].get('use_legacy_check_version', False)): - from .check import ActiveDirectoryCheckV2 - + if not is_affirmative(instances[0].get('use_legacy_check_version', False)): return ActiveDirectoryCheckV2(name, init_config, instances) else: return super(ActiveDirectoryCheck, cls).__new__(cls) diff --git a/aerospike/datadog_checks/aerospike/aerospike.py b/aerospike/datadog_checks/aerospike/aerospike.py index 126b69303cdda..f52814865f090 100644 --- a/aerospike/datadog_checks/aerospike/aerospike.py +++ b/aerospike/datadog_checks/aerospike/aerospike.py @@ -10,11 +10,11 @@ from collections import defaultdict from typing import List # noqa: F401 -from six import PY2, iteritems - -from datadog_checks.base import AgentCheck, ConfigurationError +from datadog_checks.base import AgentCheck from datadog_checks.base.errors import CheckException +from .check import AerospikeCheckV2 + try: import aerospike except ImportError as e: @@ -78,15 +78,6 @@ def __new__(cls, name, init_config, instances): instance = instances[0] if 'openmetrics_endpoint' in instance: - if PY2: - raise ConfigurationError( - "This version of the integration is only available when using py3. " - "Check https://docs.datadoghq.com/agent/guide/agent-v6-python-3 " - "for more information or use the older style config." - ) - # TODO: when we drop Python 2 move this import up top - from .check import AerospikeCheckV2 - return AerospikeCheckV2(name, init_config, instances) else: @@ -232,7 +223,7 @@ def collect_info(self, command, metric_type, separator=';', required_keys=None, self.log.warning('Exceeded cap `%s` for metric type `%s` - please contact support', cap, metric_type) return - for key, value in iteritems(required_data): + for key, value in required_data.items(): self.send(metric_type, key, value, tags) def get_namespaces(self): @@ -491,7 +482,7 @@ def collect_latency(self, namespaces): ns_latencies[ns].setdefault("metric_names", []).extend(metric_names) - for ns, v in iteritems(ns_latencies): + for ns, v in ns_latencies.items(): metric_names = v.get("metric_names", []) metric_values = v.get("metric_values", []) namespace_tags = ['namespace:{}'.format(ns)] if ns else [] diff --git a/amazon_msk/datadog_checks/amazon_msk/amazon_msk.py b/amazon_msk/datadog_checks/amazon_msk/amazon_msk.py index a00b8c4647443..ef580ae06f756 100644 --- a/amazon_msk/datadog_checks/amazon_msk/amazon_msk.py +++ b/amazon_msk/datadog_checks/amazon_msk/amazon_msk.py @@ -4,10 +4,10 @@ import json import boto3 -from six import PY2 from datadog_checks.base import ConfigurationError, OpenMetricsBaseCheck, is_affirmative +from .check import AmazonMskCheckV2 from .metrics import JMX_METRICS_MAP, JMX_METRICS_OVERRIDES, NODE_METRICS_MAP, NODE_METRICS_OVERRIDES from .utils import construct_boto_config @@ -29,15 +29,6 @@ def __new__(cls, name, init_config, instances): instance = instances[0] if is_affirmative(instance.get('use_openmetrics', False)): - if PY2: - raise ConfigurationError( - "Openmetrics on this integration is only available when using py3. " - "Check https://docs.datadoghq.com/agent/guide/agent-v6-python-3 " - "for more information" - ) - # TODO: when we drop Python 2 move this import up top - from .check import AmazonMskCheckV2 - return AmazonMskCheckV2(name, init_config, instances) else: return super(AmazonMskCheck, cls).__new__(cls) diff --git a/argocd/datadog_checks/argocd/check.py b/argocd/datadog_checks/argocd/check.py index 08e6b766d5a8b..fa57212e829cc 100644 --- a/argocd/datadog_checks/argocd/check.py +++ b/argocd/datadog_checks/argocd/check.py @@ -3,8 +3,6 @@ # Licensed under a 3-clause BSD style license (see LICENSE) from collections import defaultdict -from six import PY2 - from datadog_checks.base import ConfigurationError, OpenMetricsBaseCheckV2 from datadog_checks.base.constants import ServiceCheck @@ -36,12 +34,6 @@ class ArgocdCheck(OpenMetricsBaseCheckV2, ConfigMixin): DEFAULT_METRIC_LIMIT = 0 def __init__(self, name, init_config, instances): - if PY2: - raise ConfigurationError( - "This version of the integration is only available when using py3. " - "Check https://docs.datadoghq.com/agent/guide/agent-v6-python-3 " - "for more information." - ) super(ArgocdCheck, self).__init__(name, init_config, instances) self.check_initializations.appendleft(self.parse_config) self.check_initializations.append(self.configure_additional_transformers) diff --git a/argocd/tests/test_argocd.py b/argocd/tests/test_argocd.py index e0f71d51c1012..e53d0cb8a9549 100644 --- a/argocd/tests/test_argocd.py +++ b/argocd/tests/test_argocd.py @@ -3,11 +3,9 @@ # Licensed under a 3-clause BSD style license (see LICENSE) import pytest -from mock import patch from datadog_checks.argocd import ArgocdCheck from datadog_checks.base.constants import ServiceCheck -from datadog_checks.base.errors import ConfigurationError from datadog_checks.dev.utils import get_metadata_metrics from .common import ( @@ -96,10 +94,3 @@ def test_app_controller_service_check(dd_run_check, aggregator, mock_http_respon ServiceCheck.UNKNOWN, tags=['endpoint:http://app_controller:8082', 'name:faz'], ) - - -@patch('datadog_checks.argocd.check.PY2', True) -def test_py2(): - # Test to ensure that a ConfigurationError is raised when running with Python 2. - with pytest.raises(ConfigurationError, match="This version of the integration is only available when using py3."): - ArgocdCheck('argocd', {}, [MOCKED_APP_CONTROLLER_INSTANCE]) diff --git a/aspdotnet/datadog_checks/aspdotnet/aspdotnet.py b/aspdotnet/datadog_checks/aspdotnet/aspdotnet.py index 582aa7091e7bf..e21c137b5fff9 100644 --- a/aspdotnet/datadog_checks/aspdotnet/aspdotnet.py +++ b/aspdotnet/datadog_checks/aspdotnet/aspdotnet.py @@ -1,10 +1,9 @@ # (C) Datadog, Inc. 2013-present # All rights reserved # Licensed under Simplified BSD License (see LICENSE) -from six import PY3 - from datadog_checks.base import PDHBaseCheck, is_affirmative +from .check import AspdotnetCheckV2 from .metrics import DEFAULT_COUNTERS EVENT_TYPE = SOURCE_TYPE_NAME = 'aspdotnet' @@ -13,9 +12,7 @@ class AspdotnetCheck(PDHBaseCheck): def __new__(cls, name, init_config, instances): - if PY3 and not is_affirmative(instances[0].get('use_legacy_check_version', False)): - from .check import AspdotnetCheckV2 - + if not is_affirmative(instances[0].get('use_legacy_check_version', False)): return AspdotnetCheckV2(name, init_config, instances) else: return super(AspdotnetCheck, cls).__new__(cls) diff --git a/cilium/datadog_checks/cilium/cilium.py b/cilium/datadog_checks/cilium/cilium.py index 2a614a7a9aad4..3d0eab6e09339 100644 --- a/cilium/datadog_checks/cilium/cilium.py +++ b/cilium/datadog_checks/cilium/cilium.py @@ -1,10 +1,9 @@ # (C) Datadog, Inc. 2019-present # All rights reserved # Licensed under a 3-clause BSD style license (see LICENSE) -from six import PY2 - from datadog_checks.base import ConfigurationError, OpenMetricsBaseCheck, is_affirmative +from .check import CiliumCheckV2 from .metrics import AGENT_METRICS, OPERATOR_METRICS @@ -19,15 +18,6 @@ def __new__(cls, name, init_config, instances): instance = instances[0] if is_affirmative(instance.get('use_openmetrics', False)): - if PY2: - raise ConfigurationError( - "This version of the integration is only available when using py3. " - "Check https://docs.datadoghq.com/agent/guide/agent-v6-python-3 " - "for more information or use the older style config." - ) - # TODO: when we drop Python 2 move this import up top - from .check import CiliumCheckV2 - return CiliumCheckV2(name, init_config, instances) else: return super(CiliumCheck, cls).__new__(cls) diff --git a/cockroachdb/datadog_checks/cockroachdb/cockroachdb.py b/cockroachdb/datadog_checks/cockroachdb/cockroachdb.py index f45fb633b579b..936b6f567db1b 100644 --- a/cockroachdb/datadog_checks/cockroachdb/cockroachdb.py +++ b/cockroachdb/datadog_checks/cockroachdb/cockroachdb.py @@ -1,10 +1,9 @@ # (C) Datadog, Inc. 2018-present # All rights reserved # Licensed under a 3-clause BSD style license (see LICENSE) -from six import PY2 - -from datadog_checks.base import ConfigurationError, OpenMetricsBaseCheck +from datadog_checks.base import OpenMetricsBaseCheck +from .check import CockroachdbCheckV2 from .metrics import METRIC_MAP @@ -16,15 +15,6 @@ def __new__(cls, name, init_config, instances): instance = instances[0] if 'openmetrics_endpoint' in instance: - if PY2: - raise ConfigurationError( - 'This version of the integration is only available when using Python 3. ' - 'Check https://docs.datadoghq.com/agent/guide/agent-v6-python-3/ ' - 'for more information or use the older style config.' - ) - # TODO: when we drop Python 2 move this import up top - from .check import CockroachdbCheckV2 - return CockroachdbCheckV2(name, init_config, instances) else: return super(CockroachdbCheck, cls).__new__(cls) diff --git a/coredns/datadog_checks/coredns/coredns.py b/coredns/datadog_checks/coredns/coredns.py index 771bd74d03b7b..02b9a5ed8ef04 100644 --- a/coredns/datadog_checks/coredns/coredns.py +++ b/coredns/datadog_checks/coredns/coredns.py @@ -1,10 +1,9 @@ # (C) Datadog, Inc. 2018-present # All rights reserved # Licensed under a 3-clause BSD style license (see LICENSE) -from six import PY2 - from datadog_checks.base import ConfigurationError, OpenMetricsBaseCheck +from .check import CoreDNS from .metrics import DEFAULT_METRICS, GO_METRICS @@ -25,15 +24,6 @@ def __new__(cls, name, init_config, instances): instance = instances[0] if 'openmetrics_endpoint' in instance: - if PY2: - raise ConfigurationError( - "This version of the integration is only available when using py3. " - "Check https://docs.datadoghq.com/agent/guide/agent-v6-python-3 " - "for more information or use the older style config." - ) - # TODO: when we drop Python 2 move this import up top - from .check import CoreDNS - return CoreDNS(name, init_config, instances) else: return super(CoreDNSCheck, cls).__new__(cls) diff --git a/dotnetclr/datadog_checks/dotnetclr/dotnetclr.py b/dotnetclr/datadog_checks/dotnetclr/dotnetclr.py index 7aa466b3fb6ef..502768e8651ba 100644 --- a/dotnetclr/datadog_checks/dotnetclr/dotnetclr.py +++ b/dotnetclr/datadog_checks/dotnetclr/dotnetclr.py @@ -1,10 +1,9 @@ # (C) Datadog, Inc. 2013-present # All rights reserved # Licensed under Simplified BSD License (see LICENSE) -from six import PY3 - from datadog_checks.base import PDHBaseCheck, is_affirmative +from .check import DotnetclrCheckV2 from .metrics import DEFAULT_COUNTERS EVENT_TYPE = SOURCE_TYPE_NAME = 'dotnetclr' @@ -12,9 +11,7 @@ class DotnetclrCheck(PDHBaseCheck): def __new__(cls, name, init_config, instances): - if PY3 and not is_affirmative(instances[0].get('use_legacy_check_version', False)): - from .check import DotnetclrCheckV2 - + if not is_affirmative(instances[0].get('use_legacy_check_version', False)): return DotnetclrCheckV2(name, init_config, instances) else: return super(DotnetclrCheck, cls).__new__(cls) diff --git a/envoy/datadog_checks/envoy/envoy.py b/envoy/datadog_checks/envoy/envoy.py index b656fa34cfa2a..4790da50656fe 100644 --- a/envoy/datadog_checks/envoy/envoy.py +++ b/envoy/datadog_checks/envoy/envoy.py @@ -6,10 +6,10 @@ from urllib.parse import urljoin import requests -from six import PY2 from datadog_checks.base import AgentCheck, ConfigurationError, is_affirmative +from .check import EnvoyCheckV2 from .errors import UnknownMetric, UnknownTags from .parser import parse_histogram, parse_metric from .utils import _get_server_info @@ -27,15 +27,6 @@ def __new__(cls, name, init_config, instances): instance = instances[0] if 'openmetrics_endpoint' in instance: - if PY2: - raise ConfigurationError( - "This version of the integration is only available when using py3. " - "Check https://docs.datadoghq.com/agent/guide/agent-v6-python-3 " - "for more information or use the older style config." - ) - # TODO: when we drop Python 2 move this import up top - from .check import EnvoyCheckV2 - return EnvoyCheckV2(name, init_config, instances) else: return super(Envoy, cls).__new__(cls) diff --git a/exchange_server/datadog_checks/exchange_server/exchange_server.py b/exchange_server/datadog_checks/exchange_server/exchange_server.py index fd70a43678a55..2e2231b8b2676 100644 --- a/exchange_server/datadog_checks/exchange_server/exchange_server.py +++ b/exchange_server/datadog_checks/exchange_server/exchange_server.py @@ -1,18 +1,15 @@ # (C) Datadog, Inc. 2013-present # All rights reserved # Licensed under Simplified BSD License (see LICENSE) -from six import PY3 - from datadog_checks.base import PDHBaseCheck, is_affirmative +from .check import ExchangeCheckV2 from .metrics import DEFAULT_COUNTERS class ExchangeCheck(PDHBaseCheck): def __new__(cls, name, init_config, instances): - if PY3 and not is_affirmative(instances[0].get('use_legacy_check_version', False)): - from .check import ExchangeCheckV2 - + if not is_affirmative(instances[0].get('use_legacy_check_version', False)): return ExchangeCheckV2(name, init_config, instances) else: return super(ExchangeCheck, cls).__new__(cls) diff --git a/gitlab/datadog_checks/gitlab/gitlab.py b/gitlab/datadog_checks/gitlab/gitlab.py index 1c527ec52f67f..0346b2f9f5532 100644 --- a/gitlab/datadog_checks/gitlab/gitlab.py +++ b/gitlab/datadog_checks/gitlab/gitlab.py @@ -4,13 +4,13 @@ from copy import deepcopy import requests -from six import PY2 from datadog_checks.base import AgentCheck from datadog_checks.base.checks.openmetrics import OpenMetricsBaseCheck -from datadog_checks.base.errors import CheckException, ConfigurationError +from datadog_checks.base.errors import CheckException from .common import get_gitlab_version, get_tags +from .gitlab_v2 import GitlabCheckV2 from .metrics import METRICS_MAP @@ -41,15 +41,6 @@ def __new__(cls, name, init_config, instances): instance = instances[0] if instance.get('openmetrics_endpoint'): - if PY2: - raise ConfigurationError( - 'This version of the integration is only available when using Python 3. ' - 'Check https://docs.datadoghq.com/agent/guide/agent-v6-python-3/ ' - 'for more information or use the older style config.' - ) - # TODO: when we drop Python 2 move this import up top - from .gitlab_v2 import GitlabCheckV2 - return GitlabCheckV2(name, init_config, instances) return super(GitlabCheck, cls).__new__(cls) diff --git a/haproxy/datadog_checks/haproxy/check.py b/haproxy/datadog_checks/haproxy/check.py index d1a6f34e8c072..c0bc67199ec58 100644 --- a/haproxy/datadog_checks/haproxy/check.py +++ b/haproxy/datadog_checks/haproxy/check.py @@ -1,10 +1,9 @@ # (C) Datadog, Inc. 2020-present # All rights reserved # Licensed under a 3-clause BSD style license (see LICENSE) -from six import PY2 - -from datadog_checks.base import ConfigurationError, OpenMetricsBaseCheck, is_affirmative +from datadog_checks.base import OpenMetricsBaseCheck, is_affirmative +from .checkv2 import HaproxyCheckV2 from .legacy.haproxy import HAProxyCheckLegacy from .metrics import METRIC_MAP @@ -16,14 +15,6 @@ def __new__(cls, name, init_config, instances): instance = instances[0] if is_affirmative(instance.get('use_openmetrics', False)): - if PY2: - raise ConfigurationError( - "Openmetrics on this integration is only available when using py3. " - "Check https://docs.datadoghq.com/agent/guide/agent-v6-python-3 " - "for more information" - ) - from .checkv2 import HaproxyCheckV2 - return HaproxyCheckV2(name, init_config, instances) elif is_affirmative(instance.get('use_prometheus', False)): return super(HAProxyCheck, cls).__new__(cls) diff --git a/hyperv/datadog_checks/hyperv/hyperv.py b/hyperv/datadog_checks/hyperv/hyperv.py index bb8a2067c636f..2f61bdfd00054 100644 --- a/hyperv/datadog_checks/hyperv/hyperv.py +++ b/hyperv/datadog_checks/hyperv/hyperv.py @@ -1,18 +1,15 @@ # (C) Datadog, Inc. 2018-present # All rights reserved # Licensed under a 3-clause BSD style license (see LICENSE) -from six import PY3 - from datadog_checks.base import PDHBaseCheck, is_affirmative +from .check import HypervCheckV2 from .metrics import DEFAULT_COUNTERS class HypervCheck(PDHBaseCheck): def __new__(cls, name, init_config, instances): - if PY3 and not is_affirmative(instances[0].get('use_legacy_check_version', False)): - from .check import HypervCheckV2 - + if not is_affirmative(instances[0].get('use_legacy_check_version', False)): return HypervCheckV2(name, init_config, instances) else: return super(HypervCheck, cls).__new__(cls) diff --git a/kong/datadog_checks/kong/kong.py b/kong/datadog_checks/kong/kong.py index ff5eb45768710..8e3ab12f52609 100644 --- a/kong/datadog_checks/kong/kong.py +++ b/kong/datadog_checks/kong/kong.py @@ -4,9 +4,10 @@ from urllib.parse import urlparse import simplejson as json -from six import PY2 -from datadog_checks.base import AgentCheck, ConfigurationError +from datadog_checks.base import AgentCheck + +from .check import KongCheck class Kong(AgentCheck): @@ -24,15 +25,6 @@ def __new__(cls, name, init_config, instances): instance = instances[0] if 'openmetrics_endpoint' in instance: - if PY2: - raise ConfigurationError( - "This version of the integration is only available when using py3. " - "Check https://docs.datadoghq.com/agent/guide/agent-v6-python-3 " - "for more information or use the older style config." - ) - # TODO: when we drop Python 2 move this import up top - from .check import KongCheck - return KongCheck(name, init_config, instances) else: return super(Kong, cls).__new__(cls) diff --git a/linkerd/datadog_checks/linkerd/linkerd.py b/linkerd/datadog_checks/linkerd/linkerd.py index d6ee881b9886a..68709690d66ea 100644 --- a/linkerd/datadog_checks/linkerd/linkerd.py +++ b/linkerd/datadog_checks/linkerd/linkerd.py @@ -1,11 +1,9 @@ # (C) Datadog, Inc. 2018-present # All rights reserved # Licensed under a 3-clause BSD style license (see LICENSE) -from six import PY2 - -from datadog_checks.base import ConfigurationError from datadog_checks.base.checks.openmetrics import OpenMetricsBaseCheck +from .check import LinkerdCheckV2 from .metrics import METRIC_MAP, TYPE_OVERRIDES @@ -22,15 +20,6 @@ def __new__(cls, name, init_config, instances): instance = instances[0] if 'openmetrics_endpoint' in instance: - if PY2: - raise ConfigurationError( - "This version of the integration is only available when using py3. " - "Check https://docs.datadoghq.com/agent/guide/agent-v6-python-3 " - "for more information or use the older style config." - ) - # TODO: when we drop Python 2 move this import up top - from .check import LinkerdCheckV2 - return LinkerdCheckV2(name, init_config, instances) else: return super(LinkerdCheck, cls).__new__(cls) diff --git a/openmetrics/datadog_checks/openmetrics/config_models/validators.py b/openmetrics/datadog_checks/openmetrics/config_models/validators.py index 2172fc6190425..05bb72362bc80 100644 --- a/openmetrics/datadog_checks/openmetrics/config_models/validators.py +++ b/openmetrics/datadog_checks/openmetrics/config_models/validators.py @@ -1,19 +1,13 @@ # (C) Datadog, Inc. 2021-present # All rights reserved # Licensed under a 3-clause BSD style license (see LICENSE) -from six import PY2 def initialize_instance(values, **kwargs): v2_endpoint = values.get('openmetrics_endpoint') v1_endpoint = values.get('prometheus_url') - if PY2 and v2_endpoint: - raise ValueError('`openmetrics_endpoint` cannot be used if the agent is running Python 2') if v1_endpoint and v2_endpoint: raise ValueError('`prometheus_url` cannot be used along `openmetrics_endpoint`') if not v1_endpoint and not v2_endpoint: - if PY2: - raise ValueError('`prometheus_url` is required') - else: - raise ValueError('`openmetrics_endpoint` is required') + raise ValueError('`openmetrics_endpoint` is required') return values diff --git a/openmetrics/datadog_checks/openmetrics/openmetrics.py b/openmetrics/datadog_checks/openmetrics/openmetrics.py index cb5fd21c46c3b..6e4575ac21fc6 100644 --- a/openmetrics/datadog_checks/openmetrics/openmetrics.py +++ b/openmetrics/datadog_checks/openmetrics/openmetrics.py @@ -1,8 +1,6 @@ # (C) Datadog, Inc. 2018-present # All rights reserved # Licensed under a 3-clause BSD style license (see LICENSE) -from six import PY2 - from datadog_checks.base import OpenMetricsBaseCheck, OpenMetricsBaseCheckV2 @@ -10,7 +8,7 @@ class OpenMetricsCheck(OpenMetricsBaseCheck): def __new__(cls, name, init_config, instances): instance = instances[0] - if not PY2 and 'openmetrics_endpoint' in instance: + if 'openmetrics_endpoint' in instance: return OpenMetricsBaseCheckV2(name, init_config, instances) else: return super(OpenMetricsBaseCheck, cls).__new__(cls) diff --git a/rabbitmq/datadog_checks/rabbitmq/check.py b/rabbitmq/datadog_checks/rabbitmq/check.py index 3cd41f7d566b5..c122fdc0deb76 100644 --- a/rabbitmq/datadog_checks/rabbitmq/check.py +++ b/rabbitmq/datadog_checks/rabbitmq/check.py @@ -1,9 +1,9 @@ # (C) Datadog, Inc. 2023-present # All rights reserved # Licensed under a 3-clause BSD style license (see LICENSE) -from six import PY2 +from datadog_checks.base import AgentCheck -from datadog_checks.base import AgentCheck, ConfigurationError +from .openmetrics import RabbitMQOpenMetrics class RabbitMQ(AgentCheck): @@ -13,13 +13,6 @@ def __new__(cls, name, init_config, instances): instance = instances[0] if 'prometheus_plugin' in instance: - if PY2: - raise ConfigurationError( - "This version of the integration is only available when using py3. " - "Check https://docs.datadoghq.com/agent/guide/agent-v6-python-3 " - "for more information or use the `rabbitmq_api_url` config." - ) - from .openmetrics import RabbitMQOpenMetrics return RabbitMQOpenMetrics(name, init_config, instances) else: diff --git a/scylla/datadog_checks/scylla/scylla.py b/scylla/datadog_checks/scylla/scylla.py index f2354729967ee..c7ade2d966725 100644 --- a/scylla/datadog_checks/scylla/scylla.py +++ b/scylla/datadog_checks/scylla/scylla.py @@ -3,11 +3,10 @@ # Licensed under a 3-clause BSD style license (see LICENSE) from urllib.parse import urlparse -from six import PY2 - from datadog_checks.base import ConfigurationError, OpenMetricsBaseCheck from .metrics import ADDITIONAL_METRICS_MAP, INSTANCE_DEFAULT_METRICS +from .scylla_v2 import ScyllaCheckV2 class ScyllaCheck(OpenMetricsBaseCheck): @@ -27,15 +26,6 @@ def __new__(cls, name, init_config, instances): instance = instances[0] if instance.get('openmetrics_endpoint'): - if PY2: - raise ConfigurationError( - 'This version of the integration is only available when using Python 3. ' - 'Check https://docs.datadoghq.com/agent/guide/agent-v6-python-3/ ' - 'for more information or use the older style config.' - ) - # TODO: when we drop Python 2 move this import up top - from .scylla_v2 import ScyllaCheckV2 - return ScyllaCheckV2(name, init_config, instances) else: return super(ScyllaCheck, cls).__new__(cls) diff --git a/teamcity/datadog_checks/teamcity/check.py b/teamcity/datadog_checks/teamcity/check.py index 118cfba5a2302..8efbcd16ea001 100644 --- a/teamcity/datadog_checks/teamcity/check.py +++ b/teamcity/datadog_checks/teamcity/check.py @@ -1,9 +1,9 @@ # (C) Datadog, Inc. 2014-present # All rights reserved # Licensed under a 3-clause BSD style license (see LICENSE) -from six import PY2 +from datadog_checks.base import AgentCheck, is_affirmative -from datadog_checks.base import AgentCheck, ConfigurationError, is_affirmative +from .teamcity_openmetrics import TeamCityOpenMetrics class TeamCityCheck(AgentCheck): @@ -13,15 +13,6 @@ def __new__(cls, name, init_config, instances): instance = instances[0] if is_affirmative(instance.get('use_openmetrics', False)): - if PY2: - raise ConfigurationError( - "This version of the integration is only available when using py3. " - "Check https://docs.datadoghq.com/agent/guide/agent-v6-python-3 " - "for more information or use the older style config." - ) - # TODO: when we drop Python 2 move this import up top - from .teamcity_openmetrics import TeamCityOpenMetrics - return TeamCityOpenMetrics(name, init_config, instances) else: from .teamcity_rest import TeamCityRest diff --git a/teamcity/datadog_checks/teamcity/teamcity_rest.py b/teamcity/datadog_checks/teamcity/teamcity_rest.py index 83ddff680332f..e078cc6af2683 100644 --- a/teamcity/datadog_checks/teamcity/teamcity_rest.py +++ b/teamcity/datadog_checks/teamcity/teamcity_rest.py @@ -4,7 +4,6 @@ from copy import deepcopy from requests.exceptions import HTTPError -from six import PY2 from datadog_checks.base import AgentCheck, ConfigurationError, is_affirmative from datadog_checks.base.utils.time import get_precise_time @@ -82,9 +81,6 @@ def __init__(self, name, init_config, instances): self.bc_store = BuildConfigs() - if PY2: - self.check_initializations.append(self._validate_config) - def _validate_config(self): if self.instance.get('projects'): raise ConfigurationError( @@ -172,10 +168,6 @@ def _initialize(self): self._initialize_single_build_config() else: - if PY2: - raise self.CheckException( - 'Multi-build configuration monitoring is not currently supported in Python 2.' - ) self.log.debug("Initializing multi-build configuration monitoring.") self._initialize_multi_build_config() diff --git a/vault/datadog_checks/vault/vault.py b/vault/datadog_checks/vault/vault.py index bf81969ed4a53..56e0ffbf5a224 100644 --- a/vault/datadog_checks/vault/vault.py +++ b/vault/datadog_checks/vault/vault.py @@ -4,10 +4,10 @@ import time import requests -from six import PY2 -from datadog_checks.base import ConfigurationError, OpenMetricsBaseCheck, is_affirmative +from datadog_checks.base import OpenMetricsBaseCheck, is_affirmative +from .check import VaultCheckV2 from .common import API_METHODS, DEFAULT_API_VERSION, SYS_HEALTH_DEFAULT_CODES, SYS_LEADER_DEFAULT_CODES, Api, Leader from .errors import ApiUnreachable from .metrics import METRIC_MAP, METRIC_ROLLBACK_COMPAT_MAP, ROUTE_METRICS_TO_TRANSFORM @@ -38,14 +38,6 @@ def __new__(cls, name, init_config, instances): instance = instances[0] if is_affirmative(instance.get('use_openmetrics', False)): - if PY2: - raise ConfigurationError( - 'This version of the integration is only available when using Python 3. ' - 'Check https://docs.datadoghq.com/agent/guide/agent-v6-python-3/ ' - 'for more information or use the older style config.' - ) - # TODO: when we drop Python 2 move this import up top - from .check import VaultCheckV2 return VaultCheckV2(name, init_config, instances) else: @@ -278,9 +270,6 @@ def access_api(self, url, ignore_status_codes=None): return json_data def parse_config(self): - if PY2 and not self._api_url: - raise ConfigurationError('Vault setting `api_url` is required') - api_version = self._api_url[-1] if api_version not in ('1',): self.log.warning('Unknown Vault API version `%s`, using version `%s`', api_version, DEFAULT_API_VERSION) diff --git a/win32_event_log/datadog_checks/win32_event_log/check.py b/win32_event_log/datadog_checks/win32_event_log/check.py index f7c1d82fc3165..b75cf2c9f06ae 100644 --- a/win32_event_log/datadog_checks/win32_event_log/check.py +++ b/win32_event_log/datadog_checks/win32_event_log/check.py @@ -8,21 +8,16 @@ import win32event import win32evtlog import win32security -from six import PY2 from datadog_checks.base import AgentCheck, ConfigurationError, is_affirmative from datadog_checks.base.errors import SkipInstanceError from datadog_checks.base.utils.common import exclude_undefined_keys from datadog_checks.base.utils.time import get_timestamp +from .config_models import ConfigMixin from .filters import construct_xpath_query from .legacy import Win32EventLogWMI -if PY2: - ConfigMixin = object -else: - from .config_models import ConfigMixin - class Win32EventLogCheck(AgentCheck, ConfigMixin): # The lower cased version of the `API SOURCE ATTRIBUTE` column from the table located here: @@ -91,12 +86,8 @@ def __new__(cls, name, init_config, instances): ) if use_legacy_mode: - # Supports PY2 and PY3 return Win32EventLogWMI(name, init_config, instances) elif use_legacy_mode_v2: - # Supports PY3 - if PY2: - raise ConfigurationError("legacy_mode_v2 is not supported on Python2") return super(Win32EventLogCheck, cls).__new__(cls) else: raise SkipInstanceError( From ddf05101441e3c105be374abc22e2f94da86a8f5 Mon Sep 17 00:00:00 2001 From: Erica Ho <109224841+eho1307@users.noreply.github.com> Date: Thu, 19 Sep 2024 10:11:21 -0400 Subject: [PATCH 16/34] update kafka readme (#18607) * update kafka readme Remove the note about DSM now that we display the product connection as a UI component in app * update kafka_consumer readme Remove the note about DSM now that we display the product connection as a UI component in app * remove trailing link * remove trailing link --- kafka/README.md | 3 --- kafka_consumer/README.md | 3 --- 2 files changed, 6 deletions(-) diff --git a/kafka/README.md b/kafka/README.md index 0cb3d2b09c054..bf20e29bc8ba8 100644 --- a/kafka/README.md +++ b/kafka/README.md @@ -6,8 +6,6 @@ View Kafka broker metrics collected for a 360-view of the health and performance of your Kafka clusters in real time. With this integration, you can collect metrics and logs from your Kafka deployment to visualize telemetry and alert on the performance of your Kafka stack. -If you would benefit from visualizing the topology of your streaming data pipelines and identifying the root cause of bottlenecks, learn more about [Data Streams Monitoring][24]. - **Note**: - This check has a limit of 350 metrics per instance. The number of returned metrics is indicated in the Agent status output. Specify the metrics you are interested in by editing the configuration below. For more detailed instructions on customizing the metrics to collect, see the [JMX Checks documentation][2]. - This integration attached sample configuration works only for Kafka >= 0.8.2. @@ -175,5 +173,4 @@ See [service_checks.json][15] for a list of service checks provided by this inte [21]: https://www.datadoghq.com/blog/monitor-kafka-with-datadog [22]: https://raw.githubusercontent.com/DataDog/dd-agent/5.2.1/conf.d/kafka.yaml.example [23]: https://www.datadoghq.com/knowledge-center/apache-kafka/ -[24]: https://www.datadoghq.com/product/data-streams-monitoring/ [25]: https://app.datadoghq.com/data-streams diff --git a/kafka_consumer/README.md b/kafka_consumer/README.md index 36fa9a9da85c1..296ace6effe75 100644 --- a/kafka_consumer/README.md +++ b/kafka_consumer/README.md @@ -6,8 +6,6 @@ This Agent integration collects message offset metrics from your Kafka consumers. This check fetches the highwater offsets from the Kafka brokers, consumer offsets that are stored in Kafka (or Zookeeper for old-style consumers), and then calculates consumer lag (which is the difference between the broker offset and the consumer offset). -If you would benefit from visualizing the topology of your streaming data pipelines and identifying the root cause of bottlenecks, learn more about [Data Streams Monitoring][16]. - **Note:** - This integration ensures that consumer offsets are checked before broker offsets; in the worst case, consumer lag may be a little overstated. Checking these offsets in the reverse order can understate consumer lag to the point of having negative values, which is a dire scenario usually indicating messages are being skipped. - If you want to collect JMX metrics from your Kafka brokers or Java-based consumers/producers, see the [Kafka Broker integration][19]. @@ -143,7 +141,6 @@ sudo service datadog-agent restart [13]: https://www.datadoghq.com/blog/monitoring-kafka-performance-metrics [14]: https://www.datadoghq.com/blog/collecting-kafka-performance-metrics [15]: https://www.datadoghq.com/blog/monitor-kafka-with-datadog -[16]: https://www.datadoghq.com/product/data-streams-monitoring/ [17]: https://docs.datadoghq.com/containers/kubernetes/integrations/ [18]: https://app.datadoghq.com/data-streams [19]: https://app.datadoghq.com/integrations/kafka?search=kafka From 8da9ca86eb300aaba96e5fc6213fe0890c9eac04 Mon Sep 17 00:00:00 2001 From: Ilia Kurenkov Date: Thu, 19 Sep 2024 16:33:56 +0200 Subject: [PATCH 17/34] Remove last batch of easy 'six' dependencies (#18624) --- .../datadog_checks/citrix_hypervisor/check.py | 6 +++--- citrix_hypervisor/tests/test_citrix_hypervisor.py | 4 ++-- citrix_hypervisor/tests/test_metadata.py | 2 +- .../datadog_test_libs/win/pdh_mocks.py | 13 +++---------- gearmand/datadog_checks/gearmand/gearmand.py | 10 +--------- network/datadog_checks/network/ethtool.py | 4 +--- .../datadog_checks/postgres/statement_samples.py | 3 +-- 7 files changed, 12 insertions(+), 30 deletions(-) diff --git a/citrix_hypervisor/datadog_checks/citrix_hypervisor/check.py b/citrix_hypervisor/datadog_checks/citrix_hypervisor/check.py index 5323b270bad13..02c96fcf00676 100644 --- a/citrix_hypervisor/datadog_checks/citrix_hypervisor/check.py +++ b/citrix_hypervisor/datadog_checks/citrix_hypervisor/check.py @@ -2,9 +2,9 @@ # All rights reserved # Licensed under a 3-clause BSD style license (see LICENSE) from typing import Any, Dict, List # noqa: F401 +from xmlrpc.client import ServerProxy import yaml -from six.moves import xmlrpc_client as xmlrpclib from datadog_checks.base import AgentCheck, ConfigurationError from datadog_checks.base.ddyaml import yaml_load_force_loader @@ -94,7 +94,7 @@ def _get_master_session(self, session): master_address = session['ErrorDescription'][1] if not master_address.startswith('http://'): master_address = 'http://' + master_address - master_xenserver = xmlrpclib.Server(master_address) + master_xenserver = ServerProxy(master_address) # Master credentials can be different, we could specify new `master_username` and # `master_password` options later if requested @@ -109,7 +109,7 @@ def _get_master_session(self, session): def open_session(self): # type: () -> Dict[str, str] try: - self.xenserver = xmlrpclib.Server(self._base_url) + self.xenserver = ServerProxy(self._base_url) except Exception as e: self.log.warning(str(e)) return {} diff --git a/citrix_hypervisor/tests/test_citrix_hypervisor.py b/citrix_hypervisor/tests/test_citrix_hypervisor.py index 50ba95bc68430..a0b9271b72a01 100644 --- a/citrix_hypervisor/tests/test_citrix_hypervisor.py +++ b/citrix_hypervisor/tests/test_citrix_hypervisor.py @@ -27,7 +27,7 @@ ], ) def test_open_session(instance, side_effect, expected_session, tag): - with mock.patch('six.moves.xmlrpc_client.Server', side_effect=side_effect): + with mock.patch('datadog_checks.citrix_hypervisor.check.ServerProxy', side_effect=side_effect): check = CitrixHypervisorCheck('citrix_hypervisor', {}, [instance]) session = check.open_session() @@ -38,7 +38,7 @@ def test_open_session(instance, side_effect, expected_session, tag): @pytest.mark.usefixtures('mock_responses') @pytest.mark.parametrize('server_type', [pytest.param('master'), pytest.param('slave')]) def test_check(aggregator, dd_run_check, instance, server_type): - with mock.patch('six.moves.xmlrpc_client.Server', return_value=mocked_xenserver(server_type)): + with mock.patch('datadog_checks.citrix_hypervisor.check.ServerProxy', return_value=mocked_xenserver(server_type)): check = CitrixHypervisorCheck('citrix_hypervisor', {}, [instance]) dd_run_check(check) diff --git a/citrix_hypervisor/tests/test_metadata.py b/citrix_hypervisor/tests/test_metadata.py index 490710a1c10e3..f4160301e424e 100644 --- a/citrix_hypervisor/tests/test_metadata.py +++ b/citrix_hypervisor/tests/test_metadata.py @@ -31,7 +31,7 @@ def test_collect_metadata(datadog_agent, instance): xenserver.session.get_this_host.return_value = {'Status': 'Success', 'Value': 'hostref'} xenserver.host.get_software_version.return_value = content - with mock.patch('six.moves.xmlrpc_client.Server', return_value=xenserver): + with mock.patch('datadog_checks.citrix_hypervisor.check.ServerProxy', return_value=xenserver): check.check(None) datadog_agent.assert_metadata('test:123', version_metadata) datadog_agent.assert_metadata_count(len(version_metadata)) diff --git a/datadog_checks_tests_helper/datadog_test_libs/win/pdh_mocks.py b/datadog_checks_tests_helper/datadog_test_libs/win/pdh_mocks.py index 7b1d1cf4d0ad9..1b1d6d990c34e 100644 --- a/datadog_checks_tests_helper/datadog_test_libs/win/pdh_mocks.py +++ b/datadog_checks_tests_helper/datadog_test_libs/win/pdh_mocks.py @@ -2,12 +2,11 @@ # All rights reserved # Licensed under a 3-clause BSD style license (see LICENSE) import os +import winreg from collections import defaultdict import mock import pytest -from six import PY3 -from six.moves import winreg HERE = os.path.abspath(os.path.dirname(__file__)) @@ -19,10 +18,7 @@ @pytest.fixture def pdh_mocks_fixture(): - if PY3: - regqueryvalueex = mock.patch('winreg.QueryValueEx', mock_QueryValueEx) - else: - regqueryvalueex = mock.patch('_winreg.QueryValueEx', mock_QueryValueEx) + regqueryvalueex = mock.patch('winreg.QueryValueEx', mock_QueryValueEx) pdhlookupbyindex = mock.patch('win32pdh.LookupPerfNameByIndex', mock_LookupPerfNameByIndex) pdhenumobjectitems = mock.patch('win32pdh.EnumObjectItems', mock_EnumObjectItems) @@ -46,10 +42,7 @@ def pdh_mocks_fixture(): @pytest.fixture def pdh_mocks_fixture_bad_perf_strings(): - if PY3: - regqueryvalueex = mock.patch('winreg.QueryValueEx', mock_QueryValueExWithRaise) - else: - regqueryvalueex = mock.patch('_winreg.QueryValueEx', mock_QueryValueExWithRaise) + regqueryvalueex = mock.patch('winreg.QueryValueEx', mock_QueryValueExWithRaise) pdhlookupbyindex = mock.patch('win32pdh.LookupPerfNameByIndex', mock_LookupPerfNameByIndex) pdhenumobjectitems = mock.patch('win32pdh.EnumObjectItems', mock_EnumObjectItems) diff --git a/gearmand/datadog_checks/gearmand/gearmand.py b/gearmand/datadog_checks/gearmand/gearmand.py index 20768e418324b..299c955bc6378 100644 --- a/gearmand/datadog_checks/gearmand/gearmand.py +++ b/gearmand/datadog_checks/gearmand/gearmand.py @@ -3,18 +3,10 @@ # All rights reserved # Licensed under Simplified BSD License (see LICENSE) -from six import PY2 +import python3_gearman as gearman from datadog_checks.base import AgentCheck -# Python 3 compatibility is a different library -# It's a drop in replacement but has a different name -# This will enable the check to use the new library -if PY2: - import gearman -else: - import python3_gearman as gearman - MAX_NUM_TASKS = 200 diff --git a/network/datadog_checks/network/ethtool.py b/network/datadog_checks/network/ethtool.py index 35ba342bdab60..55e9dc935c780 100644 --- a/network/datadog_checks/network/ethtool.py +++ b/network/datadog_checks/network/ethtool.py @@ -11,8 +11,6 @@ import struct from collections import defaultdict -from six import PY3 - from .const import ( ENA_METRIC_NAMES, ENA_METRIC_PREFIX, @@ -41,7 +39,7 @@ def _byte_array_to_string(s): Convert a byte array to string b'hv_netvsc\x00\x00\x00\x00' -> 'hv_netvsc' """ - s = s.tobytes() if PY3 else s.tostring() + s = s.tobytes() s = s.partition(b'\x00')[0].decode('utf-8') return s diff --git a/postgres/datadog_checks/postgres/statement_samples.py b/postgres/datadog_checks/postgres/statement_samples.py index eb89ecaac7b07..0a93cad3b5616 100644 --- a/postgres/datadog_checks/postgres/statement_samples.py +++ b/postgres/datadog_checks/postgres/statement_samples.py @@ -10,7 +10,6 @@ import psycopg2 from cachetools import TTLCache -from six import PY2 from datadog_checks.postgres.cursor import CommenterCursor, CommenterDictCursor @@ -884,6 +883,6 @@ def _get_truncation_state(track_activity_query_size, statement): # multi-byte characters that fall on the limit are left out. One caveat is that if a statement's length # happens to be greater or equal to the threshold below but isn't actually truncated, this # would falsely report it as a truncated statement - statement_bytes = bytes(statement) if PY2 else bytes(statement, "utf-8") + statement_bytes = bytes(statement, "utf-8") truncated = len(statement_bytes) >= track_activity_query_size - (MAX_CHARACTER_SIZE_IN_BYTES + 1) return StatementTruncationState.truncated if truncated else StatementTruncationState.not_truncated From 4ebe692df1d793ede9b95acc0c7126bb5a10dab0 Mon Sep 17 00:00:00 2001 From: HadhemiDD <43783545+HadhemiDD@users.noreply.github.com> Date: Fri, 20 Sep 2024 11:01:01 +0200 Subject: [PATCH 18/34] fix link (#18630) --- kubernetes_cluster_autoscaler/README.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/kubernetes_cluster_autoscaler/README.md b/kubernetes_cluster_autoscaler/README.md index dcf191ee5b057..6bcba25f4241d 100644 --- a/kubernetes_cluster_autoscaler/README.md +++ b/kubernetes_cluster_autoscaler/README.md @@ -98,7 +98,7 @@ See [service_checks.json][8] for a list of service checks provided by this integ Need help? Contact [Datadog support][9]. -[1]: **LINK_TO_INTEGRATION_SITE** +[1]: https://docs.datadoghq.com/integrations/kubernetes_cluster_autoscaler/ [2]: https://app.datadoghq.com/account/settings/agent/latest [3]: https://docs.datadoghq.com/agent/kubernetes/integrations/ [4]: https://github.com/DataDog/integrations-core/blob/master/kubernetes_cluster_autoscaler/datadog_checks/kubernetes_cluster_autoscaler/data/conf.yaml.example From 1091fca604df38f5b0dc152304c744be4660c8a5 Mon Sep 17 00:00:00 2001 From: Kyle Neale Date: Fri, 20 Sep 2024 10:05:52 -0400 Subject: [PATCH 19/34] Remove py3.12 bump from Oracle changelog (#18629) --- oracle/changelog.d/18207.added | 1 - 1 file changed, 1 deletion(-) delete mode 100644 oracle/changelog.d/18207.added diff --git a/oracle/changelog.d/18207.added b/oracle/changelog.d/18207.added deleted file mode 100644 index 624cd9836c9fe..0000000000000 --- a/oracle/changelog.d/18207.added +++ /dev/null @@ -1 +0,0 @@ -Bump the python version from 3.11 to 3.12 \ No newline at end of file From a1f350b9d9c0c8cf601a8638a13102c03def8e19 Mon Sep 17 00:00:00 2001 From: Kyle Neale Date: Fri, 20 Sep 2024 10:06:13 -0400 Subject: [PATCH 20/34] Remove Oracle from ddev CI scripts and codecov (#18628) * remove Oracle from ddev CI scripts and labeler * add back labeler since it breaks validations --- .codecov.yml | 9 --------- .../scripts/oracle/linux/55_docker_login.sh | 20 ------------------- .github/workflows/config/labeler.yml | 4 ++-- 3 files changed, 2 insertions(+), 31 deletions(-) delete mode 100755 .ddev/ci/scripts/oracle/linux/55_docker_login.sh diff --git a/.codecov.yml b/.codecov.yml index 8649e2e30051a..f60066db3ad0a 100644 --- a/.codecov.yml +++ b/.codecov.yml @@ -450,10 +450,6 @@ coverage: target: 75 flags: - openstack_controller - Oracle_Database: - target: 75 - flags: - - oracle PDH: target: 75 flags: @@ -1296,11 +1292,6 @@ flags: paths: - openstack_controller/datadog_checks/openstack_controller - openstack_controller/tests - oracle: - carryforward: true - paths: - - oracle/datadog_checks/oracle - - oracle/tests pdh_check: carryforward: true paths: diff --git a/.ddev/ci/scripts/oracle/linux/55_docker_login.sh b/.ddev/ci/scripts/oracle/linux/55_docker_login.sh deleted file mode 100755 index 59112e7029185..0000000000000 --- a/.ddev/ci/scripts/oracle/linux/55_docker_login.sh +++ /dev/null @@ -1,20 +0,0 @@ -#!/bin/bash -set -euo pipefail - -set +x - -echo "[INFO] Docker login" -for i in 2 4 8 16 32; do - echo "$ORACLE_DOCKER_PASSWORD" | docker login container-registry.oracle.com --username "$ORACLE_DOCKER_USERNAME" --password-stdin && break - echo "[INFO] Wait $i seconds and retry docker login" - sleep $i -done - -set -ex - -# Allocate 3GB for oracle -# The default number of memory addresses is 65536, i.e. 512MB (Linux 64-bit). -# => To get 3GB, we multiply that amount by 6. -sudo sysctl -w vm.max_map_count=$(expr 6 \* 65536) - -set +ex diff --git a/.github/workflows/config/labeler.yml b/.github/workflows/config/labeler.yml index bebee0a7ae6f8..01e1275ed56c2 100644 --- a/.github/workflows/config/labeler.yml +++ b/.github/workflows/config/labeler.yml @@ -293,10 +293,10 @@ integration/kubernetes_state: - kubernetes_state/**/* integration/kubernetes_state_core: - kubernetes_state_core/**/* -integration/kubevirt_controller: -- kubevirt_controller/**/* integration/kubevirt_api: - kubevirt_api/**/* +integration/kubevirt_controller: +- kubevirt_controller/**/* integration/kyototycoon: - kyototycoon/**/* integration/kyverno: From 03dbcdb90a9ffb367c914c69c69c12468d0803fc Mon Sep 17 00:00:00 2001 From: dkirov-dd <166512750+dkirov-dd@users.noreply.github.com> Date: Fri, 20 Sep 2024 16:17:55 +0200 Subject: [PATCH 21/34] [AWS Neuron] Add metric units (#18632) * Add metric units * Add byte unit to memory metrics * Change percent to fraction * Move units from interval column --- aws_neuron/metadata.csv | 36 ++++++++++++++++++------------------ 1 file changed, 18 insertions(+), 18 deletions(-) diff --git a/aws_neuron/metadata.csv b/aws_neuron/metadata.csv index 6da50f2d21af1..67f1758149219 100644 --- a/aws_neuron/metadata.csv +++ b/aws_neuron/metadata.csv @@ -1,34 +1,34 @@ metric_name,metric_type,interval,unit_name,per_unit_name,description,orientation,integration,short_name,curated_metric,sample_tags aws_neuron.execution.errors.count,count,,,,Execution errors total,0,aws_neuron,,, aws_neuron.execution.errors_created,gauge,,,,Execution errors total,0,aws_neuron,,, -aws_neuron.execution.latency_seconds,gauge,,,,Execution latency in seconds,0,aws_neuron,,, +aws_neuron.execution.latency_seconds,gauge,,second,,Execution latency in seconds,0,aws_neuron,,, aws_neuron.execution.status.count,count,,,,Execution status total,0,aws_neuron,,, aws_neuron.execution.status_created,gauge,,,,Execution status total,0,aws_neuron,,, aws_neuron.hardware_ecc_events.count,count,,,,Hardware ecc events total,0,aws_neuron,,, aws_neuron.hardware_ecc_events_created,gauge,,,,Hardware ecc events total,0,aws_neuron,,, aws_neuron.instance_info,gauge,,,,EC2 instance information,0,aws_neuron,,, aws_neuron.neuron_hardware_info,gauge,,,,Neuron Hardware Information,0,aws_neuron,,, -aws_neuron.neuron_runtime.memory_used_bytes,gauge,,,,Runtime memory used bytes,0,aws_neuron,,, -aws_neuron.neuron_runtime.vcpu_usage_ratio,gauge,,,,Runtime vCPU utilization ratio,0,aws_neuron,,, -aws_neuron.neuroncore.memory_usage.constants,gauge,,,,NeuronCore memory utilization for constants,0,aws_neuron,,, -aws_neuron.neuroncore.memory_usage.model.code,gauge,,,,NeuronCore memory utilization for model_code,0,aws_neuron,,, -aws_neuron.neuroncore.memory_usage.model.shared_scratchpad,gauge,,,,NeuronCore memory utilization for model_shared_scratchpad,0,aws_neuron,,, -aws_neuron.neuroncore.memory_usage.runtime_memory,gauge,,,,NeuronCore memory utilization for runtime_memory,0,aws_neuron,,, -aws_neuron.neuroncore.memory_usage.tensors,gauge,,,,NeuronCore memory utilization for tensors,0,aws_neuron,,, -aws_neuron.neuroncore.utilization_ratio,gauge,,,,NeuronCore utilization ratio,0,aws_neuron,,, -aws_neuron.process.cpu_seconds.count,count,,,,Total user and system CPU time spent in seconds.,0,aws_neuron,,, +aws_neuron.neuron_runtime.memory_used_bytes,gauge,,byte,,Runtime memory used bytes,0,aws_neuron,,, +aws_neuron.neuron_runtime.vcpu_usage_ratio,gauge,,fraction,,Runtime vCPU utilization ratio,0,aws_neuron,,, +aws_neuron.neuroncore.memory_usage.constants,gauge,,byte,,NeuronCore memory utilization for constants,0,aws_neuron,,, +aws_neuron.neuroncore.memory_usage.model.code,gauge,,byte,,NeuronCore memory utilization for model_code,0,aws_neuron,,, +aws_neuron.neuroncore.memory_usage.model.shared_scratchpad,gauge,,byte,,NeuronCore memory utilization for model_shared_scratchpad,0,aws_neuron,,, +aws_neuron.neuroncore.memory_usage.runtime_memory,gauge,,byte,,NeuronCore memory utilization for runtime_memory,0,aws_neuron,,, +aws_neuron.neuroncore.memory_usage.tensors,gauge,,byte,,NeuronCore memory utilization for tensors,0,aws_neuron,,, +aws_neuron.neuroncore.utilization_ratio,gauge,,fraction,,NeuronCore utilization ratio,0,aws_neuron,,, +aws_neuron.process.cpu_seconds.count,count,,second,,Total user and system CPU time spent in seconds.,0,aws_neuron,,, aws_neuron.process.max_fds,gauge,,,,Maximum number of open file descriptors.,0,aws_neuron,,, aws_neuron.process.open_fds,gauge,,,,Number of open file descriptors.,0,aws_neuron,,, -aws_neuron.process.resident_memory_bytes,gauge,,,,Resident memory size in bytes.,0,aws_neuron,,, -aws_neuron.process.start_time_seconds,gauge,,,,Start time of the process since unix epoch in seconds.,0,aws_neuron,,, -aws_neuron.process.virtual_memory_bytes,gauge,,,,Virtual memory size in bytes.,0,aws_neuron,,, +aws_neuron.process.resident_memory_bytes,gauge,,byte,,Resident memory size in bytes.,0,aws_neuron,,, +aws_neuron.process.start_time_seconds,gauge,,second,,Start time of the process since unix epoch in seconds.,0,aws_neuron,,, +aws_neuron.process.virtual_memory_bytes,gauge,,byte,,Virtual memory size in bytes.,0,aws_neuron,,, aws_neuron.python_gc.collections.count,count,,,,Number of times this generation was collected,0,aws_neuron,,, aws_neuron.python_gc.objects_collected.count,count,,,,Objects collected during gc,0,aws_neuron,,, aws_neuron.python_gc.objects_uncollectable.count,count,,,,Uncollectable objects found during GC,0,aws_neuron,,, aws_neuron.python_info,gauge,,,,Python platform information,0,aws_neuron,,, -aws_neuron.system.memory.total_bytes,gauge,,,,System memory total_bytes bytes,0,aws_neuron,,, -aws_neuron.system.memory.used_bytes,gauge,,,,System memory used_bytes bytes,0,aws_neuron,,, -aws_neuron.system.swap.total_bytes,gauge,,,,System swap total_bytes bytes,0,aws_neuron,,, -aws_neuron.system.swap.used_bytes,gauge,,,,System swap used_bytes bytes,0,aws_neuron,,, +aws_neuron.system.memory.total_bytes,gauge,,byte,,System memory total_bytes bytes,0,aws_neuron,,, +aws_neuron.system.memory.used_bytes,gauge,,byte,,System memory used_bytes bytes,0,aws_neuron,,, +aws_neuron.system.swap.total_bytes,gauge,,byte,,System swap total_bytes bytes,0,aws_neuron,,, +aws_neuron.system.swap.used_bytes,gauge,,byte,,System swap used_bytes bytes,0,aws_neuron,,, aws_neuron.system.vcpu.count,gauge,,,,System vCPU count,0,aws_neuron,,, -aws_neuron.system.vcpu.usage_ratio,gauge,,,,System CPU utilization ratio,0,aws_neuron,,, +aws_neuron.system.vcpu.usage_ratio,gauge,,fraction,,System CPU utilization ratio,0,aws_neuron,,, From 4013d3c4bdd02114b1c154d1e7dd39d1268774b7 Mon Sep 17 00:00:00 2001 From: Kyle Neale Date: Fri, 20 Sep 2024 12:49:05 -0400 Subject: [PATCH 22/34] Releases new version of `datadog_checks_base` (#18627) * [Release] Bumped datadog_checks_base version to 37.0.0 * [Release] Update metadata --- .in-toto/tag.bd79a691.link | 2 +- datadog_checks_base/CHANGELOG.md | 10 ++++++++++ datadog_checks_base/changelog.d/18207.added | 1 - datadog_checks_base/changelog.d/18580.removed | 1 - datadog_checks_base/datadog_checks/base/__about__.py | 2 +- requirements-agent-release.txt | 2 +- 6 files changed, 13 insertions(+), 5 deletions(-) delete mode 100644 datadog_checks_base/changelog.d/18207.added delete mode 100644 datadog_checks_base/changelog.d/18580.removed diff --git a/.in-toto/tag.bd79a691.link b/.in-toto/tag.bd79a691.link index a68e9d8b28f8f..4310534b66bc6 100644 --- a/.in-toto/tag.bd79a691.link +++ b/.in-toto/tag.bd79a691.link @@ -1 +1 @@ -{"signatures":[{"keyid":"bd79a69194aacf097b59842ac1b563f375b07ebb","other_headers":"04000108001d162104bd79a69194aacf097b59842ac1b563f375b07ebb050266b28b39","signature":"64376f989d47bca123777fb307c947cf2997cab8dc1de09a2c6bbd7fd202637e294387ad545ff401c0614a4946ffca4ce3933238edc727fe3ad8788e585dc3d30911fae0e0df28e2a23ce2d68da82491e93a14a3a4b2f46ddf04800b6942b9ef32ca135a3d1a2ea985c666bc0b6a1d60cdfb2322acc4c8fa73830a0d070c132d43ad6112791a7af7a9789a891491b65e5f951f98e6487f30aeecab3e4638f42bd9bd17625194b7bf9dcec6d64e69caa001ccd53e3966d053577b6c76ab6bc4080827bdb3bd2d93f2ffc74bfd78bf0f39b0e8e45f2f3e0d3894a7c180b7fb17ca6533f284605c073e06633b70e7d239336d86e82023046d3a234461e6603506d1ae61b33f7d2bd1891e858fbacbd0ba2f2c4023a599b28158c91762b6295dbf254c90e82d8d16807f54bf11d21e79d0dbff5b2cf8dc4553a9e42a9db13897867dbc5f7c8f6291ca7825410ff4ad6064b8843252c825332928e8e21e63fba98ce9892d155043f1be277fe0e0f1ffd6b248b0b415751ca2efb0ed287ecd773541d6"}],"signed":{"_type":"link","byproducts":{},"command":[],"environment":{},"materials":{},"name":"tag","products":{"datadog_checks_base/datadog_checks/__init__.py":{"sha256":"9a3c64b8b00c94da4b4f34618d803d3255808caf21b8afa9195c84b61da66b6a"},"datadog_checks_base/datadog_checks/base/__about__.py":{"sha256":"96cecde7bb5d18d8cb6ff27bb5bca11394235a466fa59da0d9ac04d8af020c53"},"datadog_checks_base/datadog_checks/base/__init__.py":{"sha256":"0942da73cf6bf391501af55ba961211e917d1bd041260867c1db2078ac0735c0"},"datadog_checks_base/datadog_checks/base/agent.py":{"sha256":"9d7f71fc16188b344a18d61f5eae73ed2d0567dc5c5ffc1ddadbb1a467b7ffc9"},"datadog_checks_base/datadog_checks/base/checks/__init__.py":{"sha256":"6b45aff8e774058500e39cf7ede54ebee81f95364c8a380648eb89aa7744dc35"},"datadog_checks_base/datadog_checks/base/checks/base.py":{"sha256":"7aa44fb80996534e0f25d1f168038d486e81355fb9c6bd67024e2466d24b64d0"},"datadog_checks_base/datadog_checks/base/checks/kube_leader/__init__.py":{"sha256":"ac4335c2a324c7c24bbc9a5834730ecba39d3e60b0438e8948e7c4dd00c0a726"},"datadog_checks_base/datadog_checks/base/checks/kube_leader/base_check.py":{"sha256":"d8b21153a6b67096f86f2338437bf54955498d05bc363549affc9428e7e32a35"},"datadog_checks_base/datadog_checks/base/checks/kube_leader/mixins.py":{"sha256":"611acba9c969866f4feca52ad89b05757da1ab60393fdd2211f078da95b61400"},"datadog_checks_base/datadog_checks/base/checks/kube_leader/record.py":{"sha256":"6aa334545b055aeda90343b976cfbabf959038cee58103321b0a26e90eaa09a5"},"datadog_checks_base/datadog_checks/base/checks/kubelet_base/__init__.py":{"sha256":"e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855"},"datadog_checks_base/datadog_checks/base/checks/kubelet_base/base.py":{"sha256":"97ec3af5e262a9f1a3dcc0664f01cca4df95241771c4bf53d09fa06b4a8fbc23"},"datadog_checks_base/datadog_checks/base/checks/libs/__init__.py":{"sha256":"e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855"},"datadog_checks_base/datadog_checks/base/checks/libs/prometheus.py":{"sha256":"8bafefc35dbd077ec5511de7fd88b56c0efb564dd706dacf0fc5f27ac9c64aa6"},"datadog_checks_base/datadog_checks/base/checks/libs/thread_pool.py":{"sha256":"0005e02999a4f6c172768f3f0d26fb064abb06d639b0d5081839b6f0d5a1a4c2"},"datadog_checks_base/datadog_checks/base/checks/libs/timer.py":{"sha256":"8ac17c602136ed7a5e7a1bb39389782190afc505574dd6cd8a46c1db146780c4"},"datadog_checks_base/datadog_checks/base/checks/libs/vmware/__init__.py":{"sha256":"e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855"},"datadog_checks_base/datadog_checks/base/checks/libs/vmware/all_metrics.py":{"sha256":"4f89b8c40a8abc0f57b6abbea2227be3cd8a0a000e34a134b48800fc4a0842c6"},"datadog_checks_base/datadog_checks/base/checks/libs/vmware/basic_metrics.py":{"sha256":"37d804c9398bb121d2f89162729347394ab2b3b627ad1affb8f608eb34ab35bb"},"datadog_checks_base/datadog_checks/base/checks/logs/__init__.py":{"sha256":"ee840b22269e35c03869d3b7223c20ab5c65895ae23153a8d273c0f866d5b99d"},"datadog_checks_base/datadog_checks/base/checks/logs/crawler/__init__.py":{"sha256":"ee840b22269e35c03869d3b7223c20ab5c65895ae23153a8d273c0f866d5b99d"},"datadog_checks_base/datadog_checks/base/checks/logs/crawler/base.py":{"sha256":"90ae36a4dc349b0d0fc1ac102f3c21b1d0100b55688ce5e0149e3844da37ac8c"},"datadog_checks_base/datadog_checks/base/checks/logs/crawler/stream.py":{"sha256":"300a7c23ecc48f972f4e9ee9349345977d800ddcea359c4682425fc9dadf847e"},"datadog_checks_base/datadog_checks/base/checks/network.py":{"sha256":"5228cfd4e5410a908d28ccba6d590d6b31e0cba49d9bca82bc26063da5ae4c3a"},"datadog_checks_base/datadog_checks/base/checks/openmetrics/__init__.py":{"sha256":"3876cda6f0d3eb38d15b8d91cd85991f383e692f2a5d83984292aea2e9942771"},"datadog_checks_base/datadog_checks/base/checks/openmetrics/base_check.py":{"sha256":"6e5867daf9717220a004bc9b2910c30701c9f75a2a2b6ab56f4f1f21886a7891"},"datadog_checks_base/datadog_checks/base/checks/openmetrics/mixins.py":{"sha256":"87ec296b602d1a77c6ba2221bab39189764a9bcc30657401f962169782250777"},"datadog_checks_base/datadog_checks/base/checks/openmetrics/v2/__init__.py":{"sha256":"3fcd4506124b03d306a73e0bee8ffb0bea6f13077803ff235855906758e0d048"},"datadog_checks_base/datadog_checks/base/checks/openmetrics/v2/base.py":{"sha256":"f64ade8cfe355071e56c86180d000b052976896467b929433fb86bcddf801f3e"},"datadog_checks_base/datadog_checks/base/checks/openmetrics/v2/first_scrape_handler.py":{"sha256":"227fad65733389e49d2f6397265200162efc29b415c2e26718fd2268b1fdf7be"},"datadog_checks_base/datadog_checks/base/checks/openmetrics/v2/labels.py":{"sha256":"d05d084a1d37c12daf56c8db9ecdc5ad80e7ea0bf18f45effb67e40361e1f43f"},"datadog_checks_base/datadog_checks/base/checks/openmetrics/v2/scraper.py":{"sha256":"7783dc24cfdbef09075014396ff219a3f556ff4ee3423fe0c1381f0eb929d410"},"datadog_checks_base/datadog_checks/base/checks/openmetrics/v2/transform.py":{"sha256":"eb81688905d875914fbb6c9b246a1dc9812068b0e05a9944dd89cb949b035290"},"datadog_checks_base/datadog_checks/base/checks/openmetrics/v2/transformers/__init__.py":{"sha256":"84f667f162ef41faf32d2689c6d15b61802d2b576df084174942cbefdb2b663b"},"datadog_checks_base/datadog_checks/base/checks/openmetrics/v2/transformers/counter.py":{"sha256":"9a6362e041ad2cd605c566b4c88ff7f0d63d681222e35e0cf6945268635b71a2"},"datadog_checks_base/datadog_checks/base/checks/openmetrics/v2/transformers/counter_gauge.py":{"sha256":"a1bd42bc2747afe56b73905295a4f73972f917633a07b3866a15007a4545dc5c"},"datadog_checks_base/datadog_checks/base/checks/openmetrics/v2/transformers/gauge.py":{"sha256":"ff6a19d789bfe7f6fb94e47eb4cc49461b1e17aafa7fd0ec3bee0b6c023288f1"},"datadog_checks_base/datadog_checks/base/checks/openmetrics/v2/transformers/histogram.py":{"sha256":"872b69c3785029d57037ccb991e5ba58672adebe3efb11272431f1c167fa8e52"},"datadog_checks_base/datadog_checks/base/checks/openmetrics/v2/transformers/metadata.py":{"sha256":"069b093750fd272f78bb12deee4a472f5e042dd961530c939a5e51f3d3003aea"},"datadog_checks_base/datadog_checks/base/checks/openmetrics/v2/transformers/rate.py":{"sha256":"7beb75edc142b002a77d7810add521f79c3496c972de2b80d36322cc63ffa1c3"},"datadog_checks_base/datadog_checks/base/checks/openmetrics/v2/transformers/service_check.py":{"sha256":"e0244e3b8da63d241c593dfbe9b4c722fb0e68b0db2ce9883e197ce1c58501b5"},"datadog_checks_base/datadog_checks/base/checks/openmetrics/v2/transformers/summary.py":{"sha256":"d01d5693b79ae07da77ddb0e5fca10122a2804636aca914372304f2a31d5b52e"},"datadog_checks_base/datadog_checks/base/checks/openmetrics/v2/transformers/temporal_percent.py":{"sha256":"c02a8ea971a8550de5c99066fc04e7830a6f21d81c7ce905ff59461397e88625"},"datadog_checks_base/datadog_checks/base/checks/openmetrics/v2/transformers/time_elapsed.py":{"sha256":"c8fb3bd9478e82bd9e40e7610638c507a7add21327c034beaee516388f160db1"},"datadog_checks_base/datadog_checks/base/checks/openmetrics/v2/utils.py":{"sha256":"b6993786d240cff2b0091a85b360938da8c790b0acff64db19d069e75e2b58e4"},"datadog_checks_base/datadog_checks/base/checks/prometheus/__init__.py":{"sha256":"35c57ac8d1d9555c42ac0ac80ece6d4a459fae72f05398b195374d5c57284a30"},"datadog_checks_base/datadog_checks/base/checks/prometheus/base_check.py":{"sha256":"2d4b347b12235a4d520d0901a7191e534fa0888d68cb32e21936898ccd8b8f5d"},"datadog_checks_base/datadog_checks/base/checks/prometheus/mixins.py":{"sha256":"7cd64989b4ea2cba4925c4657677edfad6243554925184431282192411220d4b"},"datadog_checks_base/datadog_checks/base/checks/prometheus/prometheus_base.py":{"sha256":"9f35823bf488a24646a04ee8f01269a254cfa160bbfe471625f90b1c05de057e"},"datadog_checks_base/datadog_checks/base/checks/win/__init__.py":{"sha256":"9083ff7fefc6d7404110ec4ee3e1a7cb29730a8d6439ff5deb291388151a7a4a"},"datadog_checks_base/datadog_checks/base/checks/win/winpdh.py":{"sha256":"142f282601923e049811ccdc3de3b89b7e21cbaf48f08e487c34cfea1865e839"},"datadog_checks_base/datadog_checks/base/checks/win/winpdh_base.py":{"sha256":"72f82d18dfbc3bbffe8ca331d92877919630be4c7f708c1573b1b8b28c404cfd"},"datadog_checks_base/datadog_checks/base/checks/win/winpdh_stub.py":{"sha256":"3397f2064cc0b842afa19ac6f64b506a9c241ffecaf8a388605e55a52f372cc9"},"datadog_checks_base/datadog_checks/base/checks/win/wmi/__init__.py":{"sha256":"6f4f143f3ef047e807872bc2396f83a4fab9c96406d846e1a12248e43f144f37"},"datadog_checks_base/datadog_checks/base/checks/win/wmi/counter_type.py":{"sha256":"efbef65c6b56c8f2577b0caf698ca8c15983cbd0f1678da48bd583e5769bb951"},"datadog_checks_base/datadog_checks/base/checks/win/wmi/sampler.py":{"sha256":"789382e2bc921a9ba22a14766785b9d3f8d375f7773d0c36de5ad33ebc14ae19"},"datadog_checks_base/datadog_checks/base/checks/win/wmi/types.py":{"sha256":"e04f1ed72a69d8ff9e3b180bb11adfb656aeaaf6a9582b956803e872a0abc158"},"datadog_checks_base/datadog_checks/base/checks/windows/__init__.py":{"sha256":"b2e1a32eb8591a9d541a935aa5c56f20fa7ebbc3de68cf24df3a650198f2712a"},"datadog_checks_base/datadog_checks/base/checks/windows/perf_counters/__init__.py":{"sha256":"c4ced6dabda1b7e2b1fe3d22f03bae7bf94433606ffdbab7be0d04b34009e4a1"},"datadog_checks_base/datadog_checks/base/checks/windows/perf_counters/base.py":{"sha256":"bb9c0f99fb2a348fa420a6dd4e74b0f9722a26dc60604d440cd5b8f71c640770"},"datadog_checks_base/datadog_checks/base/checks/windows/perf_counters/connection.py":{"sha256":"124462f2699e89a71bb2ead225be6f014cc523f94091459c9d20bb4ce42c006e"},"datadog_checks_base/datadog_checks/base/checks/windows/perf_counters/constants.py":{"sha256":"fdaa806dab5cbbd790bfd36a6d509a43d7394be1e4922504455e68a82dc39799"},"datadog_checks_base/datadog_checks/base/checks/windows/perf_counters/counter.py":{"sha256":"8498fbe90224b5934558588a21110d2224c4d0d6334e5bd52dac213c2566f49e"},"datadog_checks_base/datadog_checks/base/checks/windows/perf_counters/transform.py":{"sha256":"6d93f17ed0f0d1dd55157e3dca21486be9da18e62529c320a6fb9e491920133f"},"datadog_checks_base/datadog_checks/base/checks/windows/perf_counters/transformers/__init__.py":{"sha256":"a8b142ebeee6817e16846d57125966018eac45ef4a9870efba31fbc9c2555e92"},"datadog_checks_base/datadog_checks/base/checks/windows/perf_counters/transformers/count.py":{"sha256":"8263467bddb648fe101243270ff9dcf30edba0a616fa65b69f9fbabe975c9a37"},"datadog_checks_base/datadog_checks/base/checks/windows/perf_counters/transformers/gauge.py":{"sha256":"73be1f652e85addc433ba64aa2fa75ee1daf85322691a351d8e2deb35af4d681"},"datadog_checks_base/datadog_checks/base/checks/windows/perf_counters/transformers/monotonic_count.py":{"sha256":"479c167c31bd2e471baab21d49ce9dce3470b40729dabe153ee5456aa3a5ce2d"},"datadog_checks_base/datadog_checks/base/checks/windows/perf_counters/transformers/rate.py":{"sha256":"3e4c739755cf6cfb68fb942b882a23361e5684c4e3c03710c2a63f8b6310052f"},"datadog_checks_base/datadog_checks/base/checks/windows/perf_counters/transformers/service_check.py":{"sha256":"c2f74b0d2b871ca2276f35bcb8cf10f764dc454b90975d70b2fb7475266dac70"},"datadog_checks_base/datadog_checks/base/checks/windows/perf_counters/transformers/temporal_percent.py":{"sha256":"2071f661338679e8b63d53790a1f7df200ea620facd4939bbfd6b44e602f3a75"},"datadog_checks_base/datadog_checks/base/checks/windows/perf_counters/transformers/time_elapsed.py":{"sha256":"85633c087612a859c562b35daf5345638eb89cc01514e88df238658594ce6fbf"},"datadog_checks_base/datadog_checks/base/checks/windows/perf_counters/utils.py":{"sha256":"a2ffb8f0936f650e8a3c4671cfa5f42187009776d50059ccd79faf9467eab18d"},"datadog_checks_base/datadog_checks/base/checks/windows/perf_counters/utils_win32pdh_fix.py":{"sha256":"9d64eb01b07d5400ff0da693ad2055d8a0e606f8a9185988d38fb00df9f151af"},"datadog_checks_base/datadog_checks/base/config.py":{"sha256":"a9c000e17f6c5d065177041ef0382219ddbdf34541a7549003477af79b57fed5"},"datadog_checks_base/datadog_checks/base/constants.py":{"sha256":"711d7db40a95cac3056dd056a88199a0720a9359064f2a91b029fd15f1503a7c"},"datadog_checks_base/datadog_checks/base/ddyaml.py":{"sha256":"d86ce592be606c30e1844e7f230e716dd894cd6839b3a70dfa826a7abb92c6ca"},"datadog_checks_base/datadog_checks/base/errors.py":{"sha256":"7cf705691df31e94a4328f1ea6f1779517bd2ab1274bc83827475733f9efcecc"},"datadog_checks_base/datadog_checks/base/log.py":{"sha256":"bba63cedd99be5b471b7bf5d34f5115e0dc7e114e121d75e760655823b46be58"},"datadog_checks_base/datadog_checks/base/stubs/__init__.py":{"sha256":"c2958047dbfb0624db6e64ceea9569b21a9aff3f8f59a613af7df049364bcf77"},"datadog_checks_base/datadog_checks/base/stubs/_util.py":{"sha256":"6431ad41af05ddc1dff3e42f4951cc0780462370bd5600bbb067061af3b46a92"},"datadog_checks_base/datadog_checks/base/stubs/aggregator.py":{"sha256":"c215913b4b31fb5d861b50bdc4060efd9c5a6e8f2b4aea9e240cec4637d7e609"},"datadog_checks_base/datadog_checks/base/stubs/common.py":{"sha256":"646cc5d9d5f2d6e545406746fdbbf3fe930c8942da05ca73adafe4f70a3d7f4e"},"datadog_checks_base/datadog_checks/base/stubs/datadog_agent.py":{"sha256":"5efbbbfb0439db0941acdd3778533732cf780e93eb2c7cd615c9227e3b28b3ad"},"datadog_checks_base/datadog_checks/base/stubs/log.py":{"sha256":"03e7969f3639813a535b8d59721f96e4255c97395d96684c4d6faf0cd15d4f5a"},"datadog_checks_base/datadog_checks/base/stubs/similar.py":{"sha256":"cd9d5bab9c0f690fbc70163f1d2fbad76b29151dd4277bf214069756c19c7013"},"datadog_checks_base/datadog_checks/base/stubs/tagging.py":{"sha256":"cf12dd3c2e04a87c46892fc71216da3ac2ffb399d922137c043931d810133aab"},"datadog_checks_base/datadog_checks/base/types.py":{"sha256":"6a76a3652d16d13b31507250c3e24738fd8d49eb82f418ac5d2cbd9804ad9714"},"datadog_checks_base/datadog_checks/base/utils/__init__.py":{"sha256":"4c2d2aee209b36a7188df5396b304da429e2f0b01060e7d8e8500313749910f0"},"datadog_checks_base/datadog_checks/base/utils/agent/__init__.py":{"sha256":"a37696bf2dcf872903fe1ed84f7b3adbc3b45b66291e2b3436542c495d4f234e"},"datadog_checks_base/datadog_checks/base/utils/agent/common.py":{"sha256":"d5746c3c959da666f100a2815471813d49b858776f8aad28bee820b69c7b2d6a"},"datadog_checks_base/datadog_checks/base/utils/agent/debug.py":{"sha256":"cde05b34bb7763f5b1a5ff4e74092595d2f2d6098bd14e9b30398e1d20c63373"},"datadog_checks_base/datadog_checks/base/utils/agent/memory.py":{"sha256":"986d743c51608d608b7ad17a8a10b8eb7550cf55774b849332cfc5f1b00392f6"},"datadog_checks_base/datadog_checks/base/utils/agent/packages.py":{"sha256":"ab88bbece8f9d0a38db114f121b3aa7babf43239debbe4069dd236cfc742d8f9"},"datadog_checks_base/datadog_checks/base/utils/agent/utils.py":{"sha256":"155fe8eab71c53907432b5f299afb8c80aa62a08649734de39fd6785872663ba"},"datadog_checks_base/datadog_checks/base/utils/aws.py":{"sha256":"c3114b5a5545b6fe7f11445db17cc384e45c4e93348c1940a2470c88f575c43f"},"datadog_checks_base/datadog_checks/base/utils/common.py":{"sha256":"b5b95066f10424b78e418babf395ebb80151fd09ab8bdcc0522fe08b43711e0d"},"datadog_checks_base/datadog_checks/base/utils/concurrency/__init__.py":{"sha256":"391b2c1396d766e62e6b95022deb067cfbdcad029c0031da0ec5eb5327c0445d"},"datadog_checks_base/datadog_checks/base/utils/concurrency/limiter.py":{"sha256":"8acd99a541237ec2389c11ad5c6d83d1982fd060f7f3f8b98c940291dccf5bf6"},"datadog_checks_base/datadog_checks/base/utils/constants.py":{"sha256":"4304decb8096074340c66dab703fb03d84641328257a4408ac0cc531a6c46b7f"},"datadog_checks_base/datadog_checks/base/utils/containers.py":{"sha256":"8227d931334393baecb8dcde9132740b832dcb5b26b07f847f6a9b8ebc60b24b"},"datadog_checks_base/datadog_checks/base/utils/date.py":{"sha256":"2499aa3fce0281570527472f02632ef04b4ceaff7ab48112b9c40d9bd78a7847"},"datadog_checks_base/datadog_checks/base/utils/db/__init__.py":{"sha256":"f964bc85274b6a1cf6a9b115c13d38d419f422ef40f4e96ec3a0b8ff36f17e36"},"datadog_checks_base/datadog_checks/base/utils/db/core.py":{"sha256":"b151dd639674771e9433a21667007bd07c62583213344dba4637dfc0b2acf490"},"datadog_checks_base/datadog_checks/base/utils/db/query.py":{"sha256":"001fd3d9274eaba66170d210627d928eddc4391f1f97df3e229a5b2e5200d0a5"},"datadog_checks_base/datadog_checks/base/utils/db/sql.py":{"sha256":"a0f94966a841cf408601aecc10d3dba4e83e39fb878feddbffeaefec981a344b"},"datadog_checks_base/datadog_checks/base/utils/db/sql_commenter.py":{"sha256":"aff79f2ba4c6450c4cffe423fa774425d102938ec9e270858e4572c6ed579af5"},"datadog_checks_base/datadog_checks/base/utils/db/statement_metrics.py":{"sha256":"4b49073e816a099b514d9a1323a3e2f99ba5e88b9ef18de3d7f92c347fb3128f"},"datadog_checks_base/datadog_checks/base/utils/db/timed_cache.py":{"sha256":"6bd2ace4a294bc4c41e863804d74d208b6a656d2c3f75f439fa1e9c1e18ab45c"},"datadog_checks_base/datadog_checks/base/utils/db/transform.py":{"sha256":"15a7533cee2c3dbac0e6838df549ec65064b9c77065677a575f89fcd89cf97d8"},"datadog_checks_base/datadog_checks/base/utils/db/types.py":{"sha256":"38b5f63aae7c2503c5043e28a94a422e400fee8bd118dfe2d6f164d44d0df0b8"},"datadog_checks_base/datadog_checks/base/utils/db/utils.py":{"sha256":"7d1160b13dbbe3323a3bd42b471d05ecfbecc83d4cc0648f03cf8b0054561f7e"},"datadog_checks_base/datadog_checks/base/utils/diagnose.py":{"sha256":"78b31ed2d212a64cd2df2c5547cdc81f1a2b4097c74ff5e2e82ab8cf344dc552"},"datadog_checks_base/datadog_checks/base/utils/discovery/__init__.py":{"sha256":"7c896046fefe58933a21fa392211462b829c0caac24865ff7eb5e0899f73e8c9"},"datadog_checks_base/datadog_checks/base/utils/discovery/cache.py":{"sha256":"7fd2f703b619a59fa6a59a452308ec6b969bf5c64c1a4a9d7ad76bf44a5a95b2"},"datadog_checks_base/datadog_checks/base/utils/discovery/discovery.py":{"sha256":"9bcc3d55162eefc79dd202af56ee1c84dad1752302aae24edd85359b378e734d"},"datadog_checks_base/datadog_checks/base/utils/discovery/filter.py":{"sha256":"459f13ddb2cfe2e84af0a5f01aa76860c254f4edc402f3924e119ae702d0311f"},"datadog_checks_base/datadog_checks/base/utils/functions.py":{"sha256":"8869726f147a68f3c494dc4d6f610b3b36e4df6f23f4e541031ade749c5d091c"},"datadog_checks_base/datadog_checks/base/utils/headers.py":{"sha256":"2f23e5e9cc119258e1fe0b7772f282a081f7593a8a09bc733ea84644c28e699a"},"datadog_checks_base/datadog_checks/base/utils/http.py":{"sha256":"15e51b3d9181e9a2be73655d8828a4893549d3323003ec00931b5e15124b9715"},"datadog_checks_base/datadog_checks/base/utils/limiter.py":{"sha256":"6114eb3c25f54b912d1cb55c3fff8611fcd1a2db3f2d3732d5ffee1d956cf748"},"datadog_checks_base/datadog_checks/base/utils/metadata/__init__.py":{"sha256":"6d36a6f7a190f43be4ea287c70aabc5b16b69640e48feed3b89de85875d432cb"},"datadog_checks_base/datadog_checks/base/utils/metadata/constants.py":{"sha256":"5c77cfc2f40c6f2344d8562607fed7c968862343761b17415dbb572f87839e27"},"datadog_checks_base/datadog_checks/base/utils/metadata/core.py":{"sha256":"64e5a3e1bca0d96533d7e092adff752c8ec2ad2349dcc87a510c10dd0ff03d7d"},"datadog_checks_base/datadog_checks/base/utils/metadata/utils.py":{"sha256":"4c2876f1c9b1434dcc413b9e3af4274f5ad0b604c7dadf30fde8e90901dcaa9e"},"datadog_checks_base/datadog_checks/base/utils/metadata/version.py":{"sha256":"7257bc2c7c2a72ee364ea14a24625d16d1c098e7a2b423a2ce34cd43606cc534"},"datadog_checks_base/datadog_checks/base/utils/models/__init__.py":{"sha256":"b2e1a32eb8591a9d541a935aa5c56f20fa7ebbc3de68cf24df3a650198f2712a"},"datadog_checks_base/datadog_checks/base/utils/models/types.py":{"sha256":"c9b504a7dcffac396bdbea089158c4581fa6440bd4c28103a4051c6504c4208c"},"datadog_checks_base/datadog_checks/base/utils/models/validation/__init__.py":{"sha256":"699557dfc5b5a642c793b9281e02b9267d8f3824f940a28f1b35bfc3d2e082da"},"datadog_checks_base/datadog_checks/base/utils/models/validation/core.py":{"sha256":"dab25d1f87af4729ec0530dc0b04ed788c0c09906f9e674113a736f2c8c3e5a0"},"datadog_checks_base/datadog_checks/base/utils/models/validation/utils.py":{"sha256":"748ad183c8795ee93e403b08c89285f68b6a45fc34aeeebd1f67c548dcc8b0e8"},"datadog_checks_base/datadog_checks/base/utils/network.py":{"sha256":"970452f4248a9699bed1d5be713e7b2b65fe6f026f4bcbf6afa4fb4ad2bfd6c7"},"datadog_checks_base/datadog_checks/base/utils/platform.py":{"sha256":"c16f1fe972a8e091f1beed6c3740e92c39988d2fdc0aef06a0abdf4e3223e323"},"datadog_checks_base/datadog_checks/base/utils/prometheus/__init__.py":{"sha256":"f794783ecff74f6713b846470f28eaaa841ed20c0d1681bcd18186135e2c150f"},"datadog_checks_base/datadog_checks/base/utils/prometheus/functions.py":{"sha256":"c674f283d8937acffcde65a57351acf05150d6349163e94b3e8e530801626f2b"},"datadog_checks_base/datadog_checks/base/utils/prometheus/metrics_pb2.py":{"sha256":"eab96d57e2144ab9d603d62b031c0c1a198afb41ea22fe11f1b60b0596d65fdf"},"datadog_checks_base/datadog_checks/base/utils/replay/__init__.py":{"sha256":"391b2c1396d766e62e6b95022deb067cfbdcad029c0031da0ec5eb5327c0445d"},"datadog_checks_base/datadog_checks/base/utils/replay/constants.py":{"sha256":"7b10c6b0380b23bbb5196bde7a55dd8335894e3c47a18266413ee8ef0d4509e3"},"datadog_checks_base/datadog_checks/base/utils/replay/execute.py":{"sha256":"048756ff9bafef97f21c68bcd6fe20256091059de549cc1af90ff8c36f2cb21b"},"datadog_checks_base/datadog_checks/base/utils/replay/redirect.py":{"sha256":"f21e13c07ed95637a211ddaacb1f4a68752755137bfcd1a6535745890c252697"},"datadog_checks_base/datadog_checks/base/utils/secrets.py":{"sha256":"4e3e4c04ea321975c3581dc7afeed4283cb9195d4d6499053d8e13e2ff4f1c78"},"datadog_checks_base/datadog_checks/base/utils/serialization.py":{"sha256":"0eafc28756b17e888cf64b65f6e5a02a6c242aef85a65575a3b8a79388ca596c"},"datadog_checks_base/datadog_checks/base/utils/subprocess_output.py":{"sha256":"58dc33d63b9c6e2193b43a0dc735c34ad2a080c82aa657f67959b1b06259b020"},"datadog_checks_base/datadog_checks/base/utils/tagging.py":{"sha256":"004504188c498cdbe8388110405922b7c653d8ec91c62ca6d45cc21227080acb"},"datadog_checks_base/datadog_checks/base/utils/tailfile.py":{"sha256":"c7fa4ce6982655a5b87890704ba19764a3aa89fa66a9faf01ce537816b6162d3"},"datadog_checks_base/datadog_checks/base/utils/time.py":{"sha256":"9caeb78a0273d313748990aea3dd09a6ca47119cc52671bcca42428186a9a41c"},"datadog_checks_base/datadog_checks/base/utils/timeout.py":{"sha256":"78e059a1f14dfa13aee7125e30e17769cfe87dccbd118ebe92f981bcfe101058"},"datadog_checks_base/datadog_checks/base/utils/tls.py":{"sha256":"84625ba574e57cf11bdc7845783c96fb8409b78fff535e3a99cf7bc1823e4e82"},"datadog_checks_base/datadog_checks/base/utils/tracing.py":{"sha256":"1a38b277fa0ca6fd8566969accf31129c0565a9fe2a1eaf9e64b6b76810da316"},"datadog_checks_base/datadog_checks/base/utils/tracking.py":{"sha256":"158228baabb7281f89b31831335897e48a6ffdc5a1b3ccd03933784b15ce909e"},"datadog_checks_base/datadog_checks/checks/__init__.py":{"sha256":"76381faa72acfce6863031501a53b955d2d047150023e5f828424b2e53e95483"},"datadog_checks_base/datadog_checks/checks/base.py":{"sha256":"df061b86e80d0375a5aedffe104a4198949297c9472ae52dad13fe9d5d8a05f4"},"datadog_checks_base/datadog_checks/checks/libs/__init__.py":{"sha256":"2300c3103843a8f3d4d63e0fcaf78691dbb508cbfd91b7de2bdd0802f981c777"},"datadog_checks_base/datadog_checks/checks/libs/thread_pool.py":{"sha256":"747147080f5d03c2742e233e4bd635d4c58abff8cf5c803fe5881f776c1603b4"},"datadog_checks_base/datadog_checks/checks/libs/timer.py":{"sha256":"a35b1970916e9035ae71d185b6506bbd2b798e26ef1b7b91fc3b30788b19dd49"},"datadog_checks_base/datadog_checks/checks/libs/vmware/__init__.py":{"sha256":"e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855"},"datadog_checks_base/datadog_checks/checks/libs/vmware/all_metrics.py":{"sha256":"849ca20bfd3b86aa8c8eeb88732b4026fd8a0e74a0a96ab6cbe269270b71a8c3"},"datadog_checks_base/datadog_checks/checks/libs/vmware/basic_metrics.py":{"sha256":"683af414d39b8d77d66fd97146999de4bd7a7be9ab934ed9224867a0b74c099f"},"datadog_checks_base/datadog_checks/checks/libs/wmi/__init__.py":{"sha256":"2300c3103843a8f3d4d63e0fcaf78691dbb508cbfd91b7de2bdd0802f981c777"},"datadog_checks_base/datadog_checks/checks/libs/wmi/sampler.py":{"sha256":"55310a8bffb4b55af9552268e627c83c284db55c99c47f754fa05e5bb33ccd3b"},"datadog_checks_base/datadog_checks/checks/network.py":{"sha256":"e5b7d11f19ac5286897259c2ac1e8209ee7bee94df199d5155e1b174cce6afbc"},"datadog_checks_base/datadog_checks/checks/network_checks.py":{"sha256":"cdb77741ebeaa7b6a13616873b5725009b7566a77ceef013fe5dd1c76cbdd81b"},"datadog_checks_base/datadog_checks/checks/openmetrics/__init__.py":{"sha256":"96b910b1c359a0b37a0a8753b6e1e50c803ae35bf4f1e7a31418678cf16792c0"},"datadog_checks_base/datadog_checks/checks/openmetrics/base_check.py":{"sha256":"a9ccd2133baa1058f744e494e2bb43cb5221fbbb41b6c3b8a0553ead7f6a851f"},"datadog_checks_base/datadog_checks/checks/openmetrics/mixins.py":{"sha256":"b5b9cf9d77c8dda6d752a4d37d303405fa1935ac071b17a4e8b055160b7d2a28"},"datadog_checks_base/datadog_checks/checks/prometheus/__init__.py":{"sha256":"a12ac852908b6eb9f3b594e893c41a38951a9a1e030b6286c20e76aec9965c8b"},"datadog_checks_base/datadog_checks/checks/prometheus/base_check.py":{"sha256":"882984f28171bfde4a2c34c25f609b5a758917f0c9b9e00b2ba9a0d489ba1e19"},"datadog_checks_base/datadog_checks/checks/prometheus/mixins.py":{"sha256":"aa652ec3963f90c1e7c552d35243d3863f737fa9de5e45131ca55758c5e00de5"},"datadog_checks_base/datadog_checks/checks/prometheus/prometheus_base.py":{"sha256":"434576db21c019a7366f5350b8e305c0790509aadd9cbd980f6c3ac22b87874f"},"datadog_checks_base/datadog_checks/checks/prometheus_check/__init__.py":{"sha256":"9b5434e894e03018e342ee726f635de62122bf0e1d8f59d3f0109f89a95d890d"},"datadog_checks_base/datadog_checks/checks/win/__init__.py":{"sha256":"4441d475ac7181f8e7edf1037da1310d776c647883d0362a5acfb88e44e4d45e"},"datadog_checks_base/datadog_checks/checks/win/winpdh.py":{"sha256":"864f9f437a351bb82a2962e15a1198cc6271f9cc25c3c73522402954fa42ef9e"},"datadog_checks_base/datadog_checks/checks/win/winpdh_base.py":{"sha256":"269e7c39a6562687aa6ad022b1b88c85b5f6be403c11a8412775b3c55715196e"},"datadog_checks_base/datadog_checks/checks/win/winpdh_stub.py":{"sha256":"a1895061a3be05415b331c1b17d2c02b1f912f36ed238948c46d7af68025d7ed"},"datadog_checks_base/datadog_checks/checks/win/wmi/__init__.py":{"sha256":"51008a1bf148ec1c7bc0bb8f6ad3be19ba950e6c13b5cb84a2fd2684fb66327e"},"datadog_checks_base/datadog_checks/checks/win/wmi/counter_type.py":{"sha256":"01ac5dc9d1518c74c4b621a1da0b901da5ffc954abcf81972f8e478abdb098a7"},"datadog_checks_base/datadog_checks/checks/win/wmi/sampler.py":{"sha256":"bbfeea3683c40059f5e1ff52fc459e82200727041a187328d65f58b92b445cdd"},"datadog_checks_base/datadog_checks/checks/winwmi_check.py":{"sha256":"907ea75f4055e675b732171687ce6148d6484d067b36726b942b4e9c5e1d747b"},"datadog_checks_base/datadog_checks/config.py":{"sha256":"3eb01719d94b9e857654c41f7ff9e849a489274c20e010221a7c3b8c240b4a29"},"datadog_checks_base/datadog_checks/errors.py":{"sha256":"785c269eb5fe6dab4881bbbe889cac7aa00f34eff8ae4d54b9a28af3d7af2e18"},"datadog_checks_base/datadog_checks/log.py":{"sha256":"a2bbce80c286344b2a4932da942010a563fea2ea2ba46d40ec69fe8f69910fcd"},"datadog_checks_base/datadog_checks/py.typed":{"sha256":"95aebb28195b8d737effe0df18d71d39c8d8ba6569286fd3930fbc9f9767181e"},"datadog_checks_base/datadog_checks/stubs/__init__.py":{"sha256":"331c49e4d1a353ded64777893a96906746f386f03cf32b30135aa260665377ec"},"datadog_checks_base/datadog_checks/stubs/_util.py":{"sha256":"9989e9f7fbeca50c1359c06188f60c309dc421c40f4a9a407147d18c42c180b5"},"datadog_checks_base/datadog_checks/stubs/aggregator.py":{"sha256":"ea230225c9c41fbb1885a38741cccf1809034d9b1fd696eb6efde1de6bfba762"},"datadog_checks_base/datadog_checks/stubs/datadog_agent.py":{"sha256":"b167bb8f650441e0f3762aa3e0ffe73e1921157f34ff90abe47723db963ec6bf"},"datadog_checks_base/datadog_checks/utils/__init__.py":{"sha256":"e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855"},"datadog_checks_base/datadog_checks/utils/common.py":{"sha256":"470a0220a572f2f6df6b61353fb63b82447d569274d15b6e3f5f0c12b5e7f6b9"},"datadog_checks_base/datadog_checks/utils/containers.py":{"sha256":"ca56286c527f7d5f940b0baf5a66a4eba6ad2df036759f6000b0985835c63a29"},"datadog_checks_base/datadog_checks/utils/headers.py":{"sha256":"8f92abe525d4947ba4dd41870d0367f4cda884df95e57d9fce62718f0e8fe8da"},"datadog_checks_base/datadog_checks/utils/limiter.py":{"sha256":"1aad3f848119c3aaa92394962ffbb331a22991bcef173b242347a54cffef22c1"},"datadog_checks_base/datadog_checks/utils/platform.py":{"sha256":"76ca8472c0bf63b866e28aea3bc449e4de63905a41ee9c8662f09e5eb6fef571"},"datadog_checks_base/datadog_checks/utils/prometheus/__init__.py":{"sha256":"f16c179ccf60d6c7d2e76ebb4180895ff85df0c665e6446006043f4b374d757b"},"datadog_checks_base/datadog_checks/utils/prometheus/functions.py":{"sha256":"e2f5ac4c62e0ba31f0758668d2d940424a833c7a1fa9424cde4f5e22d26a1114"},"datadog_checks_base/datadog_checks/utils/prometheus/metrics_pb2.py":{"sha256":"c60dd47541dee138de478b35dcb50a6765355954ade94eb38ec5511ba957e9d7"},"datadog_checks_base/datadog_checks/utils/proxy.py":{"sha256":"723edeaf00a1ee813aa4a6467fc34c3ae419f354c6b6172de678e39b864a8bd7"},"datadog_checks_base/datadog_checks/utils/subprocess_output.py":{"sha256":"80dea34445564d9eee264533455310e69c72b4d640e369ea8c97a365717a64c2"},"datadog_checks_base/datadog_checks/utils/tailfile.py":{"sha256":"6c4aa2725fac2e0cb0f660f545d5a8b3c1954a50a49f1e58c86ddf5cc068c137"},"datadog_checks_base/datadog_checks/utils/timeout.py":{"sha256":"42e848b0a6686bf335dfe8040b07acbb9219c12317cfdf564f0a9a974262c913"},"datadog_checks_base/datadog_checks/utils/tracing.py":{"sha256":"1d06d06a428cf8bc3be4c0e422d6982588a94ba60edb867fca60d5c43b31e435"},"datadog_checks_base/pyproject.toml":{"sha256":"e0f394312936d626c2bafbc8cb9a73f20441b676632d9638e2269b4ecfa2c47c"},"datadog_checks_base/setup.py":{"sha256":"c9c3ca1275d54fae5f5f7e1443898c8b383799d2c568d73b247d360249f1fddf"}}}} \ No newline at end of file +{"signatures":[{"keyid":"bd79a69194aacf097b59842ac1b563f375b07ebb","other_headers":"04000108001d162104bd79a69194aacf097b59842ac1b563f375b07ebb050266ecb5a4","signature":"400debb999ba49e6bac69a50ffe9a4f43540af387d902d5c2a99a45069ebd723a4bb8c803c4e6b2515f4e8079f00cace02664d77d5af366067ece3d3dff30d853d5f997505d4b277118096c781ebb50037222de16a7e8bbf0791e4fb066d636c58c57746cc79aa64f08b06bfe8f5cf13beec1df9604a24fa731456b28a8ff7b92375032331826857d8a0453770bc359cec9770bcde8eae3cb46b6de9ec1e1bca4b41c470e3d0ae826aa84c60f64ee5fb9c0906f46a568afa2cae68f5903a681c4cb9cd73bd5bc4525061eb8bfb20316587e965bd83c063a291c8cf035f1e22bab91f5c06f112009fb31395e1afbb66ced6123ffa01a69ec58163172441d59c4833f6f26412073a70a5e04f4c2dc677f8fff79bcc5321264784e08bb98638b078775785c45f4b8b1a79b0593b5fff2d9573c6740941cd3dc826c3e09fbdd51e64e18622f19d5e87966ba02015b7ec2da02bc704010b6cc73b472d1b9e1abffb85504b3c756cb27f5498b1e2d82fa4d1fd2d0bd9e00ac3b25b89238642e4fcdce5"}],"signed":{"_type":"link","byproducts":{},"command":[],"environment":{},"materials":{},"name":"tag","products":{"datadog_checks_base/datadog_checks/__init__.py":{"sha256":"9a3c64b8b00c94da4b4f34618d803d3255808caf21b8afa9195c84b61da66b6a"},"datadog_checks_base/datadog_checks/base/__about__.py":{"sha256":"14a2022ef4f16432ef8426e2576dcbd417d60c22782f0b98fee84e31b144cd20"},"datadog_checks_base/datadog_checks/base/__init__.py":{"sha256":"0942da73cf6bf391501af55ba961211e917d1bd041260867c1db2078ac0735c0"},"datadog_checks_base/datadog_checks/base/agent.py":{"sha256":"9d7f71fc16188b344a18d61f5eae73ed2d0567dc5c5ffc1ddadbb1a467b7ffc9"},"datadog_checks_base/datadog_checks/base/checks/__init__.py":{"sha256":"6b45aff8e774058500e39cf7ede54ebee81f95364c8a380648eb89aa7744dc35"},"datadog_checks_base/datadog_checks/base/checks/base.py":{"sha256":"471972dac4dfd7b59aff29957a553b8ebb0d70edc9506aa28f3a13e58c0a9256"},"datadog_checks_base/datadog_checks/base/checks/kube_leader/__init__.py":{"sha256":"ac4335c2a324c7c24bbc9a5834730ecba39d3e60b0438e8948e7c4dd00c0a726"},"datadog_checks_base/datadog_checks/base/checks/kube_leader/base_check.py":{"sha256":"d8b21153a6b67096f86f2338437bf54955498d05bc363549affc9428e7e32a35"},"datadog_checks_base/datadog_checks/base/checks/kube_leader/mixins.py":{"sha256":"611acba9c969866f4feca52ad89b05757da1ab60393fdd2211f078da95b61400"},"datadog_checks_base/datadog_checks/base/checks/kube_leader/record.py":{"sha256":"6aa334545b055aeda90343b976cfbabf959038cee58103321b0a26e90eaa09a5"},"datadog_checks_base/datadog_checks/base/checks/kubelet_base/__init__.py":{"sha256":"e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855"},"datadog_checks_base/datadog_checks/base/checks/kubelet_base/base.py":{"sha256":"97ec3af5e262a9f1a3dcc0664f01cca4df95241771c4bf53d09fa06b4a8fbc23"},"datadog_checks_base/datadog_checks/base/checks/libs/__init__.py":{"sha256":"e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855"},"datadog_checks_base/datadog_checks/base/checks/libs/prometheus.py":{"sha256":"8bafefc35dbd077ec5511de7fd88b56c0efb564dd706dacf0fc5f27ac9c64aa6"},"datadog_checks_base/datadog_checks/base/checks/libs/thread_pool.py":{"sha256":"0005e02999a4f6c172768f3f0d26fb064abb06d639b0d5081839b6f0d5a1a4c2"},"datadog_checks_base/datadog_checks/base/checks/libs/timer.py":{"sha256":"8ac17c602136ed7a5e7a1bb39389782190afc505574dd6cd8a46c1db146780c4"},"datadog_checks_base/datadog_checks/base/checks/libs/vmware/__init__.py":{"sha256":"e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855"},"datadog_checks_base/datadog_checks/base/checks/libs/vmware/all_metrics.py":{"sha256":"4f89b8c40a8abc0f57b6abbea2227be3cd8a0a000e34a134b48800fc4a0842c6"},"datadog_checks_base/datadog_checks/base/checks/libs/vmware/basic_metrics.py":{"sha256":"37d804c9398bb121d2f89162729347394ab2b3b627ad1affb8f608eb34ab35bb"},"datadog_checks_base/datadog_checks/base/checks/logs/__init__.py":{"sha256":"ee840b22269e35c03869d3b7223c20ab5c65895ae23153a8d273c0f866d5b99d"},"datadog_checks_base/datadog_checks/base/checks/logs/crawler/__init__.py":{"sha256":"ee840b22269e35c03869d3b7223c20ab5c65895ae23153a8d273c0f866d5b99d"},"datadog_checks_base/datadog_checks/base/checks/logs/crawler/base.py":{"sha256":"90ae36a4dc349b0d0fc1ac102f3c21b1d0100b55688ce5e0149e3844da37ac8c"},"datadog_checks_base/datadog_checks/base/checks/logs/crawler/stream.py":{"sha256":"300a7c23ecc48f972f4e9ee9349345977d800ddcea359c4682425fc9dadf847e"},"datadog_checks_base/datadog_checks/base/checks/network.py":{"sha256":"5228cfd4e5410a908d28ccba6d590d6b31e0cba49d9bca82bc26063da5ae4c3a"},"datadog_checks_base/datadog_checks/base/checks/openmetrics/__init__.py":{"sha256":"3876cda6f0d3eb38d15b8d91cd85991f383e692f2a5d83984292aea2e9942771"},"datadog_checks_base/datadog_checks/base/checks/openmetrics/base_check.py":{"sha256":"6e5867daf9717220a004bc9b2910c30701c9f75a2a2b6ab56f4f1f21886a7891"},"datadog_checks_base/datadog_checks/base/checks/openmetrics/mixins.py":{"sha256":"87ec296b602d1a77c6ba2221bab39189764a9bcc30657401f962169782250777"},"datadog_checks_base/datadog_checks/base/checks/openmetrics/v2/__init__.py":{"sha256":"3fcd4506124b03d306a73e0bee8ffb0bea6f13077803ff235855906758e0d048"},"datadog_checks_base/datadog_checks/base/checks/openmetrics/v2/base.py":{"sha256":"f64ade8cfe355071e56c86180d000b052976896467b929433fb86bcddf801f3e"},"datadog_checks_base/datadog_checks/base/checks/openmetrics/v2/first_scrape_handler.py":{"sha256":"227fad65733389e49d2f6397265200162efc29b415c2e26718fd2268b1fdf7be"},"datadog_checks_base/datadog_checks/base/checks/openmetrics/v2/labels.py":{"sha256":"d05d084a1d37c12daf56c8db9ecdc5ad80e7ea0bf18f45effb67e40361e1f43f"},"datadog_checks_base/datadog_checks/base/checks/openmetrics/v2/metrics.py":{"sha256":"0ae41b088a842767025046ebb3620c8811e3245f101196b2b03631b62ff62fc5"},"datadog_checks_base/datadog_checks/base/checks/openmetrics/v2/scraper.py":{"sha256":"7783dc24cfdbef09075014396ff219a3f556ff4ee3423fe0c1381f0eb929d410"},"datadog_checks_base/datadog_checks/base/checks/openmetrics/v2/transform.py":{"sha256":"eb81688905d875914fbb6c9b246a1dc9812068b0e05a9944dd89cb949b035290"},"datadog_checks_base/datadog_checks/base/checks/openmetrics/v2/transformers/__init__.py":{"sha256":"84f667f162ef41faf32d2689c6d15b61802d2b576df084174942cbefdb2b663b"},"datadog_checks_base/datadog_checks/base/checks/openmetrics/v2/transformers/counter.py":{"sha256":"9a6362e041ad2cd605c566b4c88ff7f0d63d681222e35e0cf6945268635b71a2"},"datadog_checks_base/datadog_checks/base/checks/openmetrics/v2/transformers/counter_gauge.py":{"sha256":"a1bd42bc2747afe56b73905295a4f73972f917633a07b3866a15007a4545dc5c"},"datadog_checks_base/datadog_checks/base/checks/openmetrics/v2/transformers/gauge.py":{"sha256":"ff6a19d789bfe7f6fb94e47eb4cc49461b1e17aafa7fd0ec3bee0b6c023288f1"},"datadog_checks_base/datadog_checks/base/checks/openmetrics/v2/transformers/histogram.py":{"sha256":"872b69c3785029d57037ccb991e5ba58672adebe3efb11272431f1c167fa8e52"},"datadog_checks_base/datadog_checks/base/checks/openmetrics/v2/transformers/metadata.py":{"sha256":"069b093750fd272f78bb12deee4a472f5e042dd961530c939a5e51f3d3003aea"},"datadog_checks_base/datadog_checks/base/checks/openmetrics/v2/transformers/rate.py":{"sha256":"7beb75edc142b002a77d7810add521f79c3496c972de2b80d36322cc63ffa1c3"},"datadog_checks_base/datadog_checks/base/checks/openmetrics/v2/transformers/service_check.py":{"sha256":"e0244e3b8da63d241c593dfbe9b4c722fb0e68b0db2ce9883e197ce1c58501b5"},"datadog_checks_base/datadog_checks/base/checks/openmetrics/v2/transformers/summary.py":{"sha256":"d01d5693b79ae07da77ddb0e5fca10122a2804636aca914372304f2a31d5b52e"},"datadog_checks_base/datadog_checks/base/checks/openmetrics/v2/transformers/temporal_percent.py":{"sha256":"c02a8ea971a8550de5c99066fc04e7830a6f21d81c7ce905ff59461397e88625"},"datadog_checks_base/datadog_checks/base/checks/openmetrics/v2/transformers/time_elapsed.py":{"sha256":"c8fb3bd9478e82bd9e40e7610638c507a7add21327c034beaee516388f160db1"},"datadog_checks_base/datadog_checks/base/checks/openmetrics/v2/utils.py":{"sha256":"b6993786d240cff2b0091a85b360938da8c790b0acff64db19d069e75e2b58e4"},"datadog_checks_base/datadog_checks/base/checks/prometheus/__init__.py":{"sha256":"35c57ac8d1d9555c42ac0ac80ece6d4a459fae72f05398b195374d5c57284a30"},"datadog_checks_base/datadog_checks/base/checks/prometheus/base_check.py":{"sha256":"2d4b347b12235a4d520d0901a7191e534fa0888d68cb32e21936898ccd8b8f5d"},"datadog_checks_base/datadog_checks/base/checks/prometheus/mixins.py":{"sha256":"7cd64989b4ea2cba4925c4657677edfad6243554925184431282192411220d4b"},"datadog_checks_base/datadog_checks/base/checks/prometheus/prometheus_base.py":{"sha256":"9f35823bf488a24646a04ee8f01269a254cfa160bbfe471625f90b1c05de057e"},"datadog_checks_base/datadog_checks/base/checks/win/__init__.py":{"sha256":"9083ff7fefc6d7404110ec4ee3e1a7cb29730a8d6439ff5deb291388151a7a4a"},"datadog_checks_base/datadog_checks/base/checks/win/winpdh.py":{"sha256":"142f282601923e049811ccdc3de3b89b7e21cbaf48f08e487c34cfea1865e839"},"datadog_checks_base/datadog_checks/base/checks/win/winpdh_base.py":{"sha256":"72f82d18dfbc3bbffe8ca331d92877919630be4c7f708c1573b1b8b28c404cfd"},"datadog_checks_base/datadog_checks/base/checks/win/winpdh_stub.py":{"sha256":"3397f2064cc0b842afa19ac6f64b506a9c241ffecaf8a388605e55a52f372cc9"},"datadog_checks_base/datadog_checks/base/checks/win/wmi/__init__.py":{"sha256":"6f4f143f3ef047e807872bc2396f83a4fab9c96406d846e1a12248e43f144f37"},"datadog_checks_base/datadog_checks/base/checks/win/wmi/counter_type.py":{"sha256":"efbef65c6b56c8f2577b0caf698ca8c15983cbd0f1678da48bd583e5769bb951"},"datadog_checks_base/datadog_checks/base/checks/win/wmi/sampler.py":{"sha256":"789382e2bc921a9ba22a14766785b9d3f8d375f7773d0c36de5ad33ebc14ae19"},"datadog_checks_base/datadog_checks/base/checks/win/wmi/types.py":{"sha256":"e04f1ed72a69d8ff9e3b180bb11adfb656aeaaf6a9582b956803e872a0abc158"},"datadog_checks_base/datadog_checks/base/checks/windows/__init__.py":{"sha256":"b2e1a32eb8591a9d541a935aa5c56f20fa7ebbc3de68cf24df3a650198f2712a"},"datadog_checks_base/datadog_checks/base/checks/windows/perf_counters/__init__.py":{"sha256":"c4ced6dabda1b7e2b1fe3d22f03bae7bf94433606ffdbab7be0d04b34009e4a1"},"datadog_checks_base/datadog_checks/base/checks/windows/perf_counters/base.py":{"sha256":"bb9c0f99fb2a348fa420a6dd4e74b0f9722a26dc60604d440cd5b8f71c640770"},"datadog_checks_base/datadog_checks/base/checks/windows/perf_counters/connection.py":{"sha256":"124462f2699e89a71bb2ead225be6f014cc523f94091459c9d20bb4ce42c006e"},"datadog_checks_base/datadog_checks/base/checks/windows/perf_counters/constants.py":{"sha256":"fdaa806dab5cbbd790bfd36a6d509a43d7394be1e4922504455e68a82dc39799"},"datadog_checks_base/datadog_checks/base/checks/windows/perf_counters/counter.py":{"sha256":"8498fbe90224b5934558588a21110d2224c4d0d6334e5bd52dac213c2566f49e"},"datadog_checks_base/datadog_checks/base/checks/windows/perf_counters/transform.py":{"sha256":"6d93f17ed0f0d1dd55157e3dca21486be9da18e62529c320a6fb9e491920133f"},"datadog_checks_base/datadog_checks/base/checks/windows/perf_counters/transformers/__init__.py":{"sha256":"a8b142ebeee6817e16846d57125966018eac45ef4a9870efba31fbc9c2555e92"},"datadog_checks_base/datadog_checks/base/checks/windows/perf_counters/transformers/count.py":{"sha256":"8263467bddb648fe101243270ff9dcf30edba0a616fa65b69f9fbabe975c9a37"},"datadog_checks_base/datadog_checks/base/checks/windows/perf_counters/transformers/gauge.py":{"sha256":"73be1f652e85addc433ba64aa2fa75ee1daf85322691a351d8e2deb35af4d681"},"datadog_checks_base/datadog_checks/base/checks/windows/perf_counters/transformers/monotonic_count.py":{"sha256":"479c167c31bd2e471baab21d49ce9dce3470b40729dabe153ee5456aa3a5ce2d"},"datadog_checks_base/datadog_checks/base/checks/windows/perf_counters/transformers/rate.py":{"sha256":"3e4c739755cf6cfb68fb942b882a23361e5684c4e3c03710c2a63f8b6310052f"},"datadog_checks_base/datadog_checks/base/checks/windows/perf_counters/transformers/service_check.py":{"sha256":"c2f74b0d2b871ca2276f35bcb8cf10f764dc454b90975d70b2fb7475266dac70"},"datadog_checks_base/datadog_checks/base/checks/windows/perf_counters/transformers/temporal_percent.py":{"sha256":"2071f661338679e8b63d53790a1f7df200ea620facd4939bbfd6b44e602f3a75"},"datadog_checks_base/datadog_checks/base/checks/windows/perf_counters/transformers/time_elapsed.py":{"sha256":"85633c087612a859c562b35daf5345638eb89cc01514e88df238658594ce6fbf"},"datadog_checks_base/datadog_checks/base/checks/windows/perf_counters/utils.py":{"sha256":"a2ffb8f0936f650e8a3c4671cfa5f42187009776d50059ccd79faf9467eab18d"},"datadog_checks_base/datadog_checks/base/checks/windows/perf_counters/utils_win32pdh_fix.py":{"sha256":"9d64eb01b07d5400ff0da693ad2055d8a0e606f8a9185988d38fb00df9f151af"},"datadog_checks_base/datadog_checks/base/config.py":{"sha256":"a9c000e17f6c5d065177041ef0382219ddbdf34541a7549003477af79b57fed5"},"datadog_checks_base/datadog_checks/base/constants.py":{"sha256":"711d7db40a95cac3056dd056a88199a0720a9359064f2a91b029fd15f1503a7c"},"datadog_checks_base/datadog_checks/base/ddyaml.py":{"sha256":"d86ce592be606c30e1844e7f230e716dd894cd6839b3a70dfa826a7abb92c6ca"},"datadog_checks_base/datadog_checks/base/errors.py":{"sha256":"7cf705691df31e94a4328f1ea6f1779517bd2ab1274bc83827475733f9efcecc"},"datadog_checks_base/datadog_checks/base/log.py":{"sha256":"bba63cedd99be5b471b7bf5d34f5115e0dc7e114e121d75e760655823b46be58"},"datadog_checks_base/datadog_checks/base/stubs/__init__.py":{"sha256":"c2958047dbfb0624db6e64ceea9569b21a9aff3f8f59a613af7df049364bcf77"},"datadog_checks_base/datadog_checks/base/stubs/_util.py":{"sha256":"6431ad41af05ddc1dff3e42f4951cc0780462370bd5600bbb067061af3b46a92"},"datadog_checks_base/datadog_checks/base/stubs/aggregator.py":{"sha256":"c215913b4b31fb5d861b50bdc4060efd9c5a6e8f2b4aea9e240cec4637d7e609"},"datadog_checks_base/datadog_checks/base/stubs/common.py":{"sha256":"646cc5d9d5f2d6e545406746fdbbf3fe930c8942da05ca73adafe4f70a3d7f4e"},"datadog_checks_base/datadog_checks/base/stubs/datadog_agent.py":{"sha256":"7ed89c4dd688e16a2838b6c62b75e95285f8db9664c956c5cd0ed3f53d3365bb"},"datadog_checks_base/datadog_checks/base/stubs/log.py":{"sha256":"03e7969f3639813a535b8d59721f96e4255c97395d96684c4d6faf0cd15d4f5a"},"datadog_checks_base/datadog_checks/base/stubs/similar.py":{"sha256":"cd9d5bab9c0f690fbc70163f1d2fbad76b29151dd4277bf214069756c19c7013"},"datadog_checks_base/datadog_checks/base/stubs/tagging.py":{"sha256":"cf12dd3c2e04a87c46892fc71216da3ac2ffb399d922137c043931d810133aab"},"datadog_checks_base/datadog_checks/base/types.py":{"sha256":"6a76a3652d16d13b31507250c3e24738fd8d49eb82f418ac5d2cbd9804ad9714"},"datadog_checks_base/datadog_checks/base/utils/__init__.py":{"sha256":"4c2d2aee209b36a7188df5396b304da429e2f0b01060e7d8e8500313749910f0"},"datadog_checks_base/datadog_checks/base/utils/agent/__init__.py":{"sha256":"a37696bf2dcf872903fe1ed84f7b3adbc3b45b66291e2b3436542c495d4f234e"},"datadog_checks_base/datadog_checks/base/utils/agent/common.py":{"sha256":"d5746c3c959da666f100a2815471813d49b858776f8aad28bee820b69c7b2d6a"},"datadog_checks_base/datadog_checks/base/utils/agent/debug.py":{"sha256":"cde05b34bb7763f5b1a5ff4e74092595d2f2d6098bd14e9b30398e1d20c63373"},"datadog_checks_base/datadog_checks/base/utils/agent/memory.py":{"sha256":"986d743c51608d608b7ad17a8a10b8eb7550cf55774b849332cfc5f1b00392f6"},"datadog_checks_base/datadog_checks/base/utils/agent/packages.py":{"sha256":"ab88bbece8f9d0a38db114f121b3aa7babf43239debbe4069dd236cfc742d8f9"},"datadog_checks_base/datadog_checks/base/utils/agent/utils.py":{"sha256":"155fe8eab71c53907432b5f299afb8c80aa62a08649734de39fd6785872663ba"},"datadog_checks_base/datadog_checks/base/utils/aws.py":{"sha256":"c3114b5a5545b6fe7f11445db17cc384e45c4e93348c1940a2470c88f575c43f"},"datadog_checks_base/datadog_checks/base/utils/common.py":{"sha256":"b5b95066f10424b78e418babf395ebb80151fd09ab8bdcc0522fe08b43711e0d"},"datadog_checks_base/datadog_checks/base/utils/concurrency/__init__.py":{"sha256":"391b2c1396d766e62e6b95022deb067cfbdcad029c0031da0ec5eb5327c0445d"},"datadog_checks_base/datadog_checks/base/utils/concurrency/limiter.py":{"sha256":"8acd99a541237ec2389c11ad5c6d83d1982fd060f7f3f8b98c940291dccf5bf6"},"datadog_checks_base/datadog_checks/base/utils/constants.py":{"sha256":"4304decb8096074340c66dab703fb03d84641328257a4408ac0cc531a6c46b7f"},"datadog_checks_base/datadog_checks/base/utils/containers.py":{"sha256":"8227d931334393baecb8dcde9132740b832dcb5b26b07f847f6a9b8ebc60b24b"},"datadog_checks_base/datadog_checks/base/utils/date.py":{"sha256":"2499aa3fce0281570527472f02632ef04b4ceaff7ab48112b9c40d9bd78a7847"},"datadog_checks_base/datadog_checks/base/utils/db/__init__.py":{"sha256":"f964bc85274b6a1cf6a9b115c13d38d419f422ef40f4e96ec3a0b8ff36f17e36"},"datadog_checks_base/datadog_checks/base/utils/db/core.py":{"sha256":"b151dd639674771e9433a21667007bd07c62583213344dba4637dfc0b2acf490"},"datadog_checks_base/datadog_checks/base/utils/db/query.py":{"sha256":"4fc4b08d187dfa180df6870147c791e62f20bde87fc95496b869f44548975fbf"},"datadog_checks_base/datadog_checks/base/utils/db/sql.py":{"sha256":"a0f94966a841cf408601aecc10d3dba4e83e39fb878feddbffeaefec981a344b"},"datadog_checks_base/datadog_checks/base/utils/db/sql_commenter.py":{"sha256":"aff79f2ba4c6450c4cffe423fa774425d102938ec9e270858e4572c6ed579af5"},"datadog_checks_base/datadog_checks/base/utils/db/statement_metrics.py":{"sha256":"4b49073e816a099b514d9a1323a3e2f99ba5e88b9ef18de3d7f92c347fb3128f"},"datadog_checks_base/datadog_checks/base/utils/db/timed_cache.py":{"sha256":"6bd2ace4a294bc4c41e863804d74d208b6a656d2c3f75f439fa1e9c1e18ab45c"},"datadog_checks_base/datadog_checks/base/utils/db/transform.py":{"sha256":"90d762059568666d6835135621ab6b01f80e4636e444a86cd40816fda4028f72"},"datadog_checks_base/datadog_checks/base/utils/db/types.py":{"sha256":"38b5f63aae7c2503c5043e28a94a422e400fee8bd118dfe2d6f164d44d0df0b8"},"datadog_checks_base/datadog_checks/base/utils/db/utils.py":{"sha256":"b44123564116b2b1d0bfd8c243c7ebb8ecc8a0389afd35b3236644daa23f169e"},"datadog_checks_base/datadog_checks/base/utils/diagnose.py":{"sha256":"78b31ed2d212a64cd2df2c5547cdc81f1a2b4097c74ff5e2e82ab8cf344dc552"},"datadog_checks_base/datadog_checks/base/utils/discovery/__init__.py":{"sha256":"7c896046fefe58933a21fa392211462b829c0caac24865ff7eb5e0899f73e8c9"},"datadog_checks_base/datadog_checks/base/utils/discovery/cache.py":{"sha256":"7fd2f703b619a59fa6a59a452308ec6b969bf5c64c1a4a9d7ad76bf44a5a95b2"},"datadog_checks_base/datadog_checks/base/utils/discovery/discovery.py":{"sha256":"9bcc3d55162eefc79dd202af56ee1c84dad1752302aae24edd85359b378e734d"},"datadog_checks_base/datadog_checks/base/utils/discovery/filter.py":{"sha256":"459f13ddb2cfe2e84af0a5f01aa76860c254f4edc402f3924e119ae702d0311f"},"datadog_checks_base/datadog_checks/base/utils/functions.py":{"sha256":"8869726f147a68f3c494dc4d6f610b3b36e4df6f23f4e541031ade749c5d091c"},"datadog_checks_base/datadog_checks/base/utils/headers.py":{"sha256":"2f23e5e9cc119258e1fe0b7772f282a081f7593a8a09bc733ea84644c28e699a"},"datadog_checks_base/datadog_checks/base/utils/http.py":{"sha256":"5979f14d34e6a55fc2a2ae53ee528b2e08b9463955ce3b76e164f963eab88525"},"datadog_checks_base/datadog_checks/base/utils/limiter.py":{"sha256":"6114eb3c25f54b912d1cb55c3fff8611fcd1a2db3f2d3732d5ffee1d956cf748"},"datadog_checks_base/datadog_checks/base/utils/metadata/__init__.py":{"sha256":"6d36a6f7a190f43be4ea287c70aabc5b16b69640e48feed3b89de85875d432cb"},"datadog_checks_base/datadog_checks/base/utils/metadata/constants.py":{"sha256":"5c77cfc2f40c6f2344d8562607fed7c968862343761b17415dbb572f87839e27"},"datadog_checks_base/datadog_checks/base/utils/metadata/core.py":{"sha256":"64e5a3e1bca0d96533d7e092adff752c8ec2ad2349dcc87a510c10dd0ff03d7d"},"datadog_checks_base/datadog_checks/base/utils/metadata/utils.py":{"sha256":"4c2876f1c9b1434dcc413b9e3af4274f5ad0b604c7dadf30fde8e90901dcaa9e"},"datadog_checks_base/datadog_checks/base/utils/metadata/version.py":{"sha256":"7257bc2c7c2a72ee364ea14a24625d16d1c098e7a2b423a2ce34cd43606cc534"},"datadog_checks_base/datadog_checks/base/utils/models/__init__.py":{"sha256":"b2e1a32eb8591a9d541a935aa5c56f20fa7ebbc3de68cf24df3a650198f2712a"},"datadog_checks_base/datadog_checks/base/utils/models/types.py":{"sha256":"c9b504a7dcffac396bdbea089158c4581fa6440bd4c28103a4051c6504c4208c"},"datadog_checks_base/datadog_checks/base/utils/models/validation/__init__.py":{"sha256":"699557dfc5b5a642c793b9281e02b9267d8f3824f940a28f1b35bfc3d2e082da"},"datadog_checks_base/datadog_checks/base/utils/models/validation/core.py":{"sha256":"dab25d1f87af4729ec0530dc0b04ed788c0c09906f9e674113a736f2c8c3e5a0"},"datadog_checks_base/datadog_checks/base/utils/models/validation/utils.py":{"sha256":"748ad183c8795ee93e403b08c89285f68b6a45fc34aeeebd1f67c548dcc8b0e8"},"datadog_checks_base/datadog_checks/base/utils/network.py":{"sha256":"970452f4248a9699bed1d5be713e7b2b65fe6f026f4bcbf6afa4fb4ad2bfd6c7"},"datadog_checks_base/datadog_checks/base/utils/platform.py":{"sha256":"c16f1fe972a8e091f1beed6c3740e92c39988d2fdc0aef06a0abdf4e3223e323"},"datadog_checks_base/datadog_checks/base/utils/prometheus/__init__.py":{"sha256":"f794783ecff74f6713b846470f28eaaa841ed20c0d1681bcd18186135e2c150f"},"datadog_checks_base/datadog_checks/base/utils/prometheus/functions.py":{"sha256":"c674f283d8937acffcde65a57351acf05150d6349163e94b3e8e530801626f2b"},"datadog_checks_base/datadog_checks/base/utils/prometheus/metrics_pb2.py":{"sha256":"eab96d57e2144ab9d603d62b031c0c1a198afb41ea22fe11f1b60b0596d65fdf"},"datadog_checks_base/datadog_checks/base/utils/replay/__init__.py":{"sha256":"391b2c1396d766e62e6b95022deb067cfbdcad029c0031da0ec5eb5327c0445d"},"datadog_checks_base/datadog_checks/base/utils/replay/constants.py":{"sha256":"7b10c6b0380b23bbb5196bde7a55dd8335894e3c47a18266413ee8ef0d4509e3"},"datadog_checks_base/datadog_checks/base/utils/replay/execute.py":{"sha256":"9046d9b996440a3087c9447969ec52ce4382e7c74e5d0c95fdded3d8a2c4324a"},"datadog_checks_base/datadog_checks/base/utils/replay/redirect.py":{"sha256":"f21e13c07ed95637a211ddaacb1f4a68752755137bfcd1a6535745890c252697"},"datadog_checks_base/datadog_checks/base/utils/secrets.py":{"sha256":"4e3e4c04ea321975c3581dc7afeed4283cb9195d4d6499053d8e13e2ff4f1c78"},"datadog_checks_base/datadog_checks/base/utils/serialization.py":{"sha256":"0eafc28756b17e888cf64b65f6e5a02a6c242aef85a65575a3b8a79388ca596c"},"datadog_checks_base/datadog_checks/base/utils/subprocess_output.py":{"sha256":"58dc33d63b9c6e2193b43a0dc735c34ad2a080c82aa657f67959b1b06259b020"},"datadog_checks_base/datadog_checks/base/utils/tagging.py":{"sha256":"004504188c498cdbe8388110405922b7c653d8ec91c62ca6d45cc21227080acb"},"datadog_checks_base/datadog_checks/base/utils/tailfile.py":{"sha256":"c7fa4ce6982655a5b87890704ba19764a3aa89fa66a9faf01ce537816b6162d3"},"datadog_checks_base/datadog_checks/base/utils/time.py":{"sha256":"9caeb78a0273d313748990aea3dd09a6ca47119cc52671bcca42428186a9a41c"},"datadog_checks_base/datadog_checks/base/utils/timeout.py":{"sha256":"78e059a1f14dfa13aee7125e30e17769cfe87dccbd118ebe92f981bcfe101058"},"datadog_checks_base/datadog_checks/base/utils/tls.py":{"sha256":"84625ba574e57cf11bdc7845783c96fb8409b78fff535e3a99cf7bc1823e4e82"},"datadog_checks_base/datadog_checks/base/utils/tracing.py":{"sha256":"1a38b277fa0ca6fd8566969accf31129c0565a9fe2a1eaf9e64b6b76810da316"},"datadog_checks_base/datadog_checks/base/utils/tracking.py":{"sha256":"158228baabb7281f89b31831335897e48a6ffdc5a1b3ccd03933784b15ce909e"},"datadog_checks_base/datadog_checks/checks/__init__.py":{"sha256":"76381faa72acfce6863031501a53b955d2d047150023e5f828424b2e53e95483"},"datadog_checks_base/datadog_checks/checks/base.py":{"sha256":"df061b86e80d0375a5aedffe104a4198949297c9472ae52dad13fe9d5d8a05f4"},"datadog_checks_base/datadog_checks/checks/libs/__init__.py":{"sha256":"2300c3103843a8f3d4d63e0fcaf78691dbb508cbfd91b7de2bdd0802f981c777"},"datadog_checks_base/datadog_checks/checks/libs/thread_pool.py":{"sha256":"747147080f5d03c2742e233e4bd635d4c58abff8cf5c803fe5881f776c1603b4"},"datadog_checks_base/datadog_checks/checks/libs/timer.py":{"sha256":"a35b1970916e9035ae71d185b6506bbd2b798e26ef1b7b91fc3b30788b19dd49"},"datadog_checks_base/datadog_checks/checks/libs/vmware/__init__.py":{"sha256":"e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855"},"datadog_checks_base/datadog_checks/checks/libs/vmware/all_metrics.py":{"sha256":"849ca20bfd3b86aa8c8eeb88732b4026fd8a0e74a0a96ab6cbe269270b71a8c3"},"datadog_checks_base/datadog_checks/checks/libs/vmware/basic_metrics.py":{"sha256":"683af414d39b8d77d66fd97146999de4bd7a7be9ab934ed9224867a0b74c099f"},"datadog_checks_base/datadog_checks/checks/libs/wmi/__init__.py":{"sha256":"2300c3103843a8f3d4d63e0fcaf78691dbb508cbfd91b7de2bdd0802f981c777"},"datadog_checks_base/datadog_checks/checks/libs/wmi/sampler.py":{"sha256":"55310a8bffb4b55af9552268e627c83c284db55c99c47f754fa05e5bb33ccd3b"},"datadog_checks_base/datadog_checks/checks/network.py":{"sha256":"e5b7d11f19ac5286897259c2ac1e8209ee7bee94df199d5155e1b174cce6afbc"},"datadog_checks_base/datadog_checks/checks/network_checks.py":{"sha256":"cdb77741ebeaa7b6a13616873b5725009b7566a77ceef013fe5dd1c76cbdd81b"},"datadog_checks_base/datadog_checks/checks/openmetrics/__init__.py":{"sha256":"96b910b1c359a0b37a0a8753b6e1e50c803ae35bf4f1e7a31418678cf16792c0"},"datadog_checks_base/datadog_checks/checks/openmetrics/base_check.py":{"sha256":"a9ccd2133baa1058f744e494e2bb43cb5221fbbb41b6c3b8a0553ead7f6a851f"},"datadog_checks_base/datadog_checks/checks/openmetrics/mixins.py":{"sha256":"b5b9cf9d77c8dda6d752a4d37d303405fa1935ac071b17a4e8b055160b7d2a28"},"datadog_checks_base/datadog_checks/checks/prometheus/__init__.py":{"sha256":"a12ac852908b6eb9f3b594e893c41a38951a9a1e030b6286c20e76aec9965c8b"},"datadog_checks_base/datadog_checks/checks/prometheus/base_check.py":{"sha256":"882984f28171bfde4a2c34c25f609b5a758917f0c9b9e00b2ba9a0d489ba1e19"},"datadog_checks_base/datadog_checks/checks/prometheus/mixins.py":{"sha256":"aa652ec3963f90c1e7c552d35243d3863f737fa9de5e45131ca55758c5e00de5"},"datadog_checks_base/datadog_checks/checks/prometheus/prometheus_base.py":{"sha256":"434576db21c019a7366f5350b8e305c0790509aadd9cbd980f6c3ac22b87874f"},"datadog_checks_base/datadog_checks/checks/prometheus_check/__init__.py":{"sha256":"9b5434e894e03018e342ee726f635de62122bf0e1d8f59d3f0109f89a95d890d"},"datadog_checks_base/datadog_checks/checks/win/__init__.py":{"sha256":"4441d475ac7181f8e7edf1037da1310d776c647883d0362a5acfb88e44e4d45e"},"datadog_checks_base/datadog_checks/checks/win/winpdh.py":{"sha256":"864f9f437a351bb82a2962e15a1198cc6271f9cc25c3c73522402954fa42ef9e"},"datadog_checks_base/datadog_checks/checks/win/winpdh_base.py":{"sha256":"269e7c39a6562687aa6ad022b1b88c85b5f6be403c11a8412775b3c55715196e"},"datadog_checks_base/datadog_checks/checks/win/winpdh_stub.py":{"sha256":"a1895061a3be05415b331c1b17d2c02b1f912f36ed238948c46d7af68025d7ed"},"datadog_checks_base/datadog_checks/checks/win/wmi/__init__.py":{"sha256":"51008a1bf148ec1c7bc0bb8f6ad3be19ba950e6c13b5cb84a2fd2684fb66327e"},"datadog_checks_base/datadog_checks/checks/win/wmi/counter_type.py":{"sha256":"01ac5dc9d1518c74c4b621a1da0b901da5ffc954abcf81972f8e478abdb098a7"},"datadog_checks_base/datadog_checks/checks/win/wmi/sampler.py":{"sha256":"bbfeea3683c40059f5e1ff52fc459e82200727041a187328d65f58b92b445cdd"},"datadog_checks_base/datadog_checks/checks/winwmi_check.py":{"sha256":"907ea75f4055e675b732171687ce6148d6484d067b36726b942b4e9c5e1d747b"},"datadog_checks_base/datadog_checks/config.py":{"sha256":"3eb01719d94b9e857654c41f7ff9e849a489274c20e010221a7c3b8c240b4a29"},"datadog_checks_base/datadog_checks/errors.py":{"sha256":"785c269eb5fe6dab4881bbbe889cac7aa00f34eff8ae4d54b9a28af3d7af2e18"},"datadog_checks_base/datadog_checks/log.py":{"sha256":"a2bbce80c286344b2a4932da942010a563fea2ea2ba46d40ec69fe8f69910fcd"},"datadog_checks_base/datadog_checks/py.typed":{"sha256":"95aebb28195b8d737effe0df18d71d39c8d8ba6569286fd3930fbc9f9767181e"},"datadog_checks_base/datadog_checks/stubs/__init__.py":{"sha256":"331c49e4d1a353ded64777893a96906746f386f03cf32b30135aa260665377ec"},"datadog_checks_base/datadog_checks/stubs/_util.py":{"sha256":"9989e9f7fbeca50c1359c06188f60c309dc421c40f4a9a407147d18c42c180b5"},"datadog_checks_base/datadog_checks/stubs/aggregator.py":{"sha256":"ea230225c9c41fbb1885a38741cccf1809034d9b1fd696eb6efde1de6bfba762"},"datadog_checks_base/datadog_checks/stubs/datadog_agent.py":{"sha256":"b167bb8f650441e0f3762aa3e0ffe73e1921157f34ff90abe47723db963ec6bf"},"datadog_checks_base/datadog_checks/utils/__init__.py":{"sha256":"e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855"},"datadog_checks_base/datadog_checks/utils/common.py":{"sha256":"470a0220a572f2f6df6b61353fb63b82447d569274d15b6e3f5f0c12b5e7f6b9"},"datadog_checks_base/datadog_checks/utils/containers.py":{"sha256":"ca56286c527f7d5f940b0baf5a66a4eba6ad2df036759f6000b0985835c63a29"},"datadog_checks_base/datadog_checks/utils/headers.py":{"sha256":"8f92abe525d4947ba4dd41870d0367f4cda884df95e57d9fce62718f0e8fe8da"},"datadog_checks_base/datadog_checks/utils/limiter.py":{"sha256":"1aad3f848119c3aaa92394962ffbb331a22991bcef173b242347a54cffef22c1"},"datadog_checks_base/datadog_checks/utils/platform.py":{"sha256":"76ca8472c0bf63b866e28aea3bc449e4de63905a41ee9c8662f09e5eb6fef571"},"datadog_checks_base/datadog_checks/utils/prometheus/__init__.py":{"sha256":"f16c179ccf60d6c7d2e76ebb4180895ff85df0c665e6446006043f4b374d757b"},"datadog_checks_base/datadog_checks/utils/prometheus/functions.py":{"sha256":"e2f5ac4c62e0ba31f0758668d2d940424a833c7a1fa9424cde4f5e22d26a1114"},"datadog_checks_base/datadog_checks/utils/prometheus/metrics_pb2.py":{"sha256":"c60dd47541dee138de478b35dcb50a6765355954ade94eb38ec5511ba957e9d7"},"datadog_checks_base/datadog_checks/utils/proxy.py":{"sha256":"723edeaf00a1ee813aa4a6467fc34c3ae419f354c6b6172de678e39b864a8bd7"},"datadog_checks_base/datadog_checks/utils/subprocess_output.py":{"sha256":"80dea34445564d9eee264533455310e69c72b4d640e369ea8c97a365717a64c2"},"datadog_checks_base/datadog_checks/utils/tailfile.py":{"sha256":"6c4aa2725fac2e0cb0f660f545d5a8b3c1954a50a49f1e58c86ddf5cc068c137"},"datadog_checks_base/datadog_checks/utils/timeout.py":{"sha256":"42e848b0a6686bf335dfe8040b07acbb9219c12317cfdf564f0a9a974262c913"},"datadog_checks_base/datadog_checks/utils/tracing.py":{"sha256":"1d06d06a428cf8bc3be4c0e422d6982588a94ba60edb867fca60d5c43b31e435"},"datadog_checks_base/pyproject.toml":{"sha256":"e59f9d35788ebe2c5e3e0543fa284e207522589690527ec915723fdf4e77d2f1"}}}} \ No newline at end of file diff --git a/datadog_checks_base/CHANGELOG.md b/datadog_checks_base/CHANGELOG.md index e15fc9a16816b..01544ab6ff84c 100644 --- a/datadog_checks_base/CHANGELOG.md +++ b/datadog_checks_base/CHANGELOG.md @@ -2,6 +2,16 @@ +## 37.0.0 / 2024-09-19 + +***Removed***: + +* Remove support for Python 2. ([#18580](https://github.com/DataDog/integrations-core/pull/18580)) + +***Added***: + +* Bump the python version from 3.11 to 3.12 ([#18207](https://github.com/DataDog/integrations-core/pull/18207)) + ## 36.15.0 / 2024-09-05 ***Added***: diff --git a/datadog_checks_base/changelog.d/18207.added b/datadog_checks_base/changelog.d/18207.added deleted file mode 100644 index 624cd9836c9fe..0000000000000 --- a/datadog_checks_base/changelog.d/18207.added +++ /dev/null @@ -1 +0,0 @@ -Bump the python version from 3.11 to 3.12 \ No newline at end of file diff --git a/datadog_checks_base/changelog.d/18580.removed b/datadog_checks_base/changelog.d/18580.removed deleted file mode 100644 index 1029637dfcead..0000000000000 --- a/datadog_checks_base/changelog.d/18580.removed +++ /dev/null @@ -1 +0,0 @@ -Remove support for Python 2. diff --git a/datadog_checks_base/datadog_checks/base/__about__.py b/datadog_checks_base/datadog_checks/base/__about__.py index ea6b270f57e7c..b46f3b3920f39 100644 --- a/datadog_checks_base/datadog_checks/base/__about__.py +++ b/datadog_checks_base/datadog_checks/base/__about__.py @@ -1,4 +1,4 @@ # (C) Datadog, Inc. 2018-present # All rights reserved # Licensed under a 3-clause BSD style license (see LICENSE) -__version__ = "36.15.0" +__version__ = "37.0.0" diff --git a/requirements-agent-release.txt b/requirements-agent-release.txt index ca87469758a64..1f0d38e7fbd76 100644 --- a/requirements-agent-release.txt +++ b/requirements-agent-release.txt @@ -26,7 +26,7 @@ datadog-cassandra==1.18.0 datadog-ceph==2.10.0; sys_platform != 'win32' datadog-cert-manager==4.1.2 datadog-checkpoint-quantum-firewall==1.0.0 -datadog-checks-base==36.15.0 +datadog-checks-base==37.0.0 datadog-checks-dependency-provider==1.4.0 datadog-checks-downloader==4.7.0 datadog-cilium==3.6.0 From 9a082bff746664a084f4eeaa90e2cc808fd9c8e3 Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?Jhonn=20W=2E=20Fraz=C3=A3o?= Date: Fri, 20 Sep 2024 14:31:23 -0300 Subject: [PATCH 23/34] feat: add apiserver_admission_webhook_request_total metric (#17690) Co-authored-by: Steven Blumenthal --- kube_apiserver_metrics/changelog.d/17690.added | 1 + .../kube_apiserver_metrics/kube_apiserver_metrics.py | 5 +++++ kube_apiserver_metrics/metadata.csv | 2 ++ kube_apiserver_metrics/tests/fixtures/metrics_1.23.0.txt | 7 ++++++- kube_apiserver_metrics/tests/fixtures/metrics_1.26.0.txt | 5 +++++ kube_apiserver_metrics/tests/fixtures/metrics_1.27.0.txt | 5 +++++ kube_apiserver_metrics/tests/fixtures/metrics_1.28.0.txt | 5 +++++ .../tests/test_kube_apiserver_metrics_1_23.py | 2 ++ .../tests/test_kube_apiserver_metrics_1_24.py | 2 ++ .../tests/test_kube_apiserver_metrics_1_25.py | 2 ++ .../tests/test_kube_apiserver_metrics_1_26.py | 2 ++ .../tests/test_kube_apiserver_metrics_1_27.py | 2 ++ .../tests/test_kube_apiserver_metrics_1_28.py | 2 ++ 13 files changed, 41 insertions(+), 1 deletion(-) create mode 100644 kube_apiserver_metrics/changelog.d/17690.added diff --git a/kube_apiserver_metrics/changelog.d/17690.added b/kube_apiserver_metrics/changelog.d/17690.added new file mode 100644 index 0000000000000..8db50c4db624f --- /dev/null +++ b/kube_apiserver_metrics/changelog.d/17690.added @@ -0,0 +1 @@ +Add the apiserver_admission_webhook_request_total diff --git a/kube_apiserver_metrics/datadog_checks/kube_apiserver_metrics/kube_apiserver_metrics.py b/kube_apiserver_metrics/datadog_checks/kube_apiserver_metrics/kube_apiserver_metrics.py index d07d8b168d411..eeec2cf8f0323 100644 --- a/kube_apiserver_metrics/datadog_checks/kube_apiserver_metrics/kube_apiserver_metrics.py +++ b/kube_apiserver_metrics/datadog_checks/kube_apiserver_metrics/kube_apiserver_metrics.py @@ -117,6 +117,8 @@ def __init__(self, name, init_config, instances=None): # For Kubernetes >= 1.24 # https://github.com/kubernetes/kubernetes/pull/107171 'apiserver_admission_webhook_fail_open_count': self.apiserver_admission_webhook_fail_open_count, + # https://github.com/kubernetes/kubernetes/pull/103162 + 'apiserver_admission_webhook_request_total': self.apiserver_admission_webhook_request_total, } self.kube_apiserver_config = None @@ -242,3 +244,6 @@ def apiserver_request_terminations_total(self, metric, scraper_config): def apiserver_admission_webhook_fail_open_count(self, metric, scraper_config): self.submit_metric('.apiserver_admission_webhook_fail_open_count', metric, scraper_config) + + def apiserver_admission_webhook_request_total(self, metric, scraper_config): + self.submit_metric('.apiserver_admission_webhook_request_total', metric, scraper_config) diff --git a/kube_apiserver_metrics/metadata.csv b/kube_apiserver_metrics/metadata.csv index c83454a9e5287..e76c1eefff2b2 100644 --- a/kube_apiserver_metrics/metadata.csv +++ b/kube_apiserver_metrics/metadata.csv @@ -12,6 +12,8 @@ kube_apiserver.admission_webhook_admission_latencies_seconds.sum,gauge,,second,, kube_apiserver.aggregator_unavailable_apiservice,gauge,,,,Gauge of APIServices which are marked as unavailable broken down by APIService name (alpha; Kubernetes 1.14+),0,kubernetes_api_server_metrics,unavailable apiservice, kube_apiserver.apiserver_admission_webhook_fail_open_count,gauge,,,,"Admission webhook fail open count, identified by name and broken out for each admission type (validating or mutating).",0,kubernetes_api_server_metrics,webhook fail open count, kube_apiserver.apiserver_admission_webhook_fail_open_count.count,count,,,,"Admission webhook fail open count, identified by name and broken out for each admission type (validating or mutating).",0,kubernetes_api_server_metrics,webhook fail open count, +kube_apiserver.apiserver_admission_webhook_request_total,gauge,,,,"Admission webhook request total, identified by name and broken out for each admission type (alpha; Kubernetes 1.23+)",0,kubernetes_api_server_metrics,webhook request total, +kube_apiserver.apiserver_admission_webhook_request_total.count,count,,,,"Admission webhook request total, identified by name and broken out for each admission type (alpha; Kubernetes 1.23+)",0,kubernetes_api_server_metrics,webhook request total, kube_apiserver.apiserver_dropped_requests_total,gauge,,request,,The accumulated number of requests dropped with 'Try again later' response,1,kubernetes_api_server_metrics,accumulated number of dropped requests, kube_apiserver.apiserver_dropped_requests_total.count,count,,request,,The monotonic count of requests dropped with 'Try again later' response,0,kubernetes_api_server_metrics,count of dropped requests, kube_apiserver.apiserver_request_count,gauge,,request,,The accumulated number of apiserver requests broken out for each verb API resource client and HTTP response contentType and code (deprecated in Kubernetes 1.15),1,kubernetes_api_server_metrics,accumulated number of requests, diff --git a/kube_apiserver_metrics/tests/fixtures/metrics_1.23.0.txt b/kube_apiserver_metrics/tests/fixtures/metrics_1.23.0.txt index 13f2f787e56f3..1687605fc4787 100644 --- a/kube_apiserver_metrics/tests/fixtures/metrics_1.23.0.txt +++ b/kube_apiserver_metrics/tests/fixtures/metrics_1.23.0.txt @@ -554,6 +554,11 @@ apiserver_admission_step_admission_duration_seconds_summary{operation="UPDATE",r apiserver_admission_step_admission_duration_seconds_summary{operation="UPDATE",rejected="false",type="validate",quantile="0.99"} 0.0005589 apiserver_admission_step_admission_duration_seconds_summary_sum{operation="UPDATE",rejected="false",type="validate"} 1.734694500000006 apiserver_admission_step_admission_duration_seconds_summary_count{operation="UPDATE",rejected="false",type="validate"} 7427 +# HELP apiserver_admission_webhook_request_total [ALPHA] Admission webhook request total, identified by name and broken out for each admission type (validating or mutating) and operation. Additional labels specify whether the request was rejected or not and an HTTP status code. Codes greater than 600 are truncated to 600, to keep the metrics cardinality bounded. +# TYPE apiserver_admission_webhook_request_total counter +apiserver_admission_webhook_request_total{code="200",name="datadog.webhook.auto.instrumentation",operation="CREATE",rejected="false",type="admit"} 17 +apiserver_admission_webhook_request_total{code="200",name="datadog.webhook.config",operation="CREATE",rejected="false",type="admit"} 18 +apiserver_admission_webhook_request_total{code="200",name="datadog.webhook.tags",operation="CREATE",rejected="false",type="admit"} 17 # HELP apiserver_audit_event_total [ALPHA] Counter of audit events generated and sent to the audit backend. # TYPE apiserver_audit_event_total counter apiserver_audit_event_total 0 @@ -19216,4 +19221,4 @@ workqueue_work_duration_seconds_bucket{name="priority_and_fairness_config_queue" workqueue_work_duration_seconds_bucket{name="priority_and_fairness_config_queue",le="10"} 33 workqueue_work_duration_seconds_bucket{name="priority_and_fairness_config_queue",le="+Inf"} 33 workqueue_work_duration_seconds_sum{name="priority_and_fairness_config_queue"} 0.1424415 -workqueue_work_duration_seconds_count{name="priority_and_fairness_config_queue"} 33 \ No newline at end of file +workqueue_work_duration_seconds_count{name="priority_and_fairness_config_queue"} 33 diff --git a/kube_apiserver_metrics/tests/fixtures/metrics_1.26.0.txt b/kube_apiserver_metrics/tests/fixtures/metrics_1.26.0.txt index e3f44490a4a63..2fe3c01f508b1 100644 --- a/kube_apiserver_metrics/tests/fixtures/metrics_1.26.0.txt +++ b/kube_apiserver_metrics/tests/fixtures/metrics_1.26.0.txt @@ -583,6 +583,11 @@ apiserver_admission_step_admission_duration_seconds_summary{operation="UPDATE",r apiserver_admission_step_admission_duration_seconds_summary{operation="UPDATE",rejected="false",type="validate",quantile="0.99"} 0.000180532 apiserver_admission_step_admission_duration_seconds_summary_sum{operation="UPDATE",rejected="false",type="validate"} 0.03223122800000003 apiserver_admission_step_admission_duration_seconds_summary_count{operation="UPDATE",rejected="false",type="validate"} 686 +# HELP apiserver_admission_webhook_request_total [ALPHA] Admission webhook request total, identified by name and broken out for each admission type (validating or mutating) and operation. Additional labels specify whether the request was rejected or not and an HTTP status code. Codes greater than 600 are truncated to 600, to keep the metrics cardinality bounded. +# TYPE apiserver_admission_webhook_request_total counter +apiserver_admission_webhook_request_total{code="200",name="datadog.webhook.auto.instrumentation",operation="CREATE",rejected="false",type="admit"} 17 +apiserver_admission_webhook_request_total{code="200",name="datadog.webhook.config",operation="CREATE",rejected="false",type="admit"} 18 +apiserver_admission_webhook_request_total{code="200",name="datadog.webhook.tags",operation="CREATE",rejected="false",type="admit"} 17 # HELP apiserver_audit_event_total [ALPHA] Counter of audit events generated and sent to the audit backend. # TYPE apiserver_audit_event_total counter apiserver_audit_event_total 0 diff --git a/kube_apiserver_metrics/tests/fixtures/metrics_1.27.0.txt b/kube_apiserver_metrics/tests/fixtures/metrics_1.27.0.txt index 56da2969b8ee6..86be1c0206fd5 100644 --- a/kube_apiserver_metrics/tests/fixtures/metrics_1.27.0.txt +++ b/kube_apiserver_metrics/tests/fixtures/metrics_1.27.0.txt @@ -635,6 +635,11 @@ apiserver_admission_step_admission_duration_seconds_summary{operation="UPDATE",r apiserver_admission_step_admission_duration_seconds_summary{operation="UPDATE",rejected="false",type="validate",quantile="0.99"} 0.000158667 apiserver_admission_step_admission_duration_seconds_summary_sum{operation="UPDATE",rejected="false",type="validate"} 0.38836563199999974 apiserver_admission_step_admission_duration_seconds_summary_count{operation="UPDATE",rejected="false",type="validate"} 19075 +# HELP apiserver_admission_webhook_request_total [ALPHA] Admission webhook request total, identified by name and broken out for each admission type (validating or mutating) and operation. Additional labels specify whether the request was rejected or not and an HTTP status code. Codes greater than 600 are truncated to 600, to keep the metrics cardinality bounded. +# TYPE apiserver_admission_webhook_request_total counter +apiserver_admission_webhook_request_total{code="200",name="datadog.webhook.auto.instrumentation",operation="CREATE",rejected="false",type="admit"} 17 +apiserver_admission_webhook_request_total{code="200",name="datadog.webhook.config",operation="CREATE",rejected="false",type="admit"} 18 +apiserver_admission_webhook_request_total{code="200",name="datadog.webhook.tags",operation="CREATE",rejected="false",type="admit"} 17 # HELP apiserver_audit_event_total [ALPHA] Counter of audit events generated and sent to the audit backend. # TYPE apiserver_audit_event_total counter apiserver_audit_event_total 0 diff --git a/kube_apiserver_metrics/tests/fixtures/metrics_1.28.0.txt b/kube_apiserver_metrics/tests/fixtures/metrics_1.28.0.txt index cfda122f0bca6..0a9a6e045e33e 100644 --- a/kube_apiserver_metrics/tests/fixtures/metrics_1.28.0.txt +++ b/kube_apiserver_metrics/tests/fixtures/metrics_1.28.0.txt @@ -616,6 +616,11 @@ apiserver_admission_step_admission_duration_seconds_summary{operation="UPDATE",r apiserver_admission_step_admission_duration_seconds_summary{operation="UPDATE",rejected="false",type="validate",quantile="0.99"} 0.000148958 apiserver_admission_step_admission_duration_seconds_summary_sum{operation="UPDATE",rejected="false",type="validate"} 0.43113972400000006 apiserver_admission_step_admission_duration_seconds_summary_count{operation="UPDATE",rejected="false",type="validate"} 19367 +# HELP apiserver_admission_webhook_request_total [ALPHA] Admission webhook request total, identified by name and broken out for each admission type (validating or mutating) and operation. Additional labels specify whether the request was rejected or not and an HTTP status code. Codes greater than 600 are truncated to 600, to keep the metrics cardinality bounded. +# TYPE apiserver_admission_webhook_request_total counter +apiserver_admission_webhook_request_total{code="200",name="datadog.webhook.auto.instrumentation",operation="CREATE",rejected="false",type="admit"} 17 +apiserver_admission_webhook_request_total{code="200",name="datadog.webhook.config",operation="CREATE",rejected="false",type="admit"} 18 +apiserver_admission_webhook_request_total{code="200",name="datadog.webhook.tags",operation="CREATE",rejected="false",type="admit"} 17 # HELP apiserver_audit_event_total [ALPHA] Counter of audit events generated and sent to the audit backend. # TYPE apiserver_audit_event_total counter apiserver_audit_event_total 0 diff --git a/kube_apiserver_metrics/tests/test_kube_apiserver_metrics_1_23.py b/kube_apiserver_metrics/tests/test_kube_apiserver_metrics_1_23.py index 9e67f6228da68..04ffa9147c515 100644 --- a/kube_apiserver_metrics/tests/test_kube_apiserver_metrics_1_23.py +++ b/kube_apiserver_metrics/tests/test_kube_apiserver_metrics_1_23.py @@ -38,6 +38,7 @@ class TestKubeAPIServerMetrics: 'grpc_client_started_total', 'rest_client_request_latency_seconds.sum', 'rest_client_request_latency_seconds.count', + 'apiserver_admission_webhook_request_total', 'admission_step_admission_latencies_seconds.sum', 'admission_step_admission_latencies_seconds.count', 'admission_step_admission_latencies_seconds_summary.sum', @@ -74,6 +75,7 @@ class TestKubeAPIServerMetrics: 'audit_event.count', 'rest_client_requests_total.count', 'authenticated_user_requests.count', + 'apiserver_admission_webhook_request_total.count', 'apiserver_request_total.count', 'apiserver_request_terminations_total.count', ] diff --git a/kube_apiserver_metrics/tests/test_kube_apiserver_metrics_1_24.py b/kube_apiserver_metrics/tests/test_kube_apiserver_metrics_1_24.py index e57c7ebb9e5ff..e08ee1346ee7d 100644 --- a/kube_apiserver_metrics/tests/test_kube_apiserver_metrics_1_24.py +++ b/kube_apiserver_metrics/tests/test_kube_apiserver_metrics_1_24.py @@ -63,6 +63,7 @@ class TestKubeAPIServerMetrics: 'storage_list_returned_objects_total', 'requested_deprecated_apis', 'apiserver_admission_webhook_fail_open_count', + 'apiserver_admission_webhook_request_total', 'admission_webhook_admission_latencies_seconds.sum', 'admission_webhook_admission_latencies_seconds.count', 'aggregator_unavailable_apiservice', @@ -79,6 +80,7 @@ class TestKubeAPIServerMetrics: 'apiserver_request_total.count', 'apiserver_request_terminations_total.count', 'apiserver_admission_webhook_fail_open_count.count', + 'apiserver_admission_webhook_request_total.count', ] def test_check(self, dd_run_check, aggregator, mock_http_response): diff --git a/kube_apiserver_metrics/tests/test_kube_apiserver_metrics_1_25.py b/kube_apiserver_metrics/tests/test_kube_apiserver_metrics_1_25.py index c935071a38abf..86882b5c64fc7 100644 --- a/kube_apiserver_metrics/tests/test_kube_apiserver_metrics_1_25.py +++ b/kube_apiserver_metrics/tests/test_kube_apiserver_metrics_1_25.py @@ -82,6 +82,7 @@ class TestKubeAPIServerMetrics: NAMESPACE + '.storage_list_returned_objects_total', NAMESPACE + '.requested_deprecated_apis', NAMESPACE + '.apiserver_admission_webhook_fail_open_count', + NAMESPACE + '.apiserver_admission_webhook_request_total', NAMESPACE + '.admission_webhook_admission_latencies_seconds.sum', NAMESPACE + '.admission_webhook_admission_latencies_seconds.count', NAMESPACE + '.kubernetes_feature_enabled', @@ -100,6 +101,7 @@ class TestKubeAPIServerMetrics: NAMESPACE + '.apiserver_request_total.count', NAMESPACE + '.apiserver_request_terminations_total.count', NAMESPACE + '.apiserver_admission_webhook_fail_open_count.count', + NAMESPACE + '.apiserver_admission_webhook_request_total.count', ] def test_check(self, dd_run_check, aggregator, mock_get): diff --git a/kube_apiserver_metrics/tests/test_kube_apiserver_metrics_1_26.py b/kube_apiserver_metrics/tests/test_kube_apiserver_metrics_1_26.py index b808ec865d4b3..24e6bcdaf4908 100644 --- a/kube_apiserver_metrics/tests/test_kube_apiserver_metrics_1_26.py +++ b/kube_apiserver_metrics/tests/test_kube_apiserver_metrics_1_26.py @@ -29,6 +29,7 @@ class TestKubeAPIServerMetrics: 'go_goroutines', 'rest_client_requests_total', 'authenticated_user_requests', + 'apiserver_admission_webhook_request_total', 'apiserver_request_total', 'apiserver_request_terminations_total', 'grpc_client_handled_total', @@ -76,6 +77,7 @@ class TestKubeAPIServerMetrics: 'authenticated_user_requests.count', 'apiserver_request_total.count', 'apiserver_request_terminations_total.count', + 'apiserver_admission_webhook_request_total.count', ] def test_check(self, dd_run_check, aggregator, mock_http_response): diff --git a/kube_apiserver_metrics/tests/test_kube_apiserver_metrics_1_27.py b/kube_apiserver_metrics/tests/test_kube_apiserver_metrics_1_27.py index acd14af9b34d4..b9bb828921c51 100644 --- a/kube_apiserver_metrics/tests/test_kube_apiserver_metrics_1_27.py +++ b/kube_apiserver_metrics/tests/test_kube_apiserver_metrics_1_27.py @@ -37,6 +37,7 @@ class TestKubeAPIServerMetrics: 'grpc_client_started_total', 'rest_client_request_latency_seconds.sum', 'rest_client_request_latency_seconds.count', + 'apiserver_admission_webhook_request_total', 'admission_step_admission_latencies_seconds.sum', 'admission_step_admission_latencies_seconds.count', 'admission_step_admission_latencies_seconds_summary.sum', @@ -74,6 +75,7 @@ class TestKubeAPIServerMetrics: 'audit_event.count', 'rest_client_requests_total.count', 'authenticated_user_requests.count', + 'apiserver_admission_webhook_request_total.count', 'apiserver_request_total.count', 'apiserver_request_terminations_total.count', ] diff --git a/kube_apiserver_metrics/tests/test_kube_apiserver_metrics_1_28.py b/kube_apiserver_metrics/tests/test_kube_apiserver_metrics_1_28.py index e7816e28c3b66..1a7cb1c518411 100644 --- a/kube_apiserver_metrics/tests/test_kube_apiserver_metrics_1_28.py +++ b/kube_apiserver_metrics/tests/test_kube_apiserver_metrics_1_28.py @@ -37,6 +37,7 @@ class TestKubeAPIServerMetrics: 'grpc_client_started_total', 'rest_client_request_latency_seconds.sum', 'rest_client_request_latency_seconds.count', + 'apiserver_admission_webhook_request_total', 'admission_step_admission_latencies_seconds.sum', 'admission_step_admission_latencies_seconds.count', 'admission_step_admission_latencies_seconds_summary.sum', @@ -76,6 +77,7 @@ class TestKubeAPIServerMetrics: 'audit_event.count', 'rest_client_requests_total.count', 'authenticated_user_requests.count', + 'apiserver_admission_webhook_request_total.count', 'apiserver_request_total.count', 'apiserver_request_terminations_total.count', ] From db30fc05526e70aaa73f515c83086a9d42fe70a5 Mon Sep 17 00:00:00 2001 From: Ilia Kurenkov Date: Fri, 20 Sep 2024 20:27:45 +0200 Subject: [PATCH 24/34] Drop 'six' from datadog_checks_dev (#18633) * Drop 'six' from datadog_checks_dev * drop requires_py3 from windows per counter tests --- .../tests/test_active_directory.py | 34 ------------------ active_directory/tests/test_unit.py | 3 -- aerospike/tests/test_aerospike.py | 3 -- aerospike/tests/test_unit_openmetricsv2.py | 3 +- aspdotnet/tests/test_unit.py | 3 -- cockroachdb/tests/test_bench.py | 3 +- cockroachdb/tests/test_e2e.py | 2 -- cockroachdb/tests/test_integration.py | 3 +- cockroachdb/tests/test_unit.py | 3 -- .../base/checks/logs/crawler/test_base.py | 3 -- .../openmetrics/test_legacy/test_bench.py | 3 -- .../test_legacy/test_compat_scraper.py | 4 --- .../checks/openmetrics/test_v2/test_bench.py | 3 -- .../checks/openmetrics/test_v2/test_config.py | 4 --- .../test_v2/test_first_scrape_handler.py | 4 --- .../openmetrics/test_v2/test_interface.py | 3 -- .../openmetrics/test_v2/test_options.py | 3 -- .../test_v2/test_transformers/test_counter.py | 5 --- .../test_transformers/test_counter_gauge.py | 5 --- .../test_v2/test_transformers/test_gauge.py | 5 --- .../test_transformers/test_histogram.py | 5 --- .../test_transformers/test_metadata.py | 5 --- .../test_transformers/test_native_dynamic.py | 5 --- .../test_v2/test_transformers/test_rate.py | 5 --- .../test_transformers/test_service_check.py | 5 --- .../test_v2/test_transformers/test_summary.py | 5 --- .../test_temporal_percent.py | 5 --- .../test_transformers/test_time_elapsed.py | 5 --- .../test_transformers/test_type_override.py | 4 --- .../tests/base/checks/test_agent_check.py | 3 -- .../windows/perf_counters/test_aggregation.py | 4 +-- .../windows/perf_counters/test_config.py | 4 +-- .../windows/perf_counters/test_filter.py | 4 +-- .../windows/perf_counters/test_health.py | 4 +-- .../windows/perf_counters/test_integration.py | 4 +-- .../perf_counters/test_legacy_support.py | 4 +-- .../perf_counters/test_localization.py | 4 +-- .../windows/perf_counters/test_refresh.py | 4 +-- .../windows/perf_counters/test_subclass.py | 4 +-- .../perf_counters/transformers/test_count.py | 4 +-- .../perf_counters/transformers/test_gauge.py | 4 +-- .../transformers/test_monotonic_count.py | 4 +-- .../perf_counters/transformers/test_rate.py | 4 +-- .../transformers/test_service_check.py | 4 +-- .../transformers/test_temporal_percent.py | 4 +-- .../transformers/test_time_elapsed.py | 4 +-- .../tests/base/utils/replay/test_check.py | 3 -- .../tests/models/test_interface.py | 3 -- .../tests/models/test_types.py | 3 -- .../datadog_checks/dev/conditions.py | 12 +++---- .../datadog_checks/dev/docker.py | 6 ++-- datadog_checks_dev/datadog_checks/dev/fs.py | 26 ++++---------- datadog_checks_dev/datadog_checks/dev/kind.py | 7 +--- .../datadog_checks/dev/plugin/pytest.py | 26 ++------------ .../datadog_checks/dev/ssh_tunnel.py | 7 +--- .../datadog_checks/dev/subprocess.py | 4 +-- .../datadog_checks/dev/terraform.py | 8 +---- .../datadog_checks/dev/testing.py | 4 --- .../datadog_checks/dev/utils.py | 2 +- datadog_checks_dev/pyproject.toml | 1 - dotnetclr/tests/e2e_test.py | 3 -- dotnetclr/tests/test_dotnetclr.py | 32 ----------------- dotnetclr/tests/test_unit.py | 3 -- envoy/tests/test_unit.py | 15 +------- exchange_server/tests/test_e2e.py | 2 -- exchange_server/tests/test_exchange_server.py | 36 ------------------- exchange_server/tests/test_unit.py | 3 -- gitlab/tests/test_integration.py | 3 -- gitlab/tests/test_unit.py | 15 +------- hyperv/tests/test_hyperv.py | 10 ------ ibm_mq/tests/test_ibm_mq_int.py | 4 --- ibm_mq/tests/test_ibm_mq_unit.py | 2 -- iis/tests/test_unit.py | 3 -- openstack/tests/test_openstack.py | 2 -- tls/tests/test_remote.py | 4 --- 75 files changed, 57 insertions(+), 402 deletions(-) delete mode 100644 active_directory/tests/test_active_directory.py delete mode 100644 dotnetclr/tests/test_dotnetclr.py delete mode 100644 exchange_server/tests/test_exchange_server.py diff --git a/active_directory/tests/test_active_directory.py b/active_directory/tests/test_active_directory.py deleted file mode 100644 index 4db327b084fbc..0000000000000 --- a/active_directory/tests/test_active_directory.py +++ /dev/null @@ -1,34 +0,0 @@ -# (C) Datadog, Inc. 2010-present -# All rights reserved -# Licensed under Simplified BSD License (see LICENSE) -import pytest -from datadog_test_libs.win.pdh_mocks import initialize_pdh_tests, pdh_mocks_fixture # noqa: F401 - -from datadog_checks.active_directory import ActiveDirectoryCheck -from datadog_checks.active_directory.metrics import DEFAULT_COUNTERS -from datadog_checks.base.stubs.aggregator import AggregatorStub # noqa: F401 -from datadog_checks.dev.testing import requires_py2 -from datadog_checks.dev.utils import get_metadata_metrics - -from .common import CHECK_NAME, MINIMAL_INSTANCE - -pytestmark = [requires_py2, pytest.mark.usefixtures('pdh_mocks_fixture')] - - -@pytest.fixture(autouse=True) -def setup_check(): - initialize_pdh_tests() - - -def test_basic_check(aggregator, dd_run_check): - # type: (AggregatorStub) -> None - instance = MINIMAL_INSTANCE - check = ActiveDirectoryCheck(CHECK_NAME, {}, [instance]) - dd_run_check(check) - - for metric_def in DEFAULT_COUNTERS: - metric = metric_def[3] - aggregator.assert_metric(metric, tags=None, count=1) - - aggregator.assert_all_metrics_covered() - aggregator.assert_metrics_using_metadata(get_metadata_metrics()) diff --git a/active_directory/tests/test_unit.py b/active_directory/tests/test_unit.py index 66351740a51af..97aad334c3805 100644 --- a/active_directory/tests/test_unit.py +++ b/active_directory/tests/test_unit.py @@ -4,13 +4,10 @@ from datadog_checks.active_directory import ActiveDirectoryCheck from datadog_checks.active_directory.metrics import DEFAULT_COUNTERS from datadog_checks.base.constants import ServiceCheck -from datadog_checks.dev.testing import requires_py3 from datadog_checks.dev.utils import get_metadata_metrics from .common import PERFORMANCE_OBJECTS -pytestmark = [requires_py3] - def test(aggregator, dd_default_hostname, dd_run_check, mock_performance_objects): mock_performance_objects(PERFORMANCE_OBJECTS) diff --git a/aerospike/tests/test_aerospike.py b/aerospike/tests/test_aerospike.py index b10878bbd7916..ecd3a0a688e4a 100644 --- a/aerospike/tests/test_aerospike.py +++ b/aerospike/tests/test_aerospike.py @@ -8,7 +8,6 @@ from datadog_checks.aerospike import AerospikeCheck from datadog_checks.base import AgentCheck -from datadog_checks.dev.testing import requires_py3 from datadog_checks.dev.utils import get_metadata_metrics from .common import ( @@ -70,7 +69,6 @@ def test_e2e(dd_agent_check, instance): aggregator.assert_metrics_using_metadata(get_metadata_metrics()) -@requires_py3 @pytest.mark.e2e def test_openmetrics_e2e(dd_agent_check, instance_openmetrics_v2): version_parts = [int(p) for p in VERSION.split('.')] @@ -93,7 +91,6 @@ def test_openmetrics_e2e(dd_agent_check, instance_openmetrics_v2): aggregator.assert_metrics_using_metadata(get_metadata_metrics(), check_submission_type=True) -@requires_py3 @pytest.mark.integration def test_metrics_warning(dd_run_check, instance_openmetrics_v2): instance_openmetrics_v2['metrics'] = ['migrate_rx_objs', 'migrate_tx_objs'] diff --git a/aerospike/tests/test_unit_openmetricsv2.py b/aerospike/tests/test_unit_openmetricsv2.py index f12312b46ba2c..d82f4607ff7d8 100644 --- a/aerospike/tests/test_unit_openmetricsv2.py +++ b/aerospike/tests/test_unit_openmetricsv2.py @@ -6,12 +6,11 @@ import pytest from datadog_checks.aerospike import AerospikeCheck -from datadog_checks.dev.testing import requires_py3 from datadog_checks.dev.utils import get_metadata_metrics from .common import EXPECTED_PROMETHEUS_METRICS, EXPECTED_PROMETHEUS_METRICS_5_6, HERE, PROMETHEUS_XDR_METRICS -pytestmark = [pytest.mark.unit, requires_py3] +pytestmark = [pytest.mark.unit] def get_fixture_path(filename): diff --git a/aspdotnet/tests/test_unit.py b/aspdotnet/tests/test_unit.py index c660f6efe6d20..af6dcb44eb89e 100644 --- a/aspdotnet/tests/test_unit.py +++ b/aspdotnet/tests/test_unit.py @@ -3,13 +3,10 @@ # Licensed under a 3-clause BSD style license (see LICENSE) from datadog_checks.aspdotnet import AspdotnetCheck from datadog_checks.base.constants import ServiceCheck -from datadog_checks.dev.testing import requires_py3 from datadog_checks.dev.utils import get_metadata_metrics from .common import ASP_APP_INSTANCES, ASP_APP_METRICS, ASP_METRICS, PERFORMANCE_OBJECTS -pytestmark = [requires_py3] - def test(aggregator, dd_default_hostname, dd_run_check, mock_performance_objects): mock_performance_objects(PERFORMANCE_OBJECTS) diff --git a/cockroachdb/tests/test_bench.py b/cockroachdb/tests/test_bench.py index f975b198ff7a5..da082936da0de 100644 --- a/cockroachdb/tests/test_bench.py +++ b/cockroachdb/tests/test_bench.py @@ -4,9 +4,8 @@ import pytest from datadog_checks.cockroachdb import CockroachdbCheck -from datadog_checks.dev.testing import requires_py3 -pytestmark = [pytest.mark.usefixtures('dd_environment'), requires_py3] +pytestmark = [pytest.mark.usefixtures('dd_environment')] def test_run(benchmark, dd_run_check, instance): diff --git a/cockroachdb/tests/test_e2e.py b/cockroachdb/tests/test_e2e.py index 9f2c7689b6c1d..480b8c226a1df 100644 --- a/cockroachdb/tests/test_e2e.py +++ b/cockroachdb/tests/test_e2e.py @@ -1,12 +1,10 @@ # (C) Datadog, Inc. 2021-present # All rights reserved # Licensed under a 3-clause BSD style license (see LICENSE) -from datadog_checks.dev.testing import requires_py3 from .common import assert_metrics -@requires_py3 def test_metrics(dd_agent_check, instance): aggregator = dd_agent_check(instance, rate=True) assert_metrics(aggregator) diff --git a/cockroachdb/tests/test_integration.py b/cockroachdb/tests/test_integration.py index ec75d2f2e7e28..cd018dcfa8254 100644 --- a/cockroachdb/tests/test_integration.py +++ b/cockroachdb/tests/test_integration.py @@ -5,12 +5,11 @@ from datadog_checks.base.constants import ServiceCheck from datadog_checks.cockroachdb import CockroachdbCheck -from datadog_checks.dev.testing import requires_py3 from datadog_checks.dev.utils import assert_service_checks, get_metadata_metrics from .common import COCKROACHDB_VERSION, assert_metrics -pytestmark = [pytest.mark.usefixtures('dd_environment'), requires_py3] +pytestmark = [pytest.mark.usefixtures('dd_environment')] def test_metrics(aggregator, instance, dd_run_check): diff --git a/cockroachdb/tests/test_unit.py b/cockroachdb/tests/test_unit.py index bb87ed749ebc6..88d3ea262017d 100644 --- a/cockroachdb/tests/test_unit.py +++ b/cockroachdb/tests/test_unit.py @@ -6,7 +6,6 @@ from datadog_checks.base.constants import ServiceCheck from datadog_checks.cockroachdb import CockroachdbCheck from datadog_checks.cockroachdb.metrics import METRIC_MAP, OMV2_METRIC_MAP -from datadog_checks.dev.testing import requires_py3 from datadog_checks.dev.utils import assert_service_checks, get_metadata_metrics from .common import ( @@ -23,8 +22,6 @@ get_fixture_path, ) -pytestmark = [requires_py3] - # The test below is designed to collect metrics that are not exposed in our e2e environment. # To collect security metrics, we need to enable TLS and provide certificates. In the future, diff --git a/datadog_checks_base/tests/base/checks/logs/crawler/test_base.py b/datadog_checks_base/tests/base/checks/logs/crawler/test_base.py index cc7a6aabfc740..504cdeb906b38 100644 --- a/datadog_checks_base/tests/base/checks/logs/crawler/test_base.py +++ b/datadog_checks_base/tests/base/checks/logs/crawler/test_base.py @@ -1,9 +1,6 @@ # (C) Datadog, Inc. 2024-present # All rights reserved # Licensed under a 3-clause BSD style license (see LICENSE) -from datadog_checks.dev.testing import requires_py3 - -pytestmark = [requires_py3] # TODO: Remove lazy imports and `.format` calls when we drop support for Python 2 diff --git a/datadog_checks_base/tests/base/checks/openmetrics/test_legacy/test_bench.py b/datadog_checks_base/tests/base/checks/openmetrics/test_legacy/test_bench.py index 73ca9e423c233..c540f7d47d8e7 100644 --- a/datadog_checks_base/tests/base/checks/openmetrics/test_legacy/test_bench.py +++ b/datadog_checks_base/tests/base/checks/openmetrics/test_legacy/test_bench.py @@ -7,12 +7,9 @@ from datadog_checks.base import OpenMetricsBaseCheck from datadog_checks.dev import get_here -from datadog_checks.dev.testing import requires_py3 from ..bench_utils import AMAZON_MSK_JMX_METRICS_MAP, AMAZON_MSK_JMX_METRICS_OVERRIDES -pytestmark = [requires_py3] - HERE = get_here() FIXTURE_PATH = os.path.abspath(os.path.join(os.path.dirname(HERE), '..', '..', '..', 'fixtures', 'prometheus')) diff --git a/datadog_checks_base/tests/base/checks/openmetrics/test_legacy/test_compat_scraper.py b/datadog_checks_base/tests/base/checks/openmetrics/test_legacy/test_compat_scraper.py index c5ade8a74c461..b053495cb8f25 100644 --- a/datadog_checks_base/tests/base/checks/openmetrics/test_legacy/test_compat_scraper.py +++ b/datadog_checks_base/tests/base/checks/openmetrics/test_legacy/test_compat_scraper.py @@ -3,12 +3,8 @@ # Licensed under a 3-clause BSD style license (see LICENSE) import pytest -from datadog_checks.dev.testing import requires_py3 - from .utils import get_legacy_check -pytestmark = [requires_py3] - class TestRawMetricPrefix: def test_not_string(self, dd_run_check): diff --git a/datadog_checks_base/tests/base/checks/openmetrics/test_v2/test_bench.py b/datadog_checks_base/tests/base/checks/openmetrics/test_v2/test_bench.py index e1c03e0ac08fc..678651f28f5ce 100644 --- a/datadog_checks_base/tests/base/checks/openmetrics/test_v2/test_bench.py +++ b/datadog_checks_base/tests/base/checks/openmetrics/test_v2/test_bench.py @@ -7,12 +7,9 @@ from datadog_checks.base import OpenMetricsBaseCheckV2 from datadog_checks.dev import get_here -from datadog_checks.dev.testing import requires_py3 from ..bench_utils import AMAZON_MSK_JMX_METRICS_MAP, AMAZON_MSK_JMX_METRICS_OVERRIDES -pytestmark = [requires_py3] - HERE = get_here() FIXTURE_PATH = os.path.abspath(os.path.join(os.path.dirname(HERE), '..', '..', '..', 'fixtures', 'prometheus')) diff --git a/datadog_checks_base/tests/base/checks/openmetrics/test_v2/test_config.py b/datadog_checks_base/tests/base/checks/openmetrics/test_v2/test_config.py index 029d7757c2454..0f3f6b7812e68 100644 --- a/datadog_checks_base/tests/base/checks/openmetrics/test_v2/test_config.py +++ b/datadog_checks_base/tests/base/checks/openmetrics/test_v2/test_config.py @@ -3,12 +3,8 @@ # Licensed under a 3-clause BSD style license (see LICENSE) import pytest -from datadog_checks.dev.testing import requires_py3 - from .utils import get_check -pytestmark = [requires_py3] - class TestPrometheusEndpoint: def test_not_string(self, dd_run_check): diff --git a/datadog_checks_base/tests/base/checks/openmetrics/test_v2/test_first_scrape_handler.py b/datadog_checks_base/tests/base/checks/openmetrics/test_v2/test_first_scrape_handler.py index 059fe10591942..1a710a8243bf9 100644 --- a/datadog_checks_base/tests/base/checks/openmetrics/test_v2/test_first_scrape_handler.py +++ b/datadog_checks_base/tests/base/checks/openmetrics/test_v2/test_first_scrape_handler.py @@ -4,13 +4,9 @@ import pytest from datadog_checks.base.stubs import datadog_agent -from datadog_checks.dev.testing import requires_py3 from .utils import get_check -pytestmark = [requires_py3] - - test_use_process_start_time_data = """\ # HELP go_memstats_alloc_bytes_total Total number of bytes allocated, even if freed. # TYPE go_memstats_alloc_bytes_total counter diff --git a/datadog_checks_base/tests/base/checks/openmetrics/test_v2/test_interface.py b/datadog_checks_base/tests/base/checks/openmetrics/test_v2/test_interface.py index 2e9ff176bd35d..d6071cf0226e5 100644 --- a/datadog_checks_base/tests/base/checks/openmetrics/test_v2/test_interface.py +++ b/datadog_checks_base/tests/base/checks/openmetrics/test_v2/test_interface.py @@ -3,12 +3,9 @@ # Licensed under a 3-clause BSD style license (see LICENSE) from datadog_checks.base import OpenMetricsBaseCheckV2 from datadog_checks.base.constants import ServiceCheck -from datadog_checks.dev.testing import requires_py3 from .utils import get_check -pytestmark = [requires_py3] - def test_default_config(aggregator, dd_run_check, mock_http_response): class Check(OpenMetricsBaseCheckV2): diff --git a/datadog_checks_base/tests/base/checks/openmetrics/test_v2/test_options.py b/datadog_checks_base/tests/base/checks/openmetrics/test_v2/test_options.py index 597aeda01a1b3..6ac560760069e 100644 --- a/datadog_checks_base/tests/base/checks/openmetrics/test_v2/test_options.py +++ b/datadog_checks_base/tests/base/checks/openmetrics/test_v2/test_options.py @@ -5,12 +5,9 @@ from mock import Mock from datadog_checks.base.constants import ServiceCheck -from datadog_checks.dev.testing import requires_py3 from .utils import get_check -pytestmark = [requires_py3] - class TestNamespace: def test(self, aggregator, dd_run_check, mock_http_response): diff --git a/datadog_checks_base/tests/base/checks/openmetrics/test_v2/test_transformers/test_counter.py b/datadog_checks_base/tests/base/checks/openmetrics/test_v2/test_transformers/test_counter.py index 714f882d7d274..922ee631c6304 100644 --- a/datadog_checks_base/tests/base/checks/openmetrics/test_v2/test_transformers/test_counter.py +++ b/datadog_checks_base/tests/base/checks/openmetrics/test_v2/test_transformers/test_counter.py @@ -1,14 +1,9 @@ # (C) Datadog, Inc. 2020-present # All rights reserved # Licensed under a 3-clause BSD style license (see LICENSE) -from datadog_checks.dev.testing import requires_py3 from ..utils import get_check -pytestmark = [ - requires_py3, -] - def test_basic(aggregator, dd_run_check, mock_http_response): """ diff --git a/datadog_checks_base/tests/base/checks/openmetrics/test_v2/test_transformers/test_counter_gauge.py b/datadog_checks_base/tests/base/checks/openmetrics/test_v2/test_transformers/test_counter_gauge.py index e78ffe785a5cb..ae1a6a36203b8 100644 --- a/datadog_checks_base/tests/base/checks/openmetrics/test_v2/test_transformers/test_counter_gauge.py +++ b/datadog_checks_base/tests/base/checks/openmetrics/test_v2/test_transformers/test_counter_gauge.py @@ -1,14 +1,9 @@ # (C) Datadog, Inc. 2020-present # All rights reserved # Licensed under a 3-clause BSD style license (see LICENSE) -from datadog_checks.dev.testing import requires_py3 from ..utils import get_check -pytestmark = [ - requires_py3, -] - def test(aggregator, dd_run_check, mock_http_response): mock_http_response( diff --git a/datadog_checks_base/tests/base/checks/openmetrics/test_v2/test_transformers/test_gauge.py b/datadog_checks_base/tests/base/checks/openmetrics/test_v2/test_transformers/test_gauge.py index a6e9abeadb1e8..bf0398e5d5347 100644 --- a/datadog_checks_base/tests/base/checks/openmetrics/test_v2/test_transformers/test_gauge.py +++ b/datadog_checks_base/tests/base/checks/openmetrics/test_v2/test_transformers/test_gauge.py @@ -1,14 +1,9 @@ # (C) Datadog, Inc. 2020-present # All rights reserved # Licensed under a 3-clause BSD style license (see LICENSE) -from datadog_checks.dev.testing import requires_py3 from ..utils import get_check -pytestmark = [ - requires_py3, -] - def test_basic(aggregator, dd_run_check, mock_http_response): mock_http_response( diff --git a/datadog_checks_base/tests/base/checks/openmetrics/test_v2/test_transformers/test_histogram.py b/datadog_checks_base/tests/base/checks/openmetrics/test_v2/test_transformers/test_histogram.py index 38785892e6c90..fc0bfd0ff76a7 100644 --- a/datadog_checks_base/tests/base/checks/openmetrics/test_v2/test_transformers/test_histogram.py +++ b/datadog_checks_base/tests/base/checks/openmetrics/test_v2/test_transformers/test_histogram.py @@ -1,14 +1,9 @@ # (C) Datadog, Inc. 2020-present # All rights reserved # Licensed under a 3-clause BSD style license (see LICENSE) -from datadog_checks.dev.testing import requires_py3 from ..utils import get_check -pytestmark = [ - requires_py3, -] - def assert_metric_counts(aggregator, payload): num_bucket_metrics = 0 diff --git a/datadog_checks_base/tests/base/checks/openmetrics/test_v2/test_transformers/test_metadata.py b/datadog_checks_base/tests/base/checks/openmetrics/test_v2/test_transformers/test_metadata.py index b18a9971a6d43..de5228d6ec77b 100644 --- a/datadog_checks_base/tests/base/checks/openmetrics/test_v2/test_transformers/test_metadata.py +++ b/datadog_checks_base/tests/base/checks/openmetrics/test_v2/test_transformers/test_metadata.py @@ -1,14 +1,9 @@ # (C) Datadog, Inc. 2020-present # All rights reserved # Licensed under a 3-clause BSD style license (see LICENSE) -from datadog_checks.dev.testing import requires_py3 from ..utils import get_check -pytestmark = [ - requires_py3, -] - def test_basic(aggregator, datadog_agent, dd_run_check, mock_http_response): mock_http_response( diff --git a/datadog_checks_base/tests/base/checks/openmetrics/test_v2/test_transformers/test_native_dynamic.py b/datadog_checks_base/tests/base/checks/openmetrics/test_v2/test_transformers/test_native_dynamic.py index cc3e34b668f05..3ef2462a8a6d5 100644 --- a/datadog_checks_base/tests/base/checks/openmetrics/test_v2/test_transformers/test_native_dynamic.py +++ b/datadog_checks_base/tests/base/checks/openmetrics/test_v2/test_transformers/test_native_dynamic.py @@ -1,14 +1,9 @@ # (C) Datadog, Inc. 2021-present # All rights reserved # Licensed under a 3-clause BSD style license (see LICENSE) -from datadog_checks.dev.testing import requires_py3 from ..utils import get_check -pytestmark = [ - requires_py3, -] - def test_basic(aggregator, dd_run_check, mock_http_response): mock_http_response( diff --git a/datadog_checks_base/tests/base/checks/openmetrics/test_v2/test_transformers/test_rate.py b/datadog_checks_base/tests/base/checks/openmetrics/test_v2/test_transformers/test_rate.py index dc253f82bcb69..d5912078a05a6 100644 --- a/datadog_checks_base/tests/base/checks/openmetrics/test_v2/test_transformers/test_rate.py +++ b/datadog_checks_base/tests/base/checks/openmetrics/test_v2/test_transformers/test_rate.py @@ -1,14 +1,9 @@ # (C) Datadog, Inc. 2020-present # All rights reserved # Licensed under a 3-clause BSD style license (see LICENSE) -from datadog_checks.dev.testing import requires_py3 from ..utils import get_check -pytestmark = [ - requires_py3, -] - def test(aggregator, dd_run_check, mock_http_response): mock_http_response( diff --git a/datadog_checks_base/tests/base/checks/openmetrics/test_v2/test_transformers/test_service_check.py b/datadog_checks_base/tests/base/checks/openmetrics/test_v2/test_transformers/test_service_check.py index 51d15ee69965c..d1764ac84d423 100644 --- a/datadog_checks_base/tests/base/checks/openmetrics/test_v2/test_transformers/test_service_check.py +++ b/datadog_checks_base/tests/base/checks/openmetrics/test_v2/test_transformers/test_service_check.py @@ -2,14 +2,9 @@ # All rights reserved # Licensed under a 3-clause BSD style license (see LICENSE) from datadog_checks.base.constants import ServiceCheck -from datadog_checks.dev.testing import requires_py3 from ..utils import get_check -pytestmark = [ - requires_py3, -] - def test_known(aggregator, dd_run_check, mock_http_response): mock_http_response( diff --git a/datadog_checks_base/tests/base/checks/openmetrics/test_v2/test_transformers/test_summary.py b/datadog_checks_base/tests/base/checks/openmetrics/test_v2/test_transformers/test_summary.py index 2a941c065a50b..37e5b79ebae2c 100644 --- a/datadog_checks_base/tests/base/checks/openmetrics/test_v2/test_transformers/test_summary.py +++ b/datadog_checks_base/tests/base/checks/openmetrics/test_v2/test_transformers/test_summary.py @@ -1,14 +1,9 @@ # (C) Datadog, Inc. 2020-present # All rights reserved # Licensed under a 3-clause BSD style license (see LICENSE) -from datadog_checks.dev.testing import requires_py3 from ..utils import get_check -pytestmark = [ - requires_py3, -] - def assert_metric_counts(aggregator, payload): num_quantile_metrics = 0 diff --git a/datadog_checks_base/tests/base/checks/openmetrics/test_v2/test_transformers/test_temporal_percent.py b/datadog_checks_base/tests/base/checks/openmetrics/test_v2/test_transformers/test_temporal_percent.py index 49eeeb6ca3841..23140edb50b29 100644 --- a/datadog_checks_base/tests/base/checks/openmetrics/test_v2/test_transformers/test_temporal_percent.py +++ b/datadog_checks_base/tests/base/checks/openmetrics/test_v2/test_transformers/test_temporal_percent.py @@ -1,14 +1,9 @@ # (C) Datadog, Inc. 2020-present # All rights reserved # Licensed under a 3-clause BSD style license (see LICENSE) -from datadog_checks.dev.testing import requires_py3 from ..utils import get_check -pytestmark = [ - requires_py3, -] - def test_named(aggregator, dd_run_check, mock_http_response): mock_http_response( diff --git a/datadog_checks_base/tests/base/checks/openmetrics/test_v2/test_transformers/test_time_elapsed.py b/datadog_checks_base/tests/base/checks/openmetrics/test_v2/test_transformers/test_time_elapsed.py index 4366fb51c0f85..f23c446eaa229 100644 --- a/datadog_checks_base/tests/base/checks/openmetrics/test_v2/test_transformers/test_time_elapsed.py +++ b/datadog_checks_base/tests/base/checks/openmetrics/test_v2/test_transformers/test_time_elapsed.py @@ -2,14 +2,9 @@ # All rights reserved # Licensed under a 3-clause BSD style license (see LICENSE) from datadog_checks.base.utils.time import get_timestamp -from datadog_checks.dev.testing import requires_py3 from ..utils import get_check -pytestmark = [ - requires_py3, -] - def test(aggregator, dd_run_check, mock_http_response): mock_http_response( diff --git a/datadog_checks_base/tests/base/checks/openmetrics/test_v2/test_transformers/test_type_override.py b/datadog_checks_base/tests/base/checks/openmetrics/test_v2/test_transformers/test_type_override.py index ca79de6b0e176..a1fa39e6bfedc 100644 --- a/datadog_checks_base/tests/base/checks/openmetrics/test_v2/test_transformers/test_type_override.py +++ b/datadog_checks_base/tests/base/checks/openmetrics/test_v2/test_transformers/test_type_override.py @@ -3,12 +3,8 @@ # Licensed under a 3-clause BSD style license (see LICENSE) import pytest -from datadog_checks.dev.testing import requires_py3 - from ..utils import get_check -pytestmark = [requires_py3] - @pytest.mark.parametrize( 'metric_type', diff --git a/datadog_checks_base/tests/base/checks/test_agent_check.py b/datadog_checks_base/tests/base/checks/test_agent_check.py index 244a041582b26..aa478ceca3e17 100644 --- a/datadog_checks_base/tests/base/checks/test_agent_check.py +++ b/datadog_checks_base/tests/base/checks/test_agent_check.py @@ -13,7 +13,6 @@ from datadog_checks.base import AgentCheck, to_native_string from datadog_checks.base import __version__ as base_package_version -from datadog_checks.dev.testing import requires_py3 def test_instance(): @@ -1068,7 +1067,6 @@ def check(self, _): assert check.initialize.call_count == 2 -@requires_py3 def test_load_configuration_models(dd_run_check, mocker): instance = {'endpoint': 'url', 'tags': ['foo:bar'], 'proxy': {'http': 'http://1.2.3.4:9000'}} init_config = {'proxy': {'https': 'https://1.2.3.4:4242'}} @@ -1111,7 +1109,6 @@ def __init__(self, **kwargs): pass -@requires_py3 @pytest.mark.parametrize( "check_instance_config, default_instance_config, log_lines, unknown_options", [ diff --git a/datadog_checks_base/tests/base/checks/windows/perf_counters/test_aggregation.py b/datadog_checks_base/tests/base/checks/windows/perf_counters/test_aggregation.py index 9e8731b030593..d9967d61fef73 100644 --- a/datadog_checks_base/tests/base/checks/windows/perf_counters/test_aggregation.py +++ b/datadog_checks_base/tests/base/checks/windows/perf_counters/test_aggregation.py @@ -1,11 +1,11 @@ # (C) Datadog, Inc. 2021-present # All rights reserved # Licensed under a 3-clause BSD style license (see LICENSE) -from datadog_checks.dev.testing import requires_py3, requires_windows +from datadog_checks.dev.testing import requires_windows from .utils import GLOBAL_TAGS, get_check -pytestmark = [requires_py3, requires_windows] +pytestmark = [requires_windows] def test_single_instance(aggregator, dd_run_check, mock_performance_objects): diff --git a/datadog_checks_base/tests/base/checks/windows/perf_counters/test_config.py b/datadog_checks_base/tests/base/checks/windows/perf_counters/test_config.py index 3344b237efe41..d03346346b373 100644 --- a/datadog_checks_base/tests/base/checks/windows/perf_counters/test_config.py +++ b/datadog_checks_base/tests/base/checks/windows/perf_counters/test_config.py @@ -5,11 +5,11 @@ import pytest -from datadog_checks.dev.testing import requires_py3, requires_windows +from datadog_checks.dev.testing import requires_windows from .utils import get_check -pytestmark = [requires_py3, requires_windows] +pytestmark = [requires_windows] @pytest.fixture(autouse=True) diff --git a/datadog_checks_base/tests/base/checks/windows/perf_counters/test_filter.py b/datadog_checks_base/tests/base/checks/windows/perf_counters/test_filter.py index 71939b239bee7..443d3fdbee576 100644 --- a/datadog_checks_base/tests/base/checks/windows/perf_counters/test_filter.py +++ b/datadog_checks_base/tests/base/checks/windows/perf_counters/test_filter.py @@ -1,11 +1,11 @@ # (C) Datadog, Inc. 2021-present # All rights reserved # Licensed under a 3-clause BSD style license (see LICENSE) -from datadog_checks.dev.testing import requires_py3, requires_windows +from datadog_checks.dev.testing import requires_windows from .utils import GLOBAL_TAGS, get_check -pytestmark = [requires_py3, requires_windows] +pytestmark = [requires_windows] def test_include(aggregator, dd_run_check, mock_performance_objects): diff --git a/datadog_checks_base/tests/base/checks/windows/perf_counters/test_health.py b/datadog_checks_base/tests/base/checks/windows/perf_counters/test_health.py index 8d619380498e0..90d9764224a69 100644 --- a/datadog_checks_base/tests/base/checks/windows/perf_counters/test_health.py +++ b/datadog_checks_base/tests/base/checks/windows/perf_counters/test_health.py @@ -4,11 +4,11 @@ import logging from datadog_checks.base.constants import ServiceCheck -from datadog_checks.dev.testing import requires_py3, requires_windows +from datadog_checks.dev.testing import requires_windows from .utils import GLOBAL_TAGS, get_check -pytestmark = [requires_py3, requires_windows] +pytestmark = [requires_windows] def test_disable(aggregator, dd_run_check, mock_performance_objects): diff --git a/datadog_checks_base/tests/base/checks/windows/perf_counters/test_integration.py b/datadog_checks_base/tests/base/checks/windows/perf_counters/test_integration.py index 34a589f04c5c2..35b6ada26b9d2 100644 --- a/datadog_checks_base/tests/base/checks/windows/perf_counters/test_integration.py +++ b/datadog_checks_base/tests/base/checks/windows/perf_counters/test_integration.py @@ -6,11 +6,11 @@ import mock from datadog_checks.base.constants import ServiceCheck -from datadog_checks.dev.testing import requires_py3, requires_windows +from datadog_checks.dev.testing import requires_windows from .utils import SERVER, get_check -pytestmark = [requires_py3, requires_windows] +pytestmark = [requires_windows] def test(aggregator, dd_run_check): diff --git a/datadog_checks_base/tests/base/checks/windows/perf_counters/test_legacy_support.py b/datadog_checks_base/tests/base/checks/windows/perf_counters/test_legacy_support.py index 1c8a96c36cf90..69b186d047ed9 100644 --- a/datadog_checks_base/tests/base/checks/windows/perf_counters/test_legacy_support.py +++ b/datadog_checks_base/tests/base/checks/windows/perf_counters/test_legacy_support.py @@ -4,7 +4,7 @@ import logging from datadog_checks.base.constants import ServiceCheck -from datadog_checks.dev.testing import requires_py3, requires_windows +from datadog_checks.dev.testing import requires_windows from .utils import GLOBAL_TAGS, SERVER @@ -14,7 +14,7 @@ except Exception: PerfCountersBaseCheckWithLegacySupport = object -pytestmark = [requires_py3, requires_windows] +pytestmark = [requires_windows] class CustomCheck(PerfCountersBaseCheckWithLegacySupport): diff --git a/datadog_checks_base/tests/base/checks/windows/perf_counters/test_localization.py b/datadog_checks_base/tests/base/checks/windows/perf_counters/test_localization.py index e3e9b384bec05..cdb417185adb0 100644 --- a/datadog_checks_base/tests/base/checks/windows/perf_counters/test_localization.py +++ b/datadog_checks_base/tests/base/checks/windows/perf_counters/test_localization.py @@ -4,11 +4,11 @@ import mock from datadog_checks.base.constants import ServiceCheck -from datadog_checks.dev.testing import requires_py3, requires_windows +from datadog_checks.dev.testing import requires_windows from .utils import GLOBAL_TAGS, SERVER, get_check -pytestmark = [requires_py3, requires_windows] +pytestmark = [requires_windows] def test_default(aggregator, dd_run_check, mock_performance_objects): diff --git a/datadog_checks_base/tests/base/checks/windows/perf_counters/test_refresh.py b/datadog_checks_base/tests/base/checks/windows/perf_counters/test_refresh.py index 146dcce4f64db..df0b1e9c0eacb 100644 --- a/datadog_checks_base/tests/base/checks/windows/perf_counters/test_refresh.py +++ b/datadog_checks_base/tests/base/checks/windows/perf_counters/test_refresh.py @@ -1,11 +1,11 @@ # (C) Datadog, Inc. 2021-present # All rights reserved # Licensed under a 3-clause BSD style license (see LICENSE) -from datadog_checks.dev.testing import requires_py3, requires_windows +from datadog_checks.dev.testing import requires_windows from .utils import GLOBAL_TAGS, get_check -pytestmark = [requires_py3, requires_windows] +pytestmark = [requires_windows] def test_detection(aggregator, dd_run_check, mock_performance_objects): diff --git a/datadog_checks_base/tests/base/checks/windows/perf_counters/test_subclass.py b/datadog_checks_base/tests/base/checks/windows/perf_counters/test_subclass.py index 01da8468ea4f2..38b5e367e98a0 100644 --- a/datadog_checks_base/tests/base/checks/windows/perf_counters/test_subclass.py +++ b/datadog_checks_base/tests/base/checks/windows/perf_counters/test_subclass.py @@ -2,7 +2,7 @@ # All rights reserved # Licensed under a 3-clause BSD style license (see LICENSE) from datadog_checks.base.constants import ServiceCheck -from datadog_checks.dev.testing import requires_py3, requires_windows +from datadog_checks.dev.testing import requires_windows from .utils import GLOBAL_TAGS, SERVER @@ -14,7 +14,7 @@ PerfCountersBaseCheck = object PerfObject = object -pytestmark = [requires_py3, requires_windows] +pytestmark = [requires_windows] class CustomCheck(PerfCountersBaseCheck): diff --git a/datadog_checks_base/tests/base/checks/windows/perf_counters/transformers/test_count.py b/datadog_checks_base/tests/base/checks/windows/perf_counters/transformers/test_count.py index d8c6fe4d7e604..d00026bc25558 100644 --- a/datadog_checks_base/tests/base/checks/windows/perf_counters/transformers/test_count.py +++ b/datadog_checks_base/tests/base/checks/windows/perf_counters/transformers/test_count.py @@ -1,11 +1,11 @@ # (C) Datadog, Inc. 2021-present # All rights reserved # Licensed under a 3-clause BSD style license (see LICENSE) -from datadog_checks.dev.testing import requires_py3, requires_windows +from datadog_checks.dev.testing import requires_windows from ..utils import GLOBAL_TAGS, get_check -pytestmark = [requires_py3, requires_windows] +pytestmark = [requires_windows] def test(aggregator, dd_run_check, mock_performance_objects): diff --git a/datadog_checks_base/tests/base/checks/windows/perf_counters/transformers/test_gauge.py b/datadog_checks_base/tests/base/checks/windows/perf_counters/transformers/test_gauge.py index cf47ca5777107..a7c10e4ac7647 100644 --- a/datadog_checks_base/tests/base/checks/windows/perf_counters/transformers/test_gauge.py +++ b/datadog_checks_base/tests/base/checks/windows/perf_counters/transformers/test_gauge.py @@ -1,11 +1,11 @@ # (C) Datadog, Inc. 2021-present # All rights reserved # Licensed under a 3-clause BSD style license (see LICENSE) -from datadog_checks.dev.testing import requires_py3, requires_windows +from datadog_checks.dev.testing import requires_windows from ..utils import GLOBAL_TAGS, get_check -pytestmark = [requires_py3, requires_windows] +pytestmark = [requires_windows] def test_explicit(aggregator, dd_run_check, mock_performance_objects): diff --git a/datadog_checks_base/tests/base/checks/windows/perf_counters/transformers/test_monotonic_count.py b/datadog_checks_base/tests/base/checks/windows/perf_counters/transformers/test_monotonic_count.py index 288b9789666c2..c66b2635fcfee 100644 --- a/datadog_checks_base/tests/base/checks/windows/perf_counters/transformers/test_monotonic_count.py +++ b/datadog_checks_base/tests/base/checks/windows/perf_counters/transformers/test_monotonic_count.py @@ -1,11 +1,11 @@ # (C) Datadog, Inc. 2021-present # All rights reserved # Licensed under a 3-clause BSD style license (see LICENSE) -from datadog_checks.dev.testing import requires_py3, requires_windows +from datadog_checks.dev.testing import requires_windows from ..utils import GLOBAL_TAGS, get_check -pytestmark = [requires_py3, requires_windows] +pytestmark = [requires_windows] def test(aggregator, dd_run_check, mock_performance_objects): diff --git a/datadog_checks_base/tests/base/checks/windows/perf_counters/transformers/test_rate.py b/datadog_checks_base/tests/base/checks/windows/perf_counters/transformers/test_rate.py index 323bfc2eb1a41..3a0957b43ce6d 100644 --- a/datadog_checks_base/tests/base/checks/windows/perf_counters/transformers/test_rate.py +++ b/datadog_checks_base/tests/base/checks/windows/perf_counters/transformers/test_rate.py @@ -1,11 +1,11 @@ # (C) Datadog, Inc. 2021-present # All rights reserved # Licensed under a 3-clause BSD style license (see LICENSE) -from datadog_checks.dev.testing import requires_py3, requires_windows +from datadog_checks.dev.testing import requires_windows from ..utils import GLOBAL_TAGS, get_check -pytestmark = [requires_py3, requires_windows] +pytestmark = [requires_windows] def test(aggregator, dd_run_check, mock_performance_objects): diff --git a/datadog_checks_base/tests/base/checks/windows/perf_counters/transformers/test_service_check.py b/datadog_checks_base/tests/base/checks/windows/perf_counters/transformers/test_service_check.py index 69c2402d6401e..12815bfecc6f1 100644 --- a/datadog_checks_base/tests/base/checks/windows/perf_counters/transformers/test_service_check.py +++ b/datadog_checks_base/tests/base/checks/windows/perf_counters/transformers/test_service_check.py @@ -2,11 +2,11 @@ # All rights reserved # Licensed under a 3-clause BSD style license (see LICENSE) from datadog_checks.base.constants import ServiceCheck -from datadog_checks.dev.testing import requires_py3, requires_windows +from datadog_checks.dev.testing import requires_windows from ..utils import GLOBAL_TAGS, get_check -pytestmark = [requires_py3, requires_windows] +pytestmark = [requires_windows] def test_known(aggregator, dd_run_check, mock_performance_objects): diff --git a/datadog_checks_base/tests/base/checks/windows/perf_counters/transformers/test_temporal_percent.py b/datadog_checks_base/tests/base/checks/windows/perf_counters/transformers/test_temporal_percent.py index f666fc681d9bd..59e772960bd71 100644 --- a/datadog_checks_base/tests/base/checks/windows/perf_counters/transformers/test_temporal_percent.py +++ b/datadog_checks_base/tests/base/checks/windows/perf_counters/transformers/test_temporal_percent.py @@ -1,11 +1,11 @@ # (C) Datadog, Inc. 2021-present # All rights reserved # Licensed under a 3-clause BSD style license (see LICENSE) -from datadog_checks.dev.testing import requires_py3, requires_windows +from datadog_checks.dev.testing import requires_windows from ..utils import GLOBAL_TAGS, get_check -pytestmark = [requires_py3, requires_windows] +pytestmark = [requires_windows] def test_named(aggregator, dd_run_check, mock_performance_objects): diff --git a/datadog_checks_base/tests/base/checks/windows/perf_counters/transformers/test_time_elapsed.py b/datadog_checks_base/tests/base/checks/windows/perf_counters/transformers/test_time_elapsed.py index 4380dc4e5024b..7b191063f1ea1 100644 --- a/datadog_checks_base/tests/base/checks/windows/perf_counters/transformers/test_time_elapsed.py +++ b/datadog_checks_base/tests/base/checks/windows/perf_counters/transformers/test_time_elapsed.py @@ -2,11 +2,11 @@ # All rights reserved # Licensed under a 3-clause BSD style license (see LICENSE) from datadog_checks.base.utils.time import get_timestamp -from datadog_checks.dev.testing import requires_py3, requires_windows +from datadog_checks.dev.testing import requires_windows from ..utils import GLOBAL_TAGS, get_check -pytestmark = [requires_py3, requires_windows] +pytestmark = [requires_windows] def test(aggregator, dd_run_check, mock_performance_objects): diff --git a/datadog_checks_base/tests/base/utils/replay/test_check.py b/datadog_checks_base/tests/base/utils/replay/test_check.py index 2723d4b7dc4ec..7866db15ad5de 100644 --- a/datadog_checks_base/tests/base/utils/replay/test_check.py +++ b/datadog_checks_base/tests/base/utils/replay/test_check.py @@ -7,9 +7,6 @@ from datadog_checks.base import AgentCheck from datadog_checks.base.constants import ServiceCheck -from datadog_checks.dev.testing import requires_py3 - -pytestmark = [requires_py3] class ReplayCheck(AgentCheck): diff --git a/datadog_checks_base/tests/models/test_interface.py b/datadog_checks_base/tests/models/test_interface.py index 3d189482e715a..781dcf4e0aa14 100644 --- a/datadog_checks_base/tests/models/test_interface.py +++ b/datadog_checks_base/tests/models/test_interface.py @@ -5,15 +5,12 @@ from six import PY3 from datadog_checks.base import AgentCheck -from datadog_checks.dev.testing import requires_py3 if PY3: from .config_models import ConfigMixin else: ConfigMixin = object -pytestmark = [requires_py3] - class Check(AgentCheck, ConfigMixin): def check(self, _): diff --git a/datadog_checks_base/tests/models/test_types.py b/datadog_checks_base/tests/models/test_types.py index c4ccc0f07b745..c0747072930c0 100644 --- a/datadog_checks_base/tests/models/test_types.py +++ b/datadog_checks_base/tests/models/test_types.py @@ -1,9 +1,6 @@ # (C) Datadog, Inc. 2021-present # All rights reserved # Licensed under a 3-clause BSD style license (see LICENSE) -from datadog_checks.dev.testing import requires_py3 - -pytestmark = [requires_py3] def test_make_immutable(): diff --git a/datadog_checks_dev/datadog_checks/dev/conditions.py b/datadog_checks_dev/datadog_checks/dev/conditions.py index 85d8ea474bc9d..d4c228c259780 100644 --- a/datadog_checks_dev/datadog_checks/dev/conditions.py +++ b/datadog_checks_dev/datadog_checks/dev/conditions.py @@ -6,9 +6,7 @@ import time from contextlib import closing from typing import Callable, Dict, List, Tuple, Union # noqa: F401 - -from six import string_types -from six.moves.urllib.request import urlopen +from urllib.request import urlopen from .errors import RetryError from .structures import LazyFunction @@ -69,7 +67,7 @@ def __init__( wait=1, # type: int send_request=None, ): - self.endpoints = [endpoints] if isinstance(endpoints, string_types) else endpoints + self.endpoints = [endpoints] if isinstance(endpoints, str) else endpoints self.timeout = timeout self.attempts = attempts self.wait = wait @@ -131,12 +129,10 @@ def __init__( if not (self.stdout or self.stderr): raise ValueError('Must capture stdout, stderr, or both.') - if isinstance(patterns, string_types): + if isinstance(patterns, str): patterns = [patterns] - self.patterns = [ - re.compile(pattern, re.M) if isinstance(pattern, string_types) else pattern for pattern in patterns - ] + self.patterns = [re.compile(pattern, re.M) if isinstance(pattern, str) else pattern for pattern in patterns] if matches == 'all': self.matches = len(patterns) diff --git a/datadog_checks_dev/datadog_checks/dev/docker.py b/datadog_checks_dev/datadog_checks/dev/docker.py index 044f05b9848f2..cf62dfd4c66aa 100644 --- a/datadog_checks_dev/datadog_checks/dev/docker.py +++ b/datadog_checks_dev/datadog_checks/dev/docker.py @@ -4,9 +4,7 @@ import os from contextlib import contextmanager from typing import Iterator # noqa: F401 - -from six import string_types -from six.moves.urllib.parse import urlparse +from urllib.parse import urlparse from .conditions import CheckDockerLogs from .env import environment_run, get_state, save_state @@ -168,7 +166,7 @@ def docker_run( raise TypeError('You must select either a compose file or a custom setup callable, not both.') if compose_file is not None: - if not isinstance(compose_file, string_types): + if not isinstance(compose_file, str): raise TypeError('The path to the compose file is not a string: {}'.format(repr(compose_file))) set_up = ComposeFileUp(compose_file, build=build, service_name=service_name) diff --git a/datadog_checks_dev/datadog_checks/dev/fs.py b/datadog_checks_dev/datadog_checks/dev/fs.py index c715d2ed466ff..aab0ca507ff35 100644 --- a/datadog_checks_dev/datadog_checks/dev/fs.py +++ b/datadog_checks_dev/datadog_checks/dev/fs.py @@ -11,31 +11,19 @@ from contextlib import contextmanager from io import open from tempfile import mkdtemp - -from six import PY3, text_type -from six.moves.urllib.request import urlopen +from urllib.request import urlopen from .structures import EnvVars -if PY3: - - def write_file(file, contents, encoding='utf-8'): - with open(file, 'w', encoding=encoding) as f: - f.write(contents) - def write_file_lines(file, lines, encoding='utf-8'): - with open(file, 'w', encoding=encoding) as f: - f.writelines(lines) - -else: +def write_file(file, contents, encoding='utf-8'): + with open(file, 'w', encoding=encoding) as f: + f.write(contents) - def write_file(file, contents, encoding='utf-8'): - with open(file, 'w', encoding=encoding) as f: - f.write(text_type(contents)) - def write_file_lines(file, lines, encoding='utf-8'): - with open(file, 'w', encoding=encoding) as f: - f.writelines(text_type(line) for line in lines) +def write_file_lines(file, lines, encoding='utf-8'): + with open(file, 'w', encoding=encoding) as f: + f.writelines(lines) def dir_exists(d): diff --git a/datadog_checks_dev/datadog_checks/dev/kind.py b/datadog_checks_dev/datadog_checks/dev/kind.py index 3bd61ab8ef85d..f0a7cda5d13f5 100644 --- a/datadog_checks_dev/datadog_checks/dev/kind.py +++ b/datadog_checks_dev/datadog_checks/dev/kind.py @@ -2,9 +2,9 @@ # All rights reserved # Licensed under a 3-clause BSD style license (see LICENSE) from contextlib import contextmanager +from shutil import which import pytest -from six import PY3 from .env import environment_run from .fs import create_file, file_exists, path_join @@ -12,11 +12,6 @@ from .subprocess import run_command from .utils import get_active_env, get_current_check_name -if PY3: - from shutil import which -else: - from shutilwhich import which - @contextmanager def kind_run( diff --git a/datadog_checks_dev/datadog_checks/dev/plugin/pytest.py b/datadog_checks_dev/datadog_checks/dev/plugin/pytest.py index 2345e6013cf57..f9eaaf19b38a0 100644 --- a/datadog_checks_dev/datadog_checks/dev/plugin/pytest.py +++ b/datadog_checks_dev/datadog_checks/dev/plugin/pytest.py @@ -6,14 +6,12 @@ import json import os import re -import warnings from base64 import urlsafe_b64encode from collections import namedtuple # Not using dataclasses for Py2 compatibility from io import open from typing import Dict, List, Optional, Tuple # noqa: F401 import pytest -from six import PY2, ensure_text from .._env import ( E2E_FIXTURE_NAME, @@ -138,7 +136,7 @@ def dd_environment_runner(request): # Exit testing and pass data back up to command if E2E_RESULT_FILE in os.environ: with open(os.environ[E2E_RESULT_FILE], 'w', encoding='utf-8') as f: - f.write(ensure_text(json.dumps(data))) + f.write(json.dumps(data)) # Rather than exiting we skip every test to avoid the following output: # !!!!!!!!!! _pytest.outcomes.Exit: !!!!!!!!!! @@ -391,18 +389,6 @@ def pytest_configure(config): def pytest_addoption(parser): parser.addoption("--run-latest-metrics", action="store_true", default=False, help="run check_metrics tests") - if PY2: - # Add dummy memray options to make it possible to run memray with `ddev test --memray ` - # in both py2 and 3 environments. In py2 the option is simply ignored, see pytest_collection_modifyitems. - # In py3 the option enables the memray plugin. - parser.addoption("--memray", action="store_true", default=False, help="Dummy parameter for memray") - parser.addoption( - "--hide-memray-summary", - action="store_true", - default=False, - help="Dummy parameter for memray to hide the summary", - ) - def pytest_collection_modifyitems(config, items): # at test collection time, this function gets called by pytest, see: @@ -412,20 +398,12 @@ def pytest_collection_modifyitems(config, items): # --run-check-metrics given in cli: do not skip slow tests return - if PY2: - for option in ("--memray",): - if config.getoption(option): - warnings.warn( # noqa: B028 - "`{}` option ignored as it's not supported for py2 environments.".format(option) - ) # noqa: B028, E501 - skip_latest_metrics = pytest.mark.skip(reason="need --run-latest-metrics option to run") for item in items: if "latest_metrics" in item.keywords: item.add_marker(skip_latest_metrics) - # In Python 2 we're using a much older version of pytest where the Item interface is different. - item_path = item.fspath if PY2 else item.path + item_path = item.path if item_path is None: continue for ttype in TEST_TYPES: diff --git a/datadog_checks_dev/datadog_checks/dev/ssh_tunnel.py b/datadog_checks_dev/datadog_checks/dev/ssh_tunnel.py index 77f8f7449053a..b5340b09a7d11 100644 --- a/datadog_checks_dev/datadog_checks/dev/ssh_tunnel.py +++ b/datadog_checks_dev/datadog_checks/dev/ssh_tunnel.py @@ -4,21 +4,16 @@ from __future__ import absolute_import import os +import subprocess from contextlib import contextmanager import psutil -from six import PY3 from .conditions import WaitForPortListening from .env import environment_run from .structures import LazyFunction, TempDir from .utils import ON_WINDOWS, find_free_port, get_ip -if PY3: - import subprocess -else: - import subprocess32 as subprocess - PID_FILE = 'ssh.pid' diff --git a/datadog_checks_dev/datadog_checks/dev/subprocess.py b/datadog_checks_dev/datadog_checks/dev/subprocess.py index 01d3afe513459..9dd7aae87b2d8 100644 --- a/datadog_checks_dev/datadog_checks/dev/subprocess.py +++ b/datadog_checks_dev/datadog_checks/dev/subprocess.py @@ -8,8 +8,6 @@ from subprocess import Popen from tempfile import TemporaryFile -from six import string_types - from .errors import SubprocessError from .utils import NEED_SHELL, ON_WINDOWS, mock_context_manager @@ -42,7 +40,7 @@ def run_command(command, capture=None, check=False, encoding='utf-8', shell=Fals if shell == 'detect': shell = NEED_SHELL - if isinstance(command, string_types) and not ON_WINDOWS: + if isinstance(command, str) and not ON_WINDOWS: command = shlex.split(command) if capture: diff --git a/datadog_checks_dev/datadog_checks/dev/terraform.py b/datadog_checks_dev/datadog_checks/dev/terraform.py index 11a59fc8c23a5..7720b1a1d1702 100644 --- a/datadog_checks_dev/datadog_checks/dev/terraform.py +++ b/datadog_checks_dev/datadog_checks/dev/terraform.py @@ -8,18 +8,12 @@ from contextlib import contextmanager import pytest -from six import PY3 from .env import environment_run from .fs import chdir, copy_dir_contents, copy_path, get_here, path_join from .structures import LazyFunction, TempDir from .subprocess import run_command -if PY3: - from shutil import which -else: - from shutilwhich import which - TEMPLATES_DIR = path_join(get_here(), 'tooling', 'templates', 'terraform') @@ -57,7 +51,7 @@ def terraform_run(directory, sleep=None, endpoints=None, conditions=None, env_va wrappers (list[callable]): A list of context managers to use during execution """ - if not which('terraform'): + if not shutil.which('terraform'): pytest.skip('Terraform not available') set_up = TerraformUp(directory) diff --git a/datadog_checks_dev/datadog_checks/dev/testing.py b/datadog_checks_dev/datadog_checks/dev/testing.py index 090cff1c9296b..437e873aaccd2 100644 --- a/datadog_checks_dev/datadog_checks/dev/testing.py +++ b/datadog_checks_dev/datadog_checks/dev/testing.py @@ -5,13 +5,9 @@ """ This module contains test annotations """ import pytest -from six import PY2 from .utils import ON_MACOS, ON_WINDOWS requires_windows = pytest.mark.skipif(not ON_WINDOWS, reason='Requires Windows') requires_linux = pytest.mark.skipif(ON_MACOS or ON_WINDOWS, reason='Requires Linux') requires_unix = pytest.mark.skipif(ON_WINDOWS, reason='Requires Linux or macOS') - -requires_py2 = pytest.mark.skipif(not PY2, reason='Test only available on Python 2') -requires_py3 = pytest.mark.skipif(PY2, reason='Test only available on Python 3') diff --git a/datadog_checks_dev/datadog_checks/dev/utils.py b/datadog_checks_dev/datadog_checks/dev/utils.py index 71602de528ba4..d4ffd2ce66361 100644 --- a/datadog_checks_dev/datadog_checks/dev/utils.py +++ b/datadog_checks_dev/datadog_checks/dev/utils.py @@ -13,9 +13,9 @@ import socket from contextlib import closing, contextmanager from io import open +from urllib.request import urlopen import yaml -from six.moves.urllib.request import urlopen from .ci import running_on_ci, running_on_gh_actions, running_on_windows_ci # noqa: F401 from .fs import basepath, file_exists, get_parent_dir, path_join, read_file diff --git a/datadog_checks_dev/pyproject.toml b/datadog_checks_dev/pyproject.toml index e9e8f703c9281..9df85c415a22b 100644 --- a/datadog_checks_dev/pyproject.toml +++ b/datadog_checks_dev/pyproject.toml @@ -43,7 +43,6 @@ dependencies = [ "pytest-mock", "pyyaml>=5.4.1", "requests>=2.22.0", - "six", "tenacity", ] dynamic = [ diff --git a/dotnetclr/tests/e2e_test.py b/dotnetclr/tests/e2e_test.py index 11526f7935d7c..f75b1ab2ce7c4 100644 --- a/dotnetclr/tests/e2e_test.py +++ b/dotnetclr/tests/e2e_test.py @@ -3,15 +3,12 @@ # Licensed under a 3-clause BSD style license (see LICENSE) import pytest -from datadog_checks.dev.testing import requires_py3 - from .common import MINIMAL_INSTANCE pytestmark = pytest.mark.e2e @pytest.mark.flaky -@requires_py3 def test(dd_agent_check): from datadog_checks.dotnetclr.check import DotnetclrCheckV2 diff --git a/dotnetclr/tests/test_dotnetclr.py b/dotnetclr/tests/test_dotnetclr.py deleted file mode 100644 index 0bfd9dc8e063c..0000000000000 --- a/dotnetclr/tests/test_dotnetclr.py +++ /dev/null @@ -1,32 +0,0 @@ -# (C) Datadog, Inc. 2010-present -# All rights reserved -# Licensed under Simplified BSD License (see LICENSE) -import pytest -from datadog_test_libs.win.pdh_mocks import initialize_pdh_tests, pdh_mocks_fixture # noqa: F401 - -from datadog_checks.dev.testing import requires_py2 -from datadog_checks.dotnetclr import DotnetclrCheck -from datadog_checks.dotnetclr.dotnetclr import DEFAULT_COUNTERS - -from .common import INSTANCES, MINIMAL_INSTANCE - -pytestmark = [requires_py2, pytest.mark.usefixtures('pdh_mocks_fixture')] - - -@pytest.fixture(autouse=True) -def setup_check(): - initialize_pdh_tests() - - -@pytest.mark.integration -def test_basic_check(aggregator, dd_run_check): - instance = MINIMAL_INSTANCE - c = DotnetclrCheck('dotnetclr', {}, [instance]) - dd_run_check(c) - - for metric_def in DEFAULT_COUNTERS: - metric = metric_def[3] - for inst in INSTANCES: - aggregator.assert_metric(metric, tags=["instance:%s" % inst], count=1) - - assert aggregator.metrics_asserted_pct == 100.0 diff --git a/dotnetclr/tests/test_unit.py b/dotnetclr/tests/test_unit.py index 1b0ae0615793f..326a0a3fd67d0 100644 --- a/dotnetclr/tests/test_unit.py +++ b/dotnetclr/tests/test_unit.py @@ -2,15 +2,12 @@ # All rights reserved # Licensed under a 3-clause BSD style license (see LICENSE) from datadog_checks.base.constants import ServiceCheck -from datadog_checks.dev.testing import requires_py3 from datadog_checks.dev.utils import get_metadata_metrics from datadog_checks.dotnetclr import DotnetclrCheck from datadog_checks.dotnetclr.metrics import METRICS_CONFIG from .common import PERFORMANCE_OBJECTS -pytestmark = [requires_py3] - def test(aggregator, dd_default_hostname, dd_run_check, mock_performance_objects): mock_performance_objects(PERFORMANCE_OBJECTS) diff --git a/envoy/tests/test_unit.py b/envoy/tests/test_unit.py index 03e68cc15f498..dfa88c05efa24 100644 --- a/envoy/tests/test_unit.py +++ b/envoy/tests/test_unit.py @@ -4,8 +4,7 @@ import mock import pytest -from datadog_checks.base import AgentCheck, ConfigurationError -from datadog_checks.dev.testing import requires_py2, requires_py3 +from datadog_checks.base import AgentCheck from datadog_checks.dev.utils import get_metadata_metrics from datadog_checks.envoy.metrics import PROMETHEUS_METRICS_MAP @@ -35,13 +34,6 @@ def test_unique_metrics(): duplicated_metrics.add(value) -@requires_py2 -def test_check_with_py2(aggregator, dd_run_check, check, mock_http_response): - with pytest.raises(ConfigurationError, match="This version of the integration is only available when using py3."): - check(DEFAULT_INSTANCE) - - -@requires_py3 def test_check(aggregator, dd_run_check, check, mock_http_response): mock_http_response(file_path=get_fixture_path('./openmetrics/openmetrics.txt')) @@ -70,7 +62,6 @@ def test_check(aggregator, dd_run_check, check, mock_http_response): aggregator.assert_metrics_using_metadata(get_metadata_metrics()) -@requires_py3 def test_collect_metadata(datadog_agent, fixture_path, mock_http_response, check, default_instance): c = check(default_instance) c.check_id = 'test:123' @@ -92,7 +83,6 @@ def test_collect_metadata(datadog_agent, fixture_path, mock_http_response, check datadog_agent.assert_metadata_count(len(version_metadata)) -@requires_py3 def test_collect_metadata_with_invalid_base_url( datadog_agent, fixture_path, mock_http_response, check, default_instance ): @@ -106,7 +96,6 @@ def test_collect_metadata_with_invalid_base_url( c.log.debug.assert_called_with('Skipping server info collection due to malformed url: %s', b'') -@requires_py3 @pytest.mark.parametrize( 'fixture_file', [ @@ -139,7 +128,6 @@ def test_local_rate_limit_metrics(aggregator, dd_run_check, check, mock_http_res aggregator.assert_metrics_using_metadata(get_metadata_metrics()) -@requires_py3 def test_tags_in_ssl_metrics(aggregator, dd_run_check, check, mock_http_response): mock_http_response(file_path=get_fixture_path('./openmetrics/openmetrics_ssl_metrics.txt')) @@ -157,7 +145,6 @@ def test_tags_in_ssl_metrics(aggregator, dd_run_check, check, mock_http_response aggregator.assert_metrics_using_metadata(get_metadata_metrics()) -@requires_py3 def test_collect_metadata_with_disabled_collect_server_info( datadog_agent, fixture_path, mock_http_response, check, default_instance ): diff --git a/exchange_server/tests/test_e2e.py b/exchange_server/tests/test_e2e.py index a1f9e988c91d9..86f1713554682 100644 --- a/exchange_server/tests/test_e2e.py +++ b/exchange_server/tests/test_e2e.py @@ -5,7 +5,6 @@ import pytest from datadog_checks.base import AgentCheck -from datadog_checks.dev.testing import requires_py3 from datadog_checks.exchange_server.metrics import METRICS_CONFIG @@ -24,7 +23,6 @@ def get_exchange_server_metrics(): @pytest.mark.e2e -@requires_py3 def test_e2e_py3(dd_agent_check, aggregator, instance): aggregator = dd_agent_check(instance) aggregator.assert_service_check('exchange.windows.perf.health', AgentCheck.OK) diff --git a/exchange_server/tests/test_exchange_server.py b/exchange_server/tests/test_exchange_server.py deleted file mode 100644 index 28058889c5348..0000000000000 --- a/exchange_server/tests/test_exchange_server.py +++ /dev/null @@ -1,36 +0,0 @@ -# (C) Datadog, Inc. 2010-present -# All rights reserved -# Licensed under Simplified BSD License (see LICENSE) -import pytest -from datadog_test_libs.win.pdh_mocks import initialize_pdh_tests, pdh_mocks_fixture # noqa: F401 - -from datadog_checks.dev.testing import requires_py2 -from datadog_checks.exchange_server import ExchangeCheck -from datadog_checks.exchange_server.metrics import DEFAULT_COUNTERS - -from .common import CHECK_NAME, METRIC_INSTANCES, MINIMAL_INSTANCE - -pytestmark = [requires_py2, pytest.mark.usefixtures('pdh_mocks_fixture')] - - -@pytest.fixture(autouse=True) -def setup_check(): - initialize_pdh_tests() - - -@pytest.mark.integration -def test_basic_check(aggregator, dd_run_check): - instance = MINIMAL_INSTANCE - c = ExchangeCheck(CHECK_NAME, {}, [instance]) - dd_run_check(c) - - for metric_def in DEFAULT_COUNTERS: - metric = metric_def[3] - instances = METRIC_INSTANCES.get(metric) - if instances is not None: - for inst in instances: - aggregator.assert_metric(metric, tags=["instance:%s" % inst], count=1) - else: - aggregator.assert_metric(metric, tags=None, count=1) - - assert aggregator.metrics_asserted_pct == 100.0 diff --git a/exchange_server/tests/test_unit.py b/exchange_server/tests/test_unit.py index 472009f084ebf..cef790828bb29 100644 --- a/exchange_server/tests/test_unit.py +++ b/exchange_server/tests/test_unit.py @@ -2,15 +2,12 @@ # All rights reserved # Licensed under a 3-clause BSD style license (see LICENSE) from datadog_checks.base.constants import ServiceCheck -from datadog_checks.dev.testing import requires_py3 from datadog_checks.dev.utils import get_metadata_metrics from datadog_checks.exchange_server import ExchangeCheck from datadog_checks.exchange_server.metrics import METRICS_CONFIG from .common import PERFORMANCE_OBJECTS -pytestmark = [requires_py3] - def test(aggregator, dd_default_hostname, dd_run_check, mock_performance_objects): mock_performance_objects(PERFORMANCE_OBJECTS) diff --git a/gitlab/tests/test_integration.py b/gitlab/tests/test_integration.py index 0a1ade703cb80..5aa2531a27b1e 100644 --- a/gitlab/tests/test_integration.py +++ b/gitlab/tests/test_integration.py @@ -5,7 +5,6 @@ import pytest from requests.exceptions import ConnectionError -from datadog_checks.dev.testing import requires_py3 from datadog_checks.gitlab import GitlabCheck from .common import ( @@ -30,7 +29,6 @@ def test_check(dd_run_check, aggregator, gitlab_check, get_config, use_openmetri assert_check(aggregator, METRICS_TO_TEST_V2 if use_openmetrics else METRICS_TO_TEST, use_openmetrics) -@requires_py3 def test_check_gitaly(dd_run_check, aggregator, gitlab_check, get_config): from datadog_checks.gitlab.gitlab_v2 import GitlabCheckV2 @@ -70,7 +68,6 @@ def test_connection_failure(aggregator, gitlab_check, get_bad_config): ) -@requires_py3 def test_connection_failure_openmetrics(dd_run_check, aggregator, gitlab_check, get_bad_config): check = gitlab_check(get_bad_config(True)) diff --git a/gitlab/tests/test_unit.py b/gitlab/tests/test_unit.py index 8acbb2295edbf..b0b6bd8f12cab 100644 --- a/gitlab/tests/test_unit.py +++ b/gitlab/tests/test_unit.py @@ -4,8 +4,7 @@ import pytest from mock.mock import MagicMock -from datadog_checks.base import AgentCheck, ConfigurationError -from datadog_checks.dev.testing import requires_py2, requires_py3 +from datadog_checks.base import AgentCheck from datadog_checks.dev.utils import get_metadata_metrics from datadog_checks.gitlab.common import get_gitlab_version @@ -33,7 +32,6 @@ def test_check(dd_run_check, aggregator, mock_data, gitlab_check, get_config, us aggregator.assert_metrics_using_metadata(get_metadata_metrics()) -@requires_py3 def test_check_gitaly(dd_run_check, aggregator, mock_data, gitlab_check, get_config): from datadog_checks.gitlab.gitlab_v2 import GitlabCheckV2 @@ -56,14 +54,6 @@ def test_check_gitaly(dd_run_check, aggregator, mock_data, gitlab_check, get_con ) -@requires_py2 -def test_openmetrics_with_python2(gitlab_check, get_config): - with pytest.raises( - ConfigurationError, match="This version of the integration is only available when using Python 3." - ): - gitlab_check(get_config(True)) - - @pytest.mark.parametrize( "raw_version", [ @@ -88,7 +78,6 @@ def test_get_gitlab_version_without_token(): assert version is None -@requires_py3 def test_no_gitlab_url(dd_run_check, aggregator, mock_data, gitlab_check, get_config): config = get_config(True) del config['instances'][0]['gitlab_url'] @@ -97,7 +86,6 @@ def test_no_gitlab_url(dd_run_check, aggregator, mock_data, gitlab_check, get_co aggregator.assert_service_check('gitlab.openmetrics.health', status=AgentCheck.OK) -@requires_py3 def test_parse_readiness_service_checks_with_invalid_payload( dd_run_check, aggregator, mock_data, gitlab_check, get_config ): @@ -159,7 +147,6 @@ def test_parse_readiness_service_checks_with_invalid_payload( ), ], ) -@requires_py3 def test_parse_readiness_service_checks( dd_run_check, aggregator, mock_data, gitlab_check, get_config, service_check, expected_redis_status ): diff --git a/hyperv/tests/test_hyperv.py b/hyperv/tests/test_hyperv.py index 112b26f560b6f..6d9abfe749977 100644 --- a/hyperv/tests/test_hyperv.py +++ b/hyperv/tests/test_hyperv.py @@ -4,12 +4,10 @@ import pytest from datadog_checks.base.constants import ServiceCheck -from datadog_checks.dev.testing import requires_py2, requires_py3 from datadog_checks.hyperv import HypervCheck from datadog_checks.hyperv.metrics import DEFAULT_COUNTERS -@requires_py3 def test_check(aggregator, dd_default_hostname, dd_run_check): check = HypervCheck('hyperv', {}, [{}]) check.hostname = dd_default_hostname @@ -24,14 +22,6 @@ def test_check(aggregator, dd_default_hostname, dd_run_check): _assert_metrics(aggregator) -@requires_py2 -def test_check_legacy(aggregator, instance_refresh, dd_run_check): - check = HypervCheck('hyperv', {}, [instance_refresh]) - dd_run_check(check) - - _assert_metrics(aggregator) - - @pytest.mark.e2e def test_check_e2e(dd_agent_check, instance_refresh): aggregator = dd_agent_check(instance_refresh) diff --git a/ibm_mq/tests/test_ibm_mq_int.py b/ibm_mq/tests/test_ibm_mq_int.py index 72c2655eebc8e..51927607394ad 100644 --- a/ibm_mq/tests/test_ibm_mq_int.py +++ b/ibm_mq/tests/test_ibm_mq_int.py @@ -10,7 +10,6 @@ from datadog_checks.base import AgentCheck from datadog_checks.base.utils.time import ensure_aware_datetime -from datadog_checks.dev.testing import requires_py3 from datadog_checks.dev.utils import get_metadata_metrics from . import common @@ -412,7 +411,6 @@ def test_channel_status_no_duplicates(aggregator, get_check, instance, dd_run_ch aggregator.assert_service_check("ibm_mq.channel.status", check.OK, tags=tags, count=1) -@requires_py3 def test_queue_manager_process_not_found(aggregator, get_check, instance, dd_run_check): class ProcessMock(object): @property @@ -441,7 +439,6 @@ def info(self): aggregator.assert_all_metrics_covered() -@requires_py3 def test_queue_manager_process_found(aggregator, get_check, instance, dd_run_check): class ProcessMock(object): @property @@ -468,7 +465,6 @@ def info(self): assert_all_metrics(aggregator) -@requires_py3 def test_queue_manager_process_found_cleanup(get_check, instance, dd_run_check): class ProcessMock(object): @property diff --git a/ibm_mq/tests/test_ibm_mq_unit.py b/ibm_mq/tests/test_ibm_mq_unit.py index ac9f4ef87437c..9e1d7d4cf27be 100644 --- a/ibm_mq/tests/test_ibm_mq_unit.py +++ b/ibm_mq/tests/test_ibm_mq_unit.py @@ -5,7 +5,6 @@ import pytest from datadog_checks.base import AgentCheck, ConfigurationError -from datadog_checks.dev.testing import requires_py3 from .common import skip_windows_ci @@ -265,7 +264,6 @@ def test_ssl_check_normal_connection_before_ssl_connection(instance_ssl_dummy): get_ssl_connection.assert_called_with(config, logger) -@requires_py3 def test_queue_manager_process_direct_ssl(instance): from datadog_checks.ibm_mq.config import IBMMQConfig diff --git a/iis/tests/test_unit.py b/iis/tests/test_unit.py index bced243c6e474..1f7d4b855f98c 100644 --- a/iis/tests/test_unit.py +++ b/iis/tests/test_unit.py @@ -2,15 +2,12 @@ # All rights reserved # Licensed under a 3-clause BSD style license (see LICENSE) from datadog_checks.base.constants import ServiceCheck -from datadog_checks.dev.testing import requires_py3 from datadog_checks.dev.utils import get_metadata_metrics from datadog_checks.iis import IIS from datadog_checks.iis.service_check import IIS_APPLICATION_POOL_STATE from .common import DEFAULT_COUNTERS, PERFORMANCE_OBJECTS -pytestmark = [requires_py3] - def get_metrics_data(): app_pool_metrics_data = [] diff --git a/openstack/tests/test_openstack.py b/openstack/tests/test_openstack.py index b44a1dea155b4..aef61e93d9069 100644 --- a/openstack/tests/test_openstack.py +++ b/openstack/tests/test_openstack.py @@ -9,7 +9,6 @@ import pytest from datadog_checks.base import AgentCheck -from datadog_checks.dev.testing import requires_py3 from datadog_checks.openstack.openstack import ( IncompleteAuthScope, IncompleteConfig, @@ -302,7 +301,6 @@ def test_cache_between_runs(self, *args): assert 'server_newly_added' in cached_servers -@requires_py3 @pytest.mark.parametrize( 'init_config, instances, exception', [ diff --git a/tls/tests/test_remote.py b/tls/tests/test_remote.py index c2364f61fc129..bce911e5f3017 100644 --- a/tls/tests/test_remote.py +++ b/tls/tests/test_remote.py @@ -5,7 +5,6 @@ import pytest from datadog_checks.base import ConfigurationError -from datadog_checks.dev.testing import requires_py3 from datadog_checks.tls.const import ( SERVICE_CHECK_CAN_CONNECT, SERVICE_CHECK_EXPIRATION, @@ -338,7 +337,6 @@ def test_mysql_ok(aggregator, instance_remote_mysql_valid): aggregator.assert_all_metrics_covered() -@requires_py3 def test_valid_version_with_critical_certificate_validation_and_critial_certificate_expiration( aggregator, instance_remote_ok ): @@ -370,7 +368,6 @@ def test_valid_version_with_critical_certificate_validation_and_critial_certific aggregator.assert_all_metrics_covered() -@requires_py3 def test_valid_version_and_critical_certificate_validation_due_to_socket_exception(aggregator, instance_remote_ok): c = TLSCheck('tls', {}, [instance_remote_ok]) check = TLSRemoteCheck(agent_check=c) @@ -399,7 +396,6 @@ def test_valid_version_and_critical_certificate_validation_due_to_socket_excepti aggregator.assert_all_metrics_covered() -@requires_py3 def test_valid_version_and_critical_certificate_validation_due_to_parsing_error(aggregator, instance_remote_ok): c = TLSCheck('tls', {}, [instance_remote_ok]) check = TLSRemoteCheck(agent_check=c) From 03696677b7af9570228bb572f973117ff0375bcd Mon Sep 17 00:00:00 2001 From: Kyle Neale Date: Fri, 20 Sep 2024 15:45:48 -0400 Subject: [PATCH 25/34] Bump RethinkDB version for python 3.12 (#18636) * test e2e with docker build of Agent with py 3.12.6 * test rethinkdb bump * add fix for looseversion error * add changelog and revert agent-dev image tag --- agent_requirements.in | 2 +- rethinkdb/changelog.d/18636.added | 1 + rethinkdb/pyproject.toml | 2 +- 3 files changed, 3 insertions(+), 2 deletions(-) create mode 100644 rethinkdb/changelog.d/18636.added diff --git a/agent_requirements.in b/agent_requirements.in index 2d667d7ca21ea..9d96e0e3dbc26 100644 --- a/agent_requirements.in +++ b/agent_requirements.in @@ -59,7 +59,7 @@ requests-oauthlib==2.0.0 requests-toolbelt==1.0.0 requests-unixsocket2==0.4.2 requests==2.32.3 -rethinkdb==2.4.9 +rethinkdb==2.4.10post1 scandir==1.10.0; python_version < '3.5' securesystemslib[crypto,pynacl]==0.28.0 semver==3.0.2 diff --git a/rethinkdb/changelog.d/18636.added b/rethinkdb/changelog.d/18636.added new file mode 100644 index 0000000000000..3499902d53938 --- /dev/null +++ b/rethinkdb/changelog.d/18636.added @@ -0,0 +1 @@ +Bump RethinkDB version for py3.12 E2E tests \ No newline at end of file diff --git a/rethinkdb/pyproject.toml b/rethinkdb/pyproject.toml index 722a36b217430..04c879929a958 100644 --- a/rethinkdb/pyproject.toml +++ b/rethinkdb/pyproject.toml @@ -39,7 +39,7 @@ license = "BSD-3-Clause" [project.optional-dependencies] deps = [ - "rethinkdb==2.4.9", + "rethinkdb==2.4.10post1", ] [project.urls] From a47f5c618a4e34c855dd69d72baacf7375940d9b Mon Sep 17 00:00:00 2001 From: Kyle Neale Date: Fri, 20 Sep 2024 15:47:51 -0400 Subject: [PATCH 26/34] Bump the base check version for all integration (#18617) * bump the base check version for all integration * add missed integrations * add changelog * update base check to 37.0.0 * update changelog entries --- active_directory/changelog.d/18617.fixed | 1 + active_directory/pyproject.toml | 2 +- activemq/changelog.d/18617.fixed | 1 + activemq/pyproject.toml | 2 +- activemq_xml/changelog.d/18617.fixed | 1 + activemq_xml/pyproject.toml | 2 +- aerospike/changelog.d/18617.fixed | 1 + aerospike/pyproject.toml | 2 +- airflow/changelog.d/18617.fixed | 1 + airflow/pyproject.toml | 2 +- amazon_msk/changelog.d/18617.fixed | 1 + amazon_msk/pyproject.toml | 2 +- ambari/changelog.d/18617.fixed | 1 + ambari/pyproject.toml | 2 +- apache/changelog.d/18617.fixed | 1 + apache/pyproject.toml | 2 +- arangodb/changelog.d/18617.fixed | 1 + arangodb/pyproject.toml | 2 +- argo_rollouts/changelog.d/18617.fixed | 1 + argo_rollouts/pyproject.toml | 2 +- argo_workflows/changelog.d/18617.fixed | 1 + argo_workflows/pyproject.toml | 2 +- argocd/changelog.d/18617.fixed | 1 + argocd/pyproject.toml | 2 +- aspdotnet/changelog.d/18617.fixed | 1 + aspdotnet/pyproject.toml | 2 +- avi_vantage/changelog.d/18617.fixed | 1 + avi_vantage/pyproject.toml | 2 +- aws_neuron/changelog.d/18617.fixed | 1 + aws_neuron/pyproject.toml | 2 +- azure_iot_edge/changelog.d/18617.fixed | 1 + azure_iot_edge/pyproject.toml | 2 +- boundary/changelog.d/18617.fixed | 1 + boundary/pyproject.toml | 2 +- btrfs/changelog.d/18617.fixed | 1 + btrfs/pyproject.toml | 2 +- cacti/changelog.d/18617.fixed | 1 + cacti/pyproject.toml | 2 +- calico/changelog.d/18617.fixed | 1 + calico/pyproject.toml | 2 +- cassandra/changelog.d/18617.fixed | 1 + cassandra/pyproject.toml | 2 +- cassandra_nodetool/changelog.d/18617.fixed | 1 + cassandra_nodetool/pyproject.toml | 2 +- ceph/changelog.d/18617.fixed | 1 + ceph/pyproject.toml | 2 +- cert_manager/changelog.d/18617.fixed | 1 + cert_manager/pyproject.toml | 2 +- cilium/changelog.d/18617.fixed | 1 + cilium/pyproject.toml | 2 +- cisco_aci/changelog.d/18617.fixed | 1 + cisco_aci/pyproject.toml | 2 +- citrix_hypervisor/changelog.d/18617.fixed | 1 + citrix_hypervisor/pyproject.toml | 2 +- clickhouse/changelog.d/18617.fixed | 1 + clickhouse/pyproject.toml | 2 +- cloud_foundry_api/changelog.d/18617.fixed | 1 + cloud_foundry_api/pyproject.toml | 2 +- cloudera/changelog.d/18617.fixed | 1 + cloudera/pyproject.toml | 2 +- cockroachdb/changelog.d/18617.fixed | 1 + cockroachdb/pyproject.toml | 2 +- confluent_platform/changelog.d/18617.fixed | 1 + confluent_platform/pyproject.toml | 2 +- consul/changelog.d/18617.fixed | 1 + consul/pyproject.toml | 2 +- coredns/changelog.d/18617.fixed | 1 + coredns/pyproject.toml | 2 +- couch/changelog.d/18617.fixed | 1 + couch/pyproject.toml | 2 +- couchbase/changelog.d/18617.fixed | 1 + couchbase/pyproject.toml | 2 +- crio/changelog.d/18617.fixed | 1 + crio/pyproject.toml | 2 +- datadog_checks_dev/changelog.d/18617.fixed | 1 + .../templates/integration/check/{check_name}/pyproject.toml | 2 +- .../templates/integration/jmx/{check_name}/pyproject.toml | 2 +- datadog_cluster_agent/changelog.d/18617.fixed | 1 + datadog_cluster_agent/pyproject.toml | 2 +- dcgm/changelog.d/18617.fixed | 1 + dcgm/pyproject.toml | 2 +- directory/changelog.d/18617.fixed | 1 + directory/pyproject.toml | 2 +- disk/changelog.d/18617.fixed | 1 + disk/pyproject.toml | 2 +- dns_check/changelog.d/18617.fixed | 1 + dns_check/pyproject.toml | 2 +- dotnetclr/changelog.d/18617.fixed | 1 + dotnetclr/pyproject.toml | 2 +- druid/changelog.d/18617.fixed | 1 + druid/pyproject.toml | 2 +- ecs_fargate/changelog.d/18617.fixed | 1 + ecs_fargate/pyproject.toml | 2 +- eks_fargate/changelog.d/18617.fixed | 1 + eks_fargate/pyproject.toml | 2 +- elastic/changelog.d/18617.fixed | 1 + elastic/pyproject.toml | 2 +- envoy/changelog.d/18617.fixed | 1 + envoy/pyproject.toml | 2 +- etcd/changelog.d/18617.fixed | 1 + etcd/pyproject.toml | 2 +- exchange_server/changelog.d/18617.fixed | 1 + exchange_server/pyproject.toml | 2 +- external_dns/changelog.d/18617.fixed | 1 + external_dns/pyproject.toml | 2 +- fluentd/changelog.d/18617.fixed | 1 + fluentd/pyproject.toml | 2 +- fluxcd/changelog.d/18617.fixed | 1 + fluxcd/pyproject.toml | 2 +- fly_io/changelog.d/18617.fixed | 1 + fly_io/pyproject.toml | 2 +- foundationdb/changelog.d/18617.fixed | 1 + foundationdb/pyproject.toml | 2 +- gearmand/changelog.d/18617.fixed | 1 + gearmand/pyproject.toml | 2 +- gitlab/changelog.d/18617.fixed | 1 + gitlab/pyproject.toml | 2 +- gitlab_runner/changelog.d/18617.fixed | 1 + gitlab_runner/pyproject.toml | 2 +- glusterfs/changelog.d/18617.fixed | 1 + glusterfs/pyproject.toml | 2 +- go_expvar/changelog.d/18617.fixed | 1 + go_expvar/pyproject.toml | 2 +- gunicorn/changelog.d/18617.fixed | 1 + gunicorn/pyproject.toml | 2 +- haproxy/changelog.d/18617.fixed | 1 + haproxy/pyproject.toml | 2 +- harbor/changelog.d/18617.fixed | 1 + harbor/pyproject.toml | 2 +- hazelcast/changelog.d/18617.fixed | 1 + hazelcast/pyproject.toml | 2 +- hdfs_datanode/changelog.d/18617.fixed | 1 + hdfs_datanode/pyproject.toml | 2 +- hdfs_namenode/changelog.d/18617.fixed | 1 + hdfs_namenode/pyproject.toml | 2 +- hive/changelog.d/18617.fixed | 1 + hive/pyproject.toml | 2 +- hivemq/changelog.d/18617.fixed | 1 + hivemq/pyproject.toml | 2 +- http_check/changelog.d/18617.fixed | 1 + http_check/pyproject.toml | 2 +- hudi/changelog.d/18617.fixed | 1 + hudi/pyproject.toml | 2 +- hyperv/changelog.d/18617.fixed | 1 + hyperv/pyproject.toml | 2 +- ibm_ace/changelog.d/18617.fixed | 1 + ibm_ace/pyproject.toml | 2 +- ibm_db2/changelog.d/18617.fixed | 1 + ibm_db2/pyproject.toml | 2 +- ibm_i/changelog.d/18617.fixed | 1 + ibm_i/pyproject.toml | 2 +- ibm_mq/changelog.d/18617.fixed | 1 + ibm_mq/pyproject.toml | 2 +- ibm_was/changelog.d/18617.fixed | 1 + ibm_was/pyproject.toml | 2 +- ignite/changelog.d/18617.fixed | 1 + ignite/pyproject.toml | 2 +- iis/changelog.d/18617.fixed | 1 + iis/pyproject.toml | 2 +- impala/changelog.d/18617.fixed | 1 + impala/pyproject.toml | 2 +- istio/changelog.d/18617.fixed | 1 + istio/pyproject.toml | 2 +- jboss_wildfly/changelog.d/18617.fixed | 1 + jboss_wildfly/pyproject.toml | 2 +- kafka/changelog.d/18617.fixed | 1 + kafka/pyproject.toml | 2 +- kafka_consumer/changelog.d/18617.fixed | 1 + kafka_consumer/pyproject.toml | 2 +- karpenter/changelog.d/18617.fixed | 1 + karpenter/pyproject.toml | 2 +- kong/changelog.d/18617.fixed | 1 + kong/pyproject.toml | 2 +- kube_apiserver_metrics/changelog.d/18617.fixed | 1 + kube_apiserver_metrics/pyproject.toml | 2 +- kube_controller_manager/changelog.d/18617.fixed | 1 + kube_controller_manager/pyproject.toml | 2 +- kube_dns/changelog.d/18617.fixed | 1 + kube_dns/pyproject.toml | 2 +- kube_metrics_server/changelog.d/18617.fixed | 1 + kube_metrics_server/pyproject.toml | 2 +- kube_proxy/changelog.d/18617.fixed | 1 + kube_proxy/pyproject.toml | 2 +- kube_scheduler/changelog.d/18617.fixed | 1 + kube_scheduler/pyproject.toml | 2 +- kubeflow/changelog.d/18617.fixed | 1 + kubeflow/pyproject.toml | 2 +- kubernetes_cluster_autoscaler/changelog.d/18617.fixed | 1 + kubernetes_cluster_autoscaler/pyproject.toml | 2 +- kubernetes_state/changelog.d/18617.fixed | 1 + kubernetes_state/pyproject.toml | 2 +- kubevirt_api/changelog.d/18617.fixed | 1 + kubevirt_api/pyproject.toml | 2 +- kubevirt_controller/changelog.d/18617.fixed | 1 + kubevirt_controller/pyproject.toml | 2 +- kyototycoon/changelog.d/18617.fixed | 1 + kyototycoon/pyproject.toml | 2 +- kyverno/changelog.d/18617.fixed | 1 + kyverno/pyproject.toml | 2 +- lighttpd/changelog.d/18617.fixed | 1 + lighttpd/pyproject.toml | 2 +- linkerd/changelog.d/18617.fixed | 1 + linkerd/pyproject.toml | 2 +- linux_proc_extras/changelog.d/18617.fixed | 1 + linux_proc_extras/pyproject.toml | 2 +- mapr/changelog.d/18617.fixed | 1 + mapr/pyproject.toml | 2 +- mapreduce/changelog.d/18617.fixed | 1 + mapreduce/pyproject.toml | 2 +- marathon/changelog.d/18617.fixed | 1 + marathon/pyproject.toml | 2 +- marklogic/changelog.d/18617.fixed | 1 + marklogic/pyproject.toml | 2 +- mcache/changelog.d/18617.fixed | 1 + mcache/pyproject.toml | 2 +- mesos_master/changelog.d/18617.fixed | 1 + mesos_master/pyproject.toml | 2 +- mesos_slave/changelog.d/18617.fixed | 1 + mesos_slave/pyproject.toml | 2 +- nagios/changelog.d/18617.fixed | 1 + nagios/pyproject.toml | 2 +- network/changelog.d/18617.fixed | 1 + network/pyproject.toml | 2 +- nfsstat/changelog.d/18617.fixed | 1 + nfsstat/pyproject.toml | 2 +- nginx/changelog.d/18617.fixed | 1 + nginx/pyproject.toml | 2 +- nginx_ingress_controller/changelog.d/18617.fixed | 1 + nginx_ingress_controller/pyproject.toml | 2 +- nvidia_triton/changelog.d/18617.fixed | 1 + nvidia_triton/pyproject.toml | 2 +- openldap/changelog.d/18617.fixed | 1 + openldap/pyproject.toml | 2 +- openmetrics/changelog.d/18617.fixed | 1 + openmetrics/pyproject.toml | 2 +- openstack/changelog.d/18617.fixed | 1 + openstack/pyproject.toml | 2 +- openstack_controller/changelog.d/18617.fixed | 1 + openstack_controller/pyproject.toml | 2 +- pdh_check/changelog.d/18617.fixed | 1 + pdh_check/pyproject.toml | 2 +- pgbouncer/changelog.d/18617.fixed | 1 + pgbouncer/pyproject.toml | 2 +- php_fpm/changelog.d/18617.fixed | 1 + php_fpm/pyproject.toml | 2 +- postfix/changelog.d/18617.fixed | 1 + postfix/pyproject.toml | 2 +- powerdns_recursor/changelog.d/18617.fixed | 1 + powerdns_recursor/pyproject.toml | 2 +- presto/changelog.d/18617.fixed | 1 + presto/pyproject.toml | 2 +- process/changelog.d/18617.fixed | 1 + process/pyproject.toml | 2 +- prometheus/changelog.d/18617.fixed | 1 + prometheus/pyproject.toml | 2 +- proxysql/changelog.d/18617.fixed | 1 + proxysql/pyproject.toml | 2 +- pulsar/changelog.d/18617.fixed | 1 + pulsar/pyproject.toml | 2 +- rabbitmq/changelog.d/18617.fixed | 1 + rabbitmq/pyproject.toml | 2 +- ray/changelog.d/18617.fixed | 1 + ray/pyproject.toml | 2 +- redisdb/changelog.d/18617.fixed | 1 + redisdb/pyproject.toml | 2 +- rethinkdb/changelog.d/18617.fixed | 1 + rethinkdb/pyproject.toml | 2 +- riak/changelog.d/18617.fixed | 1 + riak/pyproject.toml | 2 +- riakcs/changelog.d/18617.fixed | 1 + riakcs/pyproject.toml | 2 +- scylla/changelog.d/18617.fixed | 1 + scylla/pyproject.toml | 2 +- silk/changelog.d/18617.fixed | 1 + silk/pyproject.toml | 2 +- singlestore/changelog.d/18617.fixed | 1 + singlestore/pyproject.toml | 2 +- snmp/changelog.d/18617.fixed | 1 + snmp/pyproject.toml | 2 +- snowflake/changelog.d/18617.fixed | 1 + snowflake/pyproject.toml | 2 +- solr/changelog.d/18617.fixed | 1 + solr/pyproject.toml | 2 +- sonarqube/changelog.d/18617.fixed | 1 + sonarqube/pyproject.toml | 2 +- spark/changelog.d/18617.fixed | 1 + spark/pyproject.toml | 2 +- sqlserver/changelog.d/18617.fixed | 1 + sqlserver/pyproject.toml | 2 +- squid/changelog.d/18617.fixed | 1 + squid/pyproject.toml | 2 +- ssh_check/changelog.d/18617.fixed | 1 + ssh_check/pyproject.toml | 2 +- statsd/changelog.d/18617.fixed | 1 + statsd/pyproject.toml | 2 +- strimzi/changelog.d/18617.fixed | 1 + strimzi/pyproject.toml | 2 +- supervisord/changelog.d/18617.fixed | 1 + supervisord/pyproject.toml | 2 +- system_core/changelog.d/18617.fixed | 1 + system_core/pyproject.toml | 2 +- system_swap/changelog.d/18617.fixed | 1 + system_swap/pyproject.toml | 2 +- tcp_check/changelog.d/18617.fixed | 1 + tcp_check/pyproject.toml | 2 +- teamcity/changelog.d/18617.fixed | 1 + teamcity/pyproject.toml | 2 +- tekton/changelog.d/18617.fixed | 1 + tekton/pyproject.toml | 2 +- teleport/changelog.d/18617.fixed | 1 + teleport/pyproject.toml | 2 +- temporal/changelog.d/18617.fixed | 1 + temporal/pyproject.toml | 2 +- teradata/changelog.d/18617.fixed | 1 + teradata/pyproject.toml | 2 +- tibco_ems/changelog.d/18617.fixed | 1 + tibco_ems/pyproject.toml | 2 +- tls/changelog.d/18617.fixed | 1 + tls/pyproject.toml | 2 +- tomcat/changelog.d/18617.fixed | 1 + tomcat/pyproject.toml | 2 +- torchserve/changelog.d/18617.fixed | 1 + torchserve/pyproject.toml | 2 +- traefik_mesh/changelog.d/18617.fixed | 1 + traefik_mesh/pyproject.toml | 2 +- traffic_server/changelog.d/18617.fixed | 1 + traffic_server/pyproject.toml | 2 +- twemproxy/changelog.d/18617.fixed | 1 + twemproxy/pyproject.toml | 2 +- twistlock/changelog.d/18617.fixed | 1 + twistlock/pyproject.toml | 2 +- varnish/changelog.d/18617.fixed | 1 + varnish/pyproject.toml | 2 +- vault/changelog.d/18617.fixed | 1 + vault/pyproject.toml | 2 +- vertica/changelog.d/18617.fixed | 1 + vertica/pyproject.toml | 2 +- vllm/changelog.d/18617.fixed | 1 + vllm/pyproject.toml | 2 +- voltdb/changelog.d/18617.fixed | 1 + voltdb/pyproject.toml | 2 +- vsphere/changelog.d/18617.fixed | 1 + vsphere/pyproject.toml | 2 +- weaviate/changelog.d/18617.fixed | 1 + weaviate/pyproject.toml | 2 +- weblogic/changelog.d/18617.fixed | 1 + weblogic/pyproject.toml | 2 +- win32_event_log/changelog.d/18617.fixed | 1 + win32_event_log/pyproject.toml | 2 +- windows_performance_counters/changelog.d/18617.fixed | 1 + windows_performance_counters/pyproject.toml | 2 +- windows_service/changelog.d/18617.fixed | 1 + windows_service/pyproject.toml | 2 +- wmi_check/changelog.d/18617.fixed | 1 + wmi_check/pyproject.toml | 2 +- yarn/changelog.d/18617.fixed | 1 + yarn/pyproject.toml | 2 +- zk/changelog.d/18617.fixed | 1 + zk/pyproject.toml | 2 +- 359 files changed, 359 insertions(+), 180 deletions(-) create mode 100644 active_directory/changelog.d/18617.fixed create mode 100644 activemq/changelog.d/18617.fixed create mode 100644 activemq_xml/changelog.d/18617.fixed create mode 100644 aerospike/changelog.d/18617.fixed create mode 100644 airflow/changelog.d/18617.fixed create mode 100644 amazon_msk/changelog.d/18617.fixed create mode 100644 ambari/changelog.d/18617.fixed create mode 100644 apache/changelog.d/18617.fixed create mode 100644 arangodb/changelog.d/18617.fixed create mode 100644 argo_rollouts/changelog.d/18617.fixed create mode 100644 argo_workflows/changelog.d/18617.fixed create mode 100644 argocd/changelog.d/18617.fixed create mode 100644 aspdotnet/changelog.d/18617.fixed create mode 100644 avi_vantage/changelog.d/18617.fixed create mode 100644 aws_neuron/changelog.d/18617.fixed create mode 100644 azure_iot_edge/changelog.d/18617.fixed create mode 100644 boundary/changelog.d/18617.fixed create mode 100644 btrfs/changelog.d/18617.fixed create mode 100644 cacti/changelog.d/18617.fixed create mode 100644 calico/changelog.d/18617.fixed create mode 100644 cassandra/changelog.d/18617.fixed create mode 100644 cassandra_nodetool/changelog.d/18617.fixed create mode 100644 ceph/changelog.d/18617.fixed create mode 100644 cert_manager/changelog.d/18617.fixed create mode 100644 cilium/changelog.d/18617.fixed create mode 100644 cisco_aci/changelog.d/18617.fixed create mode 100644 citrix_hypervisor/changelog.d/18617.fixed create mode 100644 clickhouse/changelog.d/18617.fixed create mode 100644 cloud_foundry_api/changelog.d/18617.fixed create mode 100644 cloudera/changelog.d/18617.fixed create mode 100644 cockroachdb/changelog.d/18617.fixed create mode 100644 confluent_platform/changelog.d/18617.fixed create mode 100644 consul/changelog.d/18617.fixed create mode 100644 coredns/changelog.d/18617.fixed create mode 100644 couch/changelog.d/18617.fixed create mode 100644 couchbase/changelog.d/18617.fixed create mode 100644 crio/changelog.d/18617.fixed create mode 100644 datadog_checks_dev/changelog.d/18617.fixed create mode 100644 datadog_cluster_agent/changelog.d/18617.fixed create mode 100644 dcgm/changelog.d/18617.fixed create mode 100644 directory/changelog.d/18617.fixed create mode 100644 disk/changelog.d/18617.fixed create mode 100644 dns_check/changelog.d/18617.fixed create mode 100644 dotnetclr/changelog.d/18617.fixed create mode 100644 druid/changelog.d/18617.fixed create mode 100644 ecs_fargate/changelog.d/18617.fixed create mode 100644 eks_fargate/changelog.d/18617.fixed create mode 100644 elastic/changelog.d/18617.fixed create mode 100644 envoy/changelog.d/18617.fixed create mode 100644 etcd/changelog.d/18617.fixed create mode 100644 exchange_server/changelog.d/18617.fixed create mode 100644 external_dns/changelog.d/18617.fixed create mode 100644 fluentd/changelog.d/18617.fixed create mode 100644 fluxcd/changelog.d/18617.fixed create mode 100644 fly_io/changelog.d/18617.fixed create mode 100644 foundationdb/changelog.d/18617.fixed create mode 100644 gearmand/changelog.d/18617.fixed create mode 100644 gitlab/changelog.d/18617.fixed create mode 100644 gitlab_runner/changelog.d/18617.fixed create mode 100644 glusterfs/changelog.d/18617.fixed create mode 100644 go_expvar/changelog.d/18617.fixed create mode 100644 gunicorn/changelog.d/18617.fixed create mode 100644 haproxy/changelog.d/18617.fixed create mode 100644 harbor/changelog.d/18617.fixed create mode 100644 hazelcast/changelog.d/18617.fixed create mode 100644 hdfs_datanode/changelog.d/18617.fixed create mode 100644 hdfs_namenode/changelog.d/18617.fixed create mode 100644 hive/changelog.d/18617.fixed create mode 100644 hivemq/changelog.d/18617.fixed create mode 100644 http_check/changelog.d/18617.fixed create mode 100644 hudi/changelog.d/18617.fixed create mode 100644 hyperv/changelog.d/18617.fixed create mode 100644 ibm_ace/changelog.d/18617.fixed create mode 100644 ibm_db2/changelog.d/18617.fixed create mode 100644 ibm_i/changelog.d/18617.fixed create mode 100644 ibm_mq/changelog.d/18617.fixed create mode 100644 ibm_was/changelog.d/18617.fixed create mode 100644 ignite/changelog.d/18617.fixed create mode 100644 iis/changelog.d/18617.fixed create mode 100644 impala/changelog.d/18617.fixed create mode 100644 istio/changelog.d/18617.fixed create mode 100644 jboss_wildfly/changelog.d/18617.fixed create mode 100644 kafka/changelog.d/18617.fixed create mode 100644 kafka_consumer/changelog.d/18617.fixed create mode 100644 karpenter/changelog.d/18617.fixed create mode 100644 kong/changelog.d/18617.fixed create mode 100644 kube_apiserver_metrics/changelog.d/18617.fixed create mode 100644 kube_controller_manager/changelog.d/18617.fixed create mode 100644 kube_dns/changelog.d/18617.fixed create mode 100644 kube_metrics_server/changelog.d/18617.fixed create mode 100644 kube_proxy/changelog.d/18617.fixed create mode 100644 kube_scheduler/changelog.d/18617.fixed create mode 100644 kubeflow/changelog.d/18617.fixed create mode 100644 kubernetes_cluster_autoscaler/changelog.d/18617.fixed create mode 100644 kubernetes_state/changelog.d/18617.fixed create mode 100644 kubevirt_api/changelog.d/18617.fixed create mode 100644 kubevirt_controller/changelog.d/18617.fixed create mode 100644 kyototycoon/changelog.d/18617.fixed create mode 100644 kyverno/changelog.d/18617.fixed create mode 100644 lighttpd/changelog.d/18617.fixed create mode 100644 linkerd/changelog.d/18617.fixed create mode 100644 linux_proc_extras/changelog.d/18617.fixed create mode 100644 mapr/changelog.d/18617.fixed create mode 100644 mapreduce/changelog.d/18617.fixed create mode 100644 marathon/changelog.d/18617.fixed create mode 100644 marklogic/changelog.d/18617.fixed create mode 100644 mcache/changelog.d/18617.fixed create mode 100644 mesos_master/changelog.d/18617.fixed create mode 100644 mesos_slave/changelog.d/18617.fixed create mode 100644 nagios/changelog.d/18617.fixed create mode 100644 network/changelog.d/18617.fixed create mode 100644 nfsstat/changelog.d/18617.fixed create mode 100644 nginx/changelog.d/18617.fixed create mode 100644 nginx_ingress_controller/changelog.d/18617.fixed create mode 100644 nvidia_triton/changelog.d/18617.fixed create mode 100644 openldap/changelog.d/18617.fixed create mode 100644 openmetrics/changelog.d/18617.fixed create mode 100644 openstack/changelog.d/18617.fixed create mode 100644 openstack_controller/changelog.d/18617.fixed create mode 100644 pdh_check/changelog.d/18617.fixed create mode 100644 pgbouncer/changelog.d/18617.fixed create mode 100644 php_fpm/changelog.d/18617.fixed create mode 100644 postfix/changelog.d/18617.fixed create mode 100644 powerdns_recursor/changelog.d/18617.fixed create mode 100644 presto/changelog.d/18617.fixed create mode 100644 process/changelog.d/18617.fixed create mode 100644 prometheus/changelog.d/18617.fixed create mode 100644 proxysql/changelog.d/18617.fixed create mode 100644 pulsar/changelog.d/18617.fixed create mode 100644 rabbitmq/changelog.d/18617.fixed create mode 100644 ray/changelog.d/18617.fixed create mode 100644 redisdb/changelog.d/18617.fixed create mode 100644 rethinkdb/changelog.d/18617.fixed create mode 100644 riak/changelog.d/18617.fixed create mode 100644 riakcs/changelog.d/18617.fixed create mode 100644 scylla/changelog.d/18617.fixed create mode 100644 silk/changelog.d/18617.fixed create mode 100644 singlestore/changelog.d/18617.fixed create mode 100644 snmp/changelog.d/18617.fixed create mode 100644 snowflake/changelog.d/18617.fixed create mode 100644 solr/changelog.d/18617.fixed create mode 100644 sonarqube/changelog.d/18617.fixed create mode 100644 spark/changelog.d/18617.fixed create mode 100644 sqlserver/changelog.d/18617.fixed create mode 100644 squid/changelog.d/18617.fixed create mode 100644 ssh_check/changelog.d/18617.fixed create mode 100644 statsd/changelog.d/18617.fixed create mode 100644 strimzi/changelog.d/18617.fixed create mode 100644 supervisord/changelog.d/18617.fixed create mode 100644 system_core/changelog.d/18617.fixed create mode 100644 system_swap/changelog.d/18617.fixed create mode 100644 tcp_check/changelog.d/18617.fixed create mode 100644 teamcity/changelog.d/18617.fixed create mode 100644 tekton/changelog.d/18617.fixed create mode 100644 teleport/changelog.d/18617.fixed create mode 100644 temporal/changelog.d/18617.fixed create mode 100644 teradata/changelog.d/18617.fixed create mode 100644 tibco_ems/changelog.d/18617.fixed create mode 100644 tls/changelog.d/18617.fixed create mode 100644 tomcat/changelog.d/18617.fixed create mode 100644 torchserve/changelog.d/18617.fixed create mode 100644 traefik_mesh/changelog.d/18617.fixed create mode 100644 traffic_server/changelog.d/18617.fixed create mode 100644 twemproxy/changelog.d/18617.fixed create mode 100644 twistlock/changelog.d/18617.fixed create mode 100644 varnish/changelog.d/18617.fixed create mode 100644 vault/changelog.d/18617.fixed create mode 100644 vertica/changelog.d/18617.fixed create mode 100644 vllm/changelog.d/18617.fixed create mode 100644 voltdb/changelog.d/18617.fixed create mode 100644 vsphere/changelog.d/18617.fixed create mode 100644 weaviate/changelog.d/18617.fixed create mode 100644 weblogic/changelog.d/18617.fixed create mode 100644 win32_event_log/changelog.d/18617.fixed create mode 100644 windows_performance_counters/changelog.d/18617.fixed create mode 100644 windows_service/changelog.d/18617.fixed create mode 100644 wmi_check/changelog.d/18617.fixed create mode 100644 yarn/changelog.d/18617.fixed create mode 100644 zk/changelog.d/18617.fixed diff --git a/active_directory/changelog.d/18617.fixed b/active_directory/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/active_directory/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/active_directory/pyproject.toml b/active_directory/pyproject.toml index e3a3d3611e863..897819c9b3f55 100644 --- a/active_directory/pyproject.toml +++ b/active_directory/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/activemq/changelog.d/18617.fixed b/activemq/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/activemq/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/activemq/pyproject.toml b/activemq/pyproject.toml index d2aad008b6ca4..23a51b653af42 100644 --- a/activemq/pyproject.toml +++ b/activemq/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/activemq_xml/changelog.d/18617.fixed b/activemq_xml/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/activemq_xml/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/activemq_xml/pyproject.toml b/activemq_xml/pyproject.toml index 3c23903fb08c9..35777f7bf4b37 100644 --- a/activemq_xml/pyproject.toml +++ b/activemq_xml/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/aerospike/changelog.d/18617.fixed b/aerospike/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/aerospike/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/aerospike/pyproject.toml b/aerospike/pyproject.toml index 78d14d3e5d834..22fdcebb95108 100644 --- a/aerospike/pyproject.toml +++ b/aerospike/pyproject.toml @@ -29,7 +29,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/airflow/changelog.d/18617.fixed b/airflow/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/airflow/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/airflow/pyproject.toml b/airflow/pyproject.toml index 9d883d928e13c..4127c4be4f5cb 100644 --- a/airflow/pyproject.toml +++ b/airflow/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/amazon_msk/changelog.d/18617.fixed b/amazon_msk/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/amazon_msk/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/amazon_msk/pyproject.toml b/amazon_msk/pyproject.toml index 475ce4e7031de..6c1142eaad6a9 100644 --- a/amazon_msk/pyproject.toml +++ b/amazon_msk/pyproject.toml @@ -29,7 +29,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/ambari/changelog.d/18617.fixed b/ambari/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/ambari/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/ambari/pyproject.toml b/ambari/pyproject.toml index 23c506e9b8a0c..68d75059d14e7 100644 --- a/ambari/pyproject.toml +++ b/ambari/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=34.1.2", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/apache/changelog.d/18617.fixed b/apache/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/apache/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/apache/pyproject.toml b/apache/pyproject.toml index c5eb9434f534e..6f7b6eacef9f5 100644 --- a/apache/pyproject.toml +++ b/apache/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/arangodb/changelog.d/18617.fixed b/arangodb/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/arangodb/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/arangodb/pyproject.toml b/arangodb/pyproject.toml index 58702df689362..76033616da230 100644 --- a/arangodb/pyproject.toml +++ b/arangodb/pyproject.toml @@ -29,7 +29,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/argo_rollouts/changelog.d/18617.fixed b/argo_rollouts/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/argo_rollouts/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/argo_rollouts/pyproject.toml b/argo_rollouts/pyproject.toml index d05e7a5c0c01b..8532164f09729 100644 --- a/argo_rollouts/pyproject.toml +++ b/argo_rollouts/pyproject.toml @@ -29,7 +29,7 @@ classifiers = [ "Topic :: System :: Monitoring", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/argo_workflows/changelog.d/18617.fixed b/argo_workflows/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/argo_workflows/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/argo_workflows/pyproject.toml b/argo_workflows/pyproject.toml index 42bcc168aba65..dd129976260ec 100644 --- a/argo_workflows/pyproject.toml +++ b/argo_workflows/pyproject.toml @@ -29,7 +29,7 @@ classifiers = [ "Topic :: System :: Monitoring", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/argocd/changelog.d/18617.fixed b/argocd/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/argocd/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/argocd/pyproject.toml b/argocd/pyproject.toml index ce1cd506cc3f2..f3521737068e3 100644 --- a/argocd/pyproject.toml +++ b/argocd/pyproject.toml @@ -28,7 +28,7 @@ classifiers = [ "Topic :: System :: Monitoring", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/aspdotnet/changelog.d/18617.fixed b/aspdotnet/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/aspdotnet/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/aspdotnet/pyproject.toml b/aspdotnet/pyproject.toml index c73f179dea07f..e2a037b894cf3 100644 --- a/aspdotnet/pyproject.toml +++ b/aspdotnet/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/avi_vantage/changelog.d/18617.fixed b/avi_vantage/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/avi_vantage/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/avi_vantage/pyproject.toml b/avi_vantage/pyproject.toml index 39fd1d7798e3b..0d43b5b8b82e6 100644 --- a/avi_vantage/pyproject.toml +++ b/avi_vantage/pyproject.toml @@ -31,7 +31,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/aws_neuron/changelog.d/18617.fixed b/aws_neuron/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/aws_neuron/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/aws_neuron/pyproject.toml b/aws_neuron/pyproject.toml index 4a7005002add4..76a420355e9ef 100644 --- a/aws_neuron/pyproject.toml +++ b/aws_neuron/pyproject.toml @@ -29,7 +29,7 @@ classifiers = [ "Topic :: System :: Monitoring", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/azure_iot_edge/changelog.d/18617.fixed b/azure_iot_edge/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/azure_iot_edge/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/azure_iot_edge/pyproject.toml b/azure_iot_edge/pyproject.toml index 0080f809a7865..448ac919b3f72 100644 --- a/azure_iot_edge/pyproject.toml +++ b/azure_iot_edge/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/boundary/changelog.d/18617.fixed b/boundary/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/boundary/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/boundary/pyproject.toml b/boundary/pyproject.toml index 1874f346a8e66..45a3d387402d1 100644 --- a/boundary/pyproject.toml +++ b/boundary/pyproject.toml @@ -27,7 +27,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/btrfs/changelog.d/18617.fixed b/btrfs/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/btrfs/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/btrfs/pyproject.toml b/btrfs/pyproject.toml index b357e8e8da0ee..c2fbea3190d49 100644 --- a/btrfs/pyproject.toml +++ b/btrfs/pyproject.toml @@ -29,7 +29,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/cacti/changelog.d/18617.fixed b/cacti/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/cacti/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/cacti/pyproject.toml b/cacti/pyproject.toml index 45f68faa8e788..7df50a14912a5 100644 --- a/cacti/pyproject.toml +++ b/cacti/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/calico/changelog.d/18617.fixed b/calico/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/calico/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/calico/pyproject.toml b/calico/pyproject.toml index 4854fe0f4590a..efa9d8085196f 100644 --- a/calico/pyproject.toml +++ b/calico/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/cassandra/changelog.d/18617.fixed b/cassandra/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/cassandra/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/cassandra/pyproject.toml b/cassandra/pyproject.toml index 1c6714ec2aafb..3f1e1714aa050 100644 --- a/cassandra/pyproject.toml +++ b/cassandra/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/cassandra_nodetool/changelog.d/18617.fixed b/cassandra_nodetool/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/cassandra_nodetool/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/cassandra_nodetool/pyproject.toml b/cassandra_nodetool/pyproject.toml index 4d41749f33d38..162d70377a05b 100644 --- a/cassandra_nodetool/pyproject.toml +++ b/cassandra_nodetool/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/ceph/changelog.d/18617.fixed b/ceph/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/ceph/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/ceph/pyproject.toml b/ceph/pyproject.toml index 97e3696c4c478..d9e145d6a6559 100644 --- a/ceph/pyproject.toml +++ b/ceph/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/cert_manager/changelog.d/18617.fixed b/cert_manager/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/cert_manager/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/cert_manager/pyproject.toml b/cert_manager/pyproject.toml index 3855c0dcf5a30..f7bf2d4176e4e 100644 --- a/cert_manager/pyproject.toml +++ b/cert_manager/pyproject.toml @@ -31,7 +31,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/cilium/changelog.d/18617.fixed b/cilium/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/cilium/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/cilium/pyproject.toml b/cilium/pyproject.toml index 4dd8108d9a079..1d24619b23c1e 100644 --- a/cilium/pyproject.toml +++ b/cilium/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/cisco_aci/changelog.d/18617.fixed b/cisco_aci/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/cisco_aci/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/cisco_aci/pyproject.toml b/cisco_aci/pyproject.toml index eb08546ee1fa2..f487215f43317 100644 --- a/cisco_aci/pyproject.toml +++ b/cisco_aci/pyproject.toml @@ -29,7 +29,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=36.10.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/citrix_hypervisor/changelog.d/18617.fixed b/citrix_hypervisor/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/citrix_hypervisor/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/citrix_hypervisor/pyproject.toml b/citrix_hypervisor/pyproject.toml index 584263112d453..749d4a01d6dd1 100644 --- a/citrix_hypervisor/pyproject.toml +++ b/citrix_hypervisor/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/clickhouse/changelog.d/18617.fixed b/clickhouse/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/clickhouse/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/clickhouse/pyproject.toml b/clickhouse/pyproject.toml index 0944ab13d5d9d..bbf7a3460a3aa 100644 --- a/clickhouse/pyproject.toml +++ b/clickhouse/pyproject.toml @@ -29,7 +29,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/cloud_foundry_api/changelog.d/18617.fixed b/cloud_foundry_api/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/cloud_foundry_api/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/cloud_foundry_api/pyproject.toml b/cloud_foundry_api/pyproject.toml index d1909b403b19d..cedda6fcccfe0 100644 --- a/cloud_foundry_api/pyproject.toml +++ b/cloud_foundry_api/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/cloudera/changelog.d/18617.fixed b/cloudera/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/cloudera/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/cloudera/pyproject.toml b/cloudera/pyproject.toml index da6cd2de5037e..2c13d3220d144 100644 --- a/cloudera/pyproject.toml +++ b/cloudera/pyproject.toml @@ -31,7 +31,7 @@ classifiers = [ "Topic :: System :: Monitoring", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/cockroachdb/changelog.d/18617.fixed b/cockroachdb/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/cockroachdb/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/cockroachdb/pyproject.toml b/cockroachdb/pyproject.toml index 9118478397f09..114e70ab2e47d 100644 --- a/cockroachdb/pyproject.toml +++ b/cockroachdb/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/confluent_platform/changelog.d/18617.fixed b/confluent_platform/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/confluent_platform/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/confluent_platform/pyproject.toml b/confluent_platform/pyproject.toml index 9f26a28d15eda..7071ace9f7e68 100644 --- a/confluent_platform/pyproject.toml +++ b/confluent_platform/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/consul/changelog.d/18617.fixed b/consul/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/consul/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/consul/pyproject.toml b/consul/pyproject.toml index 0573106958db0..fa62e679d923b 100644 --- a/consul/pyproject.toml +++ b/consul/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/coredns/changelog.d/18617.fixed b/coredns/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/coredns/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/coredns/pyproject.toml b/coredns/pyproject.toml index 0a768df59a1b6..740650969fa90 100644 --- a/coredns/pyproject.toml +++ b/coredns/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/couch/changelog.d/18617.fixed b/couch/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/couch/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/couch/pyproject.toml b/couch/pyproject.toml index 4b7ad5f700d06..fa027dcd0f838 100644 --- a/couch/pyproject.toml +++ b/couch/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/couchbase/changelog.d/18617.fixed b/couchbase/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/couchbase/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/couchbase/pyproject.toml b/couchbase/pyproject.toml index f1538ebe2d970..a69e6de25da1d 100644 --- a/couchbase/pyproject.toml +++ b/couchbase/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/crio/changelog.d/18617.fixed b/crio/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/crio/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/crio/pyproject.toml b/crio/pyproject.toml index f5a1bf83a5c00..f1ce75103ed2f 100644 --- a/crio/pyproject.toml +++ b/crio/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/datadog_checks_dev/changelog.d/18617.fixed b/datadog_checks_dev/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/datadog_checks_dev/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/datadog_checks_dev/datadog_checks/dev/tooling/templates/integration/check/{check_name}/pyproject.toml b/datadog_checks_dev/datadog_checks/dev/tooling/templates/integration/check/{check_name}/pyproject.toml index f64b51df1418b..d705fac17339e 100644 --- a/datadog_checks_dev/datadog_checks/dev/tooling/templates/integration/check/{check_name}/pyproject.toml +++ b/datadog_checks_dev/datadog_checks/dev/tooling/templates/integration/check/{check_name}/pyproject.toml @@ -29,7 +29,7 @@ classifiers = [ "Topic :: System :: Monitoring", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/datadog_checks_dev/datadog_checks/dev/tooling/templates/integration/jmx/{check_name}/pyproject.toml b/datadog_checks_dev/datadog_checks/dev/tooling/templates/integration/jmx/{check_name}/pyproject.toml index f64b51df1418b..d705fac17339e 100644 --- a/datadog_checks_dev/datadog_checks/dev/tooling/templates/integration/jmx/{check_name}/pyproject.toml +++ b/datadog_checks_dev/datadog_checks/dev/tooling/templates/integration/jmx/{check_name}/pyproject.toml @@ -29,7 +29,7 @@ classifiers = [ "Topic :: System :: Monitoring", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/datadog_cluster_agent/changelog.d/18617.fixed b/datadog_cluster_agent/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/datadog_cluster_agent/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/datadog_cluster_agent/pyproject.toml b/datadog_cluster_agent/pyproject.toml index 4592b159c015d..d5f3a5302a18f 100644 --- a/datadog_cluster_agent/pyproject.toml +++ b/datadog_cluster_agent/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/dcgm/changelog.d/18617.fixed b/dcgm/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/dcgm/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/dcgm/pyproject.toml b/dcgm/pyproject.toml index e17cc60b915a6..ce49134e9a5cd 100644 --- a/dcgm/pyproject.toml +++ b/dcgm/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Topic :: System :: Monitoring", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/directory/changelog.d/18617.fixed b/directory/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/directory/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/directory/pyproject.toml b/directory/pyproject.toml index 77e70ff27ced5..732fe2502167b 100644 --- a/directory/pyproject.toml +++ b/directory/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/disk/changelog.d/18617.fixed b/disk/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/disk/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/disk/pyproject.toml b/disk/pyproject.toml index 7e5ed5c3f73b6..8f115e41ca7e4 100644 --- a/disk/pyproject.toml +++ b/disk/pyproject.toml @@ -29,7 +29,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/dns_check/changelog.d/18617.fixed b/dns_check/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/dns_check/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/dns_check/pyproject.toml b/dns_check/pyproject.toml index afa020b28b601..b783ec388a574 100644 --- a/dns_check/pyproject.toml +++ b/dns_check/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/dotnetclr/changelog.d/18617.fixed b/dotnetclr/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/dotnetclr/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/dotnetclr/pyproject.toml b/dotnetclr/pyproject.toml index f46f5887ec543..0f1423ce59189 100644 --- a/dotnetclr/pyproject.toml +++ b/dotnetclr/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/druid/changelog.d/18617.fixed b/druid/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/druid/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/druid/pyproject.toml b/druid/pyproject.toml index ac0fce91714fa..3b4106a662c79 100644 --- a/druid/pyproject.toml +++ b/druid/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/ecs_fargate/changelog.d/18617.fixed b/ecs_fargate/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/ecs_fargate/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/ecs_fargate/pyproject.toml b/ecs_fargate/pyproject.toml index 602ff576e1419..74b73333b4314 100644 --- a/ecs_fargate/pyproject.toml +++ b/ecs_fargate/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/eks_fargate/changelog.d/18617.fixed b/eks_fargate/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/eks_fargate/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/eks_fargate/pyproject.toml b/eks_fargate/pyproject.toml index 03620d946d5fa..473c4623800e8 100644 --- a/eks_fargate/pyproject.toml +++ b/eks_fargate/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/elastic/changelog.d/18617.fixed b/elastic/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/elastic/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/elastic/pyproject.toml b/elastic/pyproject.toml index 4280b54f62c18..bbf130aa834ae 100644 --- a/elastic/pyproject.toml +++ b/elastic/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/envoy/changelog.d/18617.fixed b/envoy/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/envoy/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/envoy/pyproject.toml b/envoy/pyproject.toml index 1d549fcae090d..b4f0eeb46ade2 100644 --- a/envoy/pyproject.toml +++ b/envoy/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/etcd/changelog.d/18617.fixed b/etcd/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/etcd/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/etcd/pyproject.toml b/etcd/pyproject.toml index 1bfbef8e3ae09..b8c9ff21dd916 100644 --- a/etcd/pyproject.toml +++ b/etcd/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/exchange_server/changelog.d/18617.fixed b/exchange_server/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/exchange_server/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/exchange_server/pyproject.toml b/exchange_server/pyproject.toml index 88b90d1f9c0f5..ed6224db76ec4 100644 --- a/exchange_server/pyproject.toml +++ b/exchange_server/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/external_dns/changelog.d/18617.fixed b/external_dns/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/external_dns/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/external_dns/pyproject.toml b/external_dns/pyproject.toml index f6ad12bc045ba..4f602024103b9 100644 --- a/external_dns/pyproject.toml +++ b/external_dns/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/fluentd/changelog.d/18617.fixed b/fluentd/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/fluentd/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/fluentd/pyproject.toml b/fluentd/pyproject.toml index cf89e04ec7ed9..87b73f42a2b8a 100644 --- a/fluentd/pyproject.toml +++ b/fluentd/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/fluxcd/changelog.d/18617.fixed b/fluxcd/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/fluxcd/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/fluxcd/pyproject.toml b/fluxcd/pyproject.toml index d3a5dc43fad68..886e529d43283 100644 --- a/fluxcd/pyproject.toml +++ b/fluxcd/pyproject.toml @@ -28,7 +28,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/fly_io/changelog.d/18617.fixed b/fly_io/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/fly_io/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/fly_io/pyproject.toml b/fly_io/pyproject.toml index ae7927991c535..bb639c95b9d39 100644 --- a/fly_io/pyproject.toml +++ b/fly_io/pyproject.toml @@ -29,7 +29,7 @@ classifiers = [ "Topic :: System :: Monitoring", ] dependencies = [ - "datadog-checks-base>=36.11.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/foundationdb/changelog.d/18617.fixed b/foundationdb/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/foundationdb/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/foundationdb/pyproject.toml b/foundationdb/pyproject.toml index 852c3de459e5a..032ced7bbd166 100644 --- a/foundationdb/pyproject.toml +++ b/foundationdb/pyproject.toml @@ -29,7 +29,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/gearmand/changelog.d/18617.fixed b/gearmand/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/gearmand/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/gearmand/pyproject.toml b/gearmand/pyproject.toml index 182308e461433..ea6bc2cabdb37 100644 --- a/gearmand/pyproject.toml +++ b/gearmand/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/gitlab/changelog.d/18617.fixed b/gitlab/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/gitlab/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/gitlab/pyproject.toml b/gitlab/pyproject.toml index 7d201fcd21a28..ebbab4db5e734 100644 --- a/gitlab/pyproject.toml +++ b/gitlab/pyproject.toml @@ -29,7 +29,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/gitlab_runner/changelog.d/18617.fixed b/gitlab_runner/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/gitlab_runner/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/gitlab_runner/pyproject.toml b/gitlab_runner/pyproject.toml index ffd85f719b3bf..fdf0a099b5c8b 100644 --- a/gitlab_runner/pyproject.toml +++ b/gitlab_runner/pyproject.toml @@ -29,7 +29,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/glusterfs/changelog.d/18617.fixed b/glusterfs/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/glusterfs/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/glusterfs/pyproject.toml b/glusterfs/pyproject.toml index e70370c454c8e..cb6613ed73130 100644 --- a/glusterfs/pyproject.toml +++ b/glusterfs/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/go_expvar/changelog.d/18617.fixed b/go_expvar/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/go_expvar/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/go_expvar/pyproject.toml b/go_expvar/pyproject.toml index 0bdea0592f989..39309d2dae0fc 100644 --- a/go_expvar/pyproject.toml +++ b/go_expvar/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/gunicorn/changelog.d/18617.fixed b/gunicorn/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/gunicorn/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/gunicorn/pyproject.toml b/gunicorn/pyproject.toml index 3b01aaeedcb86..ebd7dc8cf89f6 100644 --- a/gunicorn/pyproject.toml +++ b/gunicorn/pyproject.toml @@ -29,7 +29,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/haproxy/changelog.d/18617.fixed b/haproxy/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/haproxy/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/haproxy/pyproject.toml b/haproxy/pyproject.toml index 8b78030486b59..21da244fd7919 100644 --- a/haproxy/pyproject.toml +++ b/haproxy/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/harbor/changelog.d/18617.fixed b/harbor/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/harbor/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/harbor/pyproject.toml b/harbor/pyproject.toml index c5220e2e61b2c..4657b2da91cc0 100644 --- a/harbor/pyproject.toml +++ b/harbor/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/hazelcast/changelog.d/18617.fixed b/hazelcast/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/hazelcast/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/hazelcast/pyproject.toml b/hazelcast/pyproject.toml index 7d7226fde345c..ebab01cbd6dfb 100644 --- a/hazelcast/pyproject.toml +++ b/hazelcast/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/hdfs_datanode/changelog.d/18617.fixed b/hdfs_datanode/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/hdfs_datanode/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/hdfs_datanode/pyproject.toml b/hdfs_datanode/pyproject.toml index e22554bcdc479..f14a69c03641d 100644 --- a/hdfs_datanode/pyproject.toml +++ b/hdfs_datanode/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/hdfs_namenode/changelog.d/18617.fixed b/hdfs_namenode/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/hdfs_namenode/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/hdfs_namenode/pyproject.toml b/hdfs_namenode/pyproject.toml index 9b56ef48480d6..0bf4d20e9d38f 100644 --- a/hdfs_namenode/pyproject.toml +++ b/hdfs_namenode/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/hive/changelog.d/18617.fixed b/hive/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/hive/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/hive/pyproject.toml b/hive/pyproject.toml index 4b29a04186362..ed12047b5fe46 100644 --- a/hive/pyproject.toml +++ b/hive/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/hivemq/changelog.d/18617.fixed b/hivemq/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/hivemq/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/hivemq/pyproject.toml b/hivemq/pyproject.toml index a55b56328dd44..9c988df9a7006 100644 --- a/hivemq/pyproject.toml +++ b/hivemq/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/http_check/changelog.d/18617.fixed b/http_check/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/http_check/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/http_check/pyproject.toml b/http_check/pyproject.toml index 078d51899ecaf..4ae432a765d77 100644 --- a/http_check/pyproject.toml +++ b/http_check/pyproject.toml @@ -29,7 +29,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/hudi/changelog.d/18617.fixed b/hudi/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/hudi/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/hudi/pyproject.toml b/hudi/pyproject.toml index 83e80c2170139..6c3e21b152f42 100644 --- a/hudi/pyproject.toml +++ b/hudi/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/hyperv/changelog.d/18617.fixed b/hyperv/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/hyperv/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/hyperv/pyproject.toml b/hyperv/pyproject.toml index 82f3a96acb8ac..dbfc0c79c7c8f 100644 --- a/hyperv/pyproject.toml +++ b/hyperv/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/ibm_ace/changelog.d/18617.fixed b/ibm_ace/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/ibm_ace/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/ibm_ace/pyproject.toml b/ibm_ace/pyproject.toml index b601bac23083d..7061024682916 100644 --- a/ibm_ace/pyproject.toml +++ b/ibm_ace/pyproject.toml @@ -28,7 +28,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/ibm_db2/changelog.d/18617.fixed b/ibm_db2/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/ibm_db2/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/ibm_db2/pyproject.toml b/ibm_db2/pyproject.toml index 2078c82f10c23..786fe2ef6ae6e 100644 --- a/ibm_db2/pyproject.toml +++ b/ibm_db2/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/ibm_i/changelog.d/18617.fixed b/ibm_i/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/ibm_i/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/ibm_i/pyproject.toml b/ibm_i/pyproject.toml index 104f0f2b53af7..dcc576ad1f146 100644 --- a/ibm_i/pyproject.toml +++ b/ibm_i/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/ibm_mq/changelog.d/18617.fixed b/ibm_mq/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/ibm_mq/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/ibm_mq/pyproject.toml b/ibm_mq/pyproject.toml index 3cd5ce9f36be8..8b6ad681af2e3 100644 --- a/ibm_mq/pyproject.toml +++ b/ibm_mq/pyproject.toml @@ -27,7 +27,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/ibm_was/changelog.d/18617.fixed b/ibm_was/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/ibm_was/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/ibm_was/pyproject.toml b/ibm_was/pyproject.toml index 59dfa8bd0dfd6..246db8ba8a3c8 100644 --- a/ibm_was/pyproject.toml +++ b/ibm_was/pyproject.toml @@ -29,7 +29,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/ignite/changelog.d/18617.fixed b/ignite/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/ignite/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/ignite/pyproject.toml b/ignite/pyproject.toml index c60bab739962f..68c263e8adaee 100644 --- a/ignite/pyproject.toml +++ b/ignite/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/iis/changelog.d/18617.fixed b/iis/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/iis/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/iis/pyproject.toml b/iis/pyproject.toml index 2606590ffc384..ffbe2341cad3a 100644 --- a/iis/pyproject.toml +++ b/iis/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/impala/changelog.d/18617.fixed b/impala/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/impala/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/impala/pyproject.toml b/impala/pyproject.toml index cfbea39777d50..63030092aa80d 100644 --- a/impala/pyproject.toml +++ b/impala/pyproject.toml @@ -31,7 +31,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/istio/changelog.d/18617.fixed b/istio/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/istio/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/istio/pyproject.toml b/istio/pyproject.toml index 7bbf30c2ef909..54b73e402d7a6 100644 --- a/istio/pyproject.toml +++ b/istio/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/jboss_wildfly/changelog.d/18617.fixed b/jboss_wildfly/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/jboss_wildfly/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/jboss_wildfly/pyproject.toml b/jboss_wildfly/pyproject.toml index 52c633037058f..f87450ffdd6c3 100644 --- a/jboss_wildfly/pyproject.toml +++ b/jboss_wildfly/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/kafka/changelog.d/18617.fixed b/kafka/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/kafka/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/kafka/pyproject.toml b/kafka/pyproject.toml index e3551f1dbcdcd..7f416f1d0026d 100644 --- a/kafka/pyproject.toml +++ b/kafka/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/kafka_consumer/changelog.d/18617.fixed b/kafka_consumer/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/kafka_consumer/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/kafka_consumer/pyproject.toml b/kafka_consumer/pyproject.toml index 49cf77e4a3aea..f6f15c1917fef 100644 --- a/kafka_consumer/pyproject.toml +++ b/kafka_consumer/pyproject.toml @@ -27,7 +27,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/karpenter/changelog.d/18617.fixed b/karpenter/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/karpenter/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/karpenter/pyproject.toml b/karpenter/pyproject.toml index 3b9ed9d4acad5..d08dc980aa6c9 100644 --- a/karpenter/pyproject.toml +++ b/karpenter/pyproject.toml @@ -29,7 +29,7 @@ classifiers = [ "Topic :: System :: Monitoring", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/kong/changelog.d/18617.fixed b/kong/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/kong/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/kong/pyproject.toml b/kong/pyproject.toml index a3e1d8f6a6b8c..d43e243e5e6dc 100644 --- a/kong/pyproject.toml +++ b/kong/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/kube_apiserver_metrics/changelog.d/18617.fixed b/kube_apiserver_metrics/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/kube_apiserver_metrics/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/kube_apiserver_metrics/pyproject.toml b/kube_apiserver_metrics/pyproject.toml index 42aa982f44cdb..6024af9b9d880 100644 --- a/kube_apiserver_metrics/pyproject.toml +++ b/kube_apiserver_metrics/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/kube_controller_manager/changelog.d/18617.fixed b/kube_controller_manager/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/kube_controller_manager/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/kube_controller_manager/pyproject.toml b/kube_controller_manager/pyproject.toml index 571bae1836952..1b59e30f28a47 100644 --- a/kube_controller_manager/pyproject.toml +++ b/kube_controller_manager/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/kube_dns/changelog.d/18617.fixed b/kube_dns/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/kube_dns/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/kube_dns/pyproject.toml b/kube_dns/pyproject.toml index 6c8c2b793f826..adca7847014e8 100644 --- a/kube_dns/pyproject.toml +++ b/kube_dns/pyproject.toml @@ -29,7 +29,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/kube_metrics_server/changelog.d/18617.fixed b/kube_metrics_server/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/kube_metrics_server/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/kube_metrics_server/pyproject.toml b/kube_metrics_server/pyproject.toml index e169008a48904..f95348d19b488 100644 --- a/kube_metrics_server/pyproject.toml +++ b/kube_metrics_server/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/kube_proxy/changelog.d/18617.fixed b/kube_proxy/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/kube_proxy/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/kube_proxy/pyproject.toml b/kube_proxy/pyproject.toml index 59b2b46a4ba28..60a2668b96fcb 100644 --- a/kube_proxy/pyproject.toml +++ b/kube_proxy/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/kube_scheduler/changelog.d/18617.fixed b/kube_scheduler/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/kube_scheduler/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/kube_scheduler/pyproject.toml b/kube_scheduler/pyproject.toml index 871ea3e106fcc..a3004a7f64d0b 100644 --- a/kube_scheduler/pyproject.toml +++ b/kube_scheduler/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/kubeflow/changelog.d/18617.fixed b/kubeflow/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/kubeflow/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/kubeflow/pyproject.toml b/kubeflow/pyproject.toml index 507e230c0b3f3..69588f9f49574 100644 --- a/kubeflow/pyproject.toml +++ b/kubeflow/pyproject.toml @@ -29,7 +29,7 @@ classifiers = [ "Topic :: System :: Monitoring", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/kubernetes_cluster_autoscaler/changelog.d/18617.fixed b/kubernetes_cluster_autoscaler/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/kubernetes_cluster_autoscaler/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/kubernetes_cluster_autoscaler/pyproject.toml b/kubernetes_cluster_autoscaler/pyproject.toml index b3d94f3a08466..d43eb3200ef03 100644 --- a/kubernetes_cluster_autoscaler/pyproject.toml +++ b/kubernetes_cluster_autoscaler/pyproject.toml @@ -29,7 +29,7 @@ classifiers = [ "Topic :: System :: Monitoring", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/kubernetes_state/changelog.d/18617.fixed b/kubernetes_state/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/kubernetes_state/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/kubernetes_state/pyproject.toml b/kubernetes_state/pyproject.toml index ece60796d07fb..0f0a249f53965 100644 --- a/kubernetes_state/pyproject.toml +++ b/kubernetes_state/pyproject.toml @@ -29,7 +29,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/kubevirt_api/changelog.d/18617.fixed b/kubevirt_api/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/kubevirt_api/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/kubevirt_api/pyproject.toml b/kubevirt_api/pyproject.toml index 5b09028db4589..b697f467c0cc0 100644 --- a/kubevirt_api/pyproject.toml +++ b/kubevirt_api/pyproject.toml @@ -29,7 +29,7 @@ classifiers = [ "Topic :: System :: Monitoring", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", "kubernetes>=30.1.0", ] dynamic = [ diff --git a/kubevirt_controller/changelog.d/18617.fixed b/kubevirt_controller/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/kubevirt_controller/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/kubevirt_controller/pyproject.toml b/kubevirt_controller/pyproject.toml index dfd3797339a24..646465aa9c826 100644 --- a/kubevirt_controller/pyproject.toml +++ b/kubevirt_controller/pyproject.toml @@ -29,7 +29,7 @@ classifiers = [ "Topic :: System :: Monitoring", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/kyototycoon/changelog.d/18617.fixed b/kyototycoon/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/kyototycoon/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/kyototycoon/pyproject.toml b/kyototycoon/pyproject.toml index 7fa95a604f50a..da1097d1def25 100644 --- a/kyototycoon/pyproject.toml +++ b/kyototycoon/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/kyverno/changelog.d/18617.fixed b/kyverno/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/kyverno/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/kyverno/pyproject.toml b/kyverno/pyproject.toml index 51079fea365e8..669ff11c84ec4 100644 --- a/kyverno/pyproject.toml +++ b/kyverno/pyproject.toml @@ -29,7 +29,7 @@ classifiers = [ "Topic :: System :: Monitoring", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/lighttpd/changelog.d/18617.fixed b/lighttpd/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/lighttpd/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/lighttpd/pyproject.toml b/lighttpd/pyproject.toml index ac8ff3cac5eb6..249c13c79d2e0 100644 --- a/lighttpd/pyproject.toml +++ b/lighttpd/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/linkerd/changelog.d/18617.fixed b/linkerd/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/linkerd/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/linkerd/pyproject.toml b/linkerd/pyproject.toml index 458a3e1fb10ac..df2c9fbee83e2 100644 --- a/linkerd/pyproject.toml +++ b/linkerd/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/linux_proc_extras/changelog.d/18617.fixed b/linux_proc_extras/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/linux_proc_extras/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/linux_proc_extras/pyproject.toml b/linux_proc_extras/pyproject.toml index 9d77452d0b67f..0c7e01fb6d18b 100644 --- a/linux_proc_extras/pyproject.toml +++ b/linux_proc_extras/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/mapr/changelog.d/18617.fixed b/mapr/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/mapr/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/mapr/pyproject.toml b/mapr/pyproject.toml index dd3c0e309f8f1..7367a7801b995 100644 --- a/mapr/pyproject.toml +++ b/mapr/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/mapreduce/changelog.d/18617.fixed b/mapreduce/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/mapreduce/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/mapreduce/pyproject.toml b/mapreduce/pyproject.toml index cf57dab669bf7..f5baba953f813 100644 --- a/mapreduce/pyproject.toml +++ b/mapreduce/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/marathon/changelog.d/18617.fixed b/marathon/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/marathon/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/marathon/pyproject.toml b/marathon/pyproject.toml index 0417f459c148c..67c8b66786a92 100644 --- a/marathon/pyproject.toml +++ b/marathon/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/marklogic/changelog.d/18617.fixed b/marklogic/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/marklogic/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/marklogic/pyproject.toml b/marklogic/pyproject.toml index 32a75ba8be794..3ed9fc2121461 100644 --- a/marklogic/pyproject.toml +++ b/marklogic/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/mcache/changelog.d/18617.fixed b/mcache/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/mcache/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/mcache/pyproject.toml b/mcache/pyproject.toml index 3416ceb26b7ab..6e151da6d698d 100644 --- a/mcache/pyproject.toml +++ b/mcache/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/mesos_master/changelog.d/18617.fixed b/mesos_master/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/mesos_master/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/mesos_master/pyproject.toml b/mesos_master/pyproject.toml index ceed88f8f8ba7..80335fc3bdd1f 100644 --- a/mesos_master/pyproject.toml +++ b/mesos_master/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/mesos_slave/changelog.d/18617.fixed b/mesos_slave/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/mesos_slave/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/mesos_slave/pyproject.toml b/mesos_slave/pyproject.toml index caa42505979eb..6367dd7e7163b 100644 --- a/mesos_slave/pyproject.toml +++ b/mesos_slave/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/nagios/changelog.d/18617.fixed b/nagios/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/nagios/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/nagios/pyproject.toml b/nagios/pyproject.toml index 559e5828704ce..5acab33b20ad2 100644 --- a/nagios/pyproject.toml +++ b/nagios/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/network/changelog.d/18617.fixed b/network/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/network/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/network/pyproject.toml b/network/pyproject.toml index dc9cc589fcbac..3615536bf265b 100644 --- a/network/pyproject.toml +++ b/network/pyproject.toml @@ -29,7 +29,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/nfsstat/changelog.d/18617.fixed b/nfsstat/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/nfsstat/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/nfsstat/pyproject.toml b/nfsstat/pyproject.toml index 63ba1349c8801..a6b6da09914c1 100644 --- a/nfsstat/pyproject.toml +++ b/nfsstat/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/nginx/changelog.d/18617.fixed b/nginx/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/nginx/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/nginx/pyproject.toml b/nginx/pyproject.toml index 739baaff89c9c..a6d4f91af4197 100644 --- a/nginx/pyproject.toml +++ b/nginx/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/nginx_ingress_controller/changelog.d/18617.fixed b/nginx_ingress_controller/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/nginx_ingress_controller/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/nginx_ingress_controller/pyproject.toml b/nginx_ingress_controller/pyproject.toml index 8fe0a7fdaea79..62b2b0ece04dc 100644 --- a/nginx_ingress_controller/pyproject.toml +++ b/nginx_ingress_controller/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/nvidia_triton/changelog.d/18617.fixed b/nvidia_triton/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/nvidia_triton/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/nvidia_triton/pyproject.toml b/nvidia_triton/pyproject.toml index 2803d702469d5..136b0847ff3bb 100644 --- a/nvidia_triton/pyproject.toml +++ b/nvidia_triton/pyproject.toml @@ -29,7 +29,7 @@ classifiers = [ "Topic :: System :: Monitoring", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/openldap/changelog.d/18617.fixed b/openldap/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/openldap/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/openldap/pyproject.toml b/openldap/pyproject.toml index 49683d15481fb..4c156c54c3cae 100644 --- a/openldap/pyproject.toml +++ b/openldap/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/openmetrics/changelog.d/18617.fixed b/openmetrics/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/openmetrics/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/openmetrics/pyproject.toml b/openmetrics/pyproject.toml index 3175dc1edc5a3..c53877f79980d 100644 --- a/openmetrics/pyproject.toml +++ b/openmetrics/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/openstack/changelog.d/18617.fixed b/openstack/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/openstack/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/openstack/pyproject.toml b/openstack/pyproject.toml index 7a8e5c4b24995..3fdeb185d079f 100644 --- a/openstack/pyproject.toml +++ b/openstack/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/openstack_controller/changelog.d/18617.fixed b/openstack_controller/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/openstack_controller/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/openstack_controller/pyproject.toml b/openstack_controller/pyproject.toml index 44382e19555e2..4bfba48253c9f 100644 --- a/openstack_controller/pyproject.toml +++ b/openstack_controller/pyproject.toml @@ -28,7 +28,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/pdh_check/changelog.d/18617.fixed b/pdh_check/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/pdh_check/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/pdh_check/pyproject.toml b/pdh_check/pyproject.toml index 998815602818c..cb3fb46601bfe 100644 --- a/pdh_check/pyproject.toml +++ b/pdh_check/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/pgbouncer/changelog.d/18617.fixed b/pgbouncer/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/pgbouncer/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/pgbouncer/pyproject.toml b/pgbouncer/pyproject.toml index a79a5e653d120..2a28b83831889 100644 --- a/pgbouncer/pyproject.toml +++ b/pgbouncer/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/php_fpm/changelog.d/18617.fixed b/php_fpm/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/php_fpm/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/php_fpm/pyproject.toml b/php_fpm/pyproject.toml index 959b853eab6bb..4ddf8d0171e18 100644 --- a/php_fpm/pyproject.toml +++ b/php_fpm/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/postfix/changelog.d/18617.fixed b/postfix/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/postfix/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/postfix/pyproject.toml b/postfix/pyproject.toml index c9ff6a26814ab..43b67a636f743 100644 --- a/postfix/pyproject.toml +++ b/postfix/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/powerdns_recursor/changelog.d/18617.fixed b/powerdns_recursor/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/powerdns_recursor/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/powerdns_recursor/pyproject.toml b/powerdns_recursor/pyproject.toml index 86d4a8e7397ca..4bb1f5ec606ea 100644 --- a/powerdns_recursor/pyproject.toml +++ b/powerdns_recursor/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/presto/changelog.d/18617.fixed b/presto/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/presto/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/presto/pyproject.toml b/presto/pyproject.toml index 6e385d08d26e4..b1759f4733e7d 100644 --- a/presto/pyproject.toml +++ b/presto/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/process/changelog.d/18617.fixed b/process/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/process/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/process/pyproject.toml b/process/pyproject.toml index c55c26760190c..6d603ff3a7279 100644 --- a/process/pyproject.toml +++ b/process/pyproject.toml @@ -29,7 +29,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/prometheus/changelog.d/18617.fixed b/prometheus/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/prometheus/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/prometheus/pyproject.toml b/prometheus/pyproject.toml index 410862fb39e9c..4cbfb5e16b974 100644 --- a/prometheus/pyproject.toml +++ b/prometheus/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/proxysql/changelog.d/18617.fixed b/proxysql/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/proxysql/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/proxysql/pyproject.toml b/proxysql/pyproject.toml index b01ff51923b82..1be7a7dcfa63f 100644 --- a/proxysql/pyproject.toml +++ b/proxysql/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/pulsar/changelog.d/18617.fixed b/pulsar/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/pulsar/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/pulsar/pyproject.toml b/pulsar/pyproject.toml index 51820fbd38004..3cd3e1b66f05d 100644 --- a/pulsar/pyproject.toml +++ b/pulsar/pyproject.toml @@ -29,7 +29,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/rabbitmq/changelog.d/18617.fixed b/rabbitmq/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/rabbitmq/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/rabbitmq/pyproject.toml b/rabbitmq/pyproject.toml index a88015878d478..b5452d3a95fa6 100644 --- a/rabbitmq/pyproject.toml +++ b/rabbitmq/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/ray/changelog.d/18617.fixed b/ray/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/ray/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/ray/pyproject.toml b/ray/pyproject.toml index 2fb0787081869..9fbe6102a5fbd 100644 --- a/ray/pyproject.toml +++ b/ray/pyproject.toml @@ -29,7 +29,7 @@ classifiers = [ "Topic :: System :: Monitoring", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/redisdb/changelog.d/18617.fixed b/redisdb/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/redisdb/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/redisdb/pyproject.toml b/redisdb/pyproject.toml index 357cff34f951f..0878b830417f4 100644 --- a/redisdb/pyproject.toml +++ b/redisdb/pyproject.toml @@ -29,7 +29,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/rethinkdb/changelog.d/18617.fixed b/rethinkdb/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/rethinkdb/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/rethinkdb/pyproject.toml b/rethinkdb/pyproject.toml index 04c879929a958..648f97b304b8d 100644 --- a/rethinkdb/pyproject.toml +++ b/rethinkdb/pyproject.toml @@ -29,7 +29,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/riak/changelog.d/18617.fixed b/riak/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/riak/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/riak/pyproject.toml b/riak/pyproject.toml index 5aea83a4b044c..78cdbca34de5c 100644 --- a/riak/pyproject.toml +++ b/riak/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/riakcs/changelog.d/18617.fixed b/riakcs/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/riakcs/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/riakcs/pyproject.toml b/riakcs/pyproject.toml index 2f90c1aa18457..3187d9cbd138e 100644 --- a/riakcs/pyproject.toml +++ b/riakcs/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/scylla/changelog.d/18617.fixed b/scylla/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/scylla/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/scylla/pyproject.toml b/scylla/pyproject.toml index 18be1cee6fab3..dd9e1663f20b1 100644 --- a/scylla/pyproject.toml +++ b/scylla/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/silk/changelog.d/18617.fixed b/silk/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/silk/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/silk/pyproject.toml b/silk/pyproject.toml index 02fbc039bf98c..356f6b8e1c3ee 100644 --- a/silk/pyproject.toml +++ b/silk/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/singlestore/changelog.d/18617.fixed b/singlestore/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/singlestore/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/singlestore/pyproject.toml b/singlestore/pyproject.toml index 1016e3715cf42..3c5e25300a652 100644 --- a/singlestore/pyproject.toml +++ b/singlestore/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/snmp/changelog.d/18617.fixed b/snmp/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/snmp/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/snmp/pyproject.toml b/snmp/pyproject.toml index 6ca2c91126604..cd6b1cbbab567 100644 --- a/snmp/pyproject.toml +++ b/snmp/pyproject.toml @@ -29,7 +29,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/snowflake/changelog.d/18617.fixed b/snowflake/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/snowflake/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/snowflake/pyproject.toml b/snowflake/pyproject.toml index 735f4c2543306..4e01028620a1d 100644 --- a/snowflake/pyproject.toml +++ b/snowflake/pyproject.toml @@ -28,7 +28,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=35.0.1", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/solr/changelog.d/18617.fixed b/solr/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/solr/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/solr/pyproject.toml b/solr/pyproject.toml index 7eb592a6226e8..2dddeb2ba076b 100644 --- a/solr/pyproject.toml +++ b/solr/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/sonarqube/changelog.d/18617.fixed b/sonarqube/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/sonarqube/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/sonarqube/pyproject.toml b/sonarqube/pyproject.toml index c1c3183c7a03b..d3b1f94b775f9 100644 --- a/sonarqube/pyproject.toml +++ b/sonarqube/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/spark/changelog.d/18617.fixed b/spark/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/spark/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/spark/pyproject.toml b/spark/pyproject.toml index cf6722c7ceb09..02b6f5d326e56 100644 --- a/spark/pyproject.toml +++ b/spark/pyproject.toml @@ -29,7 +29,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/sqlserver/changelog.d/18617.fixed b/sqlserver/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/sqlserver/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/sqlserver/pyproject.toml b/sqlserver/pyproject.toml index 18c28b1749927..33c2eab09f155 100644 --- a/sqlserver/pyproject.toml +++ b/sqlserver/pyproject.toml @@ -28,7 +28,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=36.8.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/squid/changelog.d/18617.fixed b/squid/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/squid/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/squid/pyproject.toml b/squid/pyproject.toml index 8f144cfa18e63..0023f6baed252 100644 --- a/squid/pyproject.toml +++ b/squid/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/ssh_check/changelog.d/18617.fixed b/ssh_check/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/ssh_check/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/ssh_check/pyproject.toml b/ssh_check/pyproject.toml index c0131118f6a21..767f7829e7e2a 100644 --- a/ssh_check/pyproject.toml +++ b/ssh_check/pyproject.toml @@ -29,7 +29,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/statsd/changelog.d/18617.fixed b/statsd/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/statsd/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/statsd/pyproject.toml b/statsd/pyproject.toml index c2639c6b9b4ef..c8506dba0449f 100644 --- a/statsd/pyproject.toml +++ b/statsd/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/strimzi/changelog.d/18617.fixed b/strimzi/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/strimzi/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/strimzi/pyproject.toml b/strimzi/pyproject.toml index c9a760e36dfa2..6e66eed9d976d 100644 --- a/strimzi/pyproject.toml +++ b/strimzi/pyproject.toml @@ -31,7 +31,7 @@ classifiers = [ "Topic :: System :: Monitoring", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/supervisord/changelog.d/18617.fixed b/supervisord/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/supervisord/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/supervisord/pyproject.toml b/supervisord/pyproject.toml index e7eab7fbad10a..cf12235d87f63 100644 --- a/supervisord/pyproject.toml +++ b/supervisord/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/system_core/changelog.d/18617.fixed b/system_core/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/system_core/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/system_core/pyproject.toml b/system_core/pyproject.toml index 0892c8bb8766c..798e7d43a5a85 100644 --- a/system_core/pyproject.toml +++ b/system_core/pyproject.toml @@ -29,7 +29,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/system_swap/changelog.d/18617.fixed b/system_swap/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/system_swap/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/system_swap/pyproject.toml b/system_swap/pyproject.toml index 6d867738a2022..33e6da5baa862 100644 --- a/system_swap/pyproject.toml +++ b/system_swap/pyproject.toml @@ -29,7 +29,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/tcp_check/changelog.d/18617.fixed b/tcp_check/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/tcp_check/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/tcp_check/pyproject.toml b/tcp_check/pyproject.toml index abbf435a04d73..bd25ea3c22089 100644 --- a/tcp_check/pyproject.toml +++ b/tcp_check/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/teamcity/changelog.d/18617.fixed b/teamcity/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/teamcity/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/teamcity/pyproject.toml b/teamcity/pyproject.toml index e235daa243e39..e4f599ec2385b 100644 --- a/teamcity/pyproject.toml +++ b/teamcity/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/tekton/changelog.d/18617.fixed b/tekton/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/tekton/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/tekton/pyproject.toml b/tekton/pyproject.toml index cb1bf1423018b..61ee8a0bb0a84 100644 --- a/tekton/pyproject.toml +++ b/tekton/pyproject.toml @@ -29,7 +29,7 @@ classifiers = [ "Topic :: System :: Monitoring", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/teleport/changelog.d/18617.fixed b/teleport/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/teleport/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/teleport/pyproject.toml b/teleport/pyproject.toml index 6572facaf8a00..c117db3839f9a 100644 --- a/teleport/pyproject.toml +++ b/teleport/pyproject.toml @@ -29,7 +29,7 @@ classifiers = [ "Topic :: System :: Monitoring", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/temporal/changelog.d/18617.fixed b/temporal/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/temporal/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/temporal/pyproject.toml b/temporal/pyproject.toml index b9f0768aa6151..a2677edcf0bfd 100644 --- a/temporal/pyproject.toml +++ b/temporal/pyproject.toml @@ -29,7 +29,7 @@ classifiers = [ "Topic :: System :: Monitoring", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/teradata/changelog.d/18617.fixed b/teradata/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/teradata/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/teradata/pyproject.toml b/teradata/pyproject.toml index e102a5fd8a90b..698c4c02562b6 100644 --- a/teradata/pyproject.toml +++ b/teradata/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/tibco_ems/changelog.d/18617.fixed b/tibco_ems/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/tibco_ems/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/tibco_ems/pyproject.toml b/tibco_ems/pyproject.toml index 282743eef2730..e6e3d10271066 100644 --- a/tibco_ems/pyproject.toml +++ b/tibco_ems/pyproject.toml @@ -29,7 +29,7 @@ classifiers = [ "Topic :: System :: Monitoring", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/tls/changelog.d/18617.fixed b/tls/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/tls/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/tls/pyproject.toml b/tls/pyproject.toml index 42a568fe730bd..2f2ed7f180b13 100644 --- a/tls/pyproject.toml +++ b/tls/pyproject.toml @@ -29,7 +29,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/tomcat/changelog.d/18617.fixed b/tomcat/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/tomcat/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/tomcat/pyproject.toml b/tomcat/pyproject.toml index 9f0c160029382..d751760e58a1b 100644 --- a/tomcat/pyproject.toml +++ b/tomcat/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/torchserve/changelog.d/18617.fixed b/torchserve/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/torchserve/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/torchserve/pyproject.toml b/torchserve/pyproject.toml index 66bd379468471..d1ebc2e01fe8f 100644 --- a/torchserve/pyproject.toml +++ b/torchserve/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Topic :: System :: Monitoring", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/traefik_mesh/changelog.d/18617.fixed b/traefik_mesh/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/traefik_mesh/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/traefik_mesh/pyproject.toml b/traefik_mesh/pyproject.toml index 69f2dfd774b3b..f8a53bd58cf32 100644 --- a/traefik_mesh/pyproject.toml +++ b/traefik_mesh/pyproject.toml @@ -29,7 +29,7 @@ classifiers = [ "Topic :: System :: Monitoring", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/traffic_server/changelog.d/18617.fixed b/traffic_server/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/traffic_server/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/traffic_server/pyproject.toml b/traffic_server/pyproject.toml index 67931b928ec4f..a1c76d325005c 100644 --- a/traffic_server/pyproject.toml +++ b/traffic_server/pyproject.toml @@ -29,7 +29,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/twemproxy/changelog.d/18617.fixed b/twemproxy/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/twemproxy/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/twemproxy/pyproject.toml b/twemproxy/pyproject.toml index 08cbf57491673..32b327e1dd8ad 100644 --- a/twemproxy/pyproject.toml +++ b/twemproxy/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/twistlock/changelog.d/18617.fixed b/twistlock/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/twistlock/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/twistlock/pyproject.toml b/twistlock/pyproject.toml index 0329130a09502..18b4af8f763cd 100644 --- a/twistlock/pyproject.toml +++ b/twistlock/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/varnish/changelog.d/18617.fixed b/varnish/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/varnish/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/varnish/pyproject.toml b/varnish/pyproject.toml index 4411d3db54f02..d7e4550113b7c 100644 --- a/varnish/pyproject.toml +++ b/varnish/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/vault/changelog.d/18617.fixed b/vault/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/vault/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/vault/pyproject.toml b/vault/pyproject.toml index 730bb5f9aca6c..ba1e84eecd9c0 100644 --- a/vault/pyproject.toml +++ b/vault/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/vertica/changelog.d/18617.fixed b/vertica/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/vertica/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/vertica/pyproject.toml b/vertica/pyproject.toml index bd239d161cf12..76b70944f84de 100644 --- a/vertica/pyproject.toml +++ b/vertica/pyproject.toml @@ -29,7 +29,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/vllm/changelog.d/18617.fixed b/vllm/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/vllm/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/vllm/pyproject.toml b/vllm/pyproject.toml index 5df58ba7e3c8e..db7b2f8e24472 100644 --- a/vllm/pyproject.toml +++ b/vllm/pyproject.toml @@ -29,7 +29,7 @@ classifiers = [ "Topic :: System :: Monitoring", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/voltdb/changelog.d/18617.fixed b/voltdb/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/voltdb/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/voltdb/pyproject.toml b/voltdb/pyproject.toml index 1e81ac3c6f957..cb31d181314b3 100644 --- a/voltdb/pyproject.toml +++ b/voltdb/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/vsphere/changelog.d/18617.fixed b/vsphere/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/vsphere/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/vsphere/pyproject.toml b/vsphere/pyproject.toml index 399ca557e460b..84a9a2494e5bb 100644 --- a/vsphere/pyproject.toml +++ b/vsphere/pyproject.toml @@ -29,7 +29,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/weaviate/changelog.d/18617.fixed b/weaviate/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/weaviate/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/weaviate/pyproject.toml b/weaviate/pyproject.toml index 3e5cf3924bb78..cc55f66e26560 100644 --- a/weaviate/pyproject.toml +++ b/weaviate/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Topic :: System :: Monitoring", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/weblogic/changelog.d/18617.fixed b/weblogic/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/weblogic/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/weblogic/pyproject.toml b/weblogic/pyproject.toml index d59c3481ab998..a7da09a3b783b 100644 --- a/weblogic/pyproject.toml +++ b/weblogic/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/win32_event_log/changelog.d/18617.fixed b/win32_event_log/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/win32_event_log/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/win32_event_log/pyproject.toml b/win32_event_log/pyproject.toml index 5405823ebfb06..0fcecb62bc12a 100644 --- a/win32_event_log/pyproject.toml +++ b/win32_event_log/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=34.2.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/windows_performance_counters/changelog.d/18617.fixed b/windows_performance_counters/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/windows_performance_counters/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/windows_performance_counters/pyproject.toml b/windows_performance_counters/pyproject.toml index 851a85f31d463..5a6f7bcbceea5 100644 --- a/windows_performance_counters/pyproject.toml +++ b/windows_performance_counters/pyproject.toml @@ -31,7 +31,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/windows_service/changelog.d/18617.fixed b/windows_service/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/windows_service/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/windows_service/pyproject.toml b/windows_service/pyproject.toml index b3d5f1be9560c..b8ef6e45c60d7 100644 --- a/windows_service/pyproject.toml +++ b/windows_service/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/wmi_check/changelog.d/18617.fixed b/wmi_check/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/wmi_check/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/wmi_check/pyproject.toml b/wmi_check/pyproject.toml index 9637eeb6797dd..107d53c91a506 100644 --- a/wmi_check/pyproject.toml +++ b/wmi_check/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/yarn/changelog.d/18617.fixed b/yarn/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/yarn/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/yarn/pyproject.toml b/yarn/pyproject.toml index 964786604a91e..1ea6d46ca022c 100644 --- a/yarn/pyproject.toml +++ b/yarn/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", diff --git a/zk/changelog.d/18617.fixed b/zk/changelog.d/18617.fixed new file mode 100644 index 0000000000000..e759164f82784 --- /dev/null +++ b/zk/changelog.d/18617.fixed @@ -0,0 +1 @@ +Bump the version of datadog-checks-base to 37.0.0 \ No newline at end of file diff --git a/zk/pyproject.toml b/zk/pyproject.toml index 8cc1312dac750..b4e47f914c8ca 100644 --- a/zk/pyproject.toml +++ b/zk/pyproject.toml @@ -30,7 +30,7 @@ classifiers = [ "Private :: Do Not Upload", ] dependencies = [ - "datadog-checks-base>=32.6.0", + "datadog-checks-base>=37.0.0", ] dynamic = [ "version", From b357ac2ecf9f59ae4a2497b32056dd959f1dfa9d Mon Sep 17 00:00:00 2001 From: Ilia Kurenkov Date: Fri, 20 Sep 2024 22:37:57 +0200 Subject: [PATCH 27/34] Remove 'six' usage from spark test (#18640) --- spark/tests/test_spark.py | 528 +++++++++++++++----------------------- 1 file changed, 207 insertions(+), 321 deletions(-) diff --git a/spark/tests/test_spark.py b/spark/tests/test_spark.py index 27ea4d3a65adf..37b06d8906d3a 100644 --- a/spark/tests/test_spark.py +++ b/spark/tests/test_spark.py @@ -13,7 +13,6 @@ import pytest import urllib3 from requests import RequestException -from six import iteritems from datadog_checks.dev.http import MockResponse from datadog_checks.dev.utils import get_metadata_metrics @@ -655,60 +654,50 @@ def proxy_with_warning_page_mock(url, *args, **kwargs): } +def _assert(aggregator, values_and_tags): + for m_vals, tags in values_and_tags: + for metric, value in m_vals.items(): + aggregator.assert_metric(metric, value=value, tags=tags) + + @pytest.mark.unit def test_yarn(aggregator, dd_run_check): with mock.patch('requests.get', yarn_requests_get_mock): c = SparkCheck('spark', {}, [YARN_CONFIG]) dd_run_check(c) - # Check the succeeded job metrics - for metric, value in iteritems(SPARK_JOB_SUCCEEDED_METRIC_VALUES): - aggregator.assert_metric(metric, value=value, tags=SPARK_JOB_SUCCEEDED_METRIC_TAGS + CUSTOM_TAGS) - - # Check the running stage metrics - for metric, value in iteritems(SPARK_STAGE_RUNNING_METRIC_VALUES): - aggregator.assert_metric(metric, value=value, tags=SPARK_STAGE_RUNNING_METRIC_TAGS + CUSTOM_TAGS) - - # Check the complete stage metrics - for metric, value in iteritems(SPARK_STAGE_COMPLETE_METRIC_VALUES): - aggregator.assert_metric(metric, value=value, tags=SPARK_STAGE_COMPLETE_METRIC_TAGS + CUSTOM_TAGS) - - # Check the driver metrics - for metric, value in iteritems(SPARK_DRIVER_METRIC_VALUES): - aggregator.assert_metric(metric, value=value, tags=COMMON_TAGS + CUSTOM_TAGS) - - # Check the optional driver metrics - for metric, value in iteritems(SPARK_DRIVER_OPTIONAL_METRIC_VALUES): - aggregator.assert_metric(metric, value=value, tags=COMMON_TAGS + CUSTOM_TAGS) - - # Check the executor level metrics - for metric, value in iteritems(SPARK_EXECUTOR_LEVEL_METRIC_VALUES): - aggregator.assert_metric(metric, value=value, tags=SPARK_EXECUTOR_LEVEL_METRIC_TAGS + CUSTOM_TAGS) - - # Check the optional executor level metrics - for metric, value in iteritems(SPARK_EXECUTOR_LEVEL_OPTIONAL_PROCESS_TREE_METRIC_VALUES): - aggregator.assert_metric(metric, value=value, tags=SPARK_EXECUTOR_LEVEL_METRIC_TAGS + CUSTOM_TAGS) - - # Check the summary executor metrics - for metric, value in iteritems(SPARK_EXECUTOR_METRIC_VALUES): - aggregator.assert_metric(metric, value=value, tags=COMMON_TAGS + CUSTOM_TAGS) - - # Check the optional summary executor metrics - for metric, value in iteritems(SPARK_EXECUTOR_OPTIONAL_METRIC_VALUES): - aggregator.assert_metric(metric, value=value, tags=COMMON_TAGS + CUSTOM_TAGS) - - # Check the RDD metrics - for metric, value in iteritems(SPARK_RDD_METRIC_VALUES): - aggregator.assert_metric(metric, value=value, tags=COMMON_TAGS + CUSTOM_TAGS) - - # Check the streaming statistics metrics - for metric, value in iteritems(SPARK_STREAMING_STATISTICS_METRIC_VALUES): - aggregator.assert_metric(metric, value=value, tags=COMMON_TAGS + CUSTOM_TAGS) - - # Check the structured streaming metrics - for metric, value in iteritems(SPARK_STRUCTURED_STREAMING_METRIC_VALUES): - aggregator.assert_metric(metric, value=value, tags=COMMON_TAGS + CUSTOM_TAGS) - + _assert( + aggregator, + [ + # Check the succeeded job metrics + (SPARK_JOB_SUCCEEDED_METRIC_VALUES, SPARK_JOB_SUCCEEDED_METRIC_TAGS + CUSTOM_TAGS), + # Check the running stage metrics + (SPARK_STAGE_RUNNING_METRIC_VALUES, SPARK_STAGE_RUNNING_METRIC_TAGS + CUSTOM_TAGS), + # Check the complete stage metrics + (SPARK_STAGE_COMPLETE_METRIC_VALUES, SPARK_STAGE_COMPLETE_METRIC_TAGS + CUSTOM_TAGS), + # Check the driver metrics + (SPARK_DRIVER_METRIC_VALUES, COMMON_TAGS + CUSTOM_TAGS), + # Check the optional driver metrics + (SPARK_DRIVER_OPTIONAL_METRIC_VALUES, COMMON_TAGS + CUSTOM_TAGS), + # Check the executor level metrics + (SPARK_EXECUTOR_LEVEL_METRIC_VALUES, SPARK_EXECUTOR_LEVEL_METRIC_TAGS + CUSTOM_TAGS), + # Check the optional executor level metrics + ( + SPARK_EXECUTOR_LEVEL_OPTIONAL_PROCESS_TREE_METRIC_VALUES, + SPARK_EXECUTOR_LEVEL_METRIC_TAGS + CUSTOM_TAGS, + ), + # Check the summary executor metrics + (SPARK_EXECUTOR_METRIC_VALUES, COMMON_TAGS + CUSTOM_TAGS), + # Check the optional summary executor metrics + (SPARK_EXECUTOR_OPTIONAL_METRIC_VALUES, COMMON_TAGS + CUSTOM_TAGS), + # Check the RDD metrics + (SPARK_RDD_METRIC_VALUES, COMMON_TAGS + CUSTOM_TAGS), + # Check the streaming statistics metrics + (SPARK_STREAMING_STATISTICS_METRIC_VALUES, COMMON_TAGS + CUSTOM_TAGS), + # Check the structured streaming metrics + (SPARK_STRUCTURED_STREAMING_METRIC_VALUES, COMMON_TAGS + CUSTOM_TAGS), + ], + ) tags = ['url:http://localhost:8088'] + CLUSTER_TAGS + CUSTOM_TAGS tags.sort() @@ -751,58 +740,40 @@ def test_mesos(aggregator, dd_run_check): with mock.patch('requests.get', mesos_requests_get_mock): c = SparkCheck('spark', {}, [MESOS_CONFIG]) dd_run_check(c) - # Check the running job metrics - for metric, value in iteritems(SPARK_JOB_RUNNING_METRIC_VALUES): - aggregator.assert_metric(metric, value=value, tags=SPARK_JOB_RUNNING_METRIC_TAGS + CUSTOM_TAGS) - - # Check the succeeded job metrics - for metric, value in iteritems(SPARK_JOB_SUCCEEDED_METRIC_VALUES): - aggregator.assert_metric(metric, value=value, tags=SPARK_JOB_SUCCEEDED_METRIC_TAGS + CUSTOM_TAGS) - - # Check the running stage metrics - for metric, value in iteritems(SPARK_STAGE_RUNNING_METRIC_VALUES): - aggregator.assert_metric(metric, value=value, tags=SPARK_STAGE_RUNNING_METRIC_TAGS + CUSTOM_TAGS) - - # Check the complete stage metrics - for metric, value in iteritems(SPARK_STAGE_COMPLETE_METRIC_VALUES): - aggregator.assert_metric(metric, value=value, tags=SPARK_STAGE_COMPLETE_METRIC_TAGS + CUSTOM_TAGS) - - # Check the driver metrics - for metric, value in iteritems(SPARK_DRIVER_METRIC_VALUES): - aggregator.assert_metric(metric, value=value, tags=COMMON_TAGS + CUSTOM_TAGS) - - # Check the optional driver metrics - for metric, value in iteritems(SPARK_DRIVER_OPTIONAL_METRIC_VALUES): - aggregator.assert_metric(metric, value=value, tags=COMMON_TAGS + CUSTOM_TAGS) - - # Check the executor level metrics - for metric, value in iteritems(SPARK_EXECUTOR_LEVEL_METRIC_VALUES): - aggregator.assert_metric(metric, value=value, tags=SPARK_EXECUTOR_LEVEL_METRIC_TAGS + CUSTOM_TAGS) - - # Check the optional executor level metrics - for metric, value in iteritems(SPARK_EXECUTOR_LEVEL_OPTIONAL_PROCESS_TREE_METRIC_VALUES): - aggregator.assert_metric(metric, value=value, tags=SPARK_EXECUTOR_LEVEL_METRIC_TAGS + CUSTOM_TAGS) - - # Check the summary executor metrics - for metric, value in iteritems(SPARK_EXECUTOR_METRIC_VALUES): - aggregator.assert_metric(metric, value=value, tags=COMMON_TAGS + CUSTOM_TAGS) - - # Check the optional summary executor metrics - for metric, value in iteritems(SPARK_EXECUTOR_OPTIONAL_METRIC_VALUES): - aggregator.assert_metric(metric, value=value, tags=COMMON_TAGS + CUSTOM_TAGS) - - # Check the RDD metrics - for metric, value in iteritems(SPARK_RDD_METRIC_VALUES): - aggregator.assert_metric(metric, value=value, tags=COMMON_TAGS + CUSTOM_TAGS) - - # Check the streaming statistics metrics - for metric, value in iteritems(SPARK_STREAMING_STATISTICS_METRIC_VALUES): - aggregator.assert_metric(metric, value=value, tags=COMMON_TAGS + CUSTOM_TAGS) - - # Check the structured streaming metrics - for metric, value in iteritems(SPARK_STRUCTURED_STREAMING_METRIC_VALUES): - aggregator.assert_metric(metric, value=value, tags=COMMON_TAGS + CUSTOM_TAGS) - + _assert( + aggregator, + [ + # Check the running job metrics + (SPARK_JOB_RUNNING_METRIC_VALUES, SPARK_JOB_RUNNING_METRIC_TAGS + CUSTOM_TAGS), + # Check the succeeded job metrics + (SPARK_JOB_SUCCEEDED_METRIC_VALUES, SPARK_JOB_SUCCEEDED_METRIC_TAGS + CUSTOM_TAGS), + # Check the running stage metrics + (SPARK_STAGE_RUNNING_METRIC_VALUES, SPARK_STAGE_RUNNING_METRIC_TAGS + CUSTOM_TAGS), + # Check the complete stage metrics + (SPARK_STAGE_COMPLETE_METRIC_VALUES, SPARK_STAGE_COMPLETE_METRIC_TAGS + CUSTOM_TAGS), + # Check the driver metrics + (SPARK_DRIVER_METRIC_VALUES, COMMON_TAGS + CUSTOM_TAGS), + # Check the optional driver metrics + (SPARK_DRIVER_OPTIONAL_METRIC_VALUES, COMMON_TAGS + CUSTOM_TAGS), + # Check the executor level metrics + (SPARK_EXECUTOR_LEVEL_METRIC_VALUES, SPARK_EXECUTOR_LEVEL_METRIC_TAGS + CUSTOM_TAGS), + # Check the optional executor level metrics + ( + SPARK_EXECUTOR_LEVEL_OPTIONAL_PROCESS_TREE_METRIC_VALUES, + SPARK_EXECUTOR_LEVEL_METRIC_TAGS + CUSTOM_TAGS, + ), + # Check the summary executor metrics + (SPARK_EXECUTOR_METRIC_VALUES, COMMON_TAGS + CUSTOM_TAGS), + # Check the optional summary executor metrics + (SPARK_EXECUTOR_OPTIONAL_METRIC_VALUES, COMMON_TAGS + CUSTOM_TAGS), + # Check the RDD metrics + (SPARK_RDD_METRIC_VALUES, COMMON_TAGS + CUSTOM_TAGS), + # Check the streaming statistics metrics, + (SPARK_STREAMING_STATISTICS_METRIC_VALUES, COMMON_TAGS + CUSTOM_TAGS), + # Check the structured streaming metrics + (SPARK_STRUCTURED_STREAMING_METRIC_VALUES, COMMON_TAGS + CUSTOM_TAGS), + ], + ) # Check the service tests for sc in aggregator.service_checks(MESOS_SERVICE_CHECK): @@ -842,58 +813,40 @@ def test_driver_unit(aggregator, dd_run_check): c = SparkCheck('spark', {}, [DRIVER_CONFIG]) dd_run_check(c) - # Check the running job metrics - for metric, value in iteritems(SPARK_JOB_RUNNING_METRIC_VALUES): - aggregator.assert_metric(metric, value=value, tags=SPARK_JOB_RUNNING_METRIC_TAGS + CUSTOM_TAGS) - - # Check the succeeded job metrics - for metric, value in iteritems(SPARK_JOB_SUCCEEDED_METRIC_VALUES): - aggregator.assert_metric(metric, value=value, tags=SPARK_JOB_SUCCEEDED_METRIC_TAGS + CUSTOM_TAGS) - - # Check the running stage metrics - for metric, value in iteritems(SPARK_STAGE_RUNNING_METRIC_VALUES): - aggregator.assert_metric(metric, value=value, tags=SPARK_STAGE_RUNNING_METRIC_TAGS + CUSTOM_TAGS) - - # Check the complete stage metrics - for metric, value in iteritems(SPARK_STAGE_COMPLETE_METRIC_VALUES): - aggregator.assert_metric(metric, value=value, tags=SPARK_STAGE_COMPLETE_METRIC_TAGS + CUSTOM_TAGS) - - # Check the driver metrics - for metric, value in iteritems(SPARK_DRIVER_METRIC_VALUES): - aggregator.assert_metric(metric, value=value, tags=COMMON_TAGS + CUSTOM_TAGS) - - # Check the optional driver metrics - for metric, value in iteritems(SPARK_DRIVER_OPTIONAL_METRIC_VALUES): - aggregator.assert_metric(metric, value=value, tags=COMMON_TAGS + CUSTOM_TAGS) - - # Check the executor level metrics - for metric, value in iteritems(SPARK_EXECUTOR_LEVEL_METRIC_VALUES): - aggregator.assert_metric(metric, value=value, tags=SPARK_EXECUTOR_LEVEL_METRIC_TAGS + CUSTOM_TAGS) - - # Check the optional executor level metrics - for metric, value in iteritems(SPARK_EXECUTOR_LEVEL_OPTIONAL_PROCESS_TREE_METRIC_VALUES): - aggregator.assert_metric(metric, value=value, tags=SPARK_EXECUTOR_LEVEL_METRIC_TAGS + CUSTOM_TAGS) - - # Check the summary executor metrics - for metric, value in iteritems(SPARK_EXECUTOR_METRIC_VALUES): - aggregator.assert_metric(metric, value=value, tags=COMMON_TAGS + CUSTOM_TAGS) - - # Check the optional summary executor metrics - for metric, value in iteritems(SPARK_EXECUTOR_OPTIONAL_METRIC_VALUES): - aggregator.assert_metric(metric, value=value, tags=COMMON_TAGS + CUSTOM_TAGS) - - # Check the RDD metrics - for metric, value in iteritems(SPARK_RDD_METRIC_VALUES): - aggregator.assert_metric(metric, value=value, tags=COMMON_TAGS + CUSTOM_TAGS) - - # Check the streaming statistics metrics - for metric, value in iteritems(SPARK_STREAMING_STATISTICS_METRIC_VALUES): - aggregator.assert_metric(metric, value=value, tags=COMMON_TAGS + CUSTOM_TAGS) - - # Check the structured streaming metrics - for metric, value in iteritems(SPARK_STRUCTURED_STREAMING_METRIC_VALUES): - aggregator.assert_metric(metric, value=value, tags=COMMON_TAGS + CUSTOM_TAGS) - + _assert( + aggregator, + [ + # Check the running job metrics + (SPARK_JOB_RUNNING_METRIC_VALUES, SPARK_JOB_RUNNING_METRIC_TAGS + CUSTOM_TAGS), + # Check the succeeded job metrics + (SPARK_JOB_SUCCEEDED_METRIC_VALUES, SPARK_JOB_SUCCEEDED_METRIC_TAGS + CUSTOM_TAGS), + # Check the running stage metrics + (SPARK_STAGE_RUNNING_METRIC_VALUES, SPARK_STAGE_RUNNING_METRIC_TAGS + CUSTOM_TAGS), + # Check the complete stage metrics + (SPARK_STAGE_COMPLETE_METRIC_VALUES, SPARK_STAGE_COMPLETE_METRIC_TAGS + CUSTOM_TAGS), + # Check the driver metrics + (SPARK_DRIVER_METRIC_VALUES, COMMON_TAGS + CUSTOM_TAGS), + # Check the optional driver metrics + (SPARK_DRIVER_OPTIONAL_METRIC_VALUES, COMMON_TAGS + CUSTOM_TAGS), + # Check the executor level metrics + (SPARK_EXECUTOR_LEVEL_METRIC_VALUES, SPARK_EXECUTOR_LEVEL_METRIC_TAGS + CUSTOM_TAGS), + # Check the optional executor level metrics + ( + SPARK_EXECUTOR_LEVEL_OPTIONAL_PROCESS_TREE_METRIC_VALUES, + SPARK_EXECUTOR_LEVEL_METRIC_TAGS + CUSTOM_TAGS, + ), + # Check the summary executor metrics + (SPARK_EXECUTOR_METRIC_VALUES, COMMON_TAGS + CUSTOM_TAGS), + # Check the optional summary executor metrics + (SPARK_EXECUTOR_OPTIONAL_METRIC_VALUES, COMMON_TAGS + CUSTOM_TAGS), + # Check the RDD metrics + (SPARK_RDD_METRIC_VALUES, COMMON_TAGS + CUSTOM_TAGS), + # Check the streaming statistics metrics + (SPARK_STREAMING_STATISTICS_METRIC_VALUES, COMMON_TAGS + CUSTOM_TAGS), + # Check the structured streaming metrics + (SPARK_STRUCTURED_STREAMING_METRIC_VALUES, COMMON_TAGS + CUSTOM_TAGS), + ], + ) # Check the service tests for sc in aggregator.service_checks(SPARK_DRIVER_SERVICE_CHECK): @@ -920,62 +873,39 @@ def test_standalone_unit(aggregator, dd_run_check): c = SparkCheck('spark', {}, [STANDALONE_CONFIG]) dd_run_check(c) - # Check the running job metrics - for metric, value in iteritems(SPARK_JOB_RUNNING_METRIC_VALUES): - aggregator.assert_metric(metric, value=value, tags=SPARK_JOB_RUNNING_METRIC_TAGS) - - # Check the running job metrics - for metric, value in iteritems(SPARK_JOB_RUNNING_METRIC_VALUES): - aggregator.assert_metric(metric, value=value, tags=SPARK_JOB_RUNNING_METRIC_TAGS) - - # Check the succeeded job metrics - for metric, value in iteritems(SPARK_JOB_SUCCEEDED_METRIC_VALUES): - aggregator.assert_metric(metric, value=value, tags=SPARK_JOB_SUCCEEDED_METRIC_TAGS) - - # Check the running stage metrics - for metric, value in iteritems(SPARK_STAGE_RUNNING_METRIC_VALUES): - aggregator.assert_metric(metric, value=value, tags=SPARK_STAGE_RUNNING_METRIC_TAGS) - - # Check the complete stage metrics - for metric, value in iteritems(SPARK_STAGE_COMPLETE_METRIC_VALUES): - aggregator.assert_metric(metric, value=value, tags=SPARK_STAGE_COMPLETE_METRIC_TAGS) - - # Check the driver metrics - for metric, value in iteritems(SPARK_DRIVER_METRIC_VALUES): - aggregator.assert_metric(metric, value=value, tags=COMMON_TAGS) - - # Check the optional driver metrics - for metric, value in iteritems(SPARK_DRIVER_OPTIONAL_METRIC_VALUES): - aggregator.assert_metric(metric, value=value, tags=COMMON_TAGS) - - # Check the executor level metrics - for metric, value in iteritems(SPARK_EXECUTOR_LEVEL_METRIC_VALUES): - aggregator.assert_metric(metric, value=value, tags=SPARK_EXECUTOR_LEVEL_METRIC_TAGS) - - # Check the optional executor level metrics - for metric, value in iteritems(SPARK_EXECUTOR_LEVEL_OPTIONAL_PROCESS_TREE_METRIC_VALUES): - aggregator.assert_metric(metric, value=value, tags=SPARK_EXECUTOR_LEVEL_METRIC_TAGS) - - # Check the executor metrics - for metric, value in iteritems(SPARK_EXECUTOR_METRIC_VALUES): - aggregator.assert_metric(metric, value=value, tags=COMMON_TAGS) - - # Check the optional summary executor metrics - for metric, value in iteritems(SPARK_EXECUTOR_OPTIONAL_METRIC_VALUES): - aggregator.assert_metric(metric, value=value, tags=COMMON_TAGS) - - # Check the RDD metrics - for metric, value in iteritems(SPARK_RDD_METRIC_VALUES): - aggregator.assert_metric(metric, value=value, tags=COMMON_TAGS) - - # Check the streaming statistics metrics - for metric, value in iteritems(SPARK_STREAMING_STATISTICS_METRIC_VALUES): - aggregator.assert_metric(metric, value=value, tags=COMMON_TAGS) - - # Check the structured streaming metrics - for metric, value in iteritems(SPARK_STRUCTURED_STREAMING_METRIC_VALUES): - aggregator.assert_metric(metric, value=value, tags=COMMON_TAGS) - + _assert( + aggregator, + [ + # Check the running job metrics + (SPARK_JOB_RUNNING_METRIC_VALUES, SPARK_JOB_RUNNING_METRIC_TAGS), + # Check the running job metrics + (SPARK_JOB_RUNNING_METRIC_VALUES, SPARK_JOB_RUNNING_METRIC_TAGS), + # Check the succeeded job metrics + (SPARK_JOB_SUCCEEDED_METRIC_VALUES, SPARK_JOB_SUCCEEDED_METRIC_TAGS), + # Check the running stage metrics + (SPARK_STAGE_RUNNING_METRIC_VALUES, SPARK_STAGE_RUNNING_METRIC_TAGS), + # Check the complete stage metrics + (SPARK_STAGE_COMPLETE_METRIC_VALUES, SPARK_STAGE_COMPLETE_METRIC_TAGS), + # Check the driver metrics + (SPARK_DRIVER_METRIC_VALUES, COMMON_TAGS), + # Check the optional driver metrics + (SPARK_DRIVER_OPTIONAL_METRIC_VALUES, COMMON_TAGS), + # Check the executor level metrics + (SPARK_EXECUTOR_LEVEL_METRIC_VALUES, SPARK_EXECUTOR_LEVEL_METRIC_TAGS), + # Check the optional executor level metrics + (SPARK_EXECUTOR_LEVEL_OPTIONAL_PROCESS_TREE_METRIC_VALUES, SPARK_EXECUTOR_LEVEL_METRIC_TAGS), + # Check the executor metrics + (SPARK_EXECUTOR_METRIC_VALUES, COMMON_TAGS), + # Check the optional summary executor metrics + (SPARK_EXECUTOR_OPTIONAL_METRIC_VALUES, COMMON_TAGS), + # Check the RDD metrics + (SPARK_RDD_METRIC_VALUES, COMMON_TAGS), + # Check the streaming statistics metrics + (SPARK_STREAMING_STATISTICS_METRIC_VALUES, COMMON_TAGS), + # Check the structured streaming metrics + (SPARK_STRUCTURED_STREAMING_METRIC_VALUES, COMMON_TAGS), + ], + ) # Check the service tests for sc in aggregator.service_checks(STANDALONE_SERVICE_CHECK): assert sc.status == SparkCheck.OK @@ -995,61 +925,39 @@ def test_standalone_unit_with_proxy_warning_page(aggregator, dd_run_check): with mock.patch('requests.get', proxy_with_warning_page_mock): dd_run_check(c) - # Check the running job metrics - for metric, value in iteritems(SPARK_JOB_RUNNING_METRIC_VALUES): - aggregator.assert_metric(metric, value=value, tags=SPARK_JOB_RUNNING_METRIC_TAGS) - - # Check the running job metrics - for metric, value in iteritems(SPARK_JOB_RUNNING_METRIC_VALUES): - aggregator.assert_metric(metric, value=value, tags=SPARK_JOB_RUNNING_METRIC_TAGS) - - # Check the succeeded job metrics - for metric, value in iteritems(SPARK_JOB_SUCCEEDED_METRIC_VALUES): - aggregator.assert_metric(metric, value=value, tags=SPARK_JOB_SUCCEEDED_METRIC_TAGS) - - # Check the running stage metrics - for metric, value in iteritems(SPARK_STAGE_RUNNING_METRIC_VALUES): - aggregator.assert_metric(metric, value=value, tags=SPARK_STAGE_RUNNING_METRIC_TAGS) - - # Check the complete stage metrics - for metric, value in iteritems(SPARK_STAGE_COMPLETE_METRIC_VALUES): - aggregator.assert_metric(metric, value=value, tags=SPARK_STAGE_COMPLETE_METRIC_TAGS) - - # Check the driver metrics - for metric, value in iteritems(SPARK_DRIVER_METRIC_VALUES): - aggregator.assert_metric(metric, value=value, tags=COMMON_TAGS) - - # Check the optional driver metrics - for metric, value in iteritems(SPARK_DRIVER_OPTIONAL_METRIC_VALUES): - aggregator.assert_metric(metric, value=value, tags=COMMON_TAGS) - - # Check the executor level metrics - for metric, value in iteritems(SPARK_EXECUTOR_LEVEL_METRIC_VALUES): - aggregator.assert_metric(metric, value=value, tags=SPARK_EXECUTOR_LEVEL_METRIC_TAGS) - - # Check the optional executor level metrics - for metric, value in iteritems(SPARK_EXECUTOR_LEVEL_OPTIONAL_PROCESS_TREE_METRIC_VALUES): - aggregator.assert_metric(metric, value=value, tags=SPARK_EXECUTOR_LEVEL_METRIC_TAGS) - - # Check the summary executor metrics - for metric, value in iteritems(SPARK_EXECUTOR_METRIC_VALUES): - aggregator.assert_metric(metric, value=value, tags=COMMON_TAGS) - - # Check the optional summary executor metrics - for metric, value in iteritems(SPARK_EXECUTOR_OPTIONAL_METRIC_VALUES): - aggregator.assert_metric(metric, value=value, tags=COMMON_TAGS) - - # Check the RDD metrics - for metric, value in iteritems(SPARK_RDD_METRIC_VALUES): - aggregator.assert_metric(metric, value=value, tags=COMMON_TAGS) - - # Check the streaming statistics metrics - for metric, value in iteritems(SPARK_STREAMING_STATISTICS_METRIC_VALUES): - aggregator.assert_metric(metric, value=value, tags=COMMON_TAGS) - - # Check the structured streaming metrics - for metric, value in iteritems(SPARK_STRUCTURED_STREAMING_METRIC_VALUES): - aggregator.assert_metric(metric, value=value, tags=COMMON_TAGS) + _assert( + aggregator, + [ + # Check the running job metrics + (SPARK_JOB_RUNNING_METRIC_VALUES, SPARK_JOB_RUNNING_METRIC_TAGS), + # Check the running job metrics + (SPARK_JOB_RUNNING_METRIC_VALUES, SPARK_JOB_RUNNING_METRIC_TAGS), + # Check the succeeded job metrics + (SPARK_JOB_SUCCEEDED_METRIC_VALUES, SPARK_JOB_SUCCEEDED_METRIC_TAGS), + # Check the running stage metrics + (SPARK_STAGE_RUNNING_METRIC_VALUES, SPARK_STAGE_RUNNING_METRIC_TAGS), + # Check the complete stage metrics + (SPARK_STAGE_COMPLETE_METRIC_VALUES, SPARK_STAGE_COMPLETE_METRIC_TAGS), + # Check the driver metrics + (SPARK_DRIVER_METRIC_VALUES, COMMON_TAGS), + # Check the optional driver metrics + (SPARK_DRIVER_OPTIONAL_METRIC_VALUES, COMMON_TAGS), + # Check the executor level metrics + (SPARK_EXECUTOR_LEVEL_METRIC_VALUES, SPARK_EXECUTOR_LEVEL_METRIC_TAGS), + # Check the optional executor level metrics + (SPARK_EXECUTOR_LEVEL_OPTIONAL_PROCESS_TREE_METRIC_VALUES, SPARK_EXECUTOR_LEVEL_METRIC_TAGS), + # Check the summary executor metrics + (SPARK_EXECUTOR_METRIC_VALUES, COMMON_TAGS), + # Check the optional summary executor metrics + (SPARK_EXECUTOR_OPTIONAL_METRIC_VALUES, COMMON_TAGS), + # Check the RDD metrics + (SPARK_RDD_METRIC_VALUES, COMMON_TAGS), + # Check the streaming statistics metrics + (SPARK_STREAMING_STATISTICS_METRIC_VALUES, COMMON_TAGS), + # Check the structured streaming metrics + (SPARK_STRUCTURED_STREAMING_METRIC_VALUES, COMMON_TAGS), + ], + ) # Check the service tests for sc in aggregator.service_checks(STANDALONE_SERVICE_CHECK): @@ -1070,61 +978,39 @@ def test_standalone_pre20(aggregator, dd_run_check): c = SparkCheck('spark', {}, [STANDALONE_CONFIG_PRE_20]) dd_run_check(c) - # Check the running job metrics - for metric, value in iteritems(SPARK_JOB_RUNNING_METRIC_VALUES): - aggregator.assert_metric(metric, value=value, tags=SPARK_JOB_RUNNING_METRIC_TAGS) - - # Check the running job metrics - for metric, value in iteritems(SPARK_JOB_RUNNING_METRIC_VALUES): - aggregator.assert_metric(metric, value=value, tags=SPARK_JOB_RUNNING_METRIC_TAGS) - - # Check the succeeded job metrics - for metric, value in iteritems(SPARK_JOB_SUCCEEDED_METRIC_VALUES): - aggregator.assert_metric(metric, value=value, tags=SPARK_JOB_SUCCEEDED_METRIC_TAGS) - - # Check the running stage metrics - for metric, value in iteritems(SPARK_STAGE_RUNNING_METRIC_VALUES): - aggregator.assert_metric(metric, value=value, tags=SPARK_STAGE_RUNNING_METRIC_TAGS) - - # Check the complete stage metrics - for metric, value in iteritems(SPARK_STAGE_COMPLETE_METRIC_VALUES): - aggregator.assert_metric(metric, value=value, tags=SPARK_STAGE_COMPLETE_METRIC_TAGS) - - # Check the driver metrics - for metric, value in iteritems(SPARK_DRIVER_METRIC_VALUES): - aggregator.assert_metric(metric, value=value, tags=COMMON_TAGS) - - # Check the optional driver metrics - for metric, value in iteritems(SPARK_DRIVER_OPTIONAL_METRIC_VALUES): - aggregator.assert_metric(metric, value=value, tags=COMMON_TAGS) - - # Check the executor level metrics - for metric, value in iteritems(SPARK_EXECUTOR_LEVEL_METRIC_VALUES): - aggregator.assert_metric(metric, value=value, tags=SPARK_EXECUTOR_LEVEL_METRIC_TAGS) - - # Check the optional executor level metrics - for metric, value in iteritems(SPARK_EXECUTOR_LEVEL_OPTIONAL_PROCESS_TREE_METRIC_VALUES): - aggregator.assert_metric(metric, value=value, tags=SPARK_EXECUTOR_LEVEL_METRIC_TAGS) - - # Check the summary executor metrics - for metric, value in iteritems(SPARK_EXECUTOR_METRIC_VALUES): - aggregator.assert_metric(metric, value=value, tags=COMMON_TAGS) - - # Check the optional summary executor metrics - for metric, value in iteritems(SPARK_EXECUTOR_OPTIONAL_METRIC_VALUES): - aggregator.assert_metric(metric, value=value, tags=COMMON_TAGS) - - # Check the RDD metrics - for metric, value in iteritems(SPARK_RDD_METRIC_VALUES): - aggregator.assert_metric(metric, value=value, tags=COMMON_TAGS) - - # Check the streaming statistics metrics - for metric, value in iteritems(SPARK_STREAMING_STATISTICS_METRIC_VALUES): - aggregator.assert_metric(metric, value=value, tags=COMMON_TAGS) - - # Check the structured streaming metrics - for metric, value in iteritems(SPARK_STRUCTURED_STREAMING_METRIC_VALUES): - aggregator.assert_metric(metric, value=value, tags=COMMON_TAGS) + _assert( + aggregator, + [ + # Check the running job metrics + (SPARK_JOB_RUNNING_METRIC_VALUES, SPARK_JOB_RUNNING_METRIC_TAGS), + # Check the running job metrics + (SPARK_JOB_RUNNING_METRIC_VALUES, SPARK_JOB_RUNNING_METRIC_TAGS), + # Check the succeeded job metrics + (SPARK_JOB_SUCCEEDED_METRIC_VALUES, SPARK_JOB_SUCCEEDED_METRIC_TAGS), + # Check the running stage metrics + (SPARK_STAGE_RUNNING_METRIC_VALUES, SPARK_STAGE_RUNNING_METRIC_TAGS), + # Check the complete stage metrics + (SPARK_STAGE_COMPLETE_METRIC_VALUES, SPARK_STAGE_COMPLETE_METRIC_TAGS), + # Check the driver metrics + (SPARK_DRIVER_METRIC_VALUES, COMMON_TAGS), + # Check the optional driver metrics + (SPARK_DRIVER_OPTIONAL_METRIC_VALUES, COMMON_TAGS), + # Check the executor level metrics + (SPARK_EXECUTOR_LEVEL_METRIC_VALUES, SPARK_EXECUTOR_LEVEL_METRIC_TAGS), + # Check the optional executor level metrics + (SPARK_EXECUTOR_LEVEL_OPTIONAL_PROCESS_TREE_METRIC_VALUES, SPARK_EXECUTOR_LEVEL_METRIC_TAGS), + # Check the summary executor metrics + (SPARK_EXECUTOR_METRIC_VALUES, COMMON_TAGS), + # Check the optional summary executor metrics + (SPARK_EXECUTOR_OPTIONAL_METRIC_VALUES, COMMON_TAGS), + # Check the RDD metrics + (SPARK_RDD_METRIC_VALUES, COMMON_TAGS), + # Check the streaming statistics metrics + (SPARK_STREAMING_STATISTICS_METRIC_VALUES, COMMON_TAGS), + # Check the structured streaming metrics + (SPARK_STRUCTURED_STREAMING_METRIC_VALUES, COMMON_TAGS), + ], + ) # Check the service tests for sc in aggregator.service_checks(STANDALONE_SERVICE_CHECK): @@ -1200,14 +1086,14 @@ def test_enable_query_name_tag_for_structured_streaming( c = SparkCheck('spark', {}, [instance]) dd_run_check(c) - for metric, value in iteritems(SPARK_STRUCTURED_STREAMING_METRIC_VALUES): + for metric, value in SPARK_STRUCTURED_STREAMING_METRIC_VALUES.items(): tags = base_tags if metric not in SPARK_STRUCTURED_STREAMING_METRIC_NO_TAGS: tags = base_tags + ["query_name:my_named_query"] aggregator.assert_metric(metric, value=value, tags=tags) - for metric, value in iteritems(SPARK_STRUCTURED_STREAMING_METRIC_PUNCTUATED_TAGS): + for metric, value in SPARK_STRUCTURED_STREAMING_METRIC_PUNCTUATED_TAGS.items(): tags = base_tags + ["query_name:my.app.punctuation"] aggregator.assert_metric(metric, value=value, tags=tags) From 81c0ecdb3c92da0340aaf521dcedd11665a796c1 Mon Sep 17 00:00:00 2001 From: Ilia Kurenkov Date: Sat, 21 Sep 2024 00:15:09 +0200 Subject: [PATCH 28/34] Remove more esoteric uses of 'six' (#18638) --- cisco_aci/datadog_checks/cisco_aci/models.py | 397 +++++++++--------- cisco_aci/tests/fixtures/metadata.py | 5 +- cisco_aci/tests/test_fabric.py | 71 ++-- .../datadog_checks/dev/compat.py | 14 - .../datadog_checks/dev/structures.py | 5 +- .../manifest_validator/common/validator.py | 6 +- .../datadog_checks/directory/traverse.py | 20 +- iis/tests/test_e2e.py | 27 +- snmp/datadog_checks/snmp/parsing/metrics.py | 4 +- .../datadog_checks/sqlserver/sqlserver.py | 3 +- 10 files changed, 248 insertions(+), 304 deletions(-) delete mode 100644 datadog_checks_dev/datadog_checks/dev/compat.py diff --git a/cisco_aci/datadog_checks/cisco_aci/models.py b/cisco_aci/datadog_checks/cisco_aci/models.py index 45a58244ed045..3081d8ffd685c 100644 --- a/cisco_aci/datadog_checks/cisco_aci/models.py +++ b/cisco_aci/datadog_checks/cisco_aci/models.py @@ -3,199 +3,210 @@ # Licensed under a 3-clause BSD style license (see LICENSE) import re +from enum import IntEnum, StrEnum +from typing import Optional + +from pydantic import BaseModel, ConfigDict, Field, computed_field, field_validator, model_validator + +""" +Cisco ACI Response Models +""" + + +class NodeAttributes(BaseModel): + address: Optional[str] = None + fabric_st: Optional[str] = Field(default=None, alias="fabricSt") + role: Optional[str] = None + dn: Optional[str] = None + name: Optional[str] = None + model: Optional[str] = None + version: Optional[str] = None + serial: Optional[str] = None + vendor: Optional[str] = Field(default='cisco') + namespace: Optional[str] = Field(default='default') -import six - -if six.PY3: - from enum import IntEnum, StrEnum - from typing import Optional - - from pydantic import BaseModel, ConfigDict, Field, computed_field, field_validator, model_validator - - """ - Cisco ACI Response Models - """ - - class NodeAttributes(BaseModel): - address: Optional[str] = None - fabric_st: Optional[str] = Field(default=None, alias="fabricSt") - role: Optional[str] = None - dn: Optional[str] = None - name: Optional[str] = None - model: Optional[str] = None - version: Optional[str] = None - serial: Optional[str] = None - vendor: Optional[str] = Field(default='cisco') - namespace: Optional[str] = Field(default='default') - - @computed_field - @property - def device_type(self) -> str: - if self.role in ['leaf', 'spine']: - return 'switch' - return 'other' - - class Node(BaseModel): - attributes: NodeAttributes - - class EthpmPhysIfAttributes(BaseModel): - oper_st: Optional[str] = Field(default=None, alias="operSt") - oper_router_mac: Optional[str] = Field(default=None, alias="operRouterMac") - - class EthpmPhysIf(BaseModel): - attributes: EthpmPhysIfAttributes - - class L1PhysIfAttributes(BaseModel): - admin_st: Optional[str] = Field(default=None, alias="adminSt") - id: Optional[str] = None - name: Optional[str] = None - desc: Optional[str] = None - router_mac: Optional[str] = Field(default=None, alias="routerMac") - - @model_validator(mode='before') - @classmethod - def validate_name(cls, data: dict) -> dict: - if isinstance(data, dict): - name = data.get('name') - id = data.get('id') - if not name or name == '': - data['name'] = id - return data - - class PhysIf(BaseModel): - attributes: L1PhysIfAttributes - children: Optional[list] = Field(default_factory=list) - - @computed_field - @property - def ethpm_phys_if(self) -> Optional[EthpmPhysIf]: - for child in self.children: - if 'ethpmPhysIf' in child: - return EthpmPhysIf(**child['ethpmPhysIf']) - return None + @computed_field + @property + def device_type(self) -> str: + if self.role in ['leaf', 'spine']: + return 'switch' + return 'other' - """ - NDM Models - """ - - class DeviceMetadata(BaseModel): - id: Optional[str] = Field(default=None) - id_tags: list = Field(default_factory=list) - tags: list = Field(default_factory=list) - name: Optional[str] = Field(default=None) - ip_address: Optional[str] = Field(default=None) - model: Optional[str] = Field(default=None) - fabric_st: Optional[str] = Field(default=None, exclude=True) - vendor: Optional[str] = Field(default=None) - version: Optional[str] = Field(default=None) - serial_number: Optional[str] = Field(default=None) - device_type: Optional[str] = Field(default=None) - integration: Optional[str] = Field(default='cisco-aci') - - @computed_field - @property - def status(self) -> int: - mapping = { - 'active': 1, - 'inactive': 2, - 'disabled': 2, - 'discovering': 2, - 'undiscovered': 2, - 'unsupported': 2, - 'unknown': 2, - } - return mapping.get(self.fabric_st, 2) - - class DeviceMetadataList(BaseModel): - device_metadata: list = Field(default_factory=list) - - class AdminStatus(IntEnum): - UP = 1 - DOWN = 2 - - class OperStatus(IntEnum): - UP = 1 - DOWN = 2 - - class Status(StrEnum): - UP = "up" - DOWN = "down" - WARNING = "warning" - OFF = "off" - - class InterfaceMetadata(BaseModel): - device_id: Optional[str] = Field(default=None) - id_tags: list = Field(default_factory=list) - index: Optional[int] = Field(default=None) - name: Optional[str] = Field(default=None) - description: Optional[str] = Field(default=None) - mac_address: Optional[str] = Field(default=None) - admin_status: Optional[AdminStatus] = Field(default=None) - oper_status: Optional[OperStatus] = Field(default=None) - integration: Optional[str] = Field(default='cisco-aci') - - model_config = ConfigDict(validate_assignment=True, use_enum_values=True) - - @field_validator("admin_status", mode="before") - @classmethod - def parse_admin_status(cls, admin_status: AdminStatus | None) -> AdminStatus | None: - if not admin_status: - return None - if admin_status == "up" or admin_status == 1: - return AdminStatus.UP - return AdminStatus.DOWN - - @field_validator("oper_status", mode="before") - @classmethod - def parse_oper_status(cls, oper_status: OperStatus | None) -> OperStatus | None: - if not oper_status: - return None - if oper_status == "up" or oper_status == 1: - return OperStatus.UP - return OperStatus.DOWN - - @field_validator("index", mode="before") - @classmethod - def parse_index(cls, index: str | int | None) -> int | None: - if type(index) == str: - split = re.split('eth|/', index) - return int(split[-1]) - if type(index) == int: - return index - return None - @computed_field - @property - def status(self) -> Status: - if self.admin_status == AdminStatus.UP: - if self.oper_status == OperStatus.UP: - return Status.UP - if self.oper_status == OperStatus.DOWN: - return Status.DOWN - return Status.WARNING - if self.admin_status == AdminStatus.DOWN: - if self.oper_status == OperStatus.UP: - return Status.DOWN - if self.oper_status == OperStatus.DOWN: - return Status.OFF - return Status.WARNING - return Status.DOWN - - class InterfaceMetadataList(BaseModel): - interface_metadata: list = Field(default_factory=list) - - class NetworkDevicesMetadata(BaseModel): - namespace: str = None - devices: Optional[list[DeviceMetadata]] = Field(default_factory=list) - interfaces: Optional[list[InterfaceMetadata]] = Field(default_factory=list) - collect_timestamp: Optional[int] = None - size: Optional[int] = Field(default=0, exclude=True) - - model_config = ConfigDict(validate_assignment=True, use_enum_values=True) - - def append_metadata(self, metadata: DeviceMetadata | InterfaceMetadata): - if isinstance(metadata, DeviceMetadata): - self.devices.append(metadata) - if isinstance(metadata, InterfaceMetadata): - self.interfaces.append(metadata) - self.size += 1 +class Node(BaseModel): + attributes: NodeAttributes + + +class EthpmPhysIfAttributes(BaseModel): + oper_st: Optional[str] = Field(default=None, alias="operSt") + oper_router_mac: Optional[str] = Field(default=None, alias="operRouterMac") + + +class EthpmPhysIf(BaseModel): + attributes: EthpmPhysIfAttributes + + +class L1PhysIfAttributes(BaseModel): + admin_st: Optional[str] = Field(default=None, alias="adminSt") + id: Optional[str] = None + name: Optional[str] = None + desc: Optional[str] = None + router_mac: Optional[str] = Field(default=None, alias="routerMac") + + @model_validator(mode='before') + @classmethod + def validate_name(cls, data: dict) -> dict: + if isinstance(data, dict): + name = data.get('name') + id = data.get('id') + if not name or name == '': + data['name'] = id + return data + + +class PhysIf(BaseModel): + attributes: L1PhysIfAttributes + children: Optional[list] = Field(default_factory=list) + + @computed_field + @property + def ethpm_phys_if(self) -> Optional[EthpmPhysIf]: + for child in self.children: + if 'ethpmPhysIf' in child: + return EthpmPhysIf(**child['ethpmPhysIf']) + return None + + +""" +NDM Models +""" + + +class DeviceMetadata(BaseModel): + id: Optional[str] = Field(default=None) + id_tags: list = Field(default_factory=list) + tags: list = Field(default_factory=list) + name: Optional[str] = Field(default=None) + ip_address: Optional[str] = Field(default=None) + model: Optional[str] = Field(default=None) + fabric_st: Optional[str] = Field(default=None, exclude=True) + vendor: Optional[str] = Field(default=None) + version: Optional[str] = Field(default=None) + serial_number: Optional[str] = Field(default=None) + device_type: Optional[str] = Field(default=None) + integration: Optional[str] = Field(default='cisco-aci') + + @computed_field + @property + def status(self) -> int: + mapping = { + 'active': 1, + 'inactive': 2, + 'disabled': 2, + 'discovering': 2, + 'undiscovered': 2, + 'unsupported': 2, + 'unknown': 2, + } + return mapping.get(self.fabric_st, 2) + + +class DeviceMetadataList(BaseModel): + device_metadata: list = Field(default_factory=list) + + +class AdminStatus(IntEnum): + UP = 1 + DOWN = 2 + + +class OperStatus(IntEnum): + UP = 1 + DOWN = 2 + + +class Status(StrEnum): + UP = "up" + DOWN = "down" + WARNING = "warning" + OFF = "off" + + +class InterfaceMetadata(BaseModel): + device_id: Optional[str] = Field(default=None) + id_tags: list = Field(default_factory=list) + index: Optional[int] = Field(default=None) + name: Optional[str] = Field(default=None) + description: Optional[str] = Field(default=None) + mac_address: Optional[str] = Field(default=None) + admin_status: Optional[AdminStatus] = Field(default=None) + oper_status: Optional[OperStatus] = Field(default=None) + integration: Optional[str] = Field(default='cisco-aci') + + model_config = ConfigDict(validate_assignment=True, use_enum_values=True) + + @field_validator("admin_status", mode="before") + @classmethod + def parse_admin_status(cls, admin_status: AdminStatus | None) -> AdminStatus | None: + if not admin_status: + return None + if admin_status == "up" or admin_status == 1: + return AdminStatus.UP + return AdminStatus.DOWN + + @field_validator("oper_status", mode="before") + @classmethod + def parse_oper_status(cls, oper_status: OperStatus | None) -> OperStatus | None: + if not oper_status: + return None + if oper_status == "up" or oper_status == 1: + return OperStatus.UP + return OperStatus.DOWN + + @field_validator("index", mode="before") + @classmethod + def parse_index(cls, index: str | int | None) -> int | None: + if type(index) == str: + split = re.split('eth|/', index) + return int(split[-1]) + if type(index) == int: + return index + return None + + @computed_field + @property + def status(self) -> Status: + if self.admin_status == AdminStatus.UP: + if self.oper_status == OperStatus.UP: + return Status.UP + if self.oper_status == OperStatus.DOWN: + return Status.DOWN + return Status.WARNING + if self.admin_status == AdminStatus.DOWN: + if self.oper_status == OperStatus.UP: + return Status.DOWN + if self.oper_status == OperStatus.DOWN: + return Status.OFF + return Status.WARNING + return Status.DOWN + + +class InterfaceMetadataList(BaseModel): + interface_metadata: list = Field(default_factory=list) + + +class NetworkDevicesMetadata(BaseModel): + namespace: str = None + devices: Optional[list[DeviceMetadata]] = Field(default_factory=list) + interfaces: Optional[list[InterfaceMetadata]] = Field(default_factory=list) + collect_timestamp: Optional[int] = None + size: Optional[int] = Field(default=0, exclude=True) + + model_config = ConfigDict(validate_assignment=True, use_enum_values=True) + + def append_metadata(self, metadata: DeviceMetadata | InterfaceMetadata): + if isinstance(metadata, DeviceMetadata): + self.devices.append(metadata) + if isinstance(metadata, InterfaceMetadata): + self.interfaces.append(metadata) + self.size += 1 diff --git a/cisco_aci/tests/fixtures/metadata.py b/cisco_aci/tests/fixtures/metadata.py index 08cd4a837da9b..0052b27c5448e 100644 --- a/cisco_aci/tests/fixtures/metadata.py +++ b/cisco_aci/tests/fixtures/metadata.py @@ -2,10 +2,7 @@ # All rights reserved # Licensed under a 3-clause BSD style license (see LICENSE) -import six - -if six.PY3: - from datadog_checks.cisco_aci.models import DeviceMetadataList, InterfaceMetadata, NetworkDevicesMetadata +from datadog_checks.cisco_aci.models import DeviceMetadataList, InterfaceMetadata, NetworkDevicesMetadata DEVICE_METADATA = [ { diff --git a/cisco_aci/tests/test_fabric.py b/cisco_aci/tests/test_fabric.py index 567c763534602..2b9c97d7a2a0b 100644 --- a/cisco_aci/tests/test_fabric.py +++ b/cisco_aci/tests/test_fabric.py @@ -2,21 +2,17 @@ # All rights reserved # Licensed under Simplified BSD License (see LICENSE) -import six +from freezegun import freeze_time from datadog_checks.base.utils.containers import hash_mutable from datadog_checks.cisco_aci import CiscoACICheck from datadog_checks.cisco_aci.api import Api -if six.PY3: - from .fixtures.metadata import ( - EXPECTED_INTERFACE_METADATA, - EXPECTED_METADATA_EVENTS, - ) - -from freezegun import freeze_time - from . import common +from .fixtures.metadata import ( + EXPECTED_INTERFACE_METADATA, + EXPECTED_METADATA_EVENTS, +) def test_fabric_mocked(aggregator): @@ -86,38 +82,35 @@ def test_fabric_mocked(aggregator): with freeze_time("2012-01-14 03:21:34"): check.check({}) - if six.PY3: - ndm_metadata = aggregator.get_event_platform_events("network-devices-metadata") - expected_metadata = [event.model_dump(mode="json", exclude_none=True) for event in EXPECTED_METADATA_EVENTS] - assert ndm_metadata == expected_metadata + ndm_metadata = aggregator.get_event_platform_events("network-devices-metadata") + expected_metadata = [event.model_dump(mode="json", exclude_none=True) for event in EXPECTED_METADATA_EVENTS] + assert ndm_metadata == expected_metadata - interface_tag_mapping = { - 'default:10.0.200.0': (device_hn101, hn101), - 'default:10.0.200.1': (device_hn102, hn102), - 'default:10.0.200.5': (device_hn201, hn201), - 'default:10.0.200.2': (device_hn202, hn202), - } + interface_tag_mapping = { + 'default:10.0.200.0': (device_hn101, hn101), + 'default:10.0.200.1': (device_hn102, hn102), + 'default:10.0.200.5': (device_hn201, hn201), + 'default:10.0.200.2': (device_hn202, hn202), + } - for interface in EXPECTED_INTERFACE_METADATA: - device_hn, hn = interface_tag_mapping.get(interface.device_id) - device_namespace, device_ip = interface.device_id.split(':') - interface_tags = [ - 'port:{}'.format(interface.name), - 'medium:broadcast', - 'snmpTrapSt:enable', - 'node_id:{}'.format(hn.split('-')[-1]), - 'fabric_pod_id:1', - 'device_ip:{}'.format(device_ip), - 'device_namespace:{}'.format(device_namespace), - 'device_hostname:{}'.format(device_hn), - 'device_id:{}'.format(interface.device_id), - 'port.status:{}'.format(interface.status), - 'dd.internal.resource:ndm_device_user_tags:{}'.format(interface.device_id), - 'dd.internal.resource:ndm_interface_user_tags:{}:{}'.format(interface.device_id, interface.index), - ] - aggregator.assert_metric( - 'cisco_aci.fabric.port.status', value=1.0, tags=interface_tags, hostname=device_hn - ) + for interface in EXPECTED_INTERFACE_METADATA: + device_hn, hn = interface_tag_mapping.get(interface.device_id) + device_namespace, device_ip = interface.device_id.split(':') + interface_tags = [ + 'port:{}'.format(interface.name), + 'medium:broadcast', + 'snmpTrapSt:enable', + 'node_id:{}'.format(hn.split('-')[-1]), + 'fabric_pod_id:1', + 'device_ip:{}'.format(device_ip), + 'device_namespace:{}'.format(device_namespace), + 'device_hostname:{}'.format(device_hn), + 'device_id:{}'.format(interface.device_id), + 'port.status:{}'.format(interface.status), + 'dd.internal.resource:ndm_device_user_tags:{}'.format(interface.device_id), + 'dd.internal.resource:ndm_interface_user_tags:{}:{}'.format(interface.device_id, interface.index), + ] + aggregator.assert_metric('cisco_aci.fabric.port.status', value=1.0, tags=interface_tags, hostname=device_hn) metric_name = 'cisco_aci.fabric.port.ingr_total.bytes.cum' aggregator.assert_metric( diff --git a/datadog_checks_dev/datadog_checks/dev/compat.py b/datadog_checks_dev/datadog_checks/dev/compat.py deleted file mode 100644 index d890c422df626..0000000000000 --- a/datadog_checks_dev/datadog_checks/dev/compat.py +++ /dev/null @@ -1,14 +0,0 @@ -# (C) Datadog, Inc. 2018-present -# All rights reserved -# Licensed under a 3-clause BSD style license (see LICENSE) -import six - -if six.PY3: - from json import JSONDecodeError - - FileNotFoundError = FileNotFoundError - PermissionError = PermissionError -else: - JSONDecodeError = ValueError - FileNotFoundError = IOError - PermissionError = IOError diff --git a/datadog_checks_dev/datadog_checks/dev/structures.py b/datadog_checks_dev/datadog_checks/dev/structures.py index 317a3004f53bb..9834f308de087 100644 --- a/datadog_checks_dev/datadog_checks/dev/structures.py +++ b/datadog_checks_dev/datadog_checks/dev/structures.py @@ -6,14 +6,11 @@ from shutil import rmtree from tempfile import mkdtemp -import six - from ._env import e2e_active, get_env_vars, remove_env_vars, set_env_vars, tear_down_env from .warn import warning -@six.add_metaclass(abc.ABCMeta) -class LazyFunction(object): +class LazyFunction(abc.ABC): """Abstract base class for lazy function calls.""" @abc.abstractmethod diff --git a/datadog_checks_dev/datadog_checks/dev/tooling/manifest_validator/common/validator.py b/datadog_checks_dev/datadog_checks/dev/tooling/manifest_validator/common/validator.py index 20fd40589bb3d..add075a5cf969 100644 --- a/datadog_checks_dev/datadog_checks/dev/tooling/manifest_validator/common/validator.py +++ b/datadog_checks_dev/datadog_checks/dev/tooling/manifest_validator/common/validator.py @@ -7,8 +7,6 @@ import os from typing import Dict # noqa: F401 -import six - from ...datastructures import JSONDict from ...git import git_show_file from ...utils import get_metadata_file, has_logs, is_metric_in_metadata_file, read_metadata_rows @@ -29,8 +27,7 @@ def __repr__(self): return str(self) -@six.add_metaclass(abc.ABCMeta) -class BaseManifestValidator(object): +class BaseManifestValidator(abc.ABC): def __init__( self, is_extras=False, @@ -64,6 +61,7 @@ def should_validate(self): return True return False + @abc.abstractmethod def validate(self, check_name, manifest, should_fix): # type: (str, Dict, bool) -> None """Validates the decoded manifest. Will perform inline changes if fix is true""" diff --git a/directory/datadog_checks/directory/traverse.py b/directory/datadog_checks/directory/traverse.py index 09526805a0698..38db22d2c6cfe 100644 --- a/directory/datadog_checks/directory/traverse.py +++ b/directory/datadog_checks/directory/traverse.py @@ -1,18 +1,13 @@ # (C) Datadog, Inc. 2018-present # All rights reserved # Licensed under a 3-clause BSD style license (see LICENSE) -import platform -import sys - -import six - try: from os import scandir except ImportError: from scandir import scandir -def _walk(top, onerror=None, followlinks=False): +def walk(top, onerror=None, followlinks=False): """A simplified and modified version of stdlib's `os.walk` that yields the `os.DirEntry` objects that `scandir` produces during traversal instead of paths as strings. @@ -72,16 +67,3 @@ def _walk(top, onerror=None, followlinks=False): for dir_entry in dirs: for entry in walk(dir_entry.path, onerror, followlinks): yield entry - - -if six.PY3 or platform.system() != 'Windows': - walk = _walk -else: - # Fix for broken unicode handling on Windows on Python 2.x, see: - # https://github.com/benhoyt/scandir/issues/54 - file_system_encoding = sys.getfilesystemencoding() - - def walk(top, onerror, followlinks): - if isinstance(top, bytes): - top = top.decode(file_system_encoding) - return _walk(top, onerror, followlinks) diff --git a/iis/tests/test_e2e.py b/iis/tests/test_e2e.py index 8eae4525e3389..21903d36fe02b 100644 --- a/iis/tests/test_e2e.py +++ b/iis/tests/test_e2e.py @@ -2,12 +2,8 @@ # All rights reserved # Licensed under Simplified BSD License (see LICENSE) -import socket - import pytest -import six -from datadog_checks.base import AgentCheck from datadog_checks.dev.subprocess import run_command from datadog_checks.dev.utils import get_active_env from datadog_checks.iis import IIS @@ -22,31 +18,18 @@ def test_e2e(dd_agent_check, aggregator, instance): Check should run without error if IIS is installed """ aggregator = dd_agent_check(instance) - if six.PY3: - aggregator.assert_service_check('iis.windows.perf.health', IIS.OK) + aggregator.assert_service_check('iis.windows.perf.health', IIS.OK) @pytest.fixture def iis_host(aggregator): - if six.PY2: - # the python2 version of the check uses the configured hostname, - # which is the hostnane of the docker host. - return 'iis_host:{}'.format(socket.gethostname()) - else: - # the python3 version of the check uses the perf counter server field, - # which is the hostname of the container. - result = run_container_command(['hostname']) - hostname = result.stdout.strip() - return 'iis_host:{}'.format(hostname) + result = run_container_command(['hostname']) + hostname = result.stdout.strip() + return 'iis_host:{}'.format(hostname) def normalize_tags(tags): - if six.PY2: - # The python2 version of the check calls self.normalize_tag - a = AgentCheck() - return [a.normalize_tag(tag) for tag in tags] - else: - return tags + return tags def run_container_command(command): diff --git a/snmp/datadog_checks/snmp/parsing/metrics.py b/snmp/datadog_checks/snmp/parsing/metrics.py index 3b5aea30d7290..92b8707509793 100644 --- a/snmp/datadog_checks/snmp/parsing/metrics.py +++ b/snmp/datadog_checks/snmp/parsing/metrics.py @@ -8,8 +8,6 @@ from logging import Logger # noqa: F401 from typing import Dict, List, NamedTuple, Optional, Pattern, Sequence, TypedDict, Union, cast -import six - from datadog_checks.base import ConfigurationError from ..models import OID @@ -558,7 +556,7 @@ def _parse_index_slices(metric_tag): if not isinstance(rule, dict) or set(rule) != {'start', 'end'}: raise ConfigurationError('Transform rule must contain start and end. Invalid rule: {}'.format(rule)) start, end = rule['start'], rule['end'] - if not isinstance(start, six.integer_types) or not isinstance(end, six.integer_types): + if not isinstance(start, int) or not isinstance(end, int): raise ConfigurationError('Transform rule start and end must be integers. Invalid rule: {}'.format(rule)) if start > end: raise ConfigurationError( diff --git a/sqlserver/datadog_checks/sqlserver/sqlserver.py b/sqlserver/datadog_checks/sqlserver/sqlserver.py index 80e878c97833d..c2f9599b308d5 100644 --- a/sqlserver/datadog_checks/sqlserver/sqlserver.py +++ b/sqlserver/datadog_checks/sqlserver/sqlserver.py @@ -7,7 +7,6 @@ import time from collections import defaultdict -import six from cachetools import TTLCache from datadog_checks.base import AgentCheck @@ -905,7 +904,7 @@ def collect_metrics(self): instance_results = {} engine_edition = self.static_info_cache.get(STATIC_INFO_ENGINE_EDITION, "") # Execute the `fetch_all` operations first to minimize the database calls - for cls, metric_names in six.iteritems(self.instance_per_type_metrics): + for cls, metric_names in self.instance_per_type_metrics.items(): if not metric_names: instance_results[cls] = None, None else: From 56ab4b6548ecfb4d2f7ad73812acf114d1c43e76 Mon Sep 17 00:00:00 2001 From: dkirov-dd <166512750+dkirov-dd@users.noreply.github.com> Date: Sat, 21 Sep 2024 14:35:31 +0200 Subject: [PATCH 29/34] [AWS Neuron] Add dashboard (#18428) * Add dashboard * Update dashboard variables * Update dashboard according to feedback * Update aws_neuron/assets/dashboards/aws_neuron_overview.json Co-authored-by: Sandra (neko) <165049174+neko-dd@users.noreply.github.com> * Update dashboard * Add useful links to dropdown * Implement feedback * Update units and display styles --------- Co-authored-by: Sandra (neko) <165049174+neko-dd@users.noreply.github.com> --- .../dashboards/aws_neuron_overview.json | 1195 +++++++++++++++++ aws_neuron/manifest.json | 3 + 2 files changed, 1198 insertions(+) create mode 100644 aws_neuron/assets/dashboards/aws_neuron_overview.json diff --git a/aws_neuron/assets/dashboards/aws_neuron_overview.json b/aws_neuron/assets/dashboards/aws_neuron_overview.json new file mode 100644 index 0000000000000..4b30aa72d5426 --- /dev/null +++ b/aws_neuron/assets/dashboards/aws_neuron_overview.json @@ -0,0 +1,1195 @@ +{ + "title": "AWS Neuron Dashboard", + "description": "# Description\nThis dashboard provides observability for your Inferentia and Trainium instances leveraging the Neuron SDK.\n\nIt shows information about the usage and performance of your Neuron devices as well as insights into the execution and training of your models, including error rates, latency and memory usage.\n\n# Useful Links\n[AWS Neuron Integration](https://docs.datadoghq.com/integrations/aws_neuron/)\n\n[AWS Neuron Documentation](https://awsdocs-neuron.readthedocs-hosted.com/en/latest/index.html)", + "widgets": [ + { + "id": 6979231130018708, + "definition": { + "title": "", + "banner_img": "/static/images/integration_dashboard/aws-neuron_hero_1.png", + "show_title": true, + "type": "group", + "layout_type": "ordered", + "widgets": [ + { + "id": 56755139831890, + "definition": { + "type": "note", + "content": "# AWS Neuron\n\nThis dashboard provides observability for your Inferentia and Trainium instances leveraging the Neuron SDK.\n\nIt shows information about the usage and performance of your Neuron devices as well as insights into the execution and training of your models, including error rates, latency and memory usage.", + "background_color": "white", + "font_size": "14", + "text_align": "left", + "vertical_align": "top", + "show_tick": false, + "tick_pos": "50%", + "tick_edge": "left", + "has_padding": true + }, + "layout": { + "x": 0, + "y": 0, + "width": 3, + "height": 4 + } + }, + { + "id": 4721541429420236, + "definition": { + "type": "note", + "content": "# Useful Links\n\n[AWS Neuron Integration](https://docs.datadoghq.com/integrations/aws_neuron/)\n\n[AWS Neuron Documentation](https://awsdocs-neuron.readthedocs-hosted.com/en/latest/index.html)", + "background_color": "white", + "font_size": "14", + "text_align": "left", + "vertical_align": "top", + "show_tick": false, + "tick_pos": "50%", + "tick_edge": "left", + "has_padding": true + }, + "layout": { + "x": 3, + "y": 0, + "width": 2, + "height": 4 + } + } + ] + }, + "layout": { + "x": 0, + "y": 0, + "width": 5, + "height": 7 + } + }, + { + "id": 7083040833221452, + "definition": { + "title": "Overview", + "background_color": "vivid_green", + "show_title": true, + "type": "group", + "layout_type": "ordered", + "widgets": [ + { + "id": 5198931877490734, + "definition": { + "type": "note", + "content": "This overview provides a quick look into the general status and health of your Neuron instances.", + "background_color": "green", + "font_size": "14", + "text_align": "center", + "vertical_align": "top", + "show_tick": false, + "tick_pos": "50%", + "tick_edge": "left", + "has_padding": true + }, + "layout": { + "x": 0, + "y": 0, + "width": 7, + "height": 1 + } + }, + { + "id": 8720449453186756, + "definition": { + "title": "Instance Info", + "title_size": "16", + "title_align": "left", + "type": "query_table", + "requests": [ + { + "queries": [ + { + "data_source": "metrics", + "name": "query1", + "query": "avg:aws_neuron.instance_info{$instance_id,$instance_type, $availability_zone, $region, $subnet_id} by {instance_id,instance_type,availability_zone,region,subnet_id}", + "aggregator": "avg" + } + ], + "response_format": "scalar", + "text_formats": [], + "sort": { + "order_by": [ + { + "type": "group", + "name": "instance_id", + "order": "asc" + } + ], + "count": 500 + }, + "formulas": [ + { + "cell_display_mode": "number", + "alias": "Instances", + "formula": "query1" + } + ] + } + ], + "has_search_bar": "auto" + }, + "layout": { + "x": 0, + "y": 1, + "width": 7, + "height": 2 + } + }, + { + "id": 7855267612232854, + "definition": { + "title": "Execution success rate", + "title_size": "16", + "title_align": "left", + "type": "query_value", + "requests": [ + { + "response_format": "scalar", + "queries": [ + { + "data_source": "metrics", + "name": "query1", + "query": "sum:aws_neuron.execution.status.count{status_type:completed, $instance_id, $instance_type, $availability_zone, $region, $subnet_id}.as_rate()", + "aggregator": "avg" + } + ], + "conditional_formats": [ + { + "comparator": ">=", + "value": 0, + "palette": "white_on_green" + } + ], + "formulas": [ + { + "formula": "query1" + } + ] + } + ], + "autoscale": true, + "precision": 2 + }, + "layout": { + "x": 0, + "y": 3, + "width": 2, + "height": 1 + } + }, + { + "id": 2819874854268670, + "definition": { + "title": "Monitors", + "type": "manage_status", + "display_format": "countsAndList", + "color_preference": "text", + "hide_zero_counts": true, + "show_status": true, + "last_triggered_format": "relative", + "query": "tag:(integration:aws_neuron)", + "sort": "status,asc", + "count": 50, + "start": 0, + "summary_type": "monitors", + "show_priority": false, + "show_last_triggered": false + }, + "layout": { + "x": 2, + "y": 3, + "width": 5, + "height": 3 + } + }, + { + "id": 6464457537294578, + "definition": { + "title": "Execution error rate", + "title_size": "16", + "title_align": "left", + "type": "query_value", + "requests": [ + { + "response_format": "scalar", + "queries": [ + { + "data_source": "metrics", + "name": "query1", + "query": "sum:aws_neuron.execution.status.count{not status_type:completed}.as_rate()", + "aggregator": "avg" + } + ], + "conditional_formats": [ + { + "comparator": ">", + "value": 0, + "palette": "white_on_red" + }, + { + "comparator": "=", + "value": 0, + "palette": "white_on_green" + } + ], + "formulas": [ + { + "formula": "query1" + } + ] + } + ], + "autoscale": true, + "precision": 2 + }, + "layout": { + "x": 0, + "y": 4, + "width": 2, + "height": 1 + } + }, + { + "id": 4990133775519788, + "definition": { + "title": "Average execution latency", + "title_size": "16", + "title_align": "left", + "type": "query_value", + "requests": [ + { + "response_format": "scalar", + "queries": [ + { + "data_source": "metrics", + "name": "query1", + "query": "avg:aws_neuron.execution.latency_seconds{$instance_id, $instance_type, $availability_zone, $region, $subnet_id}", + "aggregator": "avg" + } + ], + "formulas": [ + { + "formula": "query1", + "number_format": { + "unit": { + "type": "canonical_unit", + "unit_name": "second" + } + } + } + ] + } + ], + "autoscale": true, + "precision": 2 + }, + "layout": { + "x": 0, + "y": 5, + "width": 2, + "height": 1 + } + } + ] + }, + "layout": { + "x": 5, + "y": 0, + "width": 7, + "height": 7 + } + }, + { + "id": 6659947028871704, + "definition": { + "title": "Execution Status", + "background_color": "vivid_green", + "show_title": true, + "type": "group", + "layout_type": "ordered", + "widgets": [ + { + "id": 468925319074014, + "definition": { + "type": "note", + "content": "Monitoring execution status helps identify the operational state of machine learning models, ensuring they are running as expected. Tracking execution latency provides insights into the performance and efficiency of model inference, helping optimize response times and resource usage.", + "background_color": "green", + "font_size": "14", + "text_align": "center", + "vertical_align": "top", + "show_tick": false, + "tick_pos": "50%", + "tick_edge": "left", + "has_padding": true + }, + "layout": { + "x": 0, + "y": 0, + "width": 12, + "height": 1 + } + }, + { + "id": 1102497452686722, + "definition": { + "title": "Execution status rates", + "title_size": "16", + "title_align": "left", + "show_legend": true, + "legend_layout": "auto", + "legend_columns": [ + "avg", + "min", + "max", + "value", + "sum" + ], + "type": "timeseries", + "requests": [ + { + "formulas": [ + { + "alias": "Executions success rate", + "style": { + "palette": "green" + }, + "formula": "query1" + }, + { + "alias": "Execution error rate", + "style": { + "palette": "red" + }, + "formula": "query2" + } + ], + "queries": [ + { + "data_source": "metrics", + "name": "query1", + "query": "avg:aws_neuron.execution.status.count{status_type:completed , $instance_id, $instance_type, $availability_zone, $region, $subnet_id} by {instance_id}.as_rate()" + }, + { + "data_source": "metrics", + "name": "query2", + "query": "avg:aws_neuron.execution.status.count{not status_type:completed and $instance_id AND $instance_type AND $availability_zone AND $region AND $subnet_id} by {instance_id}.as_rate()" + } + ], + "response_format": "timeseries", + "style": { + "palette": "dog_classic", + "order_by": "values", + "color_order": "shuffled", + "line_type": "solid", + "line_width": "normal" + }, + "display_type": "line" + } + ] + }, + "layout": { + "x": 0, + "y": 1, + "width": 6, + "height": 3 + } + }, + { + "id": 245689066689228, + "definition": { + "title": "Execution latency", + "title_size": "16", + "title_align": "left", + "show_legend": true, + "legend_layout": "auto", + "legend_columns": [ + "avg", + "min", + "max", + "value", + "sum" + ], + "type": "timeseries", + "requests": [ + { + "formulas": [ + { + "style": { + "palette": "warm" + }, + "number_format": { + "unit": { + "type": "canonical_unit", + "unit_name": "second" + } + }, + "formula": "query1" + } + ], + "queries": [ + { + "data_source": "metrics", + "name": "query1", + "query": "avg:aws_neuron.execution.latency_seconds{$instance_id, $instance_type, $availability_zone, $region, $subnet_id} by {percentile,instance_id}" + } + ], + "response_format": "timeseries", + "style": { + "palette": "dog_classic", + "order_by": "tags", + "color_order": "shuffled", + "line_type": "solid", + "line_width": "normal" + }, + "display_type": "line" + } + ] + }, + "layout": { + "x": 6, + "y": 1, + "width": 6, + "height": 3 + } + } + ] + }, + "layout": { + "x": 0, + "y": 7, + "width": 12, + "height": 5 + } + }, + { + "id": 8841094143294328, + "definition": { + "title": "Resource utilization", + "background_color": "vivid_green", + "show_title": true, + "type": "group", + "layout_type": "ordered", + "widgets": [ + { + "id": 5529167122831830, + "definition": { + "type": "note", + "content": "Make the most out of your system by monitoring resource utilization at multiple scales. Tracking NeuronCore utilization helps assess how effectively the model is using the allocated Neuron cores, guiding optimization for better performance. Additionally, the system utilization metrics provide a bird's eye view of the system resources.", + "background_color": "green", + "font_size": "14", + "text_align": "center", + "vertical_align": "top", + "show_tick": false, + "tick_pos": "50%", + "tick_edge": "left", + "has_padding": true + }, + "layout": { + "x": 0, + "y": 0, + "width": 12, + "height": 1 + } + }, + { + "id": 749112045204328, + "definition": { + "title": "NeuronCore utilization", + "title_size": "16", + "title_align": "left", + "show_legend": true, + "legend_layout": "auto", + "legend_columns": [ + "avg", + "min", + "max", + "value", + "sum" + ], + "type": "timeseries", + "requests": [ + { + "formulas": [ + { + "number_format": { + "unit": { + "type": "canonical_unit", + "unit_name": "percent" + } + }, + "formula": "100 * query1" + } + ], + "queries": [ + { + "data_source": "metrics", + "name": "query1", + "query": "avg:aws_neuron.neuroncore.utilization_ratio{$instance_id, $instance_type, $availability_zone, $region, $subnet_id} by {neuroncore,instance_id}" + } + ], + "response_format": "timeseries", + "style": { + "palette": "datadog16", + "order_by": "tags", + "color_order": "shuffled", + "line_type": "solid", + "line_width": "normal" + }, + "display_type": "line" + } + ] + }, + "layout": { + "x": 0, + "y": 1, + "width": 6, + "height": 3 + } + }, + { + "id": 96594677820004, + "definition": { + "title": "Host system utilization", + "title_size": "16", + "title_align": "left", + "show_legend": true, + "legend_layout": "auto", + "legend_columns": [ + "avg", + "min", + "max", + "value", + "sum" + ], + "type": "timeseries", + "requests": [ + { + "formulas": [ + { + "alias": "Memory usage", + "number_format": { + "unit": { + "type": "canonical_unit", + "unit_name": "percent" + } + }, + "formula": "100 * query2 / query3" + }, + { + "alias": "vCPU usage", + "number_format": { + "unit": { + "type": "canonical_unit", + "unit_name": "percent" + } + }, + "formula": "100 * query4" + }, + { + "alias": "Neuroncore usage", + "number_format": { + "unit": { + "type": "canonical_unit", + "unit_name": "percent" + } + }, + "formula": "100 * query6" + } + ], + "queries": [ + { + "data_source": "metrics", + "name": "query2", + "query": "avg:aws_neuron.system.memory.used_bytes{$instance_id, $instance_type, $availability_zone, $region, $subnet_id}" + }, + { + "data_source": "metrics", + "name": "query3", + "query": "avg:aws_neuron.system.memory.total_bytes{$instance_id, $instance_type, $availability_zone, $region, $subnet_id}" + }, + { + "data_source": "metrics", + "name": "query4", + "query": "avg:aws_neuron.system.vcpu.usage_ratio{$instance_id, $instance_type, $availability_zone, $region, $subnet_id}" + }, + { + "data_source": "metrics", + "name": "query6", + "query": "avg:aws_neuron.neuroncore.utilization_ratio{$instance_id, $instance_type, $availability_zone, $region, $subnet_id}" + } + ], + "response_format": "timeseries", + "style": { + "palette": "datadog16", + "order_by": "values", + "color_order": "shuffled", + "line_type": "solid", + "line_width": "normal" + }, + "display_type": "line" + } + ], + "yaxis": { + "include_zero": true, + "min": "0", + "max": "100" + } + }, + "layout": { + "x": 6, + "y": 1, + "width": 6, + "height": 3 + } + } + ] + }, + "layout": { + "x": 0, + "y": 12, + "width": 12, + "height": 5 + } + }, + { + "id": 7322947690579954, + "definition": { + "title": "vCPU", + "background_color": "vivid_green", + "show_title": true, + "type": "group", + "layout_type": "ordered", + "widgets": [ + { + "id": 4083451820109164, + "definition": { + "type": "note", + "content": "By tracking the Neuron runtime's vCPU usage, you can ensure that your models are efficiently utilizing the allocated resources. Meanwhile, monitoring the overall system vCPU usage helps in identifying potential bottlenecks or underutilization, enabling better resource management and capacity planning across your entire system.", + "background_color": "green", + "font_size": "14", + "text_align": "center", + "vertical_align": "top", + "show_tick": false, + "tick_pos": "50%", + "tick_edge": "left", + "has_padding": true + }, + "layout": { + "x": 0, + "y": 0, + "width": 12, + "height": 1 + } + }, + { + "id": 3715305104834630, + "definition": { + "title": "Neuron Runtime vCPU usage", + "title_size": "16", + "title_align": "left", + "show_legend": true, + "legend_layout": "auto", + "legend_columns": [ + "avg", + "min", + "max", + "value", + "sum" + ], + "type": "timeseries", + "requests": [ + { + "formulas": [ + { + "number_format": { + "unit": { + "type": "canonical_unit", + "unit_name": "percent" + } + }, + "formula": "100 * query1" + } + ], + "queries": [ + { + "data_source": "metrics", + "name": "query1", + "query": "avg:aws_neuron.neuron_runtime.vcpu_usage_ratio{$instance_id, $instance_type, $availability_zone, $region, $subnet_id} by {usage_type,instance_id}" + } + ], + "response_format": "timeseries", + "style": { + "palette": "dog_classic", + "order_by": "values", + "color_order": "shuffled", + "line_type": "solid", + "line_width": "normal" + }, + "display_type": "line" + } + ] + }, + "layout": { + "x": 0, + "y": 1, + "width": 6, + "height": 3 + } + }, + { + "id": 4535539697349626, + "definition": { + "title": "Host vCPU usage", + "title_size": "16", + "title_align": "left", + "show_legend": true, + "legend_layout": "auto", + "legend_columns": [ + "avg", + "min", + "max", + "value", + "sum" + ], + "type": "timeseries", + "requests": [ + { + "formulas": [ + { + "number_format": { + "unit": { + "type": "canonical_unit", + "unit_name": "percent" + } + }, + "formula": "100 * query1" + } + ], + "queries": [ + { + "data_source": "metrics", + "name": "query1", + "query": "avg:aws_neuron.system.vcpu.usage_ratio{$instance_id, $instance_type, $availability_zone, $region, $subnet_id} by {usage_type,instance_id}" + } + ], + "response_format": "timeseries", + "style": { + "palette": "dog_classic", + "order_by": "values", + "color_order": "shuffled", + "line_type": "solid", + "line_width": "normal" + }, + "display_type": "line" + } + ] + }, + "layout": { + "x": 6, + "y": 1, + "width": 6, + "height": 3 + } + } + ] + }, + "layout": { + "x": 0, + "y": 17, + "width": 12, + "height": 5, + "is_column_break": true + } + }, + { + "id": 4920675792552564, + "definition": { + "title": "Memory", + "background_color": "vivid_green", + "show_title": true, + "type": "group", + "layout_type": "ordered", + "widgets": [ + { + "id": 7074409510658020, + "definition": { + "type": "note", + "content": "Monitoring the Neuron runtime's memory usage helps ensure that the memory allocated for model execution is efficiently used, avoiding potential bottlenecks. Tracking system memory usage provides a broader view of overall system memory consumption, helping to prevent resource exhaustion and maintain overall system stability.", + "background_color": "green", + "font_size": "14", + "text_align": "center", + "vertical_align": "top", + "show_tick": false, + "tick_pos": "50%", + "tick_edge": "left", + "has_padding": true + }, + "layout": { + "x": 0, + "y": 0, + "width": 12, + "height": 1 + } + }, + { + "id": 8032824553753052, + "definition": { + "title": "Neuron Runtime memory usage", + "title_size": "16", + "title_align": "left", + "show_legend": true, + "legend_layout": "auto", + "legend_columns": [ + "avg", + "min", + "max", + "value", + "sum" + ], + "type": "timeseries", + "requests": [ + { + "formulas": [ + { + "style": { + "palette": "warm" + }, + "formula": "query1" + } + ], + "queries": [ + { + "data_source": "metrics", + "name": "query1", + "query": "avg:aws_neuron.neuron_runtime.memory_used_bytes{$instance_id, $instance_type, $availability_zone, $region, $subnet_id} by {memory_location,instance_id}" + } + ], + "response_format": "timeseries", + "style": { + "palette": "dog_classic", + "order_by": "values", + "color_order": "shuffled", + "line_type": "solid", + "line_width": "normal" + }, + "display_type": "bars" + } + ] + }, + "layout": { + "x": 0, + "y": 1, + "width": 6, + "height": 3 + } + }, + { + "id": 1818540329200970, + "definition": { + "title": "Host memory usage", + "title_size": "16", + "title_align": "left", + "show_legend": true, + "legend_layout": "auto", + "legend_columns": [ + "avg", + "min", + "max", + "value", + "sum" + ], + "type": "timeseries", + "requests": [ + { + "on_right_yaxis": true, + "formulas": [ + { + "alias": "Memory Usage Bytes", + "number_format": { + "unit": { + "type": "canonical_unit", + "unit_name": "byte" + } + }, + "style": { + "palette": "orange" + }, + "formula": "query1" + } + ], + "queries": [ + { + "data_source": "metrics", + "name": "query1", + "query": "avg:aws_neuron.system.memory.used_bytes{$instance_id, $instance_type, $availability_zone, $region, $subnet_id} by {instance_id}" + }, + { + "data_source": "metrics", + "name": "query2", + "query": "avg:aws_neuron.system.memory.total_bytes{$instance_id, $instance_type, $availability_zone, $region, $subnet_id}" + } + ], + "response_format": "timeseries", + "style": { + "palette": "dog_classic", + "order_by": "values", + "color_order": "shuffled", + "order_reverse": false, + "line_type": "solid", + "line_width": "normal" + }, + "display_type": "line" + }, + { + "on_right_yaxis": false, + "formulas": [ + { + "alias": "Memory Usage Percent", + "number_format": { + "unit": { + "type": "canonical_unit", + "unit_name": "percent" + } + }, + "style": { + "palette": "purple" + }, + "formula": "100 * query1 / query2" + } + ], + "queries": [ + { + "data_source": "metrics", + "name": "query1", + "query": "avg:aws_neuron.system.memory.used_bytes{$instance_id, $instance_type, $availability_zone, $region, $subnet_id} by {instance_id}" + }, + { + "data_source": "metrics", + "name": "query2", + "query": "avg:aws_neuron.system.memory.total_bytes{$instance_id, $instance_type, $availability_zone, $region, $subnet_id} by {instance_id}" + } + ], + "response_format": "timeseries", + "style": { + "palette": "dog_classic", + "order_by": "values", + "color_order": "shuffled", + "order_reverse": false, + "line_type": "solid", + "line_width": "normal" + }, + "display_type": "line" + } + ], + "yaxis": { + "include_zero": true, + "min": "0", + "max": "100" + }, + "markers": [] + }, + "layout": { + "x": 6, + "y": 1, + "width": 6, + "height": 3 + } + } + ] + }, + "layout": { + "x": 0, + "y": 22, + "width": 12, + "height": 5 + } + }, + { + "id": 6335068973644062, + "definition": { + "title": "Logs", + "background_color": "green", + "show_title": true, + "type": "group", + "layout_type": "ordered", + "widgets": [ + { + "id": 2270740140504733, + "definition": { + "type": "note", + "content": "When investigating logs, you can refer to the timeseries data to see the ratio of a certain log status to other expected statuses. If you notice an increase in the ratio of error logs, you can refer to the log stream, which is configured to filter erroneous logs.", + "background_color": "green", + "font_size": "14", + "text_align": "center", + "vertical_align": "top", + "show_tick": false, + "tick_pos": "50%", + "tick_edge": "left", + "has_padding": true + }, + "layout": { + "x": 0, + "y": 0, + "width": 12, + "height": 1 + } + }, + { + "id": 3745226166465988, + "definition": { + "title": "Error log count", + "title_size": "16", + "title_align": "left", + "show_legend": true, + "legend_layout": "horizontal", + "legend_columns": [ + "avg", + "min", + "max", + "value", + "sum" + ], + "type": "timeseries", + "requests": [ + { + "formulas": [ + { + "formula": "a" + } + ], + "queries": [ + { + "name": "a", + "data_source": "logs", + "search": { + "query": "source:aws-neuron status:ERROR " + }, + "indexes": [ + "*" + ], + "group_by": [], + "compute": { + "aggregation": "count", + "metric": "count" + }, + "storage": "online_archives" + } + ], + "response_format": "timeseries", + "style": { + "palette": "semantic", + "line_type": "solid", + "line_width": "normal" + }, + "display_type": "bars" + } + ] + }, + "layout": { + "x": 0, + "y": 1, + "width": 8, + "height": 4 + } + }, + { + "id": 5222865640613950, + "definition": { + "title": "Neuron logs", + "title_size": "16", + "title_align": "left", + "requests": [ + { + "response_format": "event_list", + "columns": [ + { + "field": "status_line", + "width": "auto" + }, + { + "field": "timestamp", + "width": "auto" + }, + { + "field": "host", + "width": "auto" + }, + { + "field": "service", + "width": "auto" + }, + { + "field": "content", + "width": "compact" + } + ], + "query": { + "data_source": "logs_stream", + "query_string": "source:aws_neuron ", + "indexes": [], + "storage": "hot" + } + } + ], + "type": "list_stream" + }, + "layout": { + "x": 8, + "y": 1, + "width": 4, + "height": 4 + } + } + ] + }, + "layout": { + "x": 0, + "y": 27, + "width": 12, + "height": 6 + } + } + ], + "template_variables": [ + { + "name": "instance_id", + "prefix": "instance_id", + "available_values": [], + "default": "*" + }, + { + "name": "instance_type", + "prefix": "instance_type", + "available_values": [], + "default": "*" + }, + { + "name": "availability_zone", + "prefix": "availability_zone", + "available_values": [], + "default": "*" + }, + { + "name": "region", + "prefix": "region", + "available_values": [], + "default": "*" + }, + { + "name": "subnet_id", + "prefix": "subnet_id", + "available_values": [], + "default": "*" + } + ], + "layout_type": "ordered", + "notify_list": [], + "reflow_type": "fixed", + "tags": [ + "team:agent-integrations" + ] +} diff --git a/aws_neuron/manifest.json b/aws_neuron/manifest.json index 1b476c21122b0..23060838a5a4d 100644 --- a/aws_neuron/manifest.json +++ b/aws_neuron/manifest.json @@ -44,6 +44,9 @@ "metadata_path": "assets/service_checks.json" } }, + "dashboards": { + "AWS Neuron Overview": "assets/dashboards/aws_neuron_overview.json" + }, "monitors": { "Execution latency": "assets/monitors/execution_latency.json", "Execution errors": "assets/monitors/execution_errors.json", From eaa0141daf8149cd920612897fd016156bffe2e5 Mon Sep 17 00:00:00 2001 From: Ilia Kurenkov Date: Sat, 21 Sep 2024 17:43:31 +0200 Subject: [PATCH 30/34] Remove setuptools build dependency (#18642) --- active_directory/pyproject.toml | 2 -- activemq/pyproject.toml | 2 -- activemq_xml/pyproject.toml | 2 -- aerospike/pyproject.toml | 2 -- airflow/pyproject.toml | 2 -- amazon_msk/pyproject.toml | 2 -- ambari/pyproject.toml | 2 -- apache/pyproject.toml | 2 -- aspdotnet/pyproject.toml | 2 -- avi_vantage/pyproject.toml | 2 -- azure_iot_edge/pyproject.toml | 2 -- btrfs/pyproject.toml | 2 -- cacti/pyproject.toml | 2 -- calico/pyproject.toml | 2 -- cassandra/pyproject.toml | 2 -- cassandra_nodetool/pyproject.toml | 2 -- ceph/pyproject.toml | 2 -- cert_manager/pyproject.toml | 2 -- checkpoint_quantum_firewall/pyproject.toml | 2 -- cilium/pyproject.toml | 2 -- cisco_aci/pyproject.toml | 2 -- citrix_hypervisor/pyproject.toml | 2 -- clickhouse/pyproject.toml | 2 -- cloud_foundry_api/pyproject.toml | 2 -- cloudera/pyproject.toml | 2 -- cockroachdb/pyproject.toml | 2 -- confluent_platform/pyproject.toml | 2 -- consul/pyproject.toml | 2 -- coredns/pyproject.toml | 2 -- couch/pyproject.toml | 2 -- couchbase/pyproject.toml | 2 -- crio/pyproject.toml | 2 -- datadog_checks_base/pyproject.toml | 2 -- datadog_checks_dependency_provider/pyproject.toml | 2 -- datadog_checks_dev/pyproject.toml | 5 ----- datadog_checks_downloader/pyproject.toml | 2 -- datadog_cluster_agent/pyproject.toml | 2 -- dcgm/pyproject.toml | 1 - directory/pyproject.toml | 2 -- disk/pyproject.toml | 2 -- dns_check/pyproject.toml | 2 -- dotnetclr/pyproject.toml | 2 -- druid/pyproject.toml | 2 -- ecs_fargate/pyproject.toml | 2 -- eks_fargate/pyproject.toml | 2 -- elastic/pyproject.toml | 2 -- envoy/pyproject.toml | 2 -- etcd/pyproject.toml | 2 -- exchange_server/pyproject.toml | 2 -- external_dns/pyproject.toml | 2 -- flink/pyproject.toml | 2 -- fluentd/pyproject.toml | 2 -- foundationdb/pyproject.toml | 2 -- gearmand/pyproject.toml | 2 -- gitlab/pyproject.toml | 2 -- gitlab_runner/pyproject.toml | 2 -- glusterfs/pyproject.toml | 2 -- go_expvar/pyproject.toml | 2 -- gunicorn/pyproject.toml | 2 -- haproxy/pyproject.toml | 2 -- harbor/pyproject.toml | 2 -- hazelcast/pyproject.toml | 2 -- hdfs_datanode/pyproject.toml | 2 -- hdfs_namenode/pyproject.toml | 2 -- hive/pyproject.toml | 2 -- hivemq/pyproject.toml | 2 -- http_check/pyproject.toml | 2 -- hudi/pyproject.toml | 2 -- hyperv/pyproject.toml | 2 -- ibm_db2/pyproject.toml | 2 -- ibm_i/pyproject.toml | 2 -- ibm_was/pyproject.toml | 2 -- ignite/pyproject.toml | 2 -- iis/pyproject.toml | 2 -- impala/pyproject.toml | 2 -- istio/pyproject.toml | 2 -- jboss_wildfly/pyproject.toml | 2 -- journald/pyproject.toml | 2 -- kafka/pyproject.toml | 2 -- kong/pyproject.toml | 2 -- kube_apiserver_metrics/pyproject.toml | 2 -- kube_controller_manager/pyproject.toml | 2 -- kube_dns/pyproject.toml | 2 -- kube_metrics_server/pyproject.toml | 2 -- kube_proxy/pyproject.toml | 2 -- kube_scheduler/pyproject.toml | 2 -- kubelet/pyproject.toml | 2 -- kubernetes_state/pyproject.toml | 2 -- kyototycoon/pyproject.toml | 2 -- lighttpd/pyproject.toml | 2 -- linkerd/pyproject.toml | 2 -- linux_proc_extras/pyproject.toml | 2 -- mapr/pyproject.toml | 2 -- mapreduce/pyproject.toml | 2 -- marathon/pyproject.toml | 2 -- marklogic/pyproject.toml | 2 -- mcache/pyproject.toml | 2 -- mesos_master/pyproject.toml | 2 -- mesos_slave/pyproject.toml | 2 -- mongo/pyproject.toml | 1 - mysql/pyproject.toml | 1 - nagios/pyproject.toml | 2 -- network/pyproject.toml | 2 -- nfsstat/pyproject.toml | 2 -- nginx/pyproject.toml | 2 -- nginx_ingress_controller/pyproject.toml | 2 -- openldap/pyproject.toml | 2 -- openmetrics/pyproject.toml | 2 -- openstack/pyproject.toml | 2 -- pan_firewall/pyproject.toml | 2 -- pdh_check/pyproject.toml | 2 -- pgbouncer/pyproject.toml | 2 -- php_fpm/pyproject.toml | 2 -- postfix/pyproject.toml | 2 -- powerdns_recursor/pyproject.toml | 2 -- presto/pyproject.toml | 2 -- process/pyproject.toml | 2 -- prometheus/pyproject.toml | 2 -- proxysql/pyproject.toml | 2 -- rabbitmq/pyproject.toml | 2 -- redisdb/pyproject.toml | 2 -- rethinkdb/pyproject.toml | 2 -- riak/pyproject.toml | 2 -- riakcs/pyproject.toml | 2 -- sap_hana/pyproject.toml | 2 -- scylla/pyproject.toml | 2 -- sidekiq/pyproject.toml | 2 -- silk/pyproject.toml | 2 -- singlestore/pyproject.toml | 2 -- snmp/pyproject.toml | 2 -- solr/pyproject.toml | 2 -- sonarqube/pyproject.toml | 2 -- spark/pyproject.toml | 2 -- squid/pyproject.toml | 2 -- ssh_check/pyproject.toml | 2 -- statsd/pyproject.toml | 2 -- strimzi/pyproject.toml | 2 -- supervisord/pyproject.toml | 2 -- system_core/pyproject.toml | 2 -- system_swap/pyproject.toml | 2 -- tcp_check/pyproject.toml | 2 -- teamcity/pyproject.toml | 2 -- tenable/pyproject.toml | 2 -- tls/pyproject.toml | 2 -- tomcat/pyproject.toml | 2 -- torchserve/pyproject.toml | 1 - twemproxy/pyproject.toml | 2 -- twistlock/pyproject.toml | 2 -- varnish/pyproject.toml | 2 -- vault/pyproject.toml | 2 -- vertica/pyproject.toml | 2 -- voltdb/pyproject.toml | 2 -- vsphere/pyproject.toml | 2 -- weaviate/pyproject.toml | 1 - weblogic/pyproject.toml | 2 -- win32_event_log/pyproject.toml | 2 -- windows_performance_counters/pyproject.toml | 2 -- windows_service/pyproject.toml | 2 -- wmi_check/pyproject.toml | 2 -- yarn/pyproject.toml | 2 -- zk/pyproject.toml | 2 -- 161 files changed, 320 deletions(-) diff --git a/active_directory/pyproject.toml b/active_directory/pyproject.toml index 897819c9b3f55..ce3d92efef704 100644 --- a/active_directory/pyproject.toml +++ b/active_directory/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/activemq/pyproject.toml b/activemq/pyproject.toml index 23a51b653af42..bfaf893598d24 100644 --- a/activemq/pyproject.toml +++ b/activemq/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/activemq_xml/pyproject.toml b/activemq_xml/pyproject.toml index 35777f7bf4b37..e42459037b302 100644 --- a/activemq_xml/pyproject.toml +++ b/activemq_xml/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/aerospike/pyproject.toml b/aerospike/pyproject.toml index 22fdcebb95108..88fb0875f2afb 100644 --- a/aerospike/pyproject.toml +++ b/aerospike/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/airflow/pyproject.toml b/airflow/pyproject.toml index 4127c4be4f5cb..eeaf7fbbfa879 100644 --- a/airflow/pyproject.toml +++ b/airflow/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/amazon_msk/pyproject.toml b/amazon_msk/pyproject.toml index 6c1142eaad6a9..ebd307eedc762 100644 --- a/amazon_msk/pyproject.toml +++ b/amazon_msk/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/ambari/pyproject.toml b/ambari/pyproject.toml index 68d75059d14e7..f413bda56b026 100644 --- a/ambari/pyproject.toml +++ b/ambari/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/apache/pyproject.toml b/apache/pyproject.toml index 6f7b6eacef9f5..b5daa1295fc55 100644 --- a/apache/pyproject.toml +++ b/apache/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/aspdotnet/pyproject.toml b/aspdotnet/pyproject.toml index e2a037b894cf3..c89e080fb62df 100644 --- a/aspdotnet/pyproject.toml +++ b/aspdotnet/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/avi_vantage/pyproject.toml b/avi_vantage/pyproject.toml index 0d43b5b8b82e6..95dc2adef0a9b 100644 --- a/avi_vantage/pyproject.toml +++ b/avi_vantage/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/azure_iot_edge/pyproject.toml b/azure_iot_edge/pyproject.toml index 448ac919b3f72..0e44b286694e1 100644 --- a/azure_iot_edge/pyproject.toml +++ b/azure_iot_edge/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/btrfs/pyproject.toml b/btrfs/pyproject.toml index c2fbea3190d49..ad329e8005800 100644 --- a/btrfs/pyproject.toml +++ b/btrfs/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/cacti/pyproject.toml b/cacti/pyproject.toml index 7df50a14912a5..d6adece58933e 100644 --- a/cacti/pyproject.toml +++ b/cacti/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/calico/pyproject.toml b/calico/pyproject.toml index efa9d8085196f..5b7ad7f187cfd 100644 --- a/calico/pyproject.toml +++ b/calico/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/cassandra/pyproject.toml b/cassandra/pyproject.toml index 3f1e1714aa050..44c998fd602e5 100644 --- a/cassandra/pyproject.toml +++ b/cassandra/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/cassandra_nodetool/pyproject.toml b/cassandra_nodetool/pyproject.toml index 162d70377a05b..f97fdb85f9dc3 100644 --- a/cassandra_nodetool/pyproject.toml +++ b/cassandra_nodetool/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/ceph/pyproject.toml b/ceph/pyproject.toml index d9e145d6a6559..21bb04c8b38a4 100644 --- a/ceph/pyproject.toml +++ b/ceph/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/cert_manager/pyproject.toml b/cert_manager/pyproject.toml index f7bf2d4176e4e..4bbae65cf6a3b 100644 --- a/cert_manager/pyproject.toml +++ b/cert_manager/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.12.0", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/checkpoint_quantum_firewall/pyproject.toml b/checkpoint_quantum_firewall/pyproject.toml index 037418c676a17..8ac7d98bd8c03 100644 --- a/checkpoint_quantum_firewall/pyproject.toml +++ b/checkpoint_quantum_firewall/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.13.0", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/cilium/pyproject.toml b/cilium/pyproject.toml index 1d24619b23c1e..15c2bafbfd99b 100644 --- a/cilium/pyproject.toml +++ b/cilium/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/cisco_aci/pyproject.toml b/cisco_aci/pyproject.toml index f487215f43317..7a0c4345b6ed9 100644 --- a/cisco_aci/pyproject.toml +++ b/cisco_aci/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/citrix_hypervisor/pyproject.toml b/citrix_hypervisor/pyproject.toml index 749d4a01d6dd1..cd321e2547518 100644 --- a/citrix_hypervisor/pyproject.toml +++ b/citrix_hypervisor/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/clickhouse/pyproject.toml b/clickhouse/pyproject.toml index bbf7a3460a3aa..0f5e57d9f085f 100644 --- a/clickhouse/pyproject.toml +++ b/clickhouse/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/cloud_foundry_api/pyproject.toml b/cloud_foundry_api/pyproject.toml index cedda6fcccfe0..6cfd63cddaafa 100644 --- a/cloud_foundry_api/pyproject.toml +++ b/cloud_foundry_api/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/cloudera/pyproject.toml b/cloudera/pyproject.toml index 2c13d3220d144..247263f9b90eb 100644 --- a/cloudera/pyproject.toml +++ b/cloudera/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.13.0", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/cockroachdb/pyproject.toml b/cockroachdb/pyproject.toml index 114e70ab2e47d..9e279b8514b32 100644 --- a/cockroachdb/pyproject.toml +++ b/cockroachdb/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/confluent_platform/pyproject.toml b/confluent_platform/pyproject.toml index 7071ace9f7e68..6a56d8658d0f6 100644 --- a/confluent_platform/pyproject.toml +++ b/confluent_platform/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/consul/pyproject.toml b/consul/pyproject.toml index fa62e679d923b..da54e2f1de7f1 100644 --- a/consul/pyproject.toml +++ b/consul/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/coredns/pyproject.toml b/coredns/pyproject.toml index 740650969fa90..9691704eb7d9e 100644 --- a/coredns/pyproject.toml +++ b/coredns/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/couch/pyproject.toml b/couch/pyproject.toml index fa027dcd0f838..a863dd3057df6 100644 --- a/couch/pyproject.toml +++ b/couch/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/couchbase/pyproject.toml b/couchbase/pyproject.toml index a69e6de25da1d..d5e690299020f 100644 --- a/couchbase/pyproject.toml +++ b/couchbase/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/crio/pyproject.toml b/crio/pyproject.toml index f1ce75103ed2f..e8a4cdedcfcd6 100644 --- a/crio/pyproject.toml +++ b/crio/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/datadog_checks_base/pyproject.toml b/datadog_checks_base/pyproject.toml index 1abe90220249a..d6b0733532c1f 100644 --- a/datadog_checks_base/pyproject.toml +++ b/datadog_checks_base/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/datadog_checks_dependency_provider/pyproject.toml b/datadog_checks_dependency_provider/pyproject.toml index 0463186ed5754..814015227d79b 100644 --- a/datadog_checks_dependency_provider/pyproject.toml +++ b/datadog_checks_dependency_provider/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/datadog_checks_dev/pyproject.toml b/datadog_checks_dev/pyproject.toml index 9df85c415a22b..260e4dc6428cd 100644 --- a/datadog_checks_dev/pyproject.toml +++ b/datadog_checks_dev/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" @@ -89,9 +87,6 @@ cli = [ "twine>=1.11.0", "virtualenv<20.22.0", # TODO: Remove once every check has a pyproject.toml - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", - "wheel>=0.31.0", ] [project.entry-points.pytest11] diff --git a/datadog_checks_downloader/pyproject.toml b/datadog_checks_downloader/pyproject.toml index 2974e65abcc60..78d294fc2c47d 100644 --- a/datadog_checks_downloader/pyproject.toml +++ b/datadog_checks_downloader/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/datadog_cluster_agent/pyproject.toml b/datadog_cluster_agent/pyproject.toml index d5f3a5302a18f..b64cd13ebbfb1 100644 --- a/datadog_cluster_agent/pyproject.toml +++ b/datadog_cluster_agent/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/dcgm/pyproject.toml b/dcgm/pyproject.toml index ce49134e9a5cd..d19f4a1129f17 100644 --- a/dcgm/pyproject.toml +++ b/dcgm/pyproject.toml @@ -1,7 +1,6 @@ [build-system] requires = [ "hatchling>=0.13.0", - "setuptools>=66; python_version > '3.0'", ] build-backend = "hatchling.build" diff --git a/directory/pyproject.toml b/directory/pyproject.toml index 732fe2502167b..a89d0b6042d98 100644 --- a/directory/pyproject.toml +++ b/directory/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/disk/pyproject.toml b/disk/pyproject.toml index 8f115e41ca7e4..1145213a95847 100644 --- a/disk/pyproject.toml +++ b/disk/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/dns_check/pyproject.toml b/dns_check/pyproject.toml index b783ec388a574..ee153851cfa73 100644 --- a/dns_check/pyproject.toml +++ b/dns_check/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/dotnetclr/pyproject.toml b/dotnetclr/pyproject.toml index 0f1423ce59189..7d6cca1c982a9 100644 --- a/dotnetclr/pyproject.toml +++ b/dotnetclr/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/druid/pyproject.toml b/druid/pyproject.toml index 3b4106a662c79..53dd0af94551b 100644 --- a/druid/pyproject.toml +++ b/druid/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/ecs_fargate/pyproject.toml b/ecs_fargate/pyproject.toml index 74b73333b4314..0c0cd54922a1a 100644 --- a/ecs_fargate/pyproject.toml +++ b/ecs_fargate/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/eks_fargate/pyproject.toml b/eks_fargate/pyproject.toml index 473c4623800e8..d86e6a8ebbec5 100644 --- a/eks_fargate/pyproject.toml +++ b/eks_fargate/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/elastic/pyproject.toml b/elastic/pyproject.toml index bbf130aa834ae..e5997716029af 100644 --- a/elastic/pyproject.toml +++ b/elastic/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/envoy/pyproject.toml b/envoy/pyproject.toml index b4f0eeb46ade2..21debd8dbc4bd 100644 --- a/envoy/pyproject.toml +++ b/envoy/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/etcd/pyproject.toml b/etcd/pyproject.toml index b8c9ff21dd916..acba1be1680e2 100644 --- a/etcd/pyproject.toml +++ b/etcd/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/exchange_server/pyproject.toml b/exchange_server/pyproject.toml index ed6224db76ec4..f7c25944b6616 100644 --- a/exchange_server/pyproject.toml +++ b/exchange_server/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/external_dns/pyproject.toml b/external_dns/pyproject.toml index 4f602024103b9..e742a5b1d076f 100644 --- a/external_dns/pyproject.toml +++ b/external_dns/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/flink/pyproject.toml b/flink/pyproject.toml index da7755c0c7ac1..1d0cd2c7a4b78 100644 --- a/flink/pyproject.toml +++ b/flink/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/fluentd/pyproject.toml b/fluentd/pyproject.toml index 87b73f42a2b8a..53518887e4404 100644 --- a/fluentd/pyproject.toml +++ b/fluentd/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/foundationdb/pyproject.toml b/foundationdb/pyproject.toml index 032ced7bbd166..bb799171099d3 100644 --- a/foundationdb/pyproject.toml +++ b/foundationdb/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/gearmand/pyproject.toml b/gearmand/pyproject.toml index ea6bc2cabdb37..89d8e9d36d260 100644 --- a/gearmand/pyproject.toml +++ b/gearmand/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/gitlab/pyproject.toml b/gitlab/pyproject.toml index ebbab4db5e734..4753ee16574c2 100644 --- a/gitlab/pyproject.toml +++ b/gitlab/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/gitlab_runner/pyproject.toml b/gitlab_runner/pyproject.toml index fdf0a099b5c8b..e3ad62a69731b 100644 --- a/gitlab_runner/pyproject.toml +++ b/gitlab_runner/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/glusterfs/pyproject.toml b/glusterfs/pyproject.toml index cb6613ed73130..531c51aaebc5f 100644 --- a/glusterfs/pyproject.toml +++ b/glusterfs/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/go_expvar/pyproject.toml b/go_expvar/pyproject.toml index 39309d2dae0fc..0632f15f1d614 100644 --- a/go_expvar/pyproject.toml +++ b/go_expvar/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/gunicorn/pyproject.toml b/gunicorn/pyproject.toml index ebd7dc8cf89f6..b62153254f0f4 100644 --- a/gunicorn/pyproject.toml +++ b/gunicorn/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/haproxy/pyproject.toml b/haproxy/pyproject.toml index 21da244fd7919..cd653fd914b2c 100644 --- a/haproxy/pyproject.toml +++ b/haproxy/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/harbor/pyproject.toml b/harbor/pyproject.toml index 4657b2da91cc0..dbfbd25cd6487 100644 --- a/harbor/pyproject.toml +++ b/harbor/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/hazelcast/pyproject.toml b/hazelcast/pyproject.toml index ebab01cbd6dfb..4564f966a80ac 100644 --- a/hazelcast/pyproject.toml +++ b/hazelcast/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/hdfs_datanode/pyproject.toml b/hdfs_datanode/pyproject.toml index f14a69c03641d..5499dcc8a74da 100644 --- a/hdfs_datanode/pyproject.toml +++ b/hdfs_datanode/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/hdfs_namenode/pyproject.toml b/hdfs_namenode/pyproject.toml index 0bf4d20e9d38f..825527f474f0a 100644 --- a/hdfs_namenode/pyproject.toml +++ b/hdfs_namenode/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/hive/pyproject.toml b/hive/pyproject.toml index ed12047b5fe46..514c8aa69e3a4 100644 --- a/hive/pyproject.toml +++ b/hive/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/hivemq/pyproject.toml b/hivemq/pyproject.toml index 9c988df9a7006..d763ef1e53457 100644 --- a/hivemq/pyproject.toml +++ b/hivemq/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/http_check/pyproject.toml b/http_check/pyproject.toml index 4ae432a765d77..b41c1f1ebba57 100644 --- a/http_check/pyproject.toml +++ b/http_check/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/hudi/pyproject.toml b/hudi/pyproject.toml index 6c3e21b152f42..0e3a3cf445fc2 100644 --- a/hudi/pyproject.toml +++ b/hudi/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/hyperv/pyproject.toml b/hyperv/pyproject.toml index dbfc0c79c7c8f..456e7061193c7 100644 --- a/hyperv/pyproject.toml +++ b/hyperv/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/ibm_db2/pyproject.toml b/ibm_db2/pyproject.toml index 786fe2ef6ae6e..d8eaebe309812 100644 --- a/ibm_db2/pyproject.toml +++ b/ibm_db2/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/ibm_i/pyproject.toml b/ibm_i/pyproject.toml index dcc576ad1f146..d462d08aca0c7 100644 --- a/ibm_i/pyproject.toml +++ b/ibm_i/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/ibm_was/pyproject.toml b/ibm_was/pyproject.toml index 246db8ba8a3c8..e2d191f7a4ba7 100644 --- a/ibm_was/pyproject.toml +++ b/ibm_was/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/ignite/pyproject.toml b/ignite/pyproject.toml index 68c263e8adaee..ca3c40df08bcc 100644 --- a/ignite/pyproject.toml +++ b/ignite/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/iis/pyproject.toml b/iis/pyproject.toml index ffbe2341cad3a..bb3a5cac571f2 100644 --- a/iis/pyproject.toml +++ b/iis/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/impala/pyproject.toml b/impala/pyproject.toml index 63030092aa80d..0dae740e828f1 100644 --- a/impala/pyproject.toml +++ b/impala/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.13.0", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/istio/pyproject.toml b/istio/pyproject.toml index 54b73e402d7a6..2b047dee6e78f 100644 --- a/istio/pyproject.toml +++ b/istio/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/jboss_wildfly/pyproject.toml b/jboss_wildfly/pyproject.toml index f87450ffdd6c3..fb2f44729534d 100644 --- a/jboss_wildfly/pyproject.toml +++ b/jboss_wildfly/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/journald/pyproject.toml b/journald/pyproject.toml index 968b34e9a8aff..6555be0cdbab1 100644 --- a/journald/pyproject.toml +++ b/journald/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/kafka/pyproject.toml b/kafka/pyproject.toml index 7f416f1d0026d..133dfb3d0d1f9 100644 --- a/kafka/pyproject.toml +++ b/kafka/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/kong/pyproject.toml b/kong/pyproject.toml index d43e243e5e6dc..357ae6aaf95d8 100644 --- a/kong/pyproject.toml +++ b/kong/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/kube_apiserver_metrics/pyproject.toml b/kube_apiserver_metrics/pyproject.toml index 6024af9b9d880..c190643a8aece 100644 --- a/kube_apiserver_metrics/pyproject.toml +++ b/kube_apiserver_metrics/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/kube_controller_manager/pyproject.toml b/kube_controller_manager/pyproject.toml index 1b59e30f28a47..31ba6db19fc52 100644 --- a/kube_controller_manager/pyproject.toml +++ b/kube_controller_manager/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/kube_dns/pyproject.toml b/kube_dns/pyproject.toml index adca7847014e8..8a9f05e1629bc 100644 --- a/kube_dns/pyproject.toml +++ b/kube_dns/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/kube_metrics_server/pyproject.toml b/kube_metrics_server/pyproject.toml index f95348d19b488..538e167839bbf 100644 --- a/kube_metrics_server/pyproject.toml +++ b/kube_metrics_server/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/kube_proxy/pyproject.toml b/kube_proxy/pyproject.toml index 60a2668b96fcb..7634d1e44c075 100644 --- a/kube_proxy/pyproject.toml +++ b/kube_proxy/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/kube_scheduler/pyproject.toml b/kube_scheduler/pyproject.toml index a3004a7f64d0b..cc5fdfec29f27 100644 --- a/kube_scheduler/pyproject.toml +++ b/kube_scheduler/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/kubelet/pyproject.toml b/kubelet/pyproject.toml index 075bd5b2e3e23..277505c22391b 100644 --- a/kubelet/pyproject.toml +++ b/kubelet/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/kubernetes_state/pyproject.toml b/kubernetes_state/pyproject.toml index 0f0a249f53965..eaf84a482535c 100644 --- a/kubernetes_state/pyproject.toml +++ b/kubernetes_state/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/kyototycoon/pyproject.toml b/kyototycoon/pyproject.toml index da1097d1def25..6fe8a56b10179 100644 --- a/kyototycoon/pyproject.toml +++ b/kyototycoon/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/lighttpd/pyproject.toml b/lighttpd/pyproject.toml index 249c13c79d2e0..c217ae3dd2215 100644 --- a/lighttpd/pyproject.toml +++ b/lighttpd/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/linkerd/pyproject.toml b/linkerd/pyproject.toml index df2c9fbee83e2..43ba293d03b71 100644 --- a/linkerd/pyproject.toml +++ b/linkerd/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/linux_proc_extras/pyproject.toml b/linux_proc_extras/pyproject.toml index 0c7e01fb6d18b..6f8ac49097474 100644 --- a/linux_proc_extras/pyproject.toml +++ b/linux_proc_extras/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/mapr/pyproject.toml b/mapr/pyproject.toml index 7367a7801b995..437f93a253c95 100644 --- a/mapr/pyproject.toml +++ b/mapr/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/mapreduce/pyproject.toml b/mapreduce/pyproject.toml index f5baba953f813..c443c83422344 100644 --- a/mapreduce/pyproject.toml +++ b/mapreduce/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/marathon/pyproject.toml b/marathon/pyproject.toml index 67c8b66786a92..49e958af5e3c0 100644 --- a/marathon/pyproject.toml +++ b/marathon/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/marklogic/pyproject.toml b/marklogic/pyproject.toml index 3ed9fc2121461..432ae428a14ca 100644 --- a/marklogic/pyproject.toml +++ b/marklogic/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/mcache/pyproject.toml b/mcache/pyproject.toml index 6e151da6d698d..1e67271f4ad21 100644 --- a/mcache/pyproject.toml +++ b/mcache/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/mesos_master/pyproject.toml b/mesos_master/pyproject.toml index 80335fc3bdd1f..a906fb0fe95fb 100644 --- a/mesos_master/pyproject.toml +++ b/mesos_master/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/mesos_slave/pyproject.toml b/mesos_slave/pyproject.toml index 6367dd7e7163b..8f538913b7693 100644 --- a/mesos_slave/pyproject.toml +++ b/mesos_slave/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/mongo/pyproject.toml b/mongo/pyproject.toml index 87eb3c784e024..58b6ac363f75b 100644 --- a/mongo/pyproject.toml +++ b/mongo/pyproject.toml @@ -1,7 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66", ] build-backend = "hatchling.build" diff --git a/mysql/pyproject.toml b/mysql/pyproject.toml index 097ffe0d70c8c..4d553d3200f48 100644 --- a/mysql/pyproject.toml +++ b/mysql/pyproject.toml @@ -1,7 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", ] build-backend = "hatchling.build" diff --git a/nagios/pyproject.toml b/nagios/pyproject.toml index 5acab33b20ad2..51a7a56c91f92 100644 --- a/nagios/pyproject.toml +++ b/nagios/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/network/pyproject.toml b/network/pyproject.toml index 3615536bf265b..4cf3786b61423 100644 --- a/network/pyproject.toml +++ b/network/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/nfsstat/pyproject.toml b/nfsstat/pyproject.toml index a6b6da09914c1..5a3d8724b65de 100644 --- a/nfsstat/pyproject.toml +++ b/nfsstat/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/nginx/pyproject.toml b/nginx/pyproject.toml index a6d4f91af4197..6c501720b7320 100644 --- a/nginx/pyproject.toml +++ b/nginx/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/nginx_ingress_controller/pyproject.toml b/nginx_ingress_controller/pyproject.toml index 62b2b0ece04dc..a16b6e3d57c1b 100644 --- a/nginx_ingress_controller/pyproject.toml +++ b/nginx_ingress_controller/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/openldap/pyproject.toml b/openldap/pyproject.toml index 4c156c54c3cae..328721da10bb4 100644 --- a/openldap/pyproject.toml +++ b/openldap/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/openmetrics/pyproject.toml b/openmetrics/pyproject.toml index c53877f79980d..197ed072d29e8 100644 --- a/openmetrics/pyproject.toml +++ b/openmetrics/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/openstack/pyproject.toml b/openstack/pyproject.toml index 3fdeb185d079f..9d5cda6e2109a 100644 --- a/openstack/pyproject.toml +++ b/openstack/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/pan_firewall/pyproject.toml b/pan_firewall/pyproject.toml index 1a856ed8e9e84..c2f2bc28e4305 100644 --- a/pan_firewall/pyproject.toml +++ b/pan_firewall/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/pdh_check/pyproject.toml b/pdh_check/pyproject.toml index cb3fb46601bfe..7d3cc45d2ab73 100644 --- a/pdh_check/pyproject.toml +++ b/pdh_check/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/pgbouncer/pyproject.toml b/pgbouncer/pyproject.toml index 2a28b83831889..ef7d3945dd460 100644 --- a/pgbouncer/pyproject.toml +++ b/pgbouncer/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/php_fpm/pyproject.toml b/php_fpm/pyproject.toml index 4ddf8d0171e18..42309b3bac449 100644 --- a/php_fpm/pyproject.toml +++ b/php_fpm/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/postfix/pyproject.toml b/postfix/pyproject.toml index 43b67a636f743..6640808536963 100644 --- a/postfix/pyproject.toml +++ b/postfix/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/powerdns_recursor/pyproject.toml b/powerdns_recursor/pyproject.toml index 4bb1f5ec606ea..ab60f4c68a028 100644 --- a/powerdns_recursor/pyproject.toml +++ b/powerdns_recursor/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/presto/pyproject.toml b/presto/pyproject.toml index b1759f4733e7d..12426a0f28157 100644 --- a/presto/pyproject.toml +++ b/presto/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/process/pyproject.toml b/process/pyproject.toml index 6d603ff3a7279..419ff8e36d0b6 100644 --- a/process/pyproject.toml +++ b/process/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/prometheus/pyproject.toml b/prometheus/pyproject.toml index 4cbfb5e16b974..82a56e598d2c0 100644 --- a/prometheus/pyproject.toml +++ b/prometheus/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/proxysql/pyproject.toml b/proxysql/pyproject.toml index 1be7a7dcfa63f..834934e69af73 100644 --- a/proxysql/pyproject.toml +++ b/proxysql/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/rabbitmq/pyproject.toml b/rabbitmq/pyproject.toml index b5452d3a95fa6..8584b833c1e83 100644 --- a/rabbitmq/pyproject.toml +++ b/rabbitmq/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/redisdb/pyproject.toml b/redisdb/pyproject.toml index 0878b830417f4..c2659bec9e30b 100644 --- a/redisdb/pyproject.toml +++ b/redisdb/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/rethinkdb/pyproject.toml b/rethinkdb/pyproject.toml index 648f97b304b8d..d6a7b2131e771 100644 --- a/rethinkdb/pyproject.toml +++ b/rethinkdb/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/riak/pyproject.toml b/riak/pyproject.toml index 78cdbca34de5c..528993df98845 100644 --- a/riak/pyproject.toml +++ b/riak/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/riakcs/pyproject.toml b/riakcs/pyproject.toml index 3187d9cbd138e..40d1320775859 100644 --- a/riakcs/pyproject.toml +++ b/riakcs/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/sap_hana/pyproject.toml b/sap_hana/pyproject.toml index 04d415ff21bb9..0070cc842f6d2 100644 --- a/sap_hana/pyproject.toml +++ b/sap_hana/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/scylla/pyproject.toml b/scylla/pyproject.toml index dd9e1663f20b1..f2d17b5a7a34e 100644 --- a/scylla/pyproject.toml +++ b/scylla/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/sidekiq/pyproject.toml b/sidekiq/pyproject.toml index c2ca2e73f32c9..aef4e2764a9e5 100644 --- a/sidekiq/pyproject.toml +++ b/sidekiq/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/silk/pyproject.toml b/silk/pyproject.toml index 356f6b8e1c3ee..94b421133d15f 100644 --- a/silk/pyproject.toml +++ b/silk/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/singlestore/pyproject.toml b/singlestore/pyproject.toml index 3c5e25300a652..516b28fb1df11 100644 --- a/singlestore/pyproject.toml +++ b/singlestore/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/snmp/pyproject.toml b/snmp/pyproject.toml index cd6b1cbbab567..2b12bbde47bfd 100644 --- a/snmp/pyproject.toml +++ b/snmp/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/solr/pyproject.toml b/solr/pyproject.toml index 2dddeb2ba076b..0d23777a0085a 100644 --- a/solr/pyproject.toml +++ b/solr/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/sonarqube/pyproject.toml b/sonarqube/pyproject.toml index d3b1f94b775f9..e9480a2a6c7e4 100644 --- a/sonarqube/pyproject.toml +++ b/sonarqube/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/spark/pyproject.toml b/spark/pyproject.toml index 02b6f5d326e56..26f51e94578a1 100644 --- a/spark/pyproject.toml +++ b/spark/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/squid/pyproject.toml b/squid/pyproject.toml index 0023f6baed252..6af7cb3842749 100644 --- a/squid/pyproject.toml +++ b/squid/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/ssh_check/pyproject.toml b/ssh_check/pyproject.toml index 767f7829e7e2a..d77643787a307 100644 --- a/ssh_check/pyproject.toml +++ b/ssh_check/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/statsd/pyproject.toml b/statsd/pyproject.toml index c8506dba0449f..ad1b1f016a4a6 100644 --- a/statsd/pyproject.toml +++ b/statsd/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/strimzi/pyproject.toml b/strimzi/pyproject.toml index 6e66eed9d976d..13347ac7f50f7 100644 --- a/strimzi/pyproject.toml +++ b/strimzi/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.13.0", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/supervisord/pyproject.toml b/supervisord/pyproject.toml index cf12235d87f63..5803a53bc4654 100644 --- a/supervisord/pyproject.toml +++ b/supervisord/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/system_core/pyproject.toml b/system_core/pyproject.toml index 798e7d43a5a85..5b520d4b36a43 100644 --- a/system_core/pyproject.toml +++ b/system_core/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/system_swap/pyproject.toml b/system_swap/pyproject.toml index 33e6da5baa862..d5a7a97cccd36 100644 --- a/system_swap/pyproject.toml +++ b/system_swap/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/tcp_check/pyproject.toml b/tcp_check/pyproject.toml index bd25ea3c22089..1fc6efb28e97b 100644 --- a/tcp_check/pyproject.toml +++ b/tcp_check/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/teamcity/pyproject.toml b/teamcity/pyproject.toml index e4f599ec2385b..279c59054f6e1 100644 --- a/teamcity/pyproject.toml +++ b/teamcity/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/tenable/pyproject.toml b/tenable/pyproject.toml index fe274a81abac0..88529d5ab3716 100644 --- a/tenable/pyproject.toml +++ b/tenable/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/tls/pyproject.toml b/tls/pyproject.toml index 2f2ed7f180b13..f0804b755040d 100644 --- a/tls/pyproject.toml +++ b/tls/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/tomcat/pyproject.toml b/tomcat/pyproject.toml index d751760e58a1b..938819a7384b6 100644 --- a/tomcat/pyproject.toml +++ b/tomcat/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/torchserve/pyproject.toml b/torchserve/pyproject.toml index d1ebc2e01fe8f..4716582e2a3f1 100644 --- a/torchserve/pyproject.toml +++ b/torchserve/pyproject.toml @@ -1,7 +1,6 @@ [build-system] requires = [ "hatchling>=0.13.0", - "setuptools>=66; python_version > '3.0'", ] build-backend = "hatchling.build" diff --git a/twemproxy/pyproject.toml b/twemproxy/pyproject.toml index 32b327e1dd8ad..ffa555b63b915 100644 --- a/twemproxy/pyproject.toml +++ b/twemproxy/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/twistlock/pyproject.toml b/twistlock/pyproject.toml index 18b4af8f763cd..9a1e38397740d 100644 --- a/twistlock/pyproject.toml +++ b/twistlock/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/varnish/pyproject.toml b/varnish/pyproject.toml index d7e4550113b7c..611985b6b73f4 100644 --- a/varnish/pyproject.toml +++ b/varnish/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/vault/pyproject.toml b/vault/pyproject.toml index ba1e84eecd9c0..6872b042fc30e 100644 --- a/vault/pyproject.toml +++ b/vault/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/vertica/pyproject.toml b/vertica/pyproject.toml index 76b70944f84de..d7deb769cce08 100644 --- a/vertica/pyproject.toml +++ b/vertica/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/voltdb/pyproject.toml b/voltdb/pyproject.toml index cb31d181314b3..802308acb89b5 100644 --- a/voltdb/pyproject.toml +++ b/voltdb/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/vsphere/pyproject.toml b/vsphere/pyproject.toml index 84a9a2494e5bb..abe0f45d27db4 100644 --- a/vsphere/pyproject.toml +++ b/vsphere/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/weaviate/pyproject.toml b/weaviate/pyproject.toml index cc55f66e26560..8317ba41feb96 100644 --- a/weaviate/pyproject.toml +++ b/weaviate/pyproject.toml @@ -1,7 +1,6 @@ [build-system] requires = [ "hatchling>=0.13.0", - "setuptools>=66; python_version > '3.0'", ] build-backend = "hatchling.build" diff --git a/weblogic/pyproject.toml b/weblogic/pyproject.toml index a7da09a3b783b..aa73c2e69717d 100644 --- a/weblogic/pyproject.toml +++ b/weblogic/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/win32_event_log/pyproject.toml b/win32_event_log/pyproject.toml index 0fcecb62bc12a..26821e04bf8ed 100644 --- a/win32_event_log/pyproject.toml +++ b/win32_event_log/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/windows_performance_counters/pyproject.toml b/windows_performance_counters/pyproject.toml index 5a6f7bcbceea5..59796f6826314 100644 --- a/windows_performance_counters/pyproject.toml +++ b/windows_performance_counters/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/windows_service/pyproject.toml b/windows_service/pyproject.toml index b8ef6e45c60d7..5bd0e589e8112 100644 --- a/windows_service/pyproject.toml +++ b/windows_service/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/wmi_check/pyproject.toml b/wmi_check/pyproject.toml index 107d53c91a506..af5e952f3371e 100644 --- a/wmi_check/pyproject.toml +++ b/wmi_check/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/yarn/pyproject.toml b/yarn/pyproject.toml index 1ea6d46ca022c..084dd1b8b76b9 100644 --- a/yarn/pyproject.toml +++ b/yarn/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" diff --git a/zk/pyproject.toml b/zk/pyproject.toml index b4e47f914c8ca..d04a399c570bf 100644 --- a/zk/pyproject.toml +++ b/zk/pyproject.toml @@ -1,8 +1,6 @@ [build-system] requires = [ "hatchling>=0.11.2", - "setuptools>=66; python_version > '3.0'", - "setuptools; python_version < '3.0'", ] build-backend = "hatchling.build" From 7178fb3513afa5de0c99b1ec6d92706e5e3ccb56 Mon Sep 17 00:00:00 2001 From: HadhemiDD <43783545+HadhemiDD@users.noreply.github.com> Date: Mon, 23 Sep 2024 10:22:52 +0200 Subject: [PATCH 31/34] add test for read_persistent_cache (#18634) --- kafka_consumer/tests/test_unit.py | 39 +++++++++++++++++++++++++++++++ 1 file changed, 39 insertions(+) diff --git a/kafka_consumer/tests/test_unit.py b/kafka_consumer/tests/test_unit.py index 569a4e8915c9e..ac593a9761e94 100644 --- a/kafka_consumer/tests/test_unit.py +++ b/kafka_consumer/tests/test_unit.py @@ -389,3 +389,42 @@ def test_get_interpolated_timestamp(): assert _get_interpolated_timestamp({10: 100, 20: 200}, 5) == 50 assert _get_interpolated_timestamp({0: 100, 10: 200}, 15) == 250 assert _get_interpolated_timestamp({10: 200}, 15) is None + + +@pytest.mark.parametrize( + 'read_persistent_cache, kafka_instance_config, consumer_lag_seconds_count', + [ + pytest.param( + "", + { + 'consumer_groups': {}, + 'data_streams_enabled': 'true', + 'monitor_unlisted_consumer_groups': True, + }, + 0, + id='Read from cache failed', + ), + ], +) +def test_load_broker_timestamps_empty( + read_persistent_cache, + kafka_instance_config, + consumer_lag_seconds_count, + kafka_instance, + dd_run_check, + caplog, + aggregator, + check, +): + + kafka_instance.update(kafka_instance_config) + check = check(kafka_instance) + check.read_persistent_cache = mock.Mock(return_value=read_persistent_cache) + dd_run_check(check) + + caplog.set_level(logging.WARN) + expected_warning = " Could not read broker timestamps from cache" + + assert expected_warning in caplog.text + aggregator.assert_metric("kafka.estimated_consumer_lag", count=consumer_lag_seconds_count) + assert check.read_persistent_cache.mock_calls == [mock.call("broker_timestamps_")] From be5a704d2028f7dbdeb4c9dfba74053f11d40c03 Mon Sep 17 00:00:00 2001 From: HadhemiDD <43783545+HadhemiDD@users.noreply.github.com> Date: Mon, 23 Sep 2024 10:55:31 +0200 Subject: [PATCH 32/34] fix link (#18645) --- kubeflow/README.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/kubeflow/README.md b/kubeflow/README.md index 634017875995c..b801a22c3dedf 100644 --- a/kubeflow/README.md +++ b/kubeflow/README.md @@ -127,7 +127,7 @@ See [service_checks.json][8] for a list of service checks provided by this integ Need help? Contact [Datadog support][9]. -[1]: **LINK_TO_INTEGRATION_SITE** +[1]: https://docs.datadoghq.com/integrations/kubeflow/ [2]: https://app.datadoghq.com/account/settings/agent/latest [3]: https://docs.datadoghq.com/agent/kubernetes/integrations/ [4]: https://github.com/DataDog/integrations-core/blob/master/kubeflow/datadog_checks/kubeflow/data/conf.yaml.example From 18a6120e5756ed6ed15bebf035c75d19e6504d5b Mon Sep 17 00:00:00 2001 From: Steven Yuen Date: Mon, 23 Sep 2024 09:30:31 -0400 Subject: [PATCH 33/34] add forgotten changelog (#18647) --- karpenter/changelog.d/18448.added | 1 + 1 file changed, 1 insertion(+) create mode 100644 karpenter/changelog.d/18448.added diff --git a/karpenter/changelog.d/18448.added b/karpenter/changelog.d/18448.added new file mode 100644 index 0000000000000..d65bcbf03b74b --- /dev/null +++ b/karpenter/changelog.d/18448.added @@ -0,0 +1 @@ +Update Karpenter Metrics for Karpenter V1 \ No newline at end of file From 56c6fe9aab983602fe8ac19f36111f698675243e Mon Sep 17 00:00:00 2001 From: Steven Yuen Date: Mon, 23 Sep 2024 10:05:31 -0400 Subject: [PATCH 34/34] [Release] Bumped karpenter version to 1.6.0 (#18651) --- karpenter/CHANGELOG.md | 6 ++++++ karpenter/changelog.d/18448.added | 1 - karpenter/datadog_checks/karpenter/__about__.py | 2 +- requirements-agent-release.txt | 2 +- 4 files changed, 8 insertions(+), 3 deletions(-) delete mode 100644 karpenter/changelog.d/18448.added diff --git a/karpenter/CHANGELOG.md b/karpenter/CHANGELOG.md index 4779057849875..1e9bb35171829 100644 --- a/karpenter/CHANGELOG.md +++ b/karpenter/CHANGELOG.md @@ -2,6 +2,12 @@ +## 1.6.0 / 2024-09-23 + +***Added***: + +* Update Karpenter Metrics for Karpenter V1 ([#18448](https://github.com/DataDog/integrations-core/pull/18448)) + ## 1.5.0 / 2024-08-09 / Agent 7.57.0 ***Added***: diff --git a/karpenter/changelog.d/18448.added b/karpenter/changelog.d/18448.added deleted file mode 100644 index d65bcbf03b74b..0000000000000 --- a/karpenter/changelog.d/18448.added +++ /dev/null @@ -1 +0,0 @@ -Update Karpenter Metrics for Karpenter V1 \ No newline at end of file diff --git a/karpenter/datadog_checks/karpenter/__about__.py b/karpenter/datadog_checks/karpenter/__about__.py index 272f7a8636293..33cc8c84a2eab 100644 --- a/karpenter/datadog_checks/karpenter/__about__.py +++ b/karpenter/datadog_checks/karpenter/__about__.py @@ -1,4 +1,4 @@ # (C) Datadog, Inc. 2023-present # All rights reserved # Licensed under a 3-clause BSD style license (see LICENSE) -__version__ = '1.5.0' +__version__ = '1.6.0' diff --git a/requirements-agent-release.txt b/requirements-agent-release.txt index 1f0d38e7fbd76..90757a9832a37 100644 --- a/requirements-agent-release.txt +++ b/requirements-agent-release.txt @@ -92,7 +92,7 @@ datadog-jboss-wildfly==2.2.0 datadog-journald==1.2.0 datadog-kafka-consumer==4.6.1 datadog-kafka==2.16.0 -datadog-karpenter==1.5.0 +datadog-karpenter==1.6.0 datadog-kong==3.2.2 datadog-kube-apiserver-metrics==4.3.1 datadog-kube-controller-manager==5.1.1