From d0b010618f96be6fb3c7e4f584b571a9489e06b9 Mon Sep 17 00:00:00 2001 From: alanty <45295374+alanty@users.noreply.github.com> Date: Thu, 31 Oct 2024 14:58:38 -0700 Subject: [PATCH] chore: Update Spark v4 addons (#692) --- .../terraform/spark-k8s-operator/addons.tf | 16 +++++++++---- .../helm-values/kubecost-values.yaml | 24 +++++++++++++++++-- 2 files changed, 33 insertions(+), 7 deletions(-) diff --git a/analytics/terraform/spark-k8s-operator/addons.tf b/analytics/terraform/spark-k8s-operator/addons.tf index ac8fb485b..9f4afdfd4 100644 --- a/analytics/terraform/spark-k8s-operator/addons.tf +++ b/analytics/terraform/spark-k8s-operator/addons.tf @@ -405,6 +405,7 @@ module "eks_data_addons" { # Spark history server is required only when EMR Spark Operator is enabled enable_spark_history_server = true spark_history_server_helm_config = { + chart_version = "1.2.0" values = [ <<-EOT sparkHistoryOpts: "-Dspark.history.fs.logDirectory=s3a://${module.s3_bucket.s3_bucket_id}/${aws_s3_object.this.key}" @@ -417,6 +418,7 @@ module "eks_data_addons" { #--------------------------------------------------------------- enable_kubecost = true kubecost_helm_config = { + chart_version = "2.3.3" values = [templatefile("${path.module}/helm-values/kubecost-values.yaml", {})] repository_username = data.aws_ecrpublic_authorization_token.token.user_name repository_password = data.aws_ecrpublic_authorization_token.token.password @@ -476,7 +478,8 @@ module "eks_blueprints_addons" { #--------------------------------------- enable_metrics_server = true metrics_server = { - values = [templatefile("${path.module}/helm-values/metrics-server-values.yaml", {})] + chart_version = "3.12.2" + values = [templatefile("${path.module}/helm-values/metrics-server-values.yaml", {})] } #--------------------------------------- @@ -484,6 +487,7 @@ module "eks_blueprints_addons" { #--------------------------------------- enable_cluster_autoscaler = true cluster_autoscaler = { + chart_version = "9.43.1" values = [templatefile("${path.module}/helm-values/cluster-autoscaler-values.yaml", { aws_region = var.region, eks_cluster_id = module.eks.cluster_name @@ -511,7 +515,8 @@ module "eks_blueprints_addons" { #--------------------------------------- enable_aws_cloudwatch_metrics = true aws_cloudwatch_metrics = { - values = [templatefile("${path.module}/helm-values/aws-cloudwatch-metrics-values.yaml", {})] + chart_version = "0.0.11" + values = [templatefile("${path.module}/helm-values/aws-cloudwatch-metrics-values.yaml", {})] } #--------------------------------------- @@ -524,6 +529,7 @@ module "eks_blueprints_addons" { retention_in_days = 30 } aws_for_fluentbit = { + chart_version = "0.1.34" s3_bucket_arns = [ module.s3_bucket.s3_bucket_arn, "${module.s3_bucket.s3_bucket_arn}/*" @@ -538,7 +544,7 @@ module "eks_blueprints_addons" { enable_aws_load_balancer_controller = true aws_load_balancer_controller = { - chart_version = "1.5.4" + chart_version = "1.9.2" set = [{ name = "enableServiceMutatorWebhook" value = "false" @@ -547,7 +553,7 @@ module "eks_blueprints_addons" { enable_ingress_nginx = true ingress_nginx = { - version = "4.5.2" + version = "4.11.3" values = [templatefile("${path.module}/helm-values/nginx-values.yaml", {})] } @@ -571,7 +577,7 @@ module "eks_blueprints_addons" { amp_url = "https://aps-workspaces.${local.region}.amazonaws.com/workspaces/${aws_prometheus_workspace.amp[0].id}" }) : templatefile("${path.module}/helm-values/kube-prometheus.yaml", {}) ] - chart_version = "48.1.1" + chart_version = "65.5.1" set_sensitive = [ { name = "grafana.adminPassword" diff --git a/analytics/terraform/spark-k8s-operator/helm-values/kubecost-values.yaml b/analytics/terraform/spark-k8s-operator/helm-values/kubecost-values.yaml index 98c2b8685..97f4a1984 100644 --- a/analytics/terraform/spark-k8s-operator/helm-values/kubecost-values.yaml +++ b/analytics/terraform/spark-k8s-operator/helm-values/kubecost-values.yaml @@ -25,8 +25,13 @@ prometheus: create: false kubeStateMetrics: enabled: false - -#imageVersion: prod-1.96.0 # commented to use the latest + server: + image: + repository: public.ecr.aws/kubecost/prometheus + configmapReload: + prometheus: + image: + repository: public.ecr.aws/kubecost/prometheus-config-reloader kubecostFrontend: image: public.ecr.aws/kubecost/frontend @@ -46,6 +51,18 @@ kubecostModel: cpu: "500m" memory: "512Mi" +forecasting: + fullImageName: public.ecr.aws/kubecost/kubecost-modeling:v0.1.18 + +networkCosts: + image: + repository: public.ecr.aws/kubecost/kubecost-network-costs + tag: v0.17.6 + +clusterController: + image: + repository: public.ecr.aws/kubecost/cluster-controller + # Set this to false if you're bringing your own service account. #serviceAccount: # create: false @@ -60,3 +77,6 @@ persistentVolume: enabled: true # Note that setting this to false means configurations will be wiped out on pod restart. storageClass: gp3 # existingClaim: kubecost-cost-analyzer # a claim in the same namespace as kubecost + +reporting: + productAnalytics: false