aboutsummaryrefslogtreecommitdiffstats
path: root/terraform-ci-infra/1n_nmd/main.tf
diff options
context:
space:
mode:
authorpmikus <pmikus@cisco.com>2021-01-07 14:27:38 +0000
committerpmikus <pmikus@cisco.com>2021-01-29 08:44:37 +0000
commita44eef233de959c5679602dc6cd1ed866d6abc14 (patch)
tree2c4d0cd5e0d1293ae86a09cf26009b869c72360d /terraform-ci-infra/1n_nmd/main.tf
parent474163bc78aa418595b227c81056987b1114104d (diff)
Infra: Monitoring capability
+ Monitoring SOA + Nomad alertmanager job + Nomad prometheus job + Nomad grafana job Signed-off-by: pmikus <pmikus@cisco.com> Change-Id: I0b32e9c87276ba1a2d4a5322816f3473c737eae2
Diffstat (limited to 'terraform-ci-infra/1n_nmd/main.tf')
-rw-r--r--terraform-ci-infra/1n_nmd/main.tf186
1 files changed, 152 insertions, 34 deletions
diff --git a/terraform-ci-infra/1n_nmd/main.tf b/terraform-ci-infra/1n_nmd/main.tf
index 4da8929e8d..4e0b7537c4 100644
--- a/terraform-ci-infra/1n_nmd/main.tf
+++ b/terraform-ci-infra/1n_nmd/main.tf
@@ -4,64 +4,182 @@
# and downstream modules can simply declare resources for that provider
# and have them automatically associated with the root provider
# configurations.
+module "alertmanager" {
+ source = "./alertmanager"
+ providers = {
+ nomad = nomad.yul1
+ }
+
+ # nomad
+ nomad_datacenters = [ "yul1" ]
+
+ # alertmanager
+ alertmanager_job_name = "prod-alertmanager"
+ alertmanager_use_canary = true
+ alertmanager_group_count = 1
+ alertmanager_vault_secret = {
+ use_vault_provider = false,
+ vault_kv_policy_name = "kv-secret",
+ vault_kv_path = "secret/data/prometheus",
+ vault_kv_field_access_key = "access_key",
+ vault_kv_field_secret_key = "secret_key"
+ }
+ alertmanager_version = "0.21.0"
+ alertmanager_cpu = 1000
+ alertmanager_mem = 1024
+ alertmanager_port = 9093
+ alertmanager_slack_api_url = "https://hooks.slack.com/services/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"
+ alertmanager_slack_channel = "fdio-infra-monitoring"
+}
+
+module "exporter" {
+ source = "./exporter"
+ providers = {
+ nomad = nomad.yul1
+ }
+
+ # nomad
+ nomad_datacenters = [ "yul1" ]
+
+ # exporter
+ exporter_job_name = "prod-exporter"
+ exporter_use_canary = false
+
+ # node
+ node_version = "1.0.1"
+ node_port = 9100
+
+ # blackbox
+ blackbox_version = "0.18.0"
+ blackbox_port = 9115
+
+ # cadvisor
+ cadvisor_image = "gcr.io/cadvisor/cadvisor:latest"
+ cadvisor_port = 8080
+}
+
+module "grafana" {
+ source = "./grafana"
+ providers = {
+ nomad = nomad.yul1
+ }
+
+ # nomad
+ nomad_datacenters = [ "yul1" ]
+
+ # grafana
+ grafana_job_name = "prod-grafana"
+ grafana_use_canary = true
+ grafana_group_count = 1
+ grafana_vault_secret = {
+ use_vault_provider = false,
+ vault_kv_policy_name = "kv-secret",
+ vault_kv_path = "secret/data/prometheus",
+ vault_kv_field_access_key = "access_key",
+ vault_kv_field_secret_key = "secret_key"
+ }
+ grafana_container_image = "grafana/grafana:7.3.7"
+ grafana_cpu = 2000
+ grafana_mem = 2048
+ grafana_port = 3000
+}
+
module "minio" {
- source = "./minio"
- providers = {
+ source = "./minio"
+ providers = {
nomad = nomad.yul1
}
# nomad
- nomad_datacenters = [ "yul1" ]
- nomad_host_volume = "prod-volume-data1-1"
+ nomad_datacenters = [ "yul1" ]
+ nomad_host_volume = "prod-volume-data1-1"
# minio
- minio_job_name = "prod-minio"
- minio_group_count = 4
- minio_service_name = "storage"
- minio_host = "http://10.32.8.1{4...7}"
- minio_port = 9000
- minio_container_image = "minio/minio:RELEASE.2020-12-03T05-49-24Z"
- minio_vault_secret = {
- use_vault_provider = false,
- vault_kv_policy_name = "kv-secret",
- vault_kv_path = "secret/data/minio",
- vault_kv_field_access_key = "access_key",
- vault_kv_field_secret_key = "secret_key"
+ minio_job_name = "prod-minio"
+ minio_group_count = 4
+ minio_service_name = "storage"
+ minio_host = "http://10.32.8.1{4...7}"
+ minio_port = 9000
+ minio_container_image = "minio/minio:RELEASE.2020-12-03T05-49-24Z"
+ minio_vault_secret = {
+ use_vault_provider = false,
+ vault_kv_policy_name = "kv-secret",
+ vault_kv_path = "secret/data/minio",
+ vault_kv_field_access_key = "access_key",
+ vault_kv_field_secret_key = "secret_key"
}
- minio_data_dir = "/data/"
- minio_use_host_volume = true
- minio_use_canary = true
- minio_envs = [ "MINIO_BROWSER=\"off\"" ]
+ minio_data_dir = "/data/"
+ minio_use_host_volume = true
+ minio_use_canary = true
+ minio_envs = [ "MINIO_BROWSER=\"off\"" ]
# minio client
- mc_job_name = "prod-mc"
- mc_container_image = "minio/mc:RELEASE.2020-12-10T01-26-17Z"
- mc_extra_commands = [
+ mc_job_name = "prod-mc"
+ mc_container_image = "minio/mc:RELEASE.2020-12-10T01-26-17Z"
+ mc_extra_commands = [
"mc policy set public LOCALMINIO/logs.fd.io",
"mc policy set public LOCALMINIO/docs.fd.io",
"mc ilm add --expiry-days '180' LOCALMINIO/logs.fd.io",
"mc admin user add LOCALMINIO storage Storage1234",
"mc admin policy set LOCALMINIO writeonly user=storage"
]
- minio_buckets = [ "logs.fd.io", "docs.fd.io" ]
+ minio_buckets = [ "logs.fd.io", "docs.fd.io" ]
}
module "nginx" {
- source = "./nginx"
- providers = {
+ source = "./nginx"
+ providers = {
nomad = nomad.yul1
}
# nomad
- nomad_datacenters = [ "yul1" ]
+ nomad_datacenters = [ "yul1" ]
# nginx
- nginx_job_name = "prod-nginx"
+ nginx_job_name = "prod-nginx"
}
-#module "vpp_device" {
-# source = "./vpp_device"
-# providers = {
-# nomad = nomad.yul1
-# }
-#} \ No newline at end of file
+module "prometheus" {
+ source = "./prometheus"
+ providers = {
+ nomad = nomad.yul1
+ }
+
+ # nomad
+ nomad_datacenters = [ "yul1" ]
+ nomad_host_volume = "prod-volume-data1-1"
+
+ # prometheus
+ prometheus_job_name = "prod-prometheus"
+ prometheus_use_canary = true
+ prometheus_group_count = 4
+ prometheus_vault_secret = {
+ use_vault_provider = false,
+ vault_kv_policy_name = "kv-secret",
+ vault_kv_path = "secret/data/prometheus",
+ vault_kv_field_access_key = "access_key",
+ vault_kv_field_secret_key = "secret_key"
+ }
+ prometheus_data_dir = "/data/"
+ prometheus_use_host_volume = true
+ prometheus_version = "2.24.0"
+ prometheus_cpu = 2000
+ prometheus_mem = 8192
+ prometheus_port = 9090
+}
+
+module "vpp_device" {
+ source = "./vpp_device"
+ providers = {
+ nomad = nomad.yul1
+ }
+
+ # nomad
+ nomad_datacenters = [ "yul1" ]
+
+ # csit_shim
+ csit_shim_job_name = "prod-device-csit-shim"
+ csit_shim_group_count = "1"
+ csit_shim_cpu = "1000"
+ csit_shim_mem = "5000"
+} \ No newline at end of file