infrastructure-as-code/openstack-tf/d4s-preprod/variables/preprod.auto.tfvars

94 lines
2.9 KiB
Plaintext
Raw Normal View History

default_security_group_name = "default_for_all"
# Provided in the output of the project setup
main_private_network_id = "23fd8a99-d551-4ada-8d3a-9859542ebb8c"
main_private_subnet_id = "cd77a2fd-4a36-4254-b1d0-70b3874c6d04"
dns_zone_id = "c1a4b4bc-f167-4387-855d-38f0f99ca05c"
octavia_information = {
main_lb_name = "d4s-pre-cloud-l4-load-balancer"
main_lb_description = "Main L4 load balancer for the D4Science PRE production"
swarm_lb_name = "d4s-pre-cloud-l4-swarm-load-balancer"
octavia_flavor = "octavia_amphora-mvcpu-ha"
octavia_flavor_id = "394988b5-6603-4a1e-a939-8e177c6681c7"
main_lb_hostname = "main-lb"
# The following aren't available when the module runs so we have to get them with the command
# openstack --os-cloud d4s-pre port list -f value | grep octavia-lb-vrrp
# This means that the execution will fail
2023-11-29 18:49:21 +01:00
octavia_vrrp_ip_1 = "10.1.33.159/32"
octavia_vrrp_ip_2 = "10.1.32.199/32"
}
os_project_data = {
id = "6fdc02e2827b405dad99f34698659742"
2023-11-03 17:12:11 +01:00
}
dns_zone = {
2023-11-03 17:12:11 +01:00
zone_name = "cloud-pre.d4science.org."
email = "postmaster@isti.cnr.it"
description = "DNS primary zone for the d4s-pre-cloud project"
ttl = 8600
2023-11-05 19:19:04 +01:00
id = "c1a4b4bc-f167-4387-855d-38f0f99ca05c"
2023-11-03 17:12:11 +01:00
}
main_private_network = {
2023-11-03 17:12:11 +01:00
name = "d4s-pre-cloud-main"
description = "D4Science Preprod private network (use this as the main network)"
}
main_private_subnet = {
2023-11-03 17:12:11 +01:00
name = "d4s-pre-cloud-main-subnet"
description = "D4Science Preprod main private subnet"
cidr = "10.1.32.0/22"
gateway_ip = "10.1.32.1"
allocation_start = "10.1.32.100"
allocation_end = "10.1.35.254"
}
external_router = {
2023-11-03 17:12:11 +01:00
name = "d4s-pre-cloud-external-router"
description = "D4Science Preprod main router"
id = "cc26064a-bb08-4c0b-929f-d0cb39f934a3"
}
basic_services_ip = {
2023-11-05 19:19:04 +01:00
ca = "10.1.32.4"
ca_cidr = "10.1.32.4/32"
2023-11-03 17:12:11 +01:00
ssh_jump = "10.1.32.5"
ssh_jump_cidr = "10.1.32.5/32"
prometheus = "10.1.32.10"
prometheus_cidr = "10.1.32.10/32"
2023-11-05 19:19:04 +01:00
haproxy_l7_1 = "10.1.32.11"
haproxy_l7_1_cidr = "10.1.32.11/32"
haproxy_l7_2 = "10.1.32.12"
haproxy_l7_2_cidr = "10.1.32.12/32"
octavia_main = "10.1.32.20"
octavia_main_cidr = "10.1.32.20/32"
2023-11-03 17:12:11 +01:00
}
main_haproxy_l7_ip = ["10.1.32.11", "10.1.32.12"]
2023-11-05 19:19:04 +01:00
# docker_swarm_data = {
# mgr_name = "swarm-mgr"
# mgr1_ip = "10.1.32.31"
# mgr1_cidr = "10.1.32.31/32"
# mgr2_ip = "10.1.32.32"
# mgr2_cidr = "10.1.32.32/32"
# mgr3_ip = "10.1.32.33"
# mgr3_cidr = "10.1.32.33/32"
# mgr_count = 3
# mgr_flavor = "m1.large"
# mgr_data_disk_size = 100
# worker_name = "swarm-worker"
# worker_count = 3
# worker_flavor = "m1.large"
# worker_data_disk_size = 200
# nfs_server_name = "swarm-nfs-server"
# nfs_server_flavor = "m1.medium"
# nfs_server_data_disk_name = "Swarm NFS server data Disk"
# nfs_server_data_disk_size = 100
# nfs_server_data_disk_device = "/dev/vdb"
# }
# swarm_managers_ip = ["10.1.32.31", "10.1.32.32", "10.1.32.33"]