Page MenuHomeSoftware Heritage
Paste P809

terraform apply
ActivePublic

Authored by ardumont on Oct 7 2020, 7:13 PM.
terraform apply
module.worker0.proxmox_vm_qemu.node: Refreshing state... [id=beaubourg/qemu/117]
module.worker1.proxmox_vm_qemu.node: Refreshing state... [id=beaubourg/qemu/118]
module.db0.proxmox_vm_qemu.node: Refreshing state... [id=orsay/qemu/115]
module.journal0.proxmox_vm_qemu.node: Refreshing state... [id=beaubourg/qemu/122]
module.webapp.proxmox_vm_qemu.node: Refreshing state... [id=branly/qemu/119]
module.deposit.proxmox_vm_qemu.node: Refreshing state... [id=beaubourg/qemu/120]
module.storage0.proxmox_vm_qemu.node: Refreshing state... [id=orsay/qemu/114]
module.scheduler0.proxmox_vm_qemu.node: Refreshing state... [id=beaubourg/qemu/116]
proxmox_vm_qemu.gateway: Refreshing state... [id=beaubourg/qemu/109]
module.worker2.proxmox_vm_qemu.node: Refreshing state... [id=branly/qemu/112]
module.vault.proxmox_vm_qemu.node: Refreshing state... [id=beaubourg/qemu/121]
An execution plan has been generated and is shown below.
Resource actions are indicated with the following symbols:
~ update in-place
Terraform will perform the following actions:
# proxmox_vm_qemu.gateway will be updated in-place
~ resource "proxmox_vm_qemu" "gateway" {
agent = 0
balloon = 0
bios = "seabios"
boot = "c"
bootdisk = "virtio0"
ciuser = "root"
clone = "template-debian-10"
+ clone_wait = 15
cores = 1
cpu = "host"
desc = "staging gateway node"
disk_gb = 0
force_create = false
full_clone = false
hotplug = "network,disk,usb"
id = "beaubourg/qemu/109"
ipconfig0 = "ip=192.168.100.125/24,gw=192.168.100.1"
ipconfig1 = "ip=192.168.128.1/24"
kvm = true
memory = 1024
name = "gateway"
nameserver = "192.168.100.29"
numa = false
onboot = true
os_type = "cloud-init"
preprovision = true
qemu_os = "other"
scsihw = "virtio-scsi-pci"
searchdomain = "internal.staging.swh.network"
sockets = 1
ssh_user = "root"
sshkeys = <<~EOT
ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABAQDVKCfpeIMg7GS3Pk03ZAcBWAeDZ+AvWk2k/pPY0z8MJ3YAbqZkRtSK7yaDgJV6Gro7nn/TxdJLo2jEzzWvlC8d8AEzhZPy5Z/qfVVjqBTBM4H5+e+TItAHFfaY5+0WvIahxcfsfaq70MWfpJhszAah3ThJ4mqzYaw+dkr42+a7Gx3Ygpb/m2dpnFnxvXdcuAJYStmHKU5AWGWWM+Fm50/fdMqUfNd8MbKhkJt5ihXQmZWMOt7ls4N8i5NZWnS9YSWow8X/ENOEqCRN9TyRkc+pPS0w9DNi0BCsWvSRJOkyvQ6caEnKWlNoywCmM1AlIQD3k4RUgRWe0vqg/UKPpH3Z root@terraform
EOT
target_node = "beaubourg"
vcpus = 0
vlan = -1
+ vmid = 109
disk {
backup = false
cache = "none"
format = "raw"
id = 0
iothread = false
mbps = 0
mbps_rd = 0
mbps_rd_max = 0
mbps_wr = 0
mbps_wr_max = 0
replicate = false
size = "20G"
ssd = false
storage = "proxmox"
storage_type = "cephfs"
type = "virtio"
}
network {
bridge = "vmbr0"
firewall = false
id = 0
link_down = false
macaddr = "6E:ED:EF:EB:3C:AA"
model = "virtio"
queues = -1
rate = -1
tag = -1
}
network {
bridge = "vmbr443"
firewall = false
id = 1
link_down = false
macaddr = "FE:95:CC:A5:EB:43"
model = "virtio"
queues = -1
rate = -1
tag = -1
}
}
# module.db0.proxmox_vm_qemu.node will be updated in-place
~ resource "proxmox_vm_qemu" "node" {
agent = 0
balloon = 1024
bios = "seabios"
boot = "c"
bootdisk = "virtio0"
ciuser = "root"
clone = "template-debian-10"
+ clone_wait = 15
cores = 4
cpu = "host"
desc = "Node to host storage/indexer/scheduler dbs"
disk_gb = 0
force_create = false
full_clone = false
hotplug = "network,disk,usb"
id = "orsay/qemu/115"
ipconfig0 = "ip=192.168.128.3/24,gw=192.168.128.1"
kvm = true
memory = 16384
name = "db0"
nameserver = "192.168.100.29"
numa = false
onboot = true
os_type = "cloud-init"
preprovision = true
qemu_os = "other"
scsihw = "virtio-scsi-pci"
searchdomain = "internal.staging.swh.network"
sockets = 1
ssh_user = "root"
sshkeys = <<~EOT
ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABAQDVKCfpeIMg7GS3Pk03ZAcBWAeDZ+AvWk2k/pPY0z8MJ3YAbqZkRtSK7yaDgJV6Gro7nn/TxdJLo2jEzzWvlC8d8AEzhZPy5Z/qfVVjqBTBM4H5+e+TItAHFfaY5+0WvIahxcfsfaq70MWfpJhszAah3ThJ4mqzYaw+dkr42+a7Gx3Ygpb/m2dpnFnxvXdcuAJYStmHKU5AWGWWM+Fm50/fdMqUfNd8MbKhkJt5ihXQmZWMOt7ls4N8i5NZWnS9YSWow8X/ENOEqCRN9TyRkc+pPS0w9DNi0BCsWvSRJOkyvQ6caEnKWlNoywCmM1AlIQD3k4RUgRWe0vqg/UKPpH3Z root@terraform
EOT
target_node = "orsay"
vcpus = 0
vlan = -1
+ vmid = 115
disk {
backup = false
cache = "none"
format = "raw"
id = 0
iothread = false
mbps = 0
mbps_rd = 0
mbps_rd_max = 0
mbps_wr = 0
mbps_wr_max = 0
replicate = false
size = "400G"
ssd = false
storage = "orsay-ssd-2018"
storage_type = "ssd"
type = "virtio"
}
network {
bridge = "vmbr443"
firewall = false
id = 0
link_down = false
macaddr = "3A:65:31:7C:24:17"
model = "virtio"
queues = -1
rate = -1
tag = -1
}
}
# module.deposit.proxmox_vm_qemu.node will be updated in-place
~ resource "proxmox_vm_qemu" "node" {
agent = 0
balloon = 1024
bios = "seabios"
boot = "c"
bootdisk = "virtio0"
ciuser = "root"
clone = "template-debian-10"
+ clone_wait = 15
cores = 4
cpu = "host"
desc = "Deposit service node"
disk_gb = 0
force_create = false
full_clone = false
hotplug = "network,disk,usb"
id = "beaubourg/qemu/120"
ipconfig0 = "ip=192.168.128.7/24,gw=192.168.128.1"
kvm = true
memory = 8192
name = "deposit"
nameserver = "192.168.100.29"
numa = false
onboot = true
os_type = "cloud-init"
preprovision = true
qemu_os = "other"
scsihw = "virtio-scsi-pci"
searchdomain = "internal.staging.swh.network"
sockets = 1
ssh_user = "root"
sshkeys = <<~EOT
ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABAQDVKCfpeIMg7GS3Pk03ZAcBWAeDZ+AvWk2k/pPY0z8MJ3YAbqZkRtSK7yaDgJV6Gro7nn/TxdJLo2jEzzWvlC8d8AEzhZPy5Z/qfVVjqBTBM4H5+e+TItAHFfaY5+0WvIahxcfsfaq70MWfpJhszAah3ThJ4mqzYaw+dkr42+a7Gx3Ygpb/m2dpnFnxvXdcuAJYStmHKU5AWGWWM+Fm50/fdMqUfNd8MbKhkJt5ihXQmZWMOt7ls4N8i5NZWnS9YSWow8X/ENOEqCRN9TyRkc+pPS0w9DNi0BCsWvSRJOkyvQ6caEnKWlNoywCmM1AlIQD3k4RUgRWe0vqg/UKPpH3Z root@terraform
EOT
target_node = "beaubourg"
vcpus = 0
vlan = -1
+ vmid = 120
disk {
backup = false
cache = "none"
format = "raw"
id = 0
iothread = false
mbps = 0
mbps_rd = 0
mbps_rd_max = 0
mbps_wr = 0
mbps_wr_max = 0
replicate = false
size = "32G"
ssd = false
storage = "proxmox"
storage_type = "cephfs"
type = "virtio"
}
network {
bridge = "vmbr443"
firewall = false
id = 0
link_down = false
macaddr = "9E:81:DD:58:15:3B"
model = "virtio"
queues = -1
rate = -1
tag = -1
}
}
# module.journal0.proxmox_vm_qemu.node will be updated in-place
~ resource "proxmox_vm_qemu" "node" {
agent = 0
balloon = 1024
bios = "seabios"
boot = "c"
bootdisk = "virtio0"
ciuser = "root"
clone = "template-debian-10"
+ clone_wait = 15
cores = 4
cpu = "host"
desc = "Journal services node"
disk_gb = 0
force_create = false
full_clone = false
hotplug = "network,disk,usb"
id = "beaubourg/qemu/122"
ipconfig0 = "ip=192.168.128.10/24,gw=192.168.128.1"
kvm = true
memory = 12288
name = "journal0"
nameserver = "192.168.100.29"
numa = false
onboot = true
os_type = "cloud-init"
preprovision = true
qemu_os = "other"
scsihw = "virtio-scsi-pci"
searchdomain = "internal.staging.swh.network"
sockets = 1
ssh_user = "root"
sshkeys = <<~EOT
ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABAQDVKCfpeIMg7GS3Pk03ZAcBWAeDZ+AvWk2k/pPY0z8MJ3YAbqZkRtSK7yaDgJV6Gro7nn/TxdJLo2jEzzWvlC8d8AEzhZPy5Z/qfVVjqBTBM4H5+e+TItAHFfaY5+0WvIahxcfsfaq70MWfpJhszAah3ThJ4mqzYaw+dkr42+a7Gx3Ygpb/m2dpnFnxvXdcuAJYStmHKU5AWGWWM+Fm50/fdMqUfNd8MbKhkJt5ihXQmZWMOt7ls4N8i5NZWnS9YSWow8X/ENOEqCRN9TyRkc+pPS0w9DNi0BCsWvSRJOkyvQ6caEnKWlNoywCmM1AlIQD3k4RUgRWe0vqg/UKPpH3Z root@terraform
EOT
target_node = "beaubourg"
vcpus = 0
vlan = -1
+ vmid = 122
disk {
backup = false
cache = "none"
format = "raw"
id = 0
iothread = false
mbps = 0
mbps_rd = 0
mbps_rd_max = 0
mbps_wr = 0
mbps_wr_max = 0
replicate = false
size = "32G"
ssd = false
storage = "proxmox"
storage_type = "cephfs"
type = "virtio"
}
network {
bridge = "vmbr443"
firewall = false
id = 0
link_down = false
macaddr = "1E:98:C2:66:BF:33"
model = "virtio"
queues = -1
rate = -1
tag = -1
}
}
# module.scheduler0.proxmox_vm_qemu.node will be updated in-place
~ resource "proxmox_vm_qemu" "node" {
agent = 0
balloon = 1024
bios = "seabios"
boot = "c"
bootdisk = "virtio0"
ciuser = "root"
clone = "template-debian-10"
+ clone_wait = 15
cores = 4
cpu = "host"
desc = "Scheduler api services"
disk_gb = 0
force_create = false
full_clone = false
hotplug = "network,disk,usb"
id = "beaubourg/qemu/116"
ipconfig0 = "ip=192.168.128.4/24,gw=192.168.128.1"
kvm = true
memory = 8192
name = "scheduler0"
nameserver = "192.168.100.29"
numa = false
onboot = true
os_type = "cloud-init"
preprovision = true
qemu_os = "other"
scsihw = "virtio-scsi-pci"
searchdomain = "internal.staging.swh.network"
sockets = 1
ssh_user = "root"
sshkeys = <<~EOT
ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABAQDVKCfpeIMg7GS3Pk03ZAcBWAeDZ+AvWk2k/pPY0z8MJ3YAbqZkRtSK7yaDgJV6Gro7nn/TxdJLo2jEzzWvlC8d8AEzhZPy5Z/qfVVjqBTBM4H5+e+TItAHFfaY5+0WvIahxcfsfaq70MWfpJhszAah3ThJ4mqzYaw+dkr42+a7Gx3Ygpb/m2dpnFnxvXdcuAJYStmHKU5AWGWWM+Fm50/fdMqUfNd8MbKhkJt5ihXQmZWMOt7ls4N8i5NZWnS9YSWow8X/ENOEqCRN9TyRkc+pPS0w9DNi0BCsWvSRJOkyvQ6caEnKWlNoywCmM1AlIQD3k4RUgRWe0vqg/UKPpH3Z root@terraform
EOT
target_node = "beaubourg"
vcpus = 0
vlan = -1
+ vmid = 116
disk {
backup = false
cache = "none"
format = "raw"
id = 0
iothread = false
mbps = 0
mbps_rd = 0
mbps_rd_max = 0
mbps_wr = 0
mbps_wr_max = 0
replicate = false
size = "32G"
ssd = false
storage = "proxmox"
storage_type = "cephfs"
type = "virtio"
}
network {
bridge = "vmbr443"
firewall = false
id = 0
link_down = false
macaddr = "92:02:7E:D0:B9:36"
model = "virtio"
queues = -1
rate = -1
tag = -1
}
}
# module.storage0.proxmox_vm_qemu.node will be updated in-place
~ resource "proxmox_vm_qemu" "node" {
agent = 0
balloon = 1024
bios = "seabios"
boot = "c"
bootdisk = "virtio0"
ciuser = "root"
clone = "template-debian-10"
+ clone_wait = 15
cores = 4
cpu = "host"
desc = "swh storage services"
disk_gb = 0
force_create = false
full_clone = false
hotplug = "network,disk,usb"
id = "orsay/qemu/114"
ipconfig0 = "ip=192.168.128.2/24,gw=192.168.128.1"
kvm = true
memory = 8192
name = "storage0"
nameserver = "192.168.100.29"
numa = false
onboot = true
os_type = "cloud-init"
preprovision = true
qemu_os = "other"
scsihw = "virtio-scsi-pci"
searchdomain = "internal.staging.swh.network"
sockets = 1
ssh_user = "root"
sshkeys = <<~EOT
ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABAQDVKCfpeIMg7GS3Pk03ZAcBWAeDZ+AvWk2k/pPY0z8MJ3YAbqZkRtSK7yaDgJV6Gro7nn/TxdJLo2jEzzWvlC8d8AEzhZPy5Z/qfVVjqBTBM4H5+e+TItAHFfaY5+0WvIahxcfsfaq70MWfpJhszAah3ThJ4mqzYaw+dkr42+a7Gx3Ygpb/m2dpnFnxvXdcuAJYStmHKU5AWGWWM+Fm50/fdMqUfNd8MbKhkJt5ihXQmZWMOt7ls4N8i5NZWnS9YSWow8X/ENOEqCRN9TyRkc+pPS0w9DNi0BCsWvSRJOkyvQ6caEnKWlNoywCmM1AlIQD3k4RUgRWe0vqg/UKPpH3Z root@terraform
EOT
target_node = "orsay"
vcpus = 0
vlan = -1
+ vmid = 114
disk {
backup = false
cache = "none"
format = "raw"
id = 0
iothread = false
mbps = 0
mbps_rd = 0
mbps_rd_max = 0
mbps_wr = 0
mbps_wr_max = 0
replicate = false
size = "32G"
ssd = false
storage = "orsay-ssd-2018"
storage_type = "ssd"
type = "virtio"
}
disk {
backup = false
cache = "none"
format = "raw"
id = 1
iothread = false
mbps = 0
mbps_rd = 0
mbps_rd_max = 0
mbps_wr = 0
mbps_wr_max = 0
replicate = false
size = "512G"
ssd = false
storage = "orsay-ssd-2018"
storage_type = "ssd"
type = "virtio"
}
network {
bridge = "vmbr443"
firewall = false
id = 0
link_down = false
macaddr = "CA:73:7F:ED:F9:01"
model = "virtio"
queues = -1
rate = -1
tag = -1
}
}
# module.vault.proxmox_vm_qemu.node will be updated in-place
~ resource "proxmox_vm_qemu" "node" {
agent = 0
balloon = 1024
bios = "seabios"
boot = "c"
bootdisk = "virtio0"
ciuser = "root"
clone = "template-debian-10"
+ clone_wait = 15
cores = 4
cpu = "host"
desc = "Vault services node"
disk_gb = 0
force_create = false
full_clone = false
hotplug = "network,disk,usb"
id = "beaubourg/qemu/121"
ipconfig0 = "ip=192.168.128.9/24,gw=192.168.128.1"
kvm = true
~ memory = 4096 -> 8192
name = "vault"
nameserver = "192.168.100.29"
numa = false
onboot = true
os_type = "cloud-init"
preprovision = true
qemu_os = "other"
scsihw = "virtio-scsi-pci"
searchdomain = "internal.staging.swh.network"
sockets = 1
ssh_user = "root"
sshkeys = <<~EOT
ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABAQDVKCfpeIMg7GS3Pk03ZAcBWAeDZ+AvWk2k/pPY0z8MJ3YAbqZkRtSK7yaDgJV6Gro7nn/TxdJLo2jEzzWvlC8d8AEzhZPy5Z/qfVVjqBTBM4H5+e+TItAHFfaY5+0WvIahxcfsfaq70MWfpJhszAah3ThJ4mqzYaw+dkr42+a7Gx3Ygpb/m2dpnFnxvXdcuAJYStmHKU5AWGWWM+Fm50/fdMqUfNd8MbKhkJt5ihXQmZWMOt7ls4N8i5NZWnS9YSWow8X/ENOEqCRN9TyRkc+pPS0w9DNi0BCsWvSRJOkyvQ6caEnKWlNoywCmM1AlIQD3k4RUgRWe0vqg/UKPpH3Z root@terraform
EOT
target_node = "beaubourg"
vcpus = 0
vlan = -1
+ vmid = 121
disk {
backup = false
cache = "none"
format = "raw"
id = 0
iothread = false
mbps = 0
mbps_rd = 0
mbps_rd_max = 0
mbps_wr = 0
mbps_wr_max = 0
replicate = false
size = "32G"
ssd = false
storage = "proxmox"
storage_type = "cephfs"
type = "virtio"
}
network {
bridge = "vmbr443"
firewall = false
id = 0
link_down = false
macaddr = "16:15:1C:79:CB:DB"
model = "virtio"
queues = -1
rate = -1
tag = -1
}
}
# module.webapp.proxmox_vm_qemu.node will be updated in-place
~ resource "proxmox_vm_qemu" "node" {
agent = 0
balloon = 1024
bios = "seabios"
boot = "c"
bootdisk = "virtio0"
ciuser = "root"
clone = "template-debian-10"
+ clone_wait = 15
cores = 4
cpu = "host"
desc = "Archive/Webapp service node"
disk_gb = 0
force_create = false
full_clone = false
hotplug = "network,disk,usb"
id = "branly/qemu/119"
ipconfig0 = "ip=192.168.128.8/24,gw=192.168.128.1"
kvm = true
memory = 16384
name = "webapp"
nameserver = "192.168.100.29"
numa = false
onboot = true
os_type = "cloud-init"
preprovision = true
qemu_os = "other"
scsihw = "virtio-scsi-pci"
searchdomain = "internal.staging.swh.network"
sockets = 1
ssh_user = "root"
sshkeys = <<~EOT
ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABAQDVKCfpeIMg7GS3Pk03ZAcBWAeDZ+AvWk2k/pPY0z8MJ3YAbqZkRtSK7yaDgJV6Gro7nn/TxdJLo2jEzzWvlC8d8AEzhZPy5Z/qfVVjqBTBM4H5+e+TItAHFfaY5+0WvIahxcfsfaq70MWfpJhszAah3ThJ4mqzYaw+dkr42+a7Gx3Ygpb/m2dpnFnxvXdcuAJYStmHKU5AWGWWM+Fm50/fdMqUfNd8MbKhkJt5ihXQmZWMOt7ls4N8i5NZWnS9YSWow8X/ENOEqCRN9TyRkc+pPS0w9DNi0BCsWvSRJOkyvQ6caEnKWlNoywCmM1AlIQD3k4RUgRWe0vqg/UKPpH3Z root@terraform
EOT
target_node = "branly"
vcpus = 0
vlan = -1
+ vmid = 119
disk {
backup = false
cache = "none"
format = "raw"
id = 0
iothread = false
mbps = 0
mbps_rd = 0
mbps_rd_max = 0
mbps_wr = 0
mbps_wr_max = 0
replicate = false
size = "32G"
ssd = false
storage = "proxmox"
storage_type = "cephfs"
type = "virtio"
}
network {
bridge = "vmbr443"
firewall = false
id = 0
link_down = false
macaddr = "1A:00:39:95:D4:5F"
model = "virtio"
queues = -1
rate = -1
tag = -1
}
}
# module.worker0.proxmox_vm_qemu.node will be updated in-place
~ resource "proxmox_vm_qemu" "node" {
agent = 0
balloon = 1024
bios = "seabios"
boot = "c"
bootdisk = "virtio0"
ciuser = "root"
clone = "template-debian-10"
+ clone_wait = 15
cores = 4
cpu = "host"
desc = "Loader/lister service node"
disk_gb = 0
force_create = false
full_clone = false
hotplug = "network,disk,usb"
id = "beaubourg/qemu/117"
ipconfig0 = "ip=192.168.128.5/24,gw=192.168.128.1"
kvm = true
memory = 12288
name = "worker0"
nameserver = "192.168.100.29"
numa = false
onboot = true
os_type = "cloud-init"
preprovision = true
qemu_os = "other"
scsihw = "virtio-scsi-pci"
searchdomain = "internal.staging.swh.network"
sockets = 1
ssh_user = "root"
sshkeys = <<~EOT
ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABAQDVKCfpeIMg7GS3Pk03ZAcBWAeDZ+AvWk2k/pPY0z8MJ3YAbqZkRtSK7yaDgJV6Gro7nn/TxdJLo2jEzzWvlC8d8AEzhZPy5Z/qfVVjqBTBM4H5+e+TItAHFfaY5+0WvIahxcfsfaq70MWfpJhszAah3ThJ4mqzYaw+dkr42+a7Gx3Ygpb/m2dpnFnxvXdcuAJYStmHKU5AWGWWM+Fm50/fdMqUfNd8MbKhkJt5ihXQmZWMOt7ls4N8i5NZWnS9YSWow8X/ENOEqCRN9TyRkc+pPS0w9DNi0BCsWvSRJOkyvQ6caEnKWlNoywCmM1AlIQD3k4RUgRWe0vqg/UKPpH3Z root@terraform
EOT
target_node = "beaubourg"
vcpus = 0
vlan = -1
+ vmid = 117
disk {
backup = false
cache = "none"
format = "raw"
id = 0
iothread = false
mbps = 0
mbps_rd = 0
mbps_rd_max = 0
mbps_wr = 0
mbps_wr_max = 0
replicate = false
size = "32G"
ssd = false
storage = "proxmox"
storage_type = "cephfs"
type = "virtio"
}
network {
bridge = "vmbr443"
firewall = false
id = 0
link_down = false
macaddr = "72:D9:03:46:B1:47"
model = "virtio"
queues = -1
rate = -1
tag = -1
}
}
# module.worker1.proxmox_vm_qemu.node will be updated in-place
~ resource "proxmox_vm_qemu" "node" {
agent = 0
balloon = 1024
bios = "seabios"
boot = "c"
bootdisk = "virtio0"
ciuser = "root"
clone = "template-debian-10"
+ clone_wait = 15
cores = 4
cpu = "host"
desc = "Loader/lister service node"
disk_gb = 0
force_create = false
full_clone = false
hotplug = "network,disk,usb"
id = "beaubourg/qemu/118"
ipconfig0 = "ip=192.168.128.6/24,gw=192.168.128.1"
kvm = true
memory = 12288
name = "worker1"
nameserver = "192.168.100.29"
numa = false
onboot = true
os_type = "cloud-init"
preprovision = true
qemu_os = "other"
scsihw = "virtio-scsi-pci"
searchdomain = "internal.staging.swh.network"
sockets = 1
ssh_user = "root"
sshkeys = <<~EOT
ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABAQDVKCfpeIMg7GS3Pk03ZAcBWAeDZ+AvWk2k/pPY0z8MJ3YAbqZkRtSK7yaDgJV6Gro7nn/TxdJLo2jEzzWvlC8d8AEzhZPy5Z/qfVVjqBTBM4H5+e+TItAHFfaY5+0WvIahxcfsfaq70MWfpJhszAah3ThJ4mqzYaw+dkr42+a7Gx3Ygpb/m2dpnFnxvXdcuAJYStmHKU5AWGWWM+Fm50/fdMqUfNd8MbKhkJt5ihXQmZWMOt7ls4N8i5NZWnS9YSWow8X/ENOEqCRN9TyRkc+pPS0w9DNi0BCsWvSRJOkyvQ6caEnKWlNoywCmM1AlIQD3k4RUgRWe0vqg/UKPpH3Z root@terraform
EOT
target_node = "beaubourg"
vcpus = 0
vlan = -1
+ vmid = 118
disk {
backup = false
cache = "none"
format = "raw"
id = 0
iothread = false
mbps = 0
mbps_rd = 0
mbps_rd_max = 0
mbps_wr = 0
mbps_wr_max = 0
replicate = false
size = "32G"
ssd = false
storage = "proxmox"
storage_type = "cephfs"
type = "virtio"
}
network {
bridge = "vmbr443"
firewall = false
id = 0
link_down = false
macaddr = "D6:A9:6F:02:E3:66"
model = "virtio"
queues = -1
rate = -1
tag = -1
}
}
# module.worker2.proxmox_vm_qemu.node will be updated in-place
~ resource "proxmox_vm_qemu" "node" {
agent = 0
balloon = 1024
bios = "seabios"
boot = "c"
bootdisk = "virtio0"
ciuser = "root"
clone = "template-debian-10"
+ clone_wait = 15
cores = 4
cpu = "host"
desc = "Loader/lister service node"
disk_gb = 0
force_create = false
full_clone = false
hotplug = "network,disk,usb"
id = "branly/qemu/112"
ipconfig0 = "ip=192.168.128.11/24,gw=192.168.128.1"
kvm = true
memory = 12288
name = "worker2"
nameserver = "192.168.100.29"
numa = false
onboot = true
os_type = "cloud-init"
preprovision = true
qemu_os = "other"
scsihw = "virtio-scsi-pci"
searchdomain = "internal.staging.swh.network"
sockets = 1
ssh_user = "root"
sshkeys = <<~EOT
ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABAQDVKCfpeIMg7GS3Pk03ZAcBWAeDZ+AvWk2k/pPY0z8MJ3YAbqZkRtSK7yaDgJV6Gro7nn/TxdJLo2jEzzWvlC8d8AEzhZPy5Z/qfVVjqBTBM4H5+e+TItAHFfaY5+0WvIahxcfsfaq70MWfpJhszAah3ThJ4mqzYaw+dkr42+a7Gx3Ygpb/m2dpnFnxvXdcuAJYStmHKU5AWGWWM+Fm50/fdMqUfNd8MbKhkJt5ihXQmZWMOt7ls4N8i5NZWnS9YSWow8X/ENOEqCRN9TyRkc+pPS0w9DNi0BCsWvSRJOkyvQ6caEnKWlNoywCmM1AlIQD3k4RUgRWe0vqg/UKPpH3Z root@terraform
EOT
target_node = "branly"
vcpus = 0
vlan = -1
+ vmid = 112
disk {
backup = false
cache = "none"
format = "raw"
id = 0
iothread = false
mbps = 0
mbps_rd = 0
mbps_rd_max = 0
mbps_wr = 0
mbps_wr_max = 0
replicate = false
size = "32G"
ssd = false
storage = "proxmox"
storage_type = "cephfs"
type = "virtio"
}
network {
bridge = "vmbr443"
firewall = false
id = 0
link_down = false
macaddr = "AA:57:27:51:75:18"
model = "virtio"
queues = -1
rate = -1
tag = -1
}
}
Plan: 0 to add, 11 to change, 0 to destroy.
Changes to Outputs:
- storage0_summary = <<~EOT
hostname: storage0
fqdn: storage0.internal.staging.swh.network
network: ip=192.168.128.2/24,gw=192.168.128.1 macaddrs=CA:73:7F:ED:F9:01
EOT -> null
Do you want to perform these actions?
Terraform will perform the actions described above.
Only 'yes' will be accepted to approve.
Enter a value: yes
module.scheduler0.proxmox_vm_qemu.node: Modifying... [id=beaubourg/qemu/116]
module.webapp.proxmox_vm_qemu.node: Modifying... [id=branly/qemu/119]
module.vault.proxmox_vm_qemu.node: Modifying... [id=beaubourg/qemu/121]
module.worker2.proxmox_vm_qemu.node: Modifying... [id=branly/qemu/112]
module.worker1.proxmox_vm_qemu.node: Modifying... [id=beaubourg/qemu/118]
module.db0.proxmox_vm_qemu.node: Modifying... [id=orsay/qemu/115]
module.journal0.proxmox_vm_qemu.node: Modifying... [id=beaubourg/qemu/122]
proxmox_vm_qemu.gateway: Modifying... [id=beaubourg/qemu/109]
module.deposit.proxmox_vm_qemu.node: Modifying... [id=beaubourg/qemu/120]
module.storage0.proxmox_vm_qemu.node: Modifying... [id=orsay/qemu/114]
module.webapp.proxmox_vm_qemu.node: Still modifying... [id=branly/qemu/119, 10s elapsed]
module.scheduler0.proxmox_vm_qemu.node: Still modifying... [id=beaubourg/qemu/116, 10s elapsed]
module.vault.proxmox_vm_qemu.node: Still modifying... [id=beaubourg/qemu/121, 10s elapsed]
module.worker2.proxmox_vm_qemu.node: Still modifying... [id=branly/qemu/112, 10s elapsed]
module.worker1.proxmox_vm_qemu.node: Still modifying... [id=beaubourg/qemu/118, 10s elapsed]
module.db0.proxmox_vm_qemu.node: Still modifying... [id=orsay/qemu/115, 10s elapsed]
module.journal0.proxmox_vm_qemu.node: Still modifying... [id=beaubourg/qemu/122, 10s elapsed]
proxmox_vm_qemu.gateway: Still modifying... [id=beaubourg/qemu/109, 10s elapsed]
module.deposit.proxmox_vm_qemu.node: Still modifying... [id=beaubourg/qemu/120, 10s elapsed]
module.storage0.proxmox_vm_qemu.node: Still modifying... [id=orsay/qemu/114, 10s elapsed]
module.scheduler0.proxmox_vm_qemu.node: Still modifying... [id=beaubourg/qemu/116, 20s elapsed]
module.webapp.proxmox_vm_qemu.node: Still modifying... [id=branly/qemu/119, 20s elapsed]
module.vault.proxmox_vm_qemu.node: Still modifying... [id=beaubourg/qemu/121, 20s elapsed]
module.worker2.proxmox_vm_qemu.node: Still modifying... [id=branly/qemu/112, 20s elapsed]
module.db0.proxmox_vm_qemu.node: Still modifying... [id=orsay/qemu/115, 20s elapsed]
module.worker1.proxmox_vm_qemu.node: Still modifying... [id=beaubourg/qemu/118, 20s elapsed]
module.journal0.proxmox_vm_qemu.node: Still modifying... [id=beaubourg/qemu/122, 20s elapsed]
proxmox_vm_qemu.gateway: Still modifying... [id=beaubourg/qemu/109, 20s elapsed]
module.deposit.proxmox_vm_qemu.node: Still modifying... [id=beaubourg/qemu/120, 20s elapsed]
module.storage0.proxmox_vm_qemu.node: Still modifying... [id=orsay/qemu/114, 20s elapsed]
module.worker0.proxmox_vm_qemu.node: Modifying... [id=beaubourg/qemu/117]
module.webapp.proxmox_vm_qemu.node: Still modifying... [id=branly/qemu/119, 30s elapsed]
module.scheduler0.proxmox_vm_qemu.node: Still modifying... [id=beaubourg/qemu/116, 30s elapsed]
module.vault.proxmox_vm_qemu.node: Still modifying... [id=beaubourg/qemu/121, 30s elapsed]
module.worker2.proxmox_vm_qemu.node: Still modifying... [id=branly/qemu/112, 30s elapsed]
module.worker1.proxmox_vm_qemu.node: Still modifying... [id=beaubourg/qemu/118, 30s elapsed]
module.journal0.proxmox_vm_qemu.node: Still modifying... [id=beaubourg/qemu/122, 30s elapsed]
proxmox_vm_qemu.gateway: Still modifying... [id=beaubourg/qemu/109, 30s elapsed]
module.deposit.proxmox_vm_qemu.node: Still modifying... [id=beaubourg/qemu/120, 30s elapsed]
module.storage0.proxmox_vm_qemu.node: Still modifying... [id=orsay/qemu/114, 30s elapsed]
module.worker0.proxmox_vm_qemu.node: Still modifying... [id=beaubourg/qemu/117, 10s elapsed]
module.scheduler0.proxmox_vm_qemu.node: Still modifying... [id=beaubourg/qemu/116, 40s elapsed]
module.webapp.proxmox_vm_qemu.node: Still modifying... [id=branly/qemu/119, 40s elapsed]
module.vault.proxmox_vm_qemu.node: Still modifying... [id=beaubourg/qemu/121, 40s elapsed]
module.worker2.proxmox_vm_qemu.node: Still modifying... [id=branly/qemu/112, 40s elapsed]
module.worker1.proxmox_vm_qemu.node: Still modifying... [id=beaubourg/qemu/118, 40s elapsed]
module.journal0.proxmox_vm_qemu.node: Still modifying... [id=beaubourg/qemu/122, 40s elapsed]
proxmox_vm_qemu.gateway: Still modifying... [id=beaubourg/qemu/109, 40s elapsed]
module.deposit.proxmox_vm_qemu.node: Still modifying... [id=beaubourg/qemu/120, 40s elapsed]
module.storage0.proxmox_vm_qemu.node: Still modifying... [id=orsay/qemu/114, 40s elapsed]
module.worker0.proxmox_vm_qemu.node: Still modifying... [id=beaubourg/qemu/117, 20s elapsed]
module.vault.proxmox_vm_qemu.node: Modifications complete after 46s [id=beaubourg/qemu/121]
module.webapp.proxmox_vm_qemu.node: Modifications complete after 47s [id=branly/qemu/119]
module.scheduler0.proxmox_vm_qemu.node: Modifications complete after 47s [id=beaubourg/qemu/116]
module.worker1.proxmox_vm_qemu.node: Modifications complete after 48s [id=beaubourg/qemu/118]
module.worker2.proxmox_vm_qemu.node: Still modifying... [id=branly/qemu/112, 50s elapsed]
proxmox_vm_qemu.gateway: Still modifying... [id=beaubourg/qemu/109, 50s elapsed]
module.journal0.proxmox_vm_qemu.node: Still modifying... [id=beaubourg/qemu/122, 50s elapsed]
module.deposit.proxmox_vm_qemu.node: Still modifying... [id=beaubourg/qemu/120, 50s elapsed]
module.worker0.proxmox_vm_qemu.node: Still modifying... [id=beaubourg/qemu/117, 30s elapsed]
module.worker2.proxmox_vm_qemu.node: Still modifying... [id=branly/qemu/112, 1m0s elapsed]
module.journal0.proxmox_vm_qemu.node: Still modifying... [id=beaubourg/qemu/122, 1m0s elapsed]
proxmox_vm_qemu.gateway: Still modifying... [id=beaubourg/qemu/109, 1m0s elapsed]
module.deposit.proxmox_vm_qemu.node: Still modifying... [id=beaubourg/qemu/120, 1m0s elapsed]
module.worker0.proxmox_vm_qemu.node: Still modifying... [id=beaubourg/qemu/117, 40s elapsed]
module.worker2.proxmox_vm_qemu.node: Modifications complete after 1m8s [id=branly/qemu/112]
module.journal0.proxmox_vm_qemu.node: Modifications complete after 1m9s [id=beaubourg/qemu/122]
proxmox_vm_qemu.gateway: Still modifying... [id=beaubourg/qemu/109, 1m10s elapsed]
module.deposit.proxmox_vm_qemu.node: Still modifying... [id=beaubourg/qemu/120, 1m10s elapsed]
module.deposit.proxmox_vm_qemu.node: Modifications complete after 1m10s [id=beaubourg/qemu/120]
proxmox_vm_qemu.gateway: Modifications complete after 1m11s [id=beaubourg/qemu/109]
module.worker0.proxmox_vm_qemu.node: Still modifying... [id=beaubourg/qemu/117, 50s elapsed]
module.worker0.proxmox_vm_qemu.node: Still modifying... [id=beaubourg/qemu/117, 1m0s elapsed]
module.worker0.proxmox_vm_qemu.node: Modifications complete after 1m9s [id=beaubourg/qemu/117]
Error: 500 lvm name '115/vm-115-disk-0.raw' contains illegal characters
on ../modules/node/main.tf line 1, in resource "proxmox_vm_qemu" "node":
1: resource "proxmox_vm_qemu" "node" {
Error: 500 lvm name '114/vm-114-disk-0.raw' contains illegal characters
on ../modules/node/main.tf line 1, in resource "proxmox_vm_qemu" "node":
1: resource "proxmox_vm_qemu" "node" {