Compare commits
12 Commits
1c4a27bca3
...
master
| Author | SHA1 | Date | |
|---|---|---|---|
| 5acb8370cc | |||
| f207f774de | |||
| 1a309cbe4f | |||
| 83d277d144 | |||
| 5e1fd2e9f3 | |||
| 3335020db5 | |||
| 9ce06671c9 | |||
| a7f68c0c4b | |||
| d1a7ccc98c | |||
| afe19041d9 | |||
| c9be2a2fc8 | |||
| 5fc58dfc98 |
@@ -5,6 +5,10 @@ on:
|
|||||||
branches:
|
branches:
|
||||||
- master
|
- master
|
||||||
|
|
||||||
|
concurrency:
|
||||||
|
group: terraform-global
|
||||||
|
cancel-in-progress: false
|
||||||
|
|
||||||
jobs:
|
jobs:
|
||||||
terraform:
|
terraform:
|
||||||
name: "Terraform Apply"
|
name: "Terraform Apply"
|
||||||
@@ -47,29 +51,172 @@ jobs:
|
|||||||
|
|
||||||
- name: Terraform Plan
|
- name: Terraform Plan
|
||||||
working-directory: terraform
|
working-directory: terraform
|
||||||
run: terraform plan
|
run: terraform plan -out=tfplan
|
||||||
|
|
||||||
|
- name: Block accidental destroy
|
||||||
|
env:
|
||||||
|
ALLOW_TF_DESTROY: ${{ secrets.ALLOW_TF_DESTROY }}
|
||||||
|
working-directory: terraform
|
||||||
|
run: |
|
||||||
|
terraform show -json -no-color tfplan > tfplan.json
|
||||||
|
DESTROY_COUNT=$(python3 -c 'import json; raw=open("tfplan.json","rb").read().decode("utf-8","ignore"); start=raw.find("{"); data=json.JSONDecoder().raw_decode(raw[start:])[0]; print(sum(1 for rc in data.get("resource_changes", []) if "delete" in rc.get("change", {}).get("actions", [])))')
|
||||||
|
echo "Planned deletes: $DESTROY_COUNT"
|
||||||
|
if [ "$DESTROY_COUNT" -gt 0 ] && [ "${ALLOW_TF_DESTROY}" != "true" ]; then
|
||||||
|
echo "Destroy actions detected. Set ALLOW_TF_DESTROY=true to allow."
|
||||||
|
exit 1
|
||||||
|
fi
|
||||||
|
|
||||||
- name: Terraform Apply
|
- name: Terraform Apply
|
||||||
working-directory: terraform
|
working-directory: terraform
|
||||||
run: terraform apply -auto-approve
|
run: terraform apply -auto-approve tfplan
|
||||||
|
|
||||||
- name: Enroll VMs in Tailscale
|
- name: Enroll VMs in Tailscale
|
||||||
env:
|
env:
|
||||||
TS_AUTHKEY: ${{ secrets.TS_AUTHKEY }}
|
TS_AUTHKEY: ${{ secrets.TS_AUTHKEY }}
|
||||||
TAILSCALE_ENROLL_HOSTS: ${{ secrets.TAILSCALE_ENROLL_HOSTS }}
|
PM_API_TOKEN_SECRET: ${{ secrets.PM_API_TOKEN_SECRET }}
|
||||||
VM_SSH_PRIVATE_KEY: ${{ secrets.VM_SSH_PRIVATE_KEY }}
|
working-directory: terraform
|
||||||
run: |
|
run: |
|
||||||
if [ -z "$TS_AUTHKEY" ] || [ -z "$TAILSCALE_ENROLL_HOSTS" ] || [ -z "$VM_SSH_PRIVATE_KEY" ]; then
|
if [ -z "$TS_AUTHKEY" ] || [ -z "$PM_API_TOKEN_SECRET" ]; then
|
||||||
echo "Skipping Tailscale enrollment (missing TS_AUTHKEY, TAILSCALE_ENROLL_HOSTS, or VM_SSH_PRIVATE_KEY)."
|
echo "Skipping Tailscale enrollment (missing TS_AUTHKEY or PM_API_TOKEN_SECRET)."
|
||||||
exit 0
|
exit 0
|
||||||
fi
|
fi
|
||||||
|
|
||||||
install -m 700 -d ~/.ssh
|
PM_API_URL=$(awk -F'"' '/^pm_api_url/{print $2}' terraform.tfvars)
|
||||||
printf '%s\n' "$VM_SSH_PRIVATE_KEY" > ~/.ssh/id_rsa
|
PM_API_TOKEN_ID=$(awk -F'"' '/^pm_api_token_id/{print $2}' terraform.tfvars)
|
||||||
chmod 600 ~/.ssh/id_rsa
|
TARGET_NODE=$(awk -F'"' '/^target_node/{print $2}' terraform.tfvars)
|
||||||
|
|
||||||
for host in $(printf '%s' "$TAILSCALE_ENROLL_HOSTS" | tr ',' ' '); do
|
export PM_API_URL PM_API_TOKEN_ID TARGET_NODE
|
||||||
echo "Enrolling $host into Tailscale"
|
|
||||||
ssh -o StrictHostKeyChecking=no -o UserKnownHostsFile=/dev/null -i ~/.ssh/id_rsa "micqdf@$host" \
|
terraform output -json > tfoutputs.json
|
||||||
"echo '$TS_AUTHKEY' | sudo tee /etc/tailscale/authkey >/dev/null && sudo chmod 600 /etc/tailscale/authkey && sudo systemctl start tailscale-firstboot.service"
|
cat > enroll_tailscale.py <<'PY'
|
||||||
done
|
import json
|
||||||
|
import os
|
||||||
|
import ssl
|
||||||
|
import sys
|
||||||
|
import time
|
||||||
|
import urllib.parse
|
||||||
|
import urllib.request
|
||||||
|
|
||||||
|
api_url = os.environ["PM_API_URL"].rstrip("/")
|
||||||
|
if api_url.endswith("/api2/json"):
|
||||||
|
api_url = api_url[: -len("/api2/json")]
|
||||||
|
token_id = os.environ["PM_API_TOKEN_ID"].strip()
|
||||||
|
token_secret = os.environ["PM_API_TOKEN_SECRET"].strip()
|
||||||
|
target_node = os.environ["TARGET_NODE"].strip()
|
||||||
|
ts_authkey = os.environ["TS_AUTHKEY"]
|
||||||
|
|
||||||
|
if not token_id or not token_secret:
|
||||||
|
raise SystemExit("Missing Proxmox token id/secret")
|
||||||
|
|
||||||
|
raw_outputs = open("tfoutputs.json", "rb").read().decode("utf-8", "ignore")
|
||||||
|
start = raw_outputs.find("{")
|
||||||
|
if start == -1:
|
||||||
|
raise SystemExit("Could not find JSON payload in terraform output")
|
||||||
|
outputs = json.JSONDecoder().raw_decode(raw_outputs[start:])[0]
|
||||||
|
|
||||||
|
targets = []
|
||||||
|
for output_name in ("alpaca_vm_ids", "llama_vm_ids"):
|
||||||
|
mapping = outputs.get(output_name, {}).get("value", {})
|
||||||
|
if isinstance(mapping, dict):
|
||||||
|
for hostname, vmid in mapping.items():
|
||||||
|
targets.append((str(hostname), int(vmid)))
|
||||||
|
|
||||||
|
if not targets:
|
||||||
|
print("No VMs found in terraform outputs; skipping tailscale enrollment")
|
||||||
|
raise SystemExit(0)
|
||||||
|
|
||||||
|
print("Tailscale enrollment targets:", ", ".join(f"{h}:{v}" for h, v in targets))
|
||||||
|
|
||||||
|
ssl_ctx = ssl._create_unverified_context()
|
||||||
|
auth_header = f"PVEAPIToken={token_id}={token_secret}"
|
||||||
|
|
||||||
|
def api_request(method, path, data=None):
|
||||||
|
url = f"{api_url}{path}"
|
||||||
|
headers = {"Authorization": auth_header}
|
||||||
|
body = None
|
||||||
|
if data is not None:
|
||||||
|
body = urllib.parse.urlencode(data, doseq=True).encode("utf-8")
|
||||||
|
headers["Content-Type"] = "application/x-www-form-urlencoded"
|
||||||
|
req = urllib.request.Request(url, data=body, headers=headers, method=method)
|
||||||
|
with urllib.request.urlopen(req, context=ssl_ctx, timeout=30) as resp:
|
||||||
|
payload = resp.read().decode("utf-8")
|
||||||
|
return json.loads(payload)
|
||||||
|
|
||||||
|
def wait_for_guest_agent(vmid, timeout_seconds=420):
|
||||||
|
deadline = time.time() + timeout_seconds
|
||||||
|
while time.time() < deadline:
|
||||||
|
try:
|
||||||
|
res = api_request("GET", f"/api2/json/nodes/{target_node}/qemu/{vmid}/agent/ping")
|
||||||
|
if res.get("data") == "pong":
|
||||||
|
return True
|
||||||
|
except Exception:
|
||||||
|
pass
|
||||||
|
time.sleep(5)
|
||||||
|
return False
|
||||||
|
|
||||||
|
def exec_guest(vmid, command):
|
||||||
|
res = api_request(
|
||||||
|
"POST",
|
||||||
|
f"/api2/json/nodes/{target_node}/qemu/{vmid}/agent/exec",
|
||||||
|
{
|
||||||
|
"command": "/run/current-system/sw/bin/sh",
|
||||||
|
"extra-args": ["-lc", command],
|
||||||
|
},
|
||||||
|
)
|
||||||
|
pid = res["data"]["pid"]
|
||||||
|
for _ in range(120):
|
||||||
|
status = api_request(
|
||||||
|
"GET",
|
||||||
|
f"/api2/json/nodes/{target_node}/qemu/{vmid}/agent/exec-status?pid={pid}",
|
||||||
|
).get("data", {})
|
||||||
|
if status.get("exited"):
|
||||||
|
return (
|
||||||
|
int(status.get("exitcode", 1)),
|
||||||
|
status.get("out-data", ""),
|
||||||
|
status.get("err-data", ""),
|
||||||
|
)
|
||||||
|
time.sleep(2)
|
||||||
|
return (124, "", "Timed out waiting for guest command")
|
||||||
|
|
||||||
|
failures = []
|
||||||
|
safe_key = ts_authkey.replace("'", "'\"'\"'")
|
||||||
|
|
||||||
|
for hostname, vmid in targets:
|
||||||
|
print(f"\n== Enrolling {hostname} (vmid {vmid}) ==")
|
||||||
|
if not wait_for_guest_agent(vmid):
|
||||||
|
failures.append(f"{hostname}: guest agent not ready")
|
||||||
|
print(f"ERROR: guest agent not ready for vmid {vmid}")
|
||||||
|
continue
|
||||||
|
|
||||||
|
safe_hostname = hostname.replace("'", "'\"'\"'")
|
||||||
|
cmd = (
|
||||||
|
"set -e; "
|
||||||
|
f"printf '%s' '{safe_key}' > /etc/tailscale/authkey; "
|
||||||
|
f"printf '%s' '{safe_hostname}' > /etc/tailscale/hostname; "
|
||||||
|
"chmod 600 /etc/tailscale/authkey; "
|
||||||
|
f"hostnamectl set-hostname '{safe_hostname}' || true; "
|
||||||
|
"systemctl restart tailscaled; "
|
||||||
|
"systemctl start tailscale-firstboot.service; "
|
||||||
|
"tailscale status || true"
|
||||||
|
)
|
||||||
|
|
||||||
|
exitcode, stdout, stderr = exec_guest(vmid, cmd)
|
||||||
|
if stdout:
|
||||||
|
print(stdout)
|
||||||
|
if stderr:
|
||||||
|
print(stderr, file=sys.stderr)
|
||||||
|
|
||||||
|
if exitcode != 0:
|
||||||
|
failures.append(f"{hostname}: command failed exit {exitcode}")
|
||||||
|
print(f"ERROR: tailscale enrollment failed for {hostname} (exit {exitcode})")
|
||||||
|
|
||||||
|
if failures:
|
||||||
|
print("\nEnrollment failures:")
|
||||||
|
for failure in failures:
|
||||||
|
print(f"- {failure}")
|
||||||
|
raise SystemExit(1)
|
||||||
|
|
||||||
|
print("\nTailscale enrollment completed for all managed VMs")
|
||||||
|
PY
|
||||||
|
|
||||||
|
python3 enroll_tailscale.py
|
||||||
|
|||||||
@@ -1,28 +1,65 @@
|
|||||||
name: Gitea Destroy Terraform
|
name: Terraform Destroy
|
||||||
run-name: ${{ gitea.actor }} triggered a Terraform Destroy 🧨
|
run-name: ${{ gitea.actor }} requested Terraform destroy
|
||||||
|
|
||||||
on:
|
on:
|
||||||
workflow_dispatch: # Manual trigger
|
workflow_dispatch:
|
||||||
|
inputs:
|
||||||
|
confirm:
|
||||||
|
description: "Type NUKE to confirm destroy"
|
||||||
|
required: true
|
||||||
|
type: string
|
||||||
|
target:
|
||||||
|
description: "Destroy scope"
|
||||||
|
required: true
|
||||||
|
default: all
|
||||||
|
type: choice
|
||||||
|
options:
|
||||||
|
- all
|
||||||
|
- alpacas
|
||||||
|
- llamas
|
||||||
|
|
||||||
|
concurrency:
|
||||||
|
group: terraform-global
|
||||||
|
cancel-in-progress: false
|
||||||
|
|
||||||
jobs:
|
jobs:
|
||||||
destroy:
|
destroy:
|
||||||
name: "Terraform Destroy"
|
name: "Terraform Destroy"
|
||||||
runs-on: ubuntu-latest
|
runs-on: ubuntu-latest
|
||||||
|
|
||||||
permissions:
|
|
||||||
contents: read
|
|
||||||
pull-requests: write
|
|
||||||
|
|
||||||
env:
|
|
||||||
TF_VAR_SSH_KEY: ${{ secrets.TF_VAR_SSH_KEY_PUBLIC }}
|
|
||||||
TF_VAR_TS_AUTHKEY: ${{ secrets.TF_VAR_TS_AUTHKEY }}
|
|
||||||
TF_VAR_PROXMOX_PASSWORD: ${{ secrets.TF_VAR_PROXMOX_PASSWORD }}
|
|
||||||
|
|
||||||
|
|
||||||
steps:
|
steps:
|
||||||
|
- name: Validate confirmation phrase
|
||||||
|
run: |
|
||||||
|
if [ "${{ inputs.confirm }}" != "NUKE" ]; then
|
||||||
|
echo "Confirmation failed. You must type NUKE."
|
||||||
|
exit 1
|
||||||
|
fi
|
||||||
|
|
||||||
- name: Checkout repository
|
- name: Checkout repository
|
||||||
uses: actions/checkout@v4
|
uses: actions/checkout@v4
|
||||||
|
|
||||||
|
- name: Create Terraform secret files
|
||||||
|
working-directory: terraform
|
||||||
|
run: |
|
||||||
|
cat > secrets.auto.tfvars << EOF
|
||||||
|
pm_api_token_secret = "${{ secrets.PM_API_TOKEN_SECRET }}"
|
||||||
|
EOF
|
||||||
|
cat > backend.hcl << EOF
|
||||||
|
bucket = "${{ secrets.B2_TF_BUCKET }}"
|
||||||
|
key = "terraform.tfstate"
|
||||||
|
region = "us-east-005"
|
||||||
|
endpoints = {
|
||||||
|
s3 = "${{ secrets.B2_TF_ENDPOINT }}"
|
||||||
|
}
|
||||||
|
access_key = "$(printf '%s' "${{ secrets.B2_KEY_ID }}" | tr -d '\r\n')"
|
||||||
|
secret_key = "$(printf '%s' "${{ secrets.B2_APPLICATION_KEY }}" | tr -d '\r\n')"
|
||||||
|
skip_credentials_validation = true
|
||||||
|
skip_metadata_api_check = true
|
||||||
|
skip_region_validation = true
|
||||||
|
skip_requesting_account_id = true
|
||||||
|
use_path_style = true
|
||||||
|
EOF
|
||||||
|
|
||||||
- name: Set up Terraform
|
- name: Set up Terraform
|
||||||
uses: hashicorp/setup-terraform@v2
|
uses: hashicorp/setup-terraform@v2
|
||||||
with:
|
with:
|
||||||
@@ -30,9 +67,27 @@ jobs:
|
|||||||
|
|
||||||
- name: Terraform Init
|
- name: Terraform Init
|
||||||
working-directory: terraform
|
working-directory: terraform
|
||||||
run: terraform init
|
run: terraform init -reconfigure -backend-config=backend.hcl
|
||||||
|
|
||||||
- name: Terraform Destroy
|
- name: Terraform Destroy Plan
|
||||||
working-directory: terraform
|
working-directory: terraform
|
||||||
run: terraform destroy -auto-approve
|
run: |
|
||||||
|
case "${{ inputs.target }}" in
|
||||||
|
all)
|
||||||
|
terraform plan -destroy -out=tfdestroy
|
||||||
|
;;
|
||||||
|
alpacas)
|
||||||
|
terraform plan -destroy -target=proxmox_vm_qemu.alpacas -out=tfdestroy
|
||||||
|
;;
|
||||||
|
llamas)
|
||||||
|
terraform plan -destroy -target=proxmox_vm_qemu.llamas -out=tfdestroy
|
||||||
|
;;
|
||||||
|
*)
|
||||||
|
echo "Invalid destroy target: ${{ inputs.target }}"
|
||||||
|
exit 1
|
||||||
|
;;
|
||||||
|
esac
|
||||||
|
|
||||||
|
- name: Terraform Destroy Apply
|
||||||
|
working-directory: terraform
|
||||||
|
run: terraform apply -auto-approve tfdestroy
|
||||||
|
|||||||
@@ -6,6 +6,10 @@ on:
|
|||||||
- stage
|
- stage
|
||||||
- test
|
- test
|
||||||
|
|
||||||
|
concurrency:
|
||||||
|
group: terraform-global
|
||||||
|
cancel-in-progress: false
|
||||||
|
|
||||||
jobs:
|
jobs:
|
||||||
terraform:
|
terraform:
|
||||||
name: "Terraform Plan"
|
name: "Terraform Plan"
|
||||||
@@ -63,6 +67,19 @@ jobs:
|
|||||||
working-directory: terraform
|
working-directory: terraform
|
||||||
run: terraform plan -out=tfplan
|
run: terraform plan -out=tfplan
|
||||||
|
|
||||||
|
- name: Block accidental destroy
|
||||||
|
env:
|
||||||
|
ALLOW_TF_DESTROY: ${{ secrets.ALLOW_TF_DESTROY }}
|
||||||
|
working-directory: terraform
|
||||||
|
run: |
|
||||||
|
terraform show -json -no-color tfplan > tfplan.json
|
||||||
|
DESTROY_COUNT=$(python3 -c 'import json; raw=open("tfplan.json","rb").read().decode("utf-8","ignore"); start=raw.find("{"); data=json.JSONDecoder().raw_decode(raw[start:])[0]; print(sum(1 for rc in data.get("resource_changes", []) if "delete" in rc.get("change", {}).get("actions", [])))')
|
||||||
|
echo "Planned deletes: $DESTROY_COUNT"
|
||||||
|
if [ "$DESTROY_COUNT" -gt 0 ] && [ "${ALLOW_TF_DESTROY}" != "true" ]; then
|
||||||
|
echo "Destroy actions detected. Set ALLOW_TF_DESTROY=true to allow."
|
||||||
|
exit 1
|
||||||
|
fi
|
||||||
|
|
||||||
- name: Upload Terraform Plan
|
- name: Upload Terraform Plan
|
||||||
uses: actions/upload-artifact@v3
|
uses: actions/upload-artifact@v3
|
||||||
with:
|
with:
|
||||||
|
|||||||
@@ -49,20 +49,21 @@
|
|||||||
RemainAfterExit = true;
|
RemainAfterExit = true;
|
||||||
};
|
};
|
||||||
script = ''
|
script = ''
|
||||||
if [ -f /var/lib/tailscale/.joined ]; then
|
|
||||||
exit 0
|
|
||||||
fi
|
|
||||||
|
|
||||||
if [ ! -s /etc/tailscale/authkey ]; then
|
if [ ! -s /etc/tailscale/authkey ]; then
|
||||||
exit 0
|
exit 0
|
||||||
fi
|
fi
|
||||||
|
|
||||||
key="$(cat /etc/tailscale/authkey)"
|
key="$(cat /etc/tailscale/authkey)"
|
||||||
${pkgs.tailscale}/bin/tailscale up --auth-key="$key" --hostname="$(hostname)"
|
ts_hostname=""
|
||||||
|
if [ -s /etc/tailscale/hostname ]; then
|
||||||
|
ts_hostname="--hostname=$(cat /etc/tailscale/hostname)"
|
||||||
|
fi
|
||||||
|
|
||||||
|
rm -f /var/lib/tailscale/tailscaled.state
|
||||||
|
${pkgs.tailscale}/bin/tailscale up --reset --auth-key="$key" $ts_hostname
|
||||||
|
|
||||||
install -d -m 0700 /var/lib/tailscale
|
|
||||||
touch /var/lib/tailscale/.joined
|
|
||||||
rm -f /etc/tailscale/authkey
|
rm -f /etc/tailscale/authkey
|
||||||
|
rm -f /etc/tailscale/hostname
|
||||||
'';
|
'';
|
||||||
};
|
};
|
||||||
|
|
||||||
|
|||||||
@@ -26,19 +26,21 @@ resource "proxmox_vm_qemu" "alpacas" {
|
|||||||
os_type = "cloud-init"
|
os_type = "cloud-init"
|
||||||
agent = 1
|
agent = 1
|
||||||
|
|
||||||
|
cpu {
|
||||||
sockets = var.sockets
|
sockets = var.sockets
|
||||||
cores = var.cores
|
cores = var.cores
|
||||||
|
}
|
||||||
memory = var.memory
|
memory = var.memory
|
||||||
scsihw = "virtio-scsi-pci"
|
scsihw = "virtio-scsi-pci"
|
||||||
boot = "order=virtio0"
|
boot = "order=scsi0"
|
||||||
bootdisk = "virtio0"
|
bootdisk = "scsi0"
|
||||||
ipconfig0 = "ip=dhcp"
|
ipconfig0 = "ip=dhcp"
|
||||||
cicustom = "user=local:snippets/cloud_init_global.yaml"
|
cicustom = "user=local:snippets/cloud_init_global.yaml"
|
||||||
|
|
||||||
|
|
||||||
disks {
|
disks {
|
||||||
virtio {
|
scsi {
|
||||||
virtio0 {
|
scsi0 {
|
||||||
disk {
|
disk {
|
||||||
size = var.disk_size
|
size = var.disk_size
|
||||||
storage = var.storage
|
storage = var.storage
|
||||||
@@ -73,18 +75,20 @@ resource "proxmox_vm_qemu" "llamas" {
|
|||||||
os_type = "cloud-init"
|
os_type = "cloud-init"
|
||||||
agent = 1
|
agent = 1
|
||||||
|
|
||||||
|
cpu {
|
||||||
sockets = var.sockets
|
sockets = var.sockets
|
||||||
cores = var.cores
|
cores = var.cores
|
||||||
|
}
|
||||||
memory = var.memory
|
memory = var.memory
|
||||||
scsihw = "virtio-scsi-pci"
|
scsihw = "virtio-scsi-pci"
|
||||||
boot = "order=virtio0"
|
boot = "order=scsi0"
|
||||||
bootdisk = "virtio0"
|
bootdisk = "scsi0"
|
||||||
ipconfig0 = "ip=dhcp"
|
ipconfig0 = "ip=dhcp"
|
||||||
cicustom = "user=local:snippets/cloud_init_global.yaml"
|
cicustom = "user=local:snippets/cloud_init_global.yaml"
|
||||||
|
|
||||||
disks {
|
disks {
|
||||||
virtio {
|
scsi {
|
||||||
virtio0 {
|
scsi0 {
|
||||||
disk {
|
disk {
|
||||||
size = var.disk_size
|
size = var.disk_size
|
||||||
storage = var.storage
|
storage = var.storage
|
||||||
|
|||||||
Reference in New Issue
Block a user