Compare commits
15 Commits
2aa9950f59
...
master
| Author | SHA1 | Date | |
|---|---|---|---|
| 5bfc135350 | |||
| 63213a4bc3 | |||
| e4243c7667 | |||
| 33bb0ffb17 | |||
| 7434a65590 | |||
| cd8e538c51 | |||
| 808c290c71 | |||
| 15e6471e7e | |||
| 79a4c941e5 | |||
| e9bac70cae | |||
| 4c167f618a | |||
| 97295a7071 | |||
| 7bc861b3e8 | |||
| 6ca189b32c | |||
| b7b364a112 |
@@ -27,7 +27,7 @@ jobs:
|
|||||||
fi
|
fi
|
||||||
|
|
||||||
- name: Checkout repository
|
- name: Checkout repository
|
||||||
uses: https://gitea.com/actions/checkout@v4
|
uses: actions/checkout@v4
|
||||||
|
|
||||||
- name: Create SSH key
|
- name: Create SSH key
|
||||||
run: |
|
run: |
|
||||||
|
|||||||
@@ -27,7 +27,7 @@ jobs:
|
|||||||
fi
|
fi
|
||||||
|
|
||||||
- name: Checkout repository
|
- name: Checkout repository
|
||||||
uses: https://gitea.com/actions/checkout@v4
|
uses: actions/checkout@v4
|
||||||
|
|
||||||
- name: Create SSH key
|
- name: Create SSH key
|
||||||
run: |
|
run: |
|
||||||
|
|||||||
@@ -16,7 +16,7 @@ jobs:
|
|||||||
|
|
||||||
steps:
|
steps:
|
||||||
- name: Checkout repository
|
- name: Checkout repository
|
||||||
uses: https://gitea.com/actions/checkout@v4
|
uses: actions/checkout@v4
|
||||||
|
|
||||||
- name: Create secrets.tfvars
|
- name: Create secrets.tfvars
|
||||||
working-directory: terraform
|
working-directory: terraform
|
||||||
|
|||||||
@@ -36,7 +36,7 @@ jobs:
|
|||||||
fi
|
fi
|
||||||
|
|
||||||
- name: Checkout repository
|
- name: Checkout repository
|
||||||
uses: https://gitea.com/actions/checkout@v4
|
uses: actions/checkout@v4
|
||||||
|
|
||||||
- name: Create Terraform secret files
|
- name: Create Terraform secret files
|
||||||
working-directory: terraform
|
working-directory: terraform
|
||||||
@@ -77,13 +77,13 @@ jobs:
|
|||||||
set -euo pipefail
|
set -euo pipefail
|
||||||
case "${{ inputs.target }}" in
|
case "${{ inputs.target }}" in
|
||||||
all)
|
all)
|
||||||
TF_PLAN_CMD="terraform plan -parallelism=1 -destroy -out=tfdestroy"
|
TF_PLAN_CMD="terraform plan -refresh=false -parallelism=1 -destroy -out=tfdestroy"
|
||||||
;;
|
;;
|
||||||
control-planes)
|
control-planes)
|
||||||
TF_PLAN_CMD="terraform plan -parallelism=1 -destroy -target=proxmox_vm_qemu.control_planes -out=tfdestroy"
|
TF_PLAN_CMD="terraform plan -refresh=false -parallelism=1 -destroy -target=proxmox_vm_qemu.control_planes -out=tfdestroy"
|
||||||
;;
|
;;
|
||||||
workers)
|
workers)
|
||||||
TF_PLAN_CMD="terraform plan -parallelism=1 -destroy -target=proxmox_vm_qemu.workers -out=tfdestroy"
|
TF_PLAN_CMD="terraform plan -refresh=false -parallelism=1 -destroy -target=proxmox_vm_qemu.workers -out=tfdestroy"
|
||||||
;;
|
;;
|
||||||
*)
|
*)
|
||||||
echo "Invalid destroy target: ${{ inputs.target }}"
|
echo "Invalid destroy target: ${{ inputs.target }}"
|
||||||
|
|||||||
@@ -17,7 +17,7 @@ jobs:
|
|||||||
|
|
||||||
steps:
|
steps:
|
||||||
- name: Checkout repository
|
- name: Checkout repository
|
||||||
uses: https://gitea.com/actions/checkout@v4
|
uses: actions/checkout@v4
|
||||||
|
|
||||||
- name: Create secrets.tfvars
|
- name: Create secrets.tfvars
|
||||||
working-directory: terraform
|
working-directory: terraform
|
||||||
|
|||||||
@@ -110,7 +110,9 @@ class Controller:
|
|||||||
"-o",
|
"-o",
|
||||||
"IdentitiesOnly=yes",
|
"IdentitiesOnly=yes",
|
||||||
"-o",
|
"-o",
|
||||||
"StrictHostKeyChecking=accept-new",
|
"StrictHostKeyChecking=no",
|
||||||
|
"-o",
|
||||||
|
"UserKnownHostsFile=/dev/null",
|
||||||
"-i",
|
"-i",
|
||||||
self.ssh_key,
|
self.ssh_key,
|
||||||
]
|
]
|
||||||
@@ -121,6 +123,8 @@ class Controller:
|
|||||||
self.fast_mode = self.env.get("FAST_MODE", "1")
|
self.fast_mode = self.env.get("FAST_MODE", "1")
|
||||||
self.skip_rebuild = self.env.get("SKIP_REBUILD", "0") == "1"
|
self.skip_rebuild = self.env.get("SKIP_REBUILD", "0") == "1"
|
||||||
self.force_reinit = True
|
self.force_reinit = True
|
||||||
|
self.ssh_ready_retries = int(self.env.get("SSH_READY_RETRIES", "20"))
|
||||||
|
self.ssh_ready_delay = int(self.env.get("SSH_READY_DELAY_SEC", "15"))
|
||||||
|
|
||||||
def log(self, msg):
|
def log(self, msg):
|
||||||
print(f"==> {msg}")
|
print(f"==> {msg}")
|
||||||
@@ -130,13 +134,26 @@ class Controller:
|
|||||||
return run_local(full, check=check, capture=True)
|
return run_local(full, check=check, capture=True)
|
||||||
|
|
||||||
def detect_user(self, ip):
|
def detect_user(self, ip):
|
||||||
for user in self.ssh_candidates:
|
for attempt in range(1, self.ssh_ready_retries + 1):
|
||||||
proc = self._ssh(user, ip, "true", check=False)
|
for user in self.ssh_candidates:
|
||||||
if proc.returncode == 0:
|
proc = self._ssh(user, ip, "true", check=False)
|
||||||
self.active_ssh_user = user
|
if proc.returncode == 0:
|
||||||
self.log(f"Using SSH user '{user}' for {ip}")
|
self.active_ssh_user = user
|
||||||
return
|
self.log(f"Using SSH user '{user}' for {ip}")
|
||||||
raise RuntimeError(f"Unable to authenticate to {ip} with users: {', '.join(self.ssh_candidates)}")
|
return
|
||||||
|
if attempt < self.ssh_ready_retries:
|
||||||
|
self.log(
|
||||||
|
f"SSH not ready on {ip} yet; retrying in {self.ssh_ready_delay}s "
|
||||||
|
f"({attempt}/{self.ssh_ready_retries})"
|
||||||
|
)
|
||||||
|
time.sleep(self.ssh_ready_delay)
|
||||||
|
raise RuntimeError(
|
||||||
|
"Unable to authenticate to "
|
||||||
|
f"{ip} with users: {', '.join(self.ssh_candidates)}. "
|
||||||
|
"If this is a freshly cloned VM, the Proxmox source template likely does not yet include the "
|
||||||
|
"current cloud-init-capable NixOS template configuration from nixos/template-base. "
|
||||||
|
"Terraform can only clone what exists in Proxmox; it cannot retrofit cloud-init support into an old template."
|
||||||
|
)
|
||||||
|
|
||||||
def remote(self, ip, cmd, check=True):
|
def remote(self, ip, cmd, check=True):
|
||||||
ordered = [self.active_ssh_user] + [u for u in self.ssh_candidates if u != self.active_ssh_user]
|
ordered = [self.active_ssh_user] + [u for u in self.ssh_candidates if u != self.active_ssh_user]
|
||||||
@@ -157,14 +174,7 @@ class Controller:
|
|||||||
return last
|
return last
|
||||||
|
|
||||||
def prepare_known_hosts(self):
|
def prepare_known_hosts(self):
|
||||||
ssh_dir = Path.home() / ".ssh"
|
pass
|
||||||
ssh_dir.mkdir(parents=True, exist_ok=True)
|
|
||||||
(ssh_dir / "known_hosts").touch()
|
|
||||||
run_local(["chmod", "700", str(ssh_dir)])
|
|
||||||
run_local(["chmod", "600", str(ssh_dir / "known_hosts")])
|
|
||||||
for ip in self.node_ips.values():
|
|
||||||
run_local(["ssh-keygen", "-R", ip], check=False)
|
|
||||||
run_local(f"ssh-keyscan -H {shlex.quote(ip)} >> {shlex.quote(str(ssh_dir / 'known_hosts'))}", check=False)
|
|
||||||
|
|
||||||
def prepare_remote_nix(self, ip):
|
def prepare_remote_nix(self, ip):
|
||||||
self.remote(ip, "sudo mkdir -p /etc/nix")
|
self.remote(ip, "sudo mkdir -p /etc/nix")
|
||||||
@@ -265,11 +275,42 @@ class Controller:
|
|||||||
|
|
||||||
def stage_install_cni(self):
|
def stage_install_cni(self):
|
||||||
self.log("Installing Flannel")
|
self.log("Installing Flannel")
|
||||||
|
manifest_path = self.script_dir.parent / "manifests" / "kube-flannel.yml"
|
||||||
|
manifest_b64 = base64.b64encode(manifest_path.read_bytes()).decode()
|
||||||
|
|
||||||
self.remote(
|
self.remote(
|
||||||
self.primary_ip,
|
self.primary_ip,
|
||||||
"sudo kubectl --kubeconfig /etc/kubernetes/admin.conf apply -f https://raw.githubusercontent.com/flannel-io/flannel/v0.25.5/Documentation/kube-flannel.yml",
|
(
|
||||||
|
"sudo mkdir -p /var/lib/terrahome && "
|
||||||
|
f"echo {shlex.quote(manifest_b64)} | base64 -d | sudo tee /var/lib/terrahome/kube-flannel.yml >/dev/null"
|
||||||
|
),
|
||||||
)
|
)
|
||||||
|
|
||||||
|
self.log("Waiting for API readiness before applying Flannel")
|
||||||
|
ready = False
|
||||||
|
for _ in range(30):
|
||||||
|
if self.cluster_ready():
|
||||||
|
ready = True
|
||||||
|
break
|
||||||
|
time.sleep(10)
|
||||||
|
if not ready:
|
||||||
|
raise RuntimeError("API server did not become ready before Flannel install")
|
||||||
|
|
||||||
|
last_error = None
|
||||||
|
for attempt in range(1, 6):
|
||||||
|
proc = self.remote(
|
||||||
|
self.primary_ip,
|
||||||
|
"sudo kubectl --kubeconfig /etc/kubernetes/admin.conf apply -f /var/lib/terrahome/kube-flannel.yml",
|
||||||
|
check=False,
|
||||||
|
)
|
||||||
|
if proc.returncode == 0:
|
||||||
|
return
|
||||||
|
last_error = (proc.stdout or "") + ("\n" if proc.stdout and proc.stderr else "") + (proc.stderr or "")
|
||||||
|
self.log(f"Flannel apply attempt {attempt}/5 failed; retrying in 15s")
|
||||||
|
time.sleep(15)
|
||||||
|
|
||||||
|
raise RuntimeError(f"Flannel apply failed after retries\n{last_error or ''}")
|
||||||
|
|
||||||
def cluster_has_node(self, name):
|
def cluster_has_node(self, name):
|
||||||
cmd = f"sudo kubectl --kubeconfig /etc/kubernetes/admin.conf get node {shlex.quote(name)} >/dev/null 2>&1"
|
cmd = f"sudo kubectl --kubeconfig /etc/kubernetes/admin.conf get node {shlex.quote(name)} >/dev/null 2>&1"
|
||||||
return self.remote(self.primary_ip, cmd, check=False).returncode == 0
|
return self.remote(self.primary_ip, cmd, check=False).returncode == 0
|
||||||
|
|||||||
212
nixos/kubeadm/manifests/kube-flannel.yml
Normal file
212
nixos/kubeadm/manifests/kube-flannel.yml
Normal file
@@ -0,0 +1,212 @@
|
|||||||
|
---
|
||||||
|
kind: Namespace
|
||||||
|
apiVersion: v1
|
||||||
|
metadata:
|
||||||
|
name: kube-flannel
|
||||||
|
labels:
|
||||||
|
k8s-app: flannel
|
||||||
|
pod-security.kubernetes.io/enforce: privileged
|
||||||
|
---
|
||||||
|
kind: ClusterRole
|
||||||
|
apiVersion: rbac.authorization.k8s.io/v1
|
||||||
|
metadata:
|
||||||
|
labels:
|
||||||
|
k8s-app: flannel
|
||||||
|
name: flannel
|
||||||
|
rules:
|
||||||
|
- apiGroups:
|
||||||
|
- ""
|
||||||
|
resources:
|
||||||
|
- pods
|
||||||
|
verbs:
|
||||||
|
- get
|
||||||
|
- apiGroups:
|
||||||
|
- ""
|
||||||
|
resources:
|
||||||
|
- nodes
|
||||||
|
verbs:
|
||||||
|
- get
|
||||||
|
- list
|
||||||
|
- watch
|
||||||
|
- apiGroups:
|
||||||
|
- ""
|
||||||
|
resources:
|
||||||
|
- nodes/status
|
||||||
|
verbs:
|
||||||
|
- patch
|
||||||
|
---
|
||||||
|
kind: ClusterRoleBinding
|
||||||
|
apiVersion: rbac.authorization.k8s.io/v1
|
||||||
|
metadata:
|
||||||
|
labels:
|
||||||
|
k8s-app: flannel
|
||||||
|
name: flannel
|
||||||
|
roleRef:
|
||||||
|
apiGroup: rbac.authorization.k8s.io
|
||||||
|
kind: ClusterRole
|
||||||
|
name: flannel
|
||||||
|
subjects:
|
||||||
|
- kind: ServiceAccount
|
||||||
|
name: flannel
|
||||||
|
namespace: kube-flannel
|
||||||
|
---
|
||||||
|
apiVersion: v1
|
||||||
|
kind: ServiceAccount
|
||||||
|
metadata:
|
||||||
|
labels:
|
||||||
|
k8s-app: flannel
|
||||||
|
name: flannel
|
||||||
|
namespace: kube-flannel
|
||||||
|
---
|
||||||
|
kind: ConfigMap
|
||||||
|
apiVersion: v1
|
||||||
|
metadata:
|
||||||
|
name: kube-flannel-cfg
|
||||||
|
namespace: kube-flannel
|
||||||
|
labels:
|
||||||
|
tier: node
|
||||||
|
k8s-app: flannel
|
||||||
|
app: flannel
|
||||||
|
data:
|
||||||
|
cni-conf.json: |
|
||||||
|
{
|
||||||
|
"name": "cbr0",
|
||||||
|
"cniVersion": "0.3.1",
|
||||||
|
"plugins": [
|
||||||
|
{
|
||||||
|
"type": "flannel",
|
||||||
|
"delegate": {
|
||||||
|
"hairpinMode": true,
|
||||||
|
"isDefaultGateway": true
|
||||||
|
}
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"type": "portmap",
|
||||||
|
"capabilities": {
|
||||||
|
"portMappings": true
|
||||||
|
}
|
||||||
|
}
|
||||||
|
]
|
||||||
|
}
|
||||||
|
net-conf.json: |
|
||||||
|
{
|
||||||
|
"Network": "10.244.0.0/16",
|
||||||
|
"EnableNFTables": false,
|
||||||
|
"Backend": {
|
||||||
|
"Type": "vxlan"
|
||||||
|
}
|
||||||
|
}
|
||||||
|
---
|
||||||
|
apiVersion: apps/v1
|
||||||
|
kind: DaemonSet
|
||||||
|
metadata:
|
||||||
|
name: kube-flannel-ds
|
||||||
|
namespace: kube-flannel
|
||||||
|
labels:
|
||||||
|
tier: node
|
||||||
|
app: flannel
|
||||||
|
k8s-app: flannel
|
||||||
|
spec:
|
||||||
|
selector:
|
||||||
|
matchLabels:
|
||||||
|
app: flannel
|
||||||
|
template:
|
||||||
|
metadata:
|
||||||
|
labels:
|
||||||
|
tier: node
|
||||||
|
app: flannel
|
||||||
|
spec:
|
||||||
|
affinity:
|
||||||
|
nodeAffinity:
|
||||||
|
requiredDuringSchedulingIgnoredDuringExecution:
|
||||||
|
nodeSelectorTerms:
|
||||||
|
- matchExpressions:
|
||||||
|
- key: kubernetes.io/os
|
||||||
|
operator: In
|
||||||
|
values:
|
||||||
|
- linux
|
||||||
|
hostNetwork: true
|
||||||
|
priorityClassName: system-node-critical
|
||||||
|
tolerations:
|
||||||
|
- operator: Exists
|
||||||
|
effect: NoSchedule
|
||||||
|
serviceAccountName: flannel
|
||||||
|
initContainers:
|
||||||
|
- name: install-cni-plugin
|
||||||
|
image: docker.io/flannel/flannel-cni-plugin:v1.5.1-flannel1
|
||||||
|
command:
|
||||||
|
- cp
|
||||||
|
args:
|
||||||
|
- -f
|
||||||
|
- /flannel
|
||||||
|
- /opt/cni/bin/flannel
|
||||||
|
volumeMounts:
|
||||||
|
- name: cni-plugin
|
||||||
|
mountPath: /opt/cni/bin
|
||||||
|
- name: install-cni
|
||||||
|
image: docker.io/flannel/flannel:v0.25.5
|
||||||
|
command:
|
||||||
|
- cp
|
||||||
|
args:
|
||||||
|
- -f
|
||||||
|
- /etc/kube-flannel/cni-conf.json
|
||||||
|
- /etc/cni/net.d/10-flannel.conflist
|
||||||
|
volumeMounts:
|
||||||
|
- name: cni
|
||||||
|
mountPath: /etc/cni/net.d
|
||||||
|
- name: flannel-cfg
|
||||||
|
mountPath: /etc/kube-flannel/
|
||||||
|
containers:
|
||||||
|
- name: kube-flannel
|
||||||
|
image: docker.io/flannel/flannel:v0.25.5
|
||||||
|
command:
|
||||||
|
- /opt/bin/flanneld
|
||||||
|
args:
|
||||||
|
- --ip-masq
|
||||||
|
- --kube-subnet-mgr
|
||||||
|
resources:
|
||||||
|
requests:
|
||||||
|
cpu: "100m"
|
||||||
|
memory: "50Mi"
|
||||||
|
securityContext:
|
||||||
|
privileged: false
|
||||||
|
capabilities:
|
||||||
|
add: ["NET_ADMIN", "NET_RAW"]
|
||||||
|
env:
|
||||||
|
- name: POD_NAME
|
||||||
|
valueFrom:
|
||||||
|
fieldRef:
|
||||||
|
fieldPath: metadata.name
|
||||||
|
- name: POD_NAMESPACE
|
||||||
|
valueFrom:
|
||||||
|
fieldRef:
|
||||||
|
fieldPath: metadata.namespace
|
||||||
|
- name: EVENT_QUEUE_DEPTH
|
||||||
|
value: "5000"
|
||||||
|
volumeMounts:
|
||||||
|
- name: run
|
||||||
|
mountPath: /run/flannel
|
||||||
|
- name: flannel-cfg
|
||||||
|
mountPath: /etc/kube-flannel/
|
||||||
|
- name: xtables-lock
|
||||||
|
mountPath: /run/xtables.lock
|
||||||
|
volumes:
|
||||||
|
- name: run
|
||||||
|
hostPath:
|
||||||
|
path: /run/flannel
|
||||||
|
type: DirectoryOrCreate
|
||||||
|
- name: cni-plugin
|
||||||
|
hostPath:
|
||||||
|
path: /opt/cni/bin
|
||||||
|
type: DirectoryOrCreate
|
||||||
|
- name: cni
|
||||||
|
hostPath:
|
||||||
|
path: /etc/cni/net.d
|
||||||
|
type: DirectoryOrCreate
|
||||||
|
- name: flannel-cfg
|
||||||
|
configMap:
|
||||||
|
name: kube-flannel-cfg
|
||||||
|
- name: xtables-lock
|
||||||
|
hostPath:
|
||||||
|
path: /run/xtables.lock
|
||||||
|
type: FileOrCreate
|
||||||
@@ -410,6 +410,9 @@ in
|
|||||||
systemd.tmpfiles.rules = [
|
systemd.tmpfiles.rules = [
|
||||||
"d /etc/kubernetes 0755 root root -"
|
"d /etc/kubernetes 0755 root root -"
|
||||||
"d /etc/kubernetes/manifests 0755 root root -"
|
"d /etc/kubernetes/manifests 0755 root root -"
|
||||||
|
"d /etc/cni/net.d 0755 root root -"
|
||||||
|
"d /opt/cni/bin 0755 root root -"
|
||||||
|
"d /run/flannel 0755 root root -"
|
||||||
"d /var/lib/kubelet 0755 root root -"
|
"d /var/lib/kubelet 0755 root root -"
|
||||||
"d /var/lib/kubelet/pki 0755 root root -"
|
"d /var/lib/kubelet/pki 0755 root root -"
|
||||||
];
|
];
|
||||||
|
|||||||
@@ -11,6 +11,7 @@ in
|
|||||||
|
|
||||||
networking.hostName = "k8s-base-template";
|
networking.hostName = "k8s-base-template";
|
||||||
networking.useDHCP = lib.mkDefault true;
|
networking.useDHCP = lib.mkDefault true;
|
||||||
|
networking.useNetworkd = true;
|
||||||
networking.nameservers = [ "1.1.1.1" "8.8.8.8" ];
|
networking.nameservers = [ "1.1.1.1" "8.8.8.8" ];
|
||||||
|
|
||||||
boot.loader.systemd-boot.enable = lib.mkForce false;
|
boot.loader.systemd-boot.enable = lib.mkForce false;
|
||||||
@@ -20,6 +21,8 @@ in
|
|||||||
};
|
};
|
||||||
|
|
||||||
services.qemuGuest.enable = true;
|
services.qemuGuest.enable = true;
|
||||||
|
services.cloud-init.enable = true;
|
||||||
|
services.cloud-init.network.enable = true;
|
||||||
services.openssh.enable = true;
|
services.openssh.enable = true;
|
||||||
services.openssh.settings = {
|
services.openssh.settings = {
|
||||||
PasswordAuthentication = false;
|
PasswordAuthentication = false;
|
||||||
|
|||||||
Reference in New Issue
Block a user