mirror of
https://github.com/ryan4yin/nix-config.git
synced 2026-04-25 10:18:37 +02:00
feat: kubevirt on k3s
This commit is contained in:
121
Justfile
121
Justfile
@@ -85,12 +85,28 @@ yabai-reload:
|
|||||||
|
|
||||||
############################################################################
|
############################################################################
|
||||||
#
|
#
|
||||||
# Homelab - Virtual Machines running on Proxmox
|
# Homelab - NixOS servers running on bare metal
|
||||||
#
|
#
|
||||||
############################################################################
|
############################################################################
|
||||||
|
|
||||||
colmena-ssh-key:
|
virt:
|
||||||
ssh-add /etc/agenix/ssh-key-romantic
|
colmena apply --on '@virt-*' --verbose --show-trace
|
||||||
|
|
||||||
|
shoryu:
|
||||||
|
colmena apply --on '@shoryu' --verbose --show-trace
|
||||||
|
|
||||||
|
shushou:
|
||||||
|
colmena apply --on '@shushou' --verbose --show-trace
|
||||||
|
|
||||||
|
youko:
|
||||||
|
colmena apply --on '@youko' --verbose --show-trace
|
||||||
|
|
||||||
|
|
||||||
|
############################################################################
|
||||||
|
#
|
||||||
|
# Homelab - Virtual Machines running on Kubevirt
|
||||||
|
#
|
||||||
|
############################################################################
|
||||||
|
|
||||||
lab:
|
lab:
|
||||||
colmena apply --on '@homelab-*' --verbose --show-trace
|
colmena apply --on '@homelab-*' --verbose --show-trace
|
||||||
@@ -109,22 +125,22 @@ kana:
|
|||||||
tsgw:
|
tsgw:
|
||||||
colmena apply --on '@tailscale-gw' --verbose --show-trace
|
colmena apply --on '@tailscale-gw' --verbose --show-trace
|
||||||
|
|
||||||
pve-aqua:
|
# pve-aqua:
|
||||||
nom build .#aquamarine
|
# nom build .#aquamarine
|
||||||
rsync -avz --progress --copy-links result root@um560:/var/lib/vz/dump/vzdump-qemu-aquamarine.vma.zst
|
# rsync -avz --progress --copy-links result root@um560:/var/lib/vz/dump/vzdump-qemu-aquamarine.vma.zst
|
||||||
|
#
|
||||||
pve-ruby:
|
# pve-ruby:
|
||||||
nom build .#ruby
|
# nom build .#ruby
|
||||||
rsync -avz --progress --copy-links result root@um560:/var/lib/vz/dump/vzdump-qemu-ruby.vma.zst
|
# rsync -avz --progress --copy-links result root@um560:/var/lib/vz/dump/vzdump-qemu-ruby.vma.zst
|
||||||
|
#
|
||||||
pve-kana:
|
# pve-kana:
|
||||||
nom build .#kana
|
# nom build .#kana
|
||||||
rsync -avz --progress --copy-links result root@gtr5:/var/lib/vz/dump/vzdump-qemu-kana.vma.zst
|
# rsync -avz --progress --copy-links result root@gtr5:/var/lib/vz/dump/vzdump-qemu-kana.vma.zst
|
||||||
|
#
|
||||||
pve-tsgw:
|
# pve-tsgw:
|
||||||
nom build .#tailscale_gw
|
# nom build .#tailscale_gw
|
||||||
rsync -avz --progress --copy-links result root@um560:/var/lib/vz/dump/vzdump-qemu-tailscale_gw.vma.zst
|
# rsync -avz --progress --copy-links result root@um560:/var/lib/vz/dump/vzdump-qemu-tailscale_gw.vma.zst
|
||||||
|
#
|
||||||
|
|
||||||
############################################################################
|
############################################################################
|
||||||
#
|
#
|
||||||
@@ -133,33 +149,33 @@ pve-tsgw:
|
|||||||
############################################################################
|
############################################################################
|
||||||
|
|
||||||
k8s:
|
k8s:
|
||||||
colmena apply --on '@k8s-*'
|
colmena apply --on '@k8s-*' --verbose --show-trace
|
||||||
|
|
||||||
master:
|
master:
|
||||||
colmena apply --on '@k8s-prod-master'
|
colmena apply --on '@k8s-prod-master-*' --verbose --show-trace
|
||||||
|
|
||||||
worker:
|
worker:
|
||||||
colmena apply --on '@k8s-prod-worker'
|
colmena apply --on '@k8s-prod-worker-*' --verbose --show-trace
|
||||||
|
|
||||||
pve-k8s:
|
|
||||||
nom build .#k3s_prod_1_master_1
|
|
||||||
rsync -avz --progress --copy-links result root@um560:/var/lib/vz/dump/vzdump-qemu-k3s_prod_1_master_1.vma.zst
|
|
||||||
|
|
||||||
nom build .#k3s_prod_1_master_2
|
|
||||||
rsync -avz --progress --copy-links result root@gtr5:/var/lib/vz/dump/vzdump-qemu-k3s_prod_1_master_2.vma.zst
|
|
||||||
|
|
||||||
nom build .#k3s_prod_1_master_3
|
|
||||||
rsync -avz --progress --copy-links result root@s500plus:/var/lib/vz/dump/vzdump-qemu-k3s_prod_1_master_3.vma.zst
|
|
||||||
|
|
||||||
nom build .#k3s_prod_1_worker_1
|
|
||||||
rsync -avz --progress --copy-links result root@gtr5:/var/lib/vz/dump/vzdump-qemu-k3s_prod_1_worker_1.vma.zst
|
|
||||||
|
|
||||||
nom build .#k3s_prod_1_worker_2
|
|
||||||
rsync -avz --progress --copy-links result root@s500plus:/var/lib/vz/dump/vzdump-qemu-k3s_prod_1_worker_2.vma.zst
|
|
||||||
|
|
||||||
nom build .#k3s_prod_1_worker_3
|
|
||||||
rsync -avz --progress --copy-links result root@s500plus:/var/lib/vz/dump/vzdump-qemu-k3s_prod_1_worker_3.vma.zst
|
|
||||||
|
|
||||||
|
# pve-k8s:
|
||||||
|
# nom build .#k3s_prod_1_master_1
|
||||||
|
# rsync -avz --progress --copy-links result root@um560:/var/lib/vz/dump/vzdump-qemu-k3s_prod_1_master_1.vma.zst
|
||||||
|
#
|
||||||
|
# nom build .#k3s_prod_1_master_2
|
||||||
|
# rsync -avz --progress --copy-links result root@gtr5:/var/lib/vz/dump/vzdump-qemu-k3s_prod_1_master_2.vma.zst
|
||||||
|
#
|
||||||
|
# nom build .#k3s_prod_1_master_3
|
||||||
|
# rsync -avz --progress --copy-links result root@s500plus:/var/lib/vz/dump/vzdump-qemu-k3s_prod_1_master_3.vma.zst
|
||||||
|
#
|
||||||
|
# nom build .#k3s_prod_1_worker_1
|
||||||
|
# rsync -avz --progress --copy-links result root@gtr5:/var/lib/vz/dump/vzdump-qemu-k3s_prod_1_worker_1.vma.zst
|
||||||
|
#
|
||||||
|
# nom build .#k3s_prod_1_worker_2
|
||||||
|
# rsync -avz --progress --copy-links result root@s500plus:/var/lib/vz/dump/vzdump-qemu-k3s_prod_1_worker_2.vma.zst
|
||||||
|
#
|
||||||
|
# nom build .#k3s_prod_1_worker_3
|
||||||
|
# rsync -avz --progress --copy-links result root@s500plus:/var/lib/vz/dump/vzdump-qemu-k3s_prod_1_worker_3.vma.zst
|
||||||
|
#
|
||||||
|
|
||||||
############################################################################
|
############################################################################
|
||||||
#
|
#
|
||||||
@@ -168,16 +184,13 @@ pve-k8s:
|
|||||||
############################################################################
|
############################################################################
|
||||||
|
|
||||||
riscv:
|
riscv:
|
||||||
colmena apply --on '@riscv'
|
|
||||||
|
|
||||||
riscv-debug:
|
|
||||||
colmena apply --on '@riscv' --verbose --show-trace
|
colmena apply --on '@riscv' --verbose --show-trace
|
||||||
|
|
||||||
nozomi:
|
nozomi:
|
||||||
colmena apply --on '@nozomi'
|
colmena apply --on '@nozomi' --verbose --show-trace
|
||||||
|
|
||||||
yukina:
|
yukina:
|
||||||
colmena apply --on '@yukina'
|
colmena apply --on '@yukina' --verbose --show-trace
|
||||||
|
|
||||||
############################################################################
|
############################################################################
|
||||||
#
|
#
|
||||||
@@ -186,13 +199,10 @@ yukina:
|
|||||||
############################################################################
|
############################################################################
|
||||||
|
|
||||||
aarch:
|
aarch:
|
||||||
colmena apply --on '@aarch'
|
|
||||||
|
|
||||||
aarch-debug:
|
|
||||||
colmena apply --on '@aarch' --verbose --show-trace
|
colmena apply --on '@aarch' --verbose --show-trace
|
||||||
|
|
||||||
suzu:
|
suzu:
|
||||||
colmena apply --on '@suzu'
|
colmena apply --on '@suzu' --verbose --show-trace
|
||||||
|
|
||||||
suzu-debug:
|
suzu-debug:
|
||||||
colmena apply --on '@suzu' --verbose --show-trace
|
colmena apply --on '@suzu' --verbose --show-trace
|
||||||
@@ -252,3 +262,14 @@ emacs-purge:
|
|||||||
emacs-reload:
|
emacs-reload:
|
||||||
doom sync
|
doom sync
|
||||||
{{reload-emacs-cmd}}
|
{{reload-emacs-cmd}}
|
||||||
|
|
||||||
|
|
||||||
|
# =================================================
|
||||||
|
#
|
||||||
|
# Kubernetes related commands
|
||||||
|
#
|
||||||
|
# =================================================
|
||||||
|
|
||||||
|
|
||||||
|
del-failed:
|
||||||
|
kubectl delete pod --all-namespaces --field-selector="status.phase==Failed"
|
||||||
|
|||||||
21
flake.lock
generated
21
flake.lock
generated
@@ -138,6 +138,26 @@
|
|||||||
"type": "github"
|
"type": "github"
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
|
"disko": {
|
||||||
|
"inputs": {
|
||||||
|
"nixpkgs": [
|
||||||
|
"nixpkgs"
|
||||||
|
]
|
||||||
|
},
|
||||||
|
"locked": {
|
||||||
|
"lastModified": 1709286488,
|
||||||
|
"narHash": "sha256-RDpTZ72zLu05djvXRzK76Ysqp9zSdh84ax/edEaJucs=",
|
||||||
|
"owner": "nix-community",
|
||||||
|
"repo": "disko",
|
||||||
|
"rev": "bde7dd352c07d43bd5b8245e6c39074a391fdd46",
|
||||||
|
"type": "github"
|
||||||
|
},
|
||||||
|
"original": {
|
||||||
|
"owner": "nix-community",
|
||||||
|
"repo": "disko",
|
||||||
|
"type": "github"
|
||||||
|
}
|
||||||
|
},
|
||||||
"doomemacs": {
|
"doomemacs": {
|
||||||
"flake": false,
|
"flake": false,
|
||||||
"locked": {
|
"locked": {
|
||||||
@@ -1209,6 +1229,7 @@
|
|||||||
"anyrun": "anyrun",
|
"anyrun": "anyrun",
|
||||||
"astronvim": "astronvim",
|
"astronvim": "astronvim",
|
||||||
"daeuniverse": "daeuniverse",
|
"daeuniverse": "daeuniverse",
|
||||||
|
"disko": "disko",
|
||||||
"doomemacs": "doomemacs",
|
"doomemacs": "doomemacs",
|
||||||
"home-manager": "home-manager_2",
|
"home-manager": "home-manager_2",
|
||||||
"hyprland": "hyprland",
|
"hyprland": "hyprland",
|
||||||
|
|||||||
@@ -156,6 +156,11 @@
|
|||||||
|
|
||||||
nix-gaming.url = "github:fufexan/nix-gaming";
|
nix-gaming.url = "github:fufexan/nix-gaming";
|
||||||
|
|
||||||
|
disko = {
|
||||||
|
url = "github:nix-community/disko";
|
||||||
|
inputs.nixpkgs.follows = "nixpkgs";
|
||||||
|
};
|
||||||
|
|
||||||
# add git hooks to format nix code before commit
|
# add git hooks to format nix code before commit
|
||||||
pre-commit-hooks = {
|
pre-commit-hooks = {
|
||||||
url = "github:cachix/pre-commit-hooks.nix";
|
url = "github:cachix/pre-commit-hooks.nix";
|
||||||
|
|||||||
@@ -56,3 +56,5 @@ When building some packages for riscv64 or aarch64, I often have no cache availa
|
|||||||
|
|
||||||

|

|
||||||

|

|
||||||
|
|
||||||
|
[List of Frieren characters](https://en.wikipedia.org/wiki/List_of_Frieren_characters)
|
||||||
|
|||||||
33
hosts/k8s/disko_config/README.md
Normal file
33
hosts/k8s/disko_config/README.md
Normal file
@@ -0,0 +1,33 @@
|
|||||||
|
# Disko Config
|
||||||
|
|
||||||
|
Generate LUKS keyfile to encrypt the root partition, it's used by disko.
|
||||||
|
|
||||||
|
```bash
|
||||||
|
# partition the usb stick
|
||||||
|
parted /dev/sdb -- mklabel gpt
|
||||||
|
parted /dev/sdb -- mkpart primary 2M 512MB
|
||||||
|
parted /dev/sdb -- mkpart primary 512MB 1024MB
|
||||||
|
mkfs.fat -F 32 -n NIXOS_DSC /dev/sdb1
|
||||||
|
mkfs.fat -F 32 -n NIXOS_K3S /dev/sdb2
|
||||||
|
|
||||||
|
# Generate a keyfile from the true random number generator
|
||||||
|
KEYFILE=./kubevirt-luks-keyfile
|
||||||
|
dd bs=8192 count=4 iflag=fullblock if=/dev/random of=$KEYFILE
|
||||||
|
|
||||||
|
# generate token for k3s
|
||||||
|
K3S_TOKEN_FILE=./kubevirt-k3s-token
|
||||||
|
K3S_TOKEN=$(grep -ao '[A-Za-z0-9]' < /dev/random | head -64 | tr -d '\n' ; echo "")
|
||||||
|
echo $K3S_TOKEN > $K3S_TOKEN_FILE
|
||||||
|
|
||||||
|
# copy the keyfile and token to the usb stick
|
||||||
|
|
||||||
|
KEYFILE=./kubevirt-luks-keyfile
|
||||||
|
DEVICE=/dev/disk/by-label/NIXOS_DSC
|
||||||
|
dd bs=8192 count=4 iflag=fullblock if=$KEYFILE of=$DEVICE
|
||||||
|
|
||||||
|
K3S_TOKEN_FILE=./kubevirt-k3s-token
|
||||||
|
USB_PATH=/run/media/ryan/NIXOS_K3S
|
||||||
|
cp $K3S_TOKEN_FILE $USB_PATH
|
||||||
|
```
|
||||||
|
|
||||||
|
|
||||||
105
hosts/k8s/disko_config/kubevirt-disko-fs.nix
Normal file
105
hosts/k8s/disko_config/kubevirt-disko-fs.nix
Normal file
@@ -0,0 +1,105 @@
|
|||||||
|
{
|
||||||
|
# contains the k3s's token
|
||||||
|
fileSystems."/run/media/nixos_k3s" = {
|
||||||
|
device = "/dev/disk/by-label/NIXOS_K3S";
|
||||||
|
fsType = "vfat";
|
||||||
|
mountOptions = [
|
||||||
|
"ro"
|
||||||
|
];
|
||||||
|
};
|
||||||
|
|
||||||
|
disko.devices = {
|
||||||
|
disk = {
|
||||||
|
sda = {
|
||||||
|
type = "disk";
|
||||||
|
device = "/dev/nvme0n1";
|
||||||
|
content = {
|
||||||
|
type = "gpt";
|
||||||
|
partitions = {
|
||||||
|
# The EFI & Boot partition
|
||||||
|
ESP = {
|
||||||
|
size = "630M";
|
||||||
|
type = "EF00";
|
||||||
|
content = {
|
||||||
|
type = "filesystem";
|
||||||
|
format = "vfat";
|
||||||
|
mountpoint = "/boot/efi";
|
||||||
|
mountOptions = [
|
||||||
|
"defaults"
|
||||||
|
];
|
||||||
|
};
|
||||||
|
};
|
||||||
|
# The root partition
|
||||||
|
luks = {
|
||||||
|
size = "100%";
|
||||||
|
content = {
|
||||||
|
type = "luks";
|
||||||
|
name = "crypted";
|
||||||
|
settings = {
|
||||||
|
keyFile = "/dev/disk/by-label/NIXOS_DSC"; # The keyfile is stored on a USB stick
|
||||||
|
keyFileSize = 8192 * 4; # The maxium size of the keyfile is 8192 bytes
|
||||||
|
keyFileOffset = 0;
|
||||||
|
fallbackToPassword = true;
|
||||||
|
allowDiscards = true;
|
||||||
|
};
|
||||||
|
# Whether to add a boot.initrd.luks.devices entry for the specified disk.
|
||||||
|
initrdUnlock = true;
|
||||||
|
|
||||||
|
# encrypt the root partition with luks2 and argon2id, will prompt for a passphrase, which will be used to unlock the partition.
|
||||||
|
# cryptsetup luksFormat
|
||||||
|
extraFormatArgs = [
|
||||||
|
"--type luks2"
|
||||||
|
"--cipher aes-xts-plain64"
|
||||||
|
"--hash sha512"
|
||||||
|
"--iter-time 5000"
|
||||||
|
"--key-size 256"
|
||||||
|
"--pbkdf argon2id"
|
||||||
|
# use true random data from /dev/random, will block until enough entropy is available
|
||||||
|
"--use-random"
|
||||||
|
];
|
||||||
|
extraOpenArgs = [
|
||||||
|
"--timeout 10"
|
||||||
|
];
|
||||||
|
content = {
|
||||||
|
type = "btrfs";
|
||||||
|
extraArgs = ["-f"];
|
||||||
|
subvolumes = {
|
||||||
|
"@root" = {
|
||||||
|
mountpoint = "/";
|
||||||
|
mountOptions = ["compress-force=zstd:1" "noatime"];
|
||||||
|
};
|
||||||
|
"@home" = {
|
||||||
|
mountpoint = "/home";
|
||||||
|
mountOptions = ["compress-force=zstd:1"];
|
||||||
|
};
|
||||||
|
"@lib" = {
|
||||||
|
mountpoint = "/var/lib";
|
||||||
|
mountOptions = ["compress-force=zstd:1"];
|
||||||
|
};
|
||||||
|
|
||||||
|
"@nix" = {
|
||||||
|
mountpoint = "/nix";
|
||||||
|
mountOptions = ["compress-force=zstd:1" "noatime"];
|
||||||
|
};
|
||||||
|
"@tmp" = {
|
||||||
|
mountpoint = "/tmp";
|
||||||
|
mountOptions = ["compress-force=zstd:1" "noatime"];
|
||||||
|
};
|
||||||
|
"@snapshots" = {
|
||||||
|
mountpoint = "/snapshots";
|
||||||
|
mountOptions = ["compress-force=zstd:1" "noatime"];
|
||||||
|
};
|
||||||
|
"@swap" = {
|
||||||
|
mountpoint = "/swap";
|
||||||
|
swap.swapfile.size = "8192M";
|
||||||
|
};
|
||||||
|
};
|
||||||
|
};
|
||||||
|
};
|
||||||
|
};
|
||||||
|
};
|
||||||
|
};
|
||||||
|
};
|
||||||
|
};
|
||||||
|
};
|
||||||
|
}
|
||||||
@@ -1,12 +1,13 @@
|
|||||||
{
|
{
|
||||||
|
pkgs,
|
||||||
vars_networking,
|
vars_networking,
|
||||||
mylib,
|
mylib,
|
||||||
...
|
...
|
||||||
}: let
|
}: let
|
||||||
hostName = "k3s-prod-1-master-1"; # Define your hostname.
|
hostName = "k3s-prod-1-master-1"; # Define your hostname.
|
||||||
k8sLib = import ../lib.nix;
|
k8sLib = import ../lib.nix;
|
||||||
coreModule = k8sLib.genCoreModule {
|
coreModule = k8sLib.gencoreModule {
|
||||||
inherit hostName vars_networking;
|
inherit pkgs hostName vars_networking;
|
||||||
};
|
};
|
||||||
in {
|
in {
|
||||||
imports =
|
imports =
|
||||||
|
|||||||
@@ -32,6 +32,8 @@ in {
|
|||||||
" --write-kubeconfig /home/${username}/.kube/config"
|
" --write-kubeconfig /home/${username}/.kube/config"
|
||||||
+ " --write-kubeconfig-mode 644"
|
+ " --write-kubeconfig-mode 644"
|
||||||
+ " --service-node-port-range 80-32767"
|
+ " --service-node-port-range 80-32767"
|
||||||
|
+ " --kube-apiserver-arg='--allow-privileged=true'" # required by kubevirt
|
||||||
|
+ " --node-taint=CriticalAddonsOnly=true:NoExecute" # prevent workloads from running on the master
|
||||||
+ " --data-dir /var/lib/rancher/k3s"
|
+ " --data-dir /var/lib/rancher/k3s"
|
||||||
+ " --disable-helm-controller"
|
+ " --disable-helm-controller"
|
||||||
+ " --etcd-expose-metrics true"
|
+ " --etcd-expose-metrics true"
|
||||||
|
|||||||
@@ -1,12 +1,13 @@
|
|||||||
{
|
{
|
||||||
|
pkgs,
|
||||||
vars_networking,
|
vars_networking,
|
||||||
mylib,
|
mylib,
|
||||||
...
|
...
|
||||||
}: let
|
}: let
|
||||||
hostName = "k3s-prod-1-master-2"; # Define your hostname.
|
hostName = "k3s-prod-1-master-2"; # define your hostname.
|
||||||
k8sLib = import ../lib.nix;
|
k8sLib = import ../lib.nix;
|
||||||
coreModule = k8sLib.genCoreModule {
|
coreModule = k8sLib.gencoreModule {
|
||||||
inherit hostName vars_networking;
|
inherit pkgs hostName vars_networking;
|
||||||
};
|
};
|
||||||
in {
|
in {
|
||||||
imports =
|
imports =
|
||||||
|
|||||||
@@ -20,6 +20,8 @@ in {
|
|||||||
" --write-kubeconfig /etc/k3s/kubeconfig.yml"
|
" --write-kubeconfig /etc/k3s/kubeconfig.yml"
|
||||||
+ " --write-kubeconfig-mode 644"
|
+ " --write-kubeconfig-mode 644"
|
||||||
+ " --service-node-port-range 80-32767"
|
+ " --service-node-port-range 80-32767"
|
||||||
|
+ " --kube-apiserver-arg='--allow-privileged=true'" # required by kubevirt
|
||||||
|
+ " --node-taint=CriticalAddonsOnly=true:NoExecute" # prevent workloads from running on the master
|
||||||
+ " --data-dir /var/lib/rancher/k3s"
|
+ " --data-dir /var/lib/rancher/k3s"
|
||||||
+ " --disable-helm-controller"
|
+ " --disable-helm-controller"
|
||||||
+ " --etcd-expose-metrics true"
|
+ " --etcd-expose-metrics true"
|
||||||
|
|||||||
@@ -1,12 +1,13 @@
|
|||||||
{
|
{
|
||||||
|
pkgs,
|
||||||
vars_networking,
|
vars_networking,
|
||||||
mylib,
|
mylib,
|
||||||
...
|
...
|
||||||
}: let
|
}: let
|
||||||
hostName = "k3s-prod-1-master-3"; # Define your hostname.
|
hostName = "k3s-prod-1-master-3"; # define your hostname.
|
||||||
k8sLib = import ../lib.nix;
|
k8sLib = import ../lib.nix;
|
||||||
coreModule = k8sLib.genCoreModule {
|
coreModule = k8sLib.gencoreModule {
|
||||||
inherit hostName vars_networking;
|
inherit pkgs hostName vars_networking;
|
||||||
};
|
};
|
||||||
in {
|
in {
|
||||||
imports =
|
imports =
|
||||||
|
|||||||
@@ -20,6 +20,8 @@ in {
|
|||||||
" --write-kubeconfig /etc/k3s/kubeconfig.yml"
|
" --write-kubeconfig /etc/k3s/kubeconfig.yml"
|
||||||
+ " --write-kubeconfig-mode 644"
|
+ " --write-kubeconfig-mode 644"
|
||||||
+ " --service-node-port-range 80-32767"
|
+ " --service-node-port-range 80-32767"
|
||||||
|
+ " --kube-apiserver-arg='--allow-privileged=true'" # required by kubevirt
|
||||||
|
+ " --node-taint=CriticalAddonsOnly=true:NoExecute" # prevent workloads from running on the master
|
||||||
+ " --data-dir /var/lib/rancher/k3s"
|
+ " --data-dir /var/lib/rancher/k3s"
|
||||||
+ " --disable-helm-controller"
|
+ " --disable-helm-controller"
|
||||||
+ " --etcd-expose-metrics true"
|
+ " --etcd-expose-metrics true"
|
||||||
|
|||||||
@@ -1,12 +1,13 @@
|
|||||||
{
|
{
|
||||||
|
pkgs,
|
||||||
vars_networking,
|
vars_networking,
|
||||||
mylib,
|
mylib,
|
||||||
...
|
...
|
||||||
}: let
|
}: let
|
||||||
hostName = "k3s-prod-1-worker-1"; # Define your hostname.
|
hostName = "k3s-prod-1-worker-1"; # define your hostname.
|
||||||
k8sLib = import ../lib.nix;
|
k8sLib = import ../lib.nix;
|
||||||
coreModule = k8sLib.genCoreModule {
|
coreModule = k8sLib.gencoreModule {
|
||||||
inherit hostName vars_networking;
|
inherit pkgs hostName vars_networking;
|
||||||
};
|
};
|
||||||
in {
|
in {
|
||||||
imports =
|
imports =
|
||||||
|
|||||||
@@ -16,6 +16,8 @@ in {
|
|||||||
serverAddr = "https://${serverIp}:6443";
|
serverAddr = "https://${serverIp}:6443";
|
||||||
tokenFile = config.age.secrets."k3s-prod-1-token".path;
|
tokenFile = config.age.secrets."k3s-prod-1-token".path;
|
||||||
# https://docs.k3s.io/cli/agent
|
# https://docs.k3s.io/cli/agent
|
||||||
extraFlags = "--data-dir /var/lib/rancher/k3s";
|
extraFlags =
|
||||||
|
" --node-label=node-type=worker"
|
||||||
|
+ " --data-dir /var/lib/rancher/k3s";
|
||||||
};
|
};
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1,12 +1,13 @@
|
|||||||
{
|
{
|
||||||
|
pkgs,
|
||||||
vars_networking,
|
vars_networking,
|
||||||
mylib,
|
mylib,
|
||||||
...
|
...
|
||||||
}: let
|
}: let
|
||||||
hostName = "k3s-prod-1-worker-2"; # Define your hostname.
|
hostName = "k3s-prod-1-worker-2"; # define your hostname.
|
||||||
k8sLib = import ../lib.nix;
|
k8sLib = import ../lib.nix;
|
||||||
coreModule = k8sLib.genCoreModule {
|
coreModule = k8sLib.gencoreModule {
|
||||||
inherit hostName vars_networking;
|
inherit pkgs hostName vars_networking;
|
||||||
};
|
};
|
||||||
in {
|
in {
|
||||||
imports =
|
imports =
|
||||||
|
|||||||
@@ -16,6 +16,8 @@ in {
|
|||||||
serverAddr = "https://${serverIp}:6443";
|
serverAddr = "https://${serverIp}:6443";
|
||||||
tokenFile = config.age.secrets."k3s-prod-1-token".path;
|
tokenFile = config.age.secrets."k3s-prod-1-token".path;
|
||||||
# https://docs.k3s.io/cli/agent
|
# https://docs.k3s.io/cli/agent
|
||||||
extraFlags = "--data-dir /var/lib/rancher/k3s";
|
extraFlags =
|
||||||
|
" --node-label=node-type=worker"
|
||||||
|
+ " --data-dir /var/lib/rancher/k3s";
|
||||||
};
|
};
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1,12 +1,13 @@
|
|||||||
{
|
{
|
||||||
|
pkgs,
|
||||||
vars_networking,
|
vars_networking,
|
||||||
mylib,
|
mylib,
|
||||||
...
|
...
|
||||||
}: let
|
}: let
|
||||||
hostName = "k3s-prod-1-worker-3"; # Define your hostname.
|
hostName = "k3s-prod-1-worker-3"; # define your hostname.
|
||||||
k8sLib = import ../lib.nix;
|
k8sLib = import ../lib.nix;
|
||||||
coreModule = k8sLib.genCoreModule {
|
coreModule = k8sLib.gencoreModule {
|
||||||
inherit hostName vars_networking;
|
inherit pkgs hostName vars_networking;
|
||||||
};
|
};
|
||||||
in {
|
in {
|
||||||
imports =
|
imports =
|
||||||
|
|||||||
@@ -16,6 +16,8 @@ in {
|
|||||||
serverAddr = "https://${serverIp}:6443";
|
serverAddr = "https://${serverIp}:6443";
|
||||||
tokenFile = config.age.secrets."k3s-prod-1-token".path;
|
tokenFile = config.age.secrets."k3s-prod-1-token".path;
|
||||||
# https://docs.k3s.io/cli/agent
|
# https://docs.k3s.io/cli/agent
|
||||||
extraFlags = "--data-dir /var/lib/rancher/k3s";
|
extraFlags =
|
||||||
|
" --node-label=node-type=worker"
|
||||||
|
+ " --data-dir /var/lib/rancher/k3s";
|
||||||
};
|
};
|
||||||
}
|
}
|
||||||
|
|||||||
22
hosts/k8s/kubevirt_shoryu/default.nix
Normal file
22
hosts/k8s/kubevirt_shoryu/default.nix
Normal file
@@ -0,0 +1,22 @@
|
|||||||
|
{
|
||||||
|
pkgs,
|
||||||
|
mylib,
|
||||||
|
vars_networking,
|
||||||
|
disko,
|
||||||
|
...
|
||||||
|
}: let
|
||||||
|
# MoreFine - S500Plus
|
||||||
|
hostName = "kubevirt-shoryu"; # Define your hostname.
|
||||||
|
k8sLib = import ../lib.nix;
|
||||||
|
coreModule = k8sLib.gencoreModule {
|
||||||
|
inherit pkgs hostName vars_networking;
|
||||||
|
};
|
||||||
|
in {
|
||||||
|
imports =
|
||||||
|
(mylib.scanPaths ./.)
|
||||||
|
++ [
|
||||||
|
coreModule
|
||||||
|
disko.nixosModules.default
|
||||||
|
../kubevirt-disko-fs.nix
|
||||||
|
];
|
||||||
|
}
|
||||||
41
hosts/k8s/kubevirt_shoryu/k3s.nix
Normal file
41
hosts/k8s/kubevirt_shoryu/k3s.nix
Normal file
@@ -0,0 +1,41 @@
|
|||||||
|
{
|
||||||
|
config,
|
||||||
|
pkgs,
|
||||||
|
username,
|
||||||
|
...
|
||||||
|
}: let
|
||||||
|
package = pkgs.k3s_1_29;
|
||||||
|
in {
|
||||||
|
environment.systemPackages = with pkgs; [
|
||||||
|
package
|
||||||
|
k9s
|
||||||
|
kubectl
|
||||||
|
istioctl
|
||||||
|
kubernetes-helm
|
||||||
|
|
||||||
|
skopeo
|
||||||
|
dive # explore docker layers
|
||||||
|
];
|
||||||
|
services.k3s = {
|
||||||
|
inherit package;
|
||||||
|
enable = true;
|
||||||
|
|
||||||
|
# Initialize HA cluster using an embedded etcd datastore.
|
||||||
|
# If you are configuring an HA cluster with an embedded etcd,
|
||||||
|
# the 1st server must have `clusterInit = true`
|
||||||
|
# and other servers must connect to it using serverAddr.
|
||||||
|
clusterInit = true;
|
||||||
|
role = "server";
|
||||||
|
tokenFile = "/run/media/nixos_k3s/kubevirt-k3s-token";
|
||||||
|
# https://docs.k3s.io/cli/server
|
||||||
|
extraFlags =
|
||||||
|
" --write-kubeconfig /etc/k3s/kubeconfig.yml"
|
||||||
|
+ " --write-kubeconfig-mode 644"
|
||||||
|
+ " --service-node-port-range 80-32767"
|
||||||
|
+ " --kube-apiserver-arg='--allow-privileged=true'" # required by kubevirt
|
||||||
|
+ " --data-dir /var/lib/rancher/k3s"
|
||||||
|
+ " --disable-helm-controller"
|
||||||
|
+ " --etcd-expose-metrics true"
|
||||||
|
+ ''--etcd-snapshot-schedule-cron "0 */12 * * *"'';
|
||||||
|
};
|
||||||
|
}
|
||||||
21
hosts/k8s/kubevirt_shushou/default.nix
Normal file
21
hosts/k8s/kubevirt_shushou/default.nix
Normal file
@@ -0,0 +1,21 @@
|
|||||||
|
{
|
||||||
|
pkgs,
|
||||||
|
mylib,
|
||||||
|
vars_networking,
|
||||||
|
disko,
|
||||||
|
...
|
||||||
|
}: let
|
||||||
|
hostName = "kubevirt-shushou"; # Define your hostname.
|
||||||
|
k8sLib = import ../lib.nix;
|
||||||
|
coreModule = k8sLib.gencoreModule {
|
||||||
|
inherit pkgs hostName vars_networking;
|
||||||
|
};
|
||||||
|
in {
|
||||||
|
imports =
|
||||||
|
(mylib.scanPaths ./.)
|
||||||
|
++ [
|
||||||
|
coreModule
|
||||||
|
disko.nixosModules.default
|
||||||
|
../kubevirt-disko-fs.nix
|
||||||
|
];
|
||||||
|
}
|
||||||
41
hosts/k8s/kubevirt_shushou/k3s.nix
Normal file
41
hosts/k8s/kubevirt_shushou/k3s.nix
Normal file
@@ -0,0 +1,41 @@
|
|||||||
|
{
|
||||||
|
config,
|
||||||
|
pkgs,
|
||||||
|
username,
|
||||||
|
...
|
||||||
|
}: let
|
||||||
|
package = pkgs.k3s_1_29;
|
||||||
|
in {
|
||||||
|
environment.systemPackages = with pkgs; [
|
||||||
|
package
|
||||||
|
k9s
|
||||||
|
kubectl
|
||||||
|
istioctl
|
||||||
|
kubernetes-helm
|
||||||
|
|
||||||
|
skopeo
|
||||||
|
dive # explore docker layers
|
||||||
|
];
|
||||||
|
services.k3s = {
|
||||||
|
inherit package;
|
||||||
|
enable = true;
|
||||||
|
|
||||||
|
# Initialize HA cluster using an embedded etcd datastore.
|
||||||
|
# If you are configuring an HA cluster with an embedded etcd,
|
||||||
|
# the 1st server must have `clusterInit = true`
|
||||||
|
# and other servers must connect to it using serverAddr.
|
||||||
|
clusterInit = true;
|
||||||
|
role = "server";
|
||||||
|
tokenFile = "/run/media/nixos_k3s/kubevirt-k3s-token";
|
||||||
|
# https://docs.k3s.io/cli/server
|
||||||
|
extraFlags =
|
||||||
|
" --write-kubeconfig /etc/k3s/kubeconfig.yml"
|
||||||
|
+ " --write-kubeconfig-mode 644"
|
||||||
|
+ " --service-node-port-range 80-32767"
|
||||||
|
+ " --kube-apiserver-arg='--allow-privileged=true'" # required by kubevirt
|
||||||
|
+ " --data-dir /var/lib/rancher/k3s"
|
||||||
|
+ " --disable-helm-controller"
|
||||||
|
+ " --etcd-expose-metrics true"
|
||||||
|
+ ''--etcd-snapshot-schedule-cron "0 */12 * * *"'';
|
||||||
|
};
|
||||||
|
}
|
||||||
21
hosts/k8s/kubevirt_youko/default.nix
Normal file
21
hosts/k8s/kubevirt_youko/default.nix
Normal file
@@ -0,0 +1,21 @@
|
|||||||
|
{
|
||||||
|
pkgs,
|
||||||
|
mylib,
|
||||||
|
vars_networking,
|
||||||
|
disko,
|
||||||
|
...
|
||||||
|
}: let
|
||||||
|
hostName = "kubevirt-youko"; # Define your hostname.
|
||||||
|
k8sLib = import ../lib.nix;
|
||||||
|
coreModule = k8sLib.gencoreModule {
|
||||||
|
inherit pkgs hostName vars_networking;
|
||||||
|
};
|
||||||
|
in {
|
||||||
|
imports =
|
||||||
|
(mylib.scanPaths ./.)
|
||||||
|
++ [
|
||||||
|
coreModule
|
||||||
|
disko.nixosModules.default
|
||||||
|
../kubevirt-disko-fs.nix
|
||||||
|
];
|
||||||
|
}
|
||||||
41
hosts/k8s/kubevirt_youko/k3s.nix
Normal file
41
hosts/k8s/kubevirt_youko/k3s.nix
Normal file
@@ -0,0 +1,41 @@
|
|||||||
|
{
|
||||||
|
config,
|
||||||
|
pkgs,
|
||||||
|
username,
|
||||||
|
...
|
||||||
|
}: let
|
||||||
|
package = pkgs.k3s_1_29;
|
||||||
|
in {
|
||||||
|
environment.systemPackages = with pkgs; [
|
||||||
|
package
|
||||||
|
k9s
|
||||||
|
kubectl
|
||||||
|
istioctl
|
||||||
|
kubernetes-helm
|
||||||
|
|
||||||
|
skopeo
|
||||||
|
dive # explore docker layers
|
||||||
|
];
|
||||||
|
services.k3s = {
|
||||||
|
inherit package;
|
||||||
|
enable = true;
|
||||||
|
|
||||||
|
# Initialize HA cluster using an embedded etcd datastore.
|
||||||
|
# If you are configuring an HA cluster with an embedded etcd,
|
||||||
|
# the 1st server must have `clusterInit = true`
|
||||||
|
# and other servers must connect to it using serverAddr.
|
||||||
|
clusterInit = true;
|
||||||
|
role = "server";
|
||||||
|
tokenFile = "/run/media/nixos_k3s/kubevirt-k3s-token";
|
||||||
|
# https://docs.k3s.io/cli/server
|
||||||
|
extraFlags =
|
||||||
|
" --write-kubeconfig /etc/k3s/kubeconfig.yml"
|
||||||
|
+ " --write-kubeconfig-mode 644"
|
||||||
|
+ " --service-node-port-range 80-32767"
|
||||||
|
+ " --kube-apiserver-arg='--allow-privileged=true'" # required by kubevirt
|
||||||
|
+ " --data-dir /var/lib/rancher/k3s"
|
||||||
|
+ " --disable-helm-controller"
|
||||||
|
+ " --etcd-expose-metrics true"
|
||||||
|
+ ''--etcd-snapshot-schedule-cron "0 */12 * * *"'';
|
||||||
|
};
|
||||||
|
}
|
||||||
@@ -1,7 +1,9 @@
|
|||||||
{
|
{
|
||||||
genCoreModule = {
|
gencoreModule = {
|
||||||
|
pkgs,
|
||||||
hostName,
|
hostName,
|
||||||
vars_networking,
|
vars_networking,
|
||||||
|
...
|
||||||
}: let
|
}: let
|
||||||
hostAddress = vars_networking.hostAddress.${hostName};
|
hostAddress = vars_networking.hostAddress.${hostName};
|
||||||
in {
|
in {
|
||||||
@@ -18,9 +20,15 @@
|
|||||||
"cifs" # mount windows share
|
"cifs" # mount windows share
|
||||||
];
|
];
|
||||||
|
|
||||||
boot.kernelModules = ["kvm-amd"];
|
boot.kernelModules = ["kvm-amd" "vfio-pci"];
|
||||||
boot.extraModprobeConfig = "options kvm_amd nested=1"; # for amd cpu
|
boot.extraModprobeConfig = "options kvm_amd nested=1"; # for amd cpu
|
||||||
|
|
||||||
|
environment.systemPackages = with pkgs; [
|
||||||
|
# Validate Hardware Virtualization Support via:
|
||||||
|
# virt-host-validate qemu
|
||||||
|
libvirt
|
||||||
|
];
|
||||||
|
|
||||||
networking = {
|
networking = {
|
||||||
inherit hostName;
|
inherit hostName;
|
||||||
inherit (vars_networking) defaultGateway nameservers;
|
inherit (vars_networking) defaultGateway nameservers;
|
||||||
|
|||||||
@@ -85,7 +85,7 @@ cryptsetup --help
|
|||||||
|
|
||||||
# NOTE: `cat shoukei.md | grep luks > luks.sh` to generate this script
|
# NOTE: `cat shoukei.md | grep luks > luks.sh` to generate this script
|
||||||
# encrypt the root partition with luks2 and argon2id, will prompt for a passphrase, which will be used to unlock the partition.
|
# encrypt the root partition with luks2 and argon2id, will prompt for a passphrase, which will be used to unlock the partition.
|
||||||
cryptsetup luksFormat --type luks2 --cipher aes-xts-plain64 --hash sha512 --iter-time 5000 --key-size 256 --pbkdf argon2id --use-urandom --verify-passphrase /dev/nvme0n1p2
|
cryptsetup luksFormat --type luks2 --cipher aes-xts-plain64 --hash sha512 --iter-time 5000 --key-size 256 --pbkdf argon2id --use-random --verify-passphrase /dev/nvme0n1p2
|
||||||
|
|
||||||
# show status
|
# show status
|
||||||
cryptsetup luksDump /dev/nvme0n1p2
|
cryptsetup luksDump /dev/nvme0n1p2
|
||||||
|
|||||||
@@ -69,7 +69,7 @@ cryptsetup --help
|
|||||||
|
|
||||||
# NOTE: `cat shoukei.md | grep luks > format.sh` to generate this script
|
# NOTE: `cat shoukei.md | grep luks > format.sh` to generate this script
|
||||||
# encrypt the root partition with luks2 and argon2id, will prompt for a passphrase, which will be used to unlock the partition.
|
# encrypt the root partition with luks2 and argon2id, will prompt for a passphrase, which will be used to unlock the partition.
|
||||||
cryptsetup luksFormat --type luks2 --cipher aes-xts-plain64 --hash sha512 --iter-time 5000 --key-size 256 --pbkdf argon2id --use-urandom --verify-passphrase /dev/nvme0n1p4
|
cryptsetup luksFormat --type luks2 --cipher aes-xts-plain64 --hash sha512 --iter-time 5000 --key-size 256 --pbkdf argon2id --use-random --verify-passphrase /dev/nvme0n1p4
|
||||||
|
|
||||||
# show status
|
# show status
|
||||||
cryptsetup luksDump /dev/nvme0n1p4
|
cryptsetup luksDump /dev/nvme0n1p4
|
||||||
|
|||||||
2
pulumi/k3s-prod-1/.gitignore
vendored
Normal file
2
pulumi/k3s-prod-1/.gitignore
vendored
Normal file
@@ -0,0 +1,2 @@
|
|||||||
|
*.pyc
|
||||||
|
venv/
|
||||||
@@ -1,3 +1,6 @@
|
|||||||
name: k3s-prod-1
|
name: k3s-prod-1
|
||||||
runtime: go
|
runtime:
|
||||||
description: A Go program to deploy all the resources for the k3s-prod-1 cluster
|
name: python
|
||||||
|
options:
|
||||||
|
virtualenv: venv
|
||||||
|
description: A Python program to deploy a Helm chart onto a Kubernetes cluster
|
||||||
|
|||||||
3
pulumi/k3s-prod-1/__main__.py
Normal file
3
pulumi/k3s-prod-1/__main__.py
Normal file
@@ -0,0 +1,3 @@
|
|||||||
|
from monitoring import *
|
||||||
|
from networking import *
|
||||||
|
from visualization import *
|
||||||
@@ -1,103 +0,0 @@
|
|||||||
module k3s-prod-1
|
|
||||||
|
|
||||||
go 1.21
|
|
||||||
|
|
||||||
toolchain go1.21.6
|
|
||||||
|
|
||||||
require (
|
|
||||||
github.com/pulumi/pulumi-kubernetes v1.6.0
|
|
||||||
github.com/pulumi/pulumi-kubernetes/sdk/v4 v4.7.1
|
|
||||||
github.com/pulumi/pulumi/sdk/v3 v3.106.0
|
|
||||||
)
|
|
||||||
|
|
||||||
require (
|
|
||||||
dario.cat/mergo v1.0.0 // indirect
|
|
||||||
github.com/Microsoft/go-winio v0.6.1 // indirect
|
|
||||||
github.com/ProtonMail/go-crypto v0.0.0-20230828082145-3c4c8a2d2371 // indirect
|
|
||||||
github.com/aead/chacha20 v0.0.0-20180709150244-8b13a72661da // indirect
|
|
||||||
github.com/agext/levenshtein v1.2.3 // indirect
|
|
||||||
github.com/apparentlymart/go-textseg/v13 v13.0.0 // indirect
|
|
||||||
github.com/atotto/clipboard v0.1.4 // indirect
|
|
||||||
github.com/aymanbagabas/go-osc52/v2 v2.0.1 // indirect
|
|
||||||
github.com/blang/semver v3.5.1+incompatible // indirect
|
|
||||||
github.com/charmbracelet/bubbles v0.16.1 // indirect
|
|
||||||
github.com/charmbracelet/bubbletea v0.24.2 // indirect
|
|
||||||
github.com/charmbracelet/lipgloss v0.7.1 // indirect
|
|
||||||
github.com/cheggaaa/pb v1.0.29 // indirect
|
|
||||||
github.com/cloudflare/circl v1.3.7 // indirect
|
|
||||||
github.com/containerd/console v1.0.4-0.20230313162750-1ae8d489ac81 // indirect
|
|
||||||
github.com/cyphar/filepath-securejoin v0.2.4 // indirect
|
|
||||||
github.com/djherbis/times v1.5.0 // indirect
|
|
||||||
github.com/emirpasic/gods v1.18.1 // indirect
|
|
||||||
github.com/go-git/gcfg v1.5.1-0.20230307220236-3a3c6141e376 // indirect
|
|
||||||
github.com/go-git/go-billy/v5 v5.5.0 // indirect
|
|
||||||
github.com/go-git/go-git/v5 v5.11.0 // indirect
|
|
||||||
github.com/gofrs/flock v0.7.1 // indirect
|
|
||||||
github.com/gogo/protobuf v1.3.2 // indirect
|
|
||||||
github.com/golang/glog v1.1.0 // indirect
|
|
||||||
github.com/golang/groupcache v0.0.0-20210331224755-41bb18bfe9da // indirect
|
|
||||||
github.com/golang/protobuf v1.5.3 // indirect
|
|
||||||
github.com/grpc-ecosystem/grpc-opentracing v0.0.0-20180507213350-8e809c8a8645 // indirect
|
|
||||||
github.com/hashicorp/errwrap v1.1.0 // indirect
|
|
||||||
github.com/hashicorp/go-multierror v1.1.1 // indirect
|
|
||||||
github.com/hashicorp/hcl/v2 v2.17.0 // indirect
|
|
||||||
github.com/inconshreveable/mousetrap v1.1.0 // indirect
|
|
||||||
github.com/jbenet/go-context v0.0.0-20150711004518-d14ea06fba99 // indirect
|
|
||||||
github.com/kevinburke/ssh_config v1.2.0 // indirect
|
|
||||||
github.com/lucasb-eyer/go-colorful v1.2.0 // indirect
|
|
||||||
github.com/mattn/go-isatty v0.0.19 // indirect
|
|
||||||
github.com/mattn/go-localereader v0.0.1 // indirect
|
|
||||||
github.com/mattn/go-runewidth v0.0.15 // indirect
|
|
||||||
github.com/mitchellh/go-homedir v1.1.0 // indirect
|
|
||||||
github.com/mitchellh/go-ps v1.0.0 // indirect
|
|
||||||
github.com/mitchellh/go-wordwrap v1.0.1 // indirect
|
|
||||||
github.com/muesli/ansi v0.0.0-20230316100256-276c6243b2f6 // indirect
|
|
||||||
github.com/muesli/cancelreader v0.2.2 // indirect
|
|
||||||
github.com/muesli/reflow v0.3.0 // indirect
|
|
||||||
github.com/muesli/termenv v0.15.2 // indirect
|
|
||||||
github.com/opentracing/basictracer-go v1.1.0 // indirect
|
|
||||||
github.com/opentracing/opentracing-go v1.2.0 // indirect
|
|
||||||
github.com/pgavlin/fx v0.1.6 // indirect
|
|
||||||
github.com/pjbgf/sha1cd v0.3.0 // indirect
|
|
||||||
github.com/pkg/errors v0.9.1 // indirect
|
|
||||||
github.com/pkg/term v1.1.0 // indirect
|
|
||||||
github.com/pulumi/appdash v0.0.0-20231130102222-75f619a67231 // indirect
|
|
||||||
github.com/pulumi/esc v0.6.2 // indirect
|
|
||||||
github.com/pulumi/pulumi/sdk v0.0.0-20200324171821-8ce10e1dfe54 // indirect
|
|
||||||
github.com/rivo/uniseg v0.4.4 // indirect
|
|
||||||
github.com/rogpeppe/go-internal v1.11.0 // indirect
|
|
||||||
github.com/sabhiram/go-gitignore v0.0.0-20210923224102-525f6e181f06 // indirect
|
|
||||||
github.com/santhosh-tekuri/jsonschema/v5 v5.0.0 // indirect
|
|
||||||
github.com/sergi/go-diff v1.3.1 // indirect
|
|
||||||
github.com/skeema/knownhosts v1.2.1 // indirect
|
|
||||||
github.com/spf13/cobra v1.7.0 // indirect
|
|
||||||
github.com/spf13/pflag v1.0.5 // indirect
|
|
||||||
github.com/src-d/gcfg v1.4.0 // indirect
|
|
||||||
github.com/texttheater/golang-levenshtein v1.0.1 // indirect
|
|
||||||
github.com/tweekmonster/luser v0.0.0-20161003172636-3fa38070dbd7 // indirect
|
|
||||||
github.com/uber/jaeger-client-go v2.30.0+incompatible // indirect
|
|
||||||
github.com/uber/jaeger-lib v2.4.1+incompatible // indirect
|
|
||||||
github.com/xanzy/ssh-agent v0.3.3 // indirect
|
|
||||||
github.com/zclconf/go-cty v1.13.2 // indirect
|
|
||||||
go.uber.org/atomic v1.9.0 // indirect
|
|
||||||
golang.org/x/crypto v0.17.0 // indirect
|
|
||||||
golang.org/x/exp v0.0.0-20231110203233-9a3e6036ecaa // indirect
|
|
||||||
golang.org/x/mod v0.14.0 // indirect
|
|
||||||
golang.org/x/net v0.19.0 // indirect
|
|
||||||
golang.org/x/sync v0.5.0 // indirect
|
|
||||||
golang.org/x/sys v0.15.0 // indirect
|
|
||||||
golang.org/x/term v0.15.0 // indirect
|
|
||||||
golang.org/x/text v0.14.0 // indirect
|
|
||||||
golang.org/x/tools v0.15.0 // indirect
|
|
||||||
google.golang.org/genproto/googleapis/rpc v0.0.0-20230706204954-ccb25ca9f130 // indirect
|
|
||||||
google.golang.org/grpc v1.57.1 // indirect
|
|
||||||
google.golang.org/protobuf v1.31.0 // indirect
|
|
||||||
gopkg.in/src-d/go-billy.v4 v4.3.2 // indirect
|
|
||||||
gopkg.in/src-d/go-git.v4 v4.13.1 // indirect
|
|
||||||
gopkg.in/warnings.v0 v0.1.2 // indirect
|
|
||||||
gopkg.in/yaml.v2 v2.4.0 // indirect
|
|
||||||
gopkg.in/yaml.v3 v3.0.1 // indirect
|
|
||||||
k8s.io/klog v1.0.0 // indirect
|
|
||||||
lukechampine.com/frand v1.4.2 // indirect
|
|
||||||
sourcegraph.com/sourcegraph/appdash v0.0.0-20190731080439-ebfcffb1b5c0 // indirect
|
|
||||||
)
|
|
||||||
File diff suppressed because it is too large
Load Diff
@@ -1,21 +0,0 @@
|
|||||||
package main
|
|
||||||
|
|
||||||
import (
|
|
||||||
"k3s-prod-1/monitoring"
|
|
||||||
"k3s-prod-1/networking"
|
|
||||||
|
|
||||||
"github.com/pulumi/pulumi/sdk/v3/go/pulumi"
|
|
||||||
)
|
|
||||||
|
|
||||||
func main() {
|
|
||||||
pulumi.Run(func(ctx *pulumi.Context) error {
|
|
||||||
if err := monitoring.NewMonitoring(ctx, "prod"); err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
if err := networking.NewNetworking(ctx, "prod"); err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
|
|
||||||
return nil
|
|
||||||
})
|
|
||||||
}
|
|
||||||
1
pulumi/k3s-prod-1/monitoring/__init__.py
Normal file
1
pulumi/k3s-prod-1/monitoring/__init__.py
Normal file
@@ -0,0 +1 @@
|
|||||||
|
from .victoria_metrics import *
|
||||||
@@ -1,28 +0,0 @@
|
|||||||
package monitoring
|
|
||||||
|
|
||||||
import (
|
|
||||||
corev1 "github.com/pulumi/pulumi-kubernetes/sdk/v4/go/kubernetes/core/v1"
|
|
||||||
metav1 "github.com/pulumi/pulumi-kubernetes/sdk/v4/go/kubernetes/meta/v1"
|
|
||||||
"github.com/pulumi/pulumi/sdk/v3/go/pulumi"
|
|
||||||
)
|
|
||||||
|
|
||||||
func NewMonitoring(ctx *pulumi.Context, env string) error {
|
|
||||||
// Create a Kubernetes Namespace
|
|
||||||
namespaceName := "monitoring"
|
|
||||||
namespace, err := corev1.NewNamespace(ctx, namespaceName, &corev1.NamespaceArgs{
|
|
||||||
Metadata: &metav1.ObjectMetaArgs{
|
|
||||||
Name: pulumi.String(namespaceName),
|
|
||||||
},
|
|
||||||
})
|
|
||||||
if err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
|
|
||||||
// Export the name of the namespace
|
|
||||||
ctx.Export("monitoringNamespaceName", namespace.Metadata.Name())
|
|
||||||
|
|
||||||
if err := NewVictoriaMetrics(ctx, env, namespace); err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
42
pulumi/k3s-prod-1/monitoring/victoria_metrics.py
Normal file
42
pulumi/k3s-prod-1/monitoring/victoria_metrics.py
Normal file
@@ -0,0 +1,42 @@
|
|||||||
|
import pulumi
|
||||||
|
import pulumi_kubernetes as kubernetes
|
||||||
|
|
||||||
|
from pathlib import Path
|
||||||
|
import yaml
|
||||||
|
|
||||||
|
config = pulumi.Config()
|
||||||
|
k8s_namespace = "monitoring"
|
||||||
|
app_labels = {
|
||||||
|
"app": "monitoring",
|
||||||
|
}
|
||||||
|
|
||||||
|
victoriaMetricsvaluesPath = Path(__file__).parent / "victoria_metrics_helm_values.yml"
|
||||||
|
|
||||||
|
# Create a namespace (user supplies the name of the namespace)
|
||||||
|
monitoring_ns = kubernetes.core.v1.Namespace(
|
||||||
|
"monitoring",
|
||||||
|
metadata=kubernetes.meta.v1.ObjectMetaArgs(
|
||||||
|
labels=app_labels,
|
||||||
|
name=k8s_namespace,
|
||||||
|
),
|
||||||
|
)
|
||||||
|
|
||||||
|
# https://github.com/VictoriaMetrics/helm-charts/tree/master/charts/victoria-metrics-k8s-stack
|
||||||
|
victoriaMetrics = kubernetes.helm.v3.Release(
|
||||||
|
"victoria-metrics-k8s-stack",
|
||||||
|
chart="victoria-metrics-k8s-stack",
|
||||||
|
namespace=monitoring_ns.metadata.name,
|
||||||
|
repository_opts=kubernetes.helm.v3.RepositoryOptsArgs(
|
||||||
|
repo="https://victoriametrics.github.io/helm-charts/",
|
||||||
|
),
|
||||||
|
version="0.19.2",
|
||||||
|
skip_crds=False,
|
||||||
|
atomic=True, # purges chart on fail
|
||||||
|
cleanup_on_fail=True, # Allow deletion of new resources created in this upgrade when upgrade fails.
|
||||||
|
dependency_update=True, # run helm dependency update before installing the chart
|
||||||
|
reset_values=True, # When upgrading, reset the values to the ones built into the chart
|
||||||
|
# verify=True, # verify the package before installing it
|
||||||
|
# recreate_pods=True, # performs pods restart for the resource if applicable
|
||||||
|
value_yaml_files=[pulumi.FileAsset(victoriaMetricsvaluesPath)],
|
||||||
|
)
|
||||||
|
|
||||||
@@ -0,0 +1,37 @@
|
|||||||
|
# https://github.com/VictoriaMetrics/helm-charts/tree/master/charts/victoria-metrics-k8s-stack
|
||||||
|
#
|
||||||
|
# Pulumi will complain ` ValueError: unexpected input of type set` if some values are not available in helm chart!
|
||||||
|
grafana:
|
||||||
|
enabled: true
|
||||||
|
defaultDashboardsTimezone: utc+8
|
||||||
|
ingress:
|
||||||
|
enabled: true
|
||||||
|
hosts:
|
||||||
|
- k8s-grafana.writefor.fun
|
||||||
|
persistence:
|
||||||
|
type: pvc
|
||||||
|
enabled: false
|
||||||
|
kube-state-metrics:
|
||||||
|
enabled: true
|
||||||
|
prometheus-node-exporter:
|
||||||
|
# install node exporter via nixos, not container
|
||||||
|
enabled: false
|
||||||
|
vmagent:
|
||||||
|
# vmagent collects metrics from targets and sends them to a remote storage
|
||||||
|
enabled: true
|
||||||
|
vmalert:
|
||||||
|
# vmalert is a Prometheus-compatible alertmanager
|
||||||
|
enabled: true
|
||||||
|
vmsingle:
|
||||||
|
# Single-node VictoriaMetrics for storing metrics.
|
||||||
|
# https://docs.victoriametrics.com/faq/#which-victoriametrics-type-is-recommended-for-use-in-production---single-node-or-cluster
|
||||||
|
# vmsingle = vmcluster(vmselect + vmstorage + vminsert)
|
||||||
|
enabled: true
|
||||||
|
ingress:
|
||||||
|
hosts:
|
||||||
|
- vm.writefor.fun
|
||||||
|
spec:
|
||||||
|
storage:
|
||||||
|
resources:
|
||||||
|
requests:
|
||||||
|
storage: 50Gi
|
||||||
@@ -1,85 +0,0 @@
|
|||||||
package monitoring
|
|
||||||
|
|
||||||
import (
|
|
||||||
corev1 "github.com/pulumi/pulumi-kubernetes/sdk/v4/go/kubernetes/core/v1"
|
|
||||||
"github.com/pulumi/pulumi-kubernetes/sdk/v4/go/kubernetes/helm/v3"
|
|
||||||
"github.com/pulumi/pulumi/sdk/v3/go/pulumi"
|
|
||||||
)
|
|
||||||
|
|
||||||
func NewVictoriaMetrics(ctx *pulumi.Context, env string, namespace corev1.Namespace) error {
|
|
||||||
var opts []pulumi.ResourceOption
|
|
||||||
opts = append(opts, pulumi.DependsOn([]pulumi.Resource{namespace}))
|
|
||||||
|
|
||||||
// https://github.com/VictoriaMetrics/helm-charts/tree/master/charts/victoria-metrics-k8s-stack
|
|
||||||
_, err := helm.NewChart(ctx, "victoria-metrics-k8s-stack", helm.ChartArgs{
|
|
||||||
Chart: pulumi.String("victoria-metrics-k8s-stack"),
|
|
||||||
Version: pulumi.String("0.19.0"),
|
|
||||||
Namespace: pulumi.String(namespace.Metadata.Name()),
|
|
||||||
FetchArgs: helm.FetchArgs{
|
|
||||||
Repo: pulumi.String("https://victoriametrics.github.io/helm-charts/"),
|
|
||||||
},
|
|
||||||
// https://github.com/VictoriaMetrics/helm-charts/blob/master/charts/victoria-metrics-k8s-stack/README.md
|
|
||||||
Values: pulumi.Map{
|
|
||||||
// grafana.ingress.enabled: true
|
|
||||||
"ingress": pulumi.Map{
|
|
||||||
"enabled": pulumi.Bool(true),
|
|
||||||
},
|
|
||||||
// grafana.defaultDashboardsTimezone: utc+8
|
|
||||||
// grafana.ingress.hosts[0].host: grafana.example.com
|
|
||||||
"grafana": pulumi.Map{
|
|
||||||
"defaultDashboardsTimezone": pulumi.String("utc+8"),
|
|
||||||
"ingress": pulumi.Map{
|
|
||||||
"hosts": pulumi.Array{
|
|
||||||
pulumi.Map{
|
|
||||||
"host": pulumi.String("k8s-grafana.writefor.fun"),
|
|
||||||
},
|
|
||||||
},
|
|
||||||
},
|
|
||||||
},
|
|
||||||
// prometheus-node-exporter.enabled: false
|
|
||||||
"nodeExporter": pulumi.Map{
|
|
||||||
"enabled": pulumi.Bool(false),
|
|
||||||
},
|
|
||||||
"vmsingle": pulumi.Map{
|
|
||||||
"enabled": pulumi.Bool(true),
|
|
||||||
"ingress": pulumi.Map{
|
|
||||||
"hosts": pulumi.Array{
|
|
||||||
pulumi.Map{
|
|
||||||
"host": pulumi.String("vm.writefor.fun"),
|
|
||||||
},
|
|
||||||
},
|
|
||||||
},
|
|
||||||
// https://docs.victoriametrics.com/operator/api/#vmsinglespec
|
|
||||||
"spec": pulumi.Map{
|
|
||||||
"affinity": pulumi.Map{
|
|
||||||
"nodeAffinity": pulumi.Map{
|
|
||||||
"requiredDuringSchedulingIgnoredDuringExecution": pulumi.Map{
|
|
||||||
"nodeSelectorTerms": pulumi.Array{
|
|
||||||
pulumi.Map{
|
|
||||||
"matchExpressions": pulumi.Array{
|
|
||||||
pulumi.Map{
|
|
||||||
"key": pulumi.String("kubernetes.io/arch"),
|
|
||||||
"operator": pulumi.String("In"),
|
|
||||||
"values": pulumi.Array{
|
|
||||||
pulumi.String("amd64"),
|
|
||||||
},
|
|
||||||
},
|
|
||||||
},
|
|
||||||
},
|
|
||||||
},
|
|
||||||
},
|
|
||||||
},
|
|
||||||
},
|
|
||||||
"storage": pulumi.Map{
|
|
||||||
"resources": pulumi.Map{
|
|
||||||
"requests": pulumi.Map{
|
|
||||||
"storage": pulumi.String("50Gi"),
|
|
||||||
},
|
|
||||||
},
|
|
||||||
},
|
|
||||||
},
|
|
||||||
},
|
|
||||||
},
|
|
||||||
}, opts...)
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
1
pulumi/k3s-prod-1/networking/__init__.py
Normal file
1
pulumi/k3s-prod-1/networking/__init__.py
Normal file
@@ -0,0 +1 @@
|
|||||||
|
from .cert_manager import *
|
||||||
@@ -1,25 +0,0 @@
|
|||||||
package networking
|
|
||||||
|
|
||||||
import (
|
|
||||||
corev1 "github.com/pulumi/pulumi-kubernetes/sdk/v4/go/kubernetes/core/v1"
|
|
||||||
"github.com/pulumi/pulumi-kubernetes/sdk/v4/go/kubernetes/helm/v3"
|
|
||||||
"github.com/pulumi/pulumi/sdk/v3/go/pulumi"
|
|
||||||
)
|
|
||||||
|
|
||||||
func NewCertManager(ctx *pulumi.Context, env string, namespace corev1.Namespace) error {
|
|
||||||
var opts []pulumi.ResourceOption
|
|
||||||
opts = append(opts, pulumi.DependsOn([]pulumi.Resource{namespace}))
|
|
||||||
|
|
||||||
_, err := helm.NewChart(ctx, "cert-manager", helm.ChartArgs{
|
|
||||||
Chart: pulumi.String("cert-manager"),
|
|
||||||
Version: pulumi.String("corev1.14.2 "),
|
|
||||||
Namespace: pulumi.String(namespace.Metadata.Name()),
|
|
||||||
FetchArgs: helm.FetchArgs{
|
|
||||||
Repo: pulumi.String("https://charts.jetstack.io"),
|
|
||||||
},
|
|
||||||
// https://cert-manager.io/docs/installation/helm/
|
|
||||||
Values: pulumi.Map{},
|
|
||||||
}, opts...)
|
|
||||||
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
15
pulumi/k3s-prod-1/networking/cert_manager.py
Normal file
15
pulumi/k3s-prod-1/networking/cert_manager.py
Normal file
@@ -0,0 +1,15 @@
|
|||||||
|
import pulumi
|
||||||
|
from pulumi_kubernetes.core.v1 import Namespace
|
||||||
|
from pulumi_kubernetes_cert_manager import CertManager, ReleaseArgs
|
||||||
|
|
||||||
|
ns_name = "cert-manager"
|
||||||
|
ns = Namespace("cert-manager", metadata={"name": ns_name})
|
||||||
|
|
||||||
|
# Install cert-manager into our cluster.
|
||||||
|
manager = CertManager(
|
||||||
|
"cert-manager",
|
||||||
|
install_crds=True,
|
||||||
|
helm_options=ReleaseArgs(
|
||||||
|
namespace=ns_name,
|
||||||
|
),
|
||||||
|
)
|
||||||
@@ -1,28 +0,0 @@
|
|||||||
package networking
|
|
||||||
|
|
||||||
import (
|
|
||||||
corev1 "github.com/pulumi/pulumi-kubernetes/sdk/v4/go/kubernetes/core/v1"
|
|
||||||
metav1 "github.com/pulumi/pulumi-kubernetes/sdk/v4/go/kubernetes/meta/v1"
|
|
||||||
"github.com/pulumi/pulumi/sdk/v3/go/pulumi"
|
|
||||||
)
|
|
||||||
|
|
||||||
func NewNetworking(ctx *pulumi.Context, env string) error {
|
|
||||||
// Create a Kubernetes Namespace
|
|
||||||
namespaceName := "networking"
|
|
||||||
namespace, err := corev1.NewNamespace(ctx, namespaceName, &corev1.NamespaceArgs{
|
|
||||||
Metadata: &metav1.ObjectMetaArgs{
|
|
||||||
Name: pulumi.String(namespaceName),
|
|
||||||
},
|
|
||||||
})
|
|
||||||
if err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
|
|
||||||
// Export the name of the namespace
|
|
||||||
ctx.Export("networkingNamespaceName", namespace.Metadata.Name())
|
|
||||||
|
|
||||||
if err := NewCertManager(ctx, env, namespace); err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
2
pulumi/k3s-prod-1/requirements.txt
Normal file
2
pulumi/k3s-prod-1/requirements.txt
Normal file
@@ -0,0 +1,2 @@
|
|||||||
|
pulumi>=3.0.0,<4.0.0
|
||||||
|
pulumi-kubernetes>=4.0.0,<5.0.0
|
||||||
2
pulumi/k3s-prod-1/visualization/__init__.py
Normal file
2
pulumi/k3s-prod-1/visualization/__init__.py
Normal file
@@ -0,0 +1,2 @@
|
|||||||
|
from .kubevirt import *
|
||||||
|
from .virtual_machines import *
|
||||||
22
pulumi/k3s-prod-1/visualization/kubevirt.py
Normal file
22
pulumi/k3s-prod-1/visualization/kubevirt.py
Normal file
@@ -0,0 +1,22 @@
|
|||||||
|
from pulumi_kubernetes.yaml import ConfigGroup
|
||||||
|
|
||||||
|
from pathlib import Path
|
||||||
|
|
||||||
|
currentDir = Path(__file__).parent
|
||||||
|
|
||||||
|
|
||||||
|
def virtHandlerNodePlacement(obj, opts):
|
||||||
|
if obj["kind"] == "KubeVirt":
|
||||||
|
obj["spec"]["workloads"] = {
|
||||||
|
"nodePlacement": {
|
||||||
|
"nodeSelector": {"node-type": "worker"}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
kubevirt = ConfigGroup(
|
||||||
|
"kubevirt",
|
||||||
|
files=[currentDir.as_posix() + "/yaml/*.yaml"],
|
||||||
|
# A set of transformations to apply to Kubernetes resource definitions before registering with engine.
|
||||||
|
transformations=[virtHandlerNodePlacement],
|
||||||
|
)
|
||||||
@@ -0,0 +1 @@
|
|||||||
|
from .test import *
|
||||||
10
pulumi/k3s-prod-1/visualization/virtual_machines/test.py
Normal file
10
pulumi/k3s-prod-1/visualization/virtual_machines/test.py
Normal file
@@ -0,0 +1,10 @@
|
|||||||
|
from pulumi_kubernetes.yaml import ConfigGroup
|
||||||
|
|
||||||
|
from pathlib import Path
|
||||||
|
|
||||||
|
currentDir = Path(__file__).parent
|
||||||
|
|
||||||
|
kubevirt = ConfigGroup(
|
||||||
|
"virtual-machines",
|
||||||
|
files=[currentDir.as_posix() + "/yaml/*.yaml"],
|
||||||
|
)
|
||||||
@@ -0,0 +1,34 @@
|
|||||||
|
apiVersion: kubevirt.io/v1
|
||||||
|
kind: VirtualMachineInstance
|
||||||
|
metadata:
|
||||||
|
name: testvmi-nocloud
|
||||||
|
spec:
|
||||||
|
terminationGracePeriodSeconds: 30
|
||||||
|
domain:
|
||||||
|
resources:
|
||||||
|
requests:
|
||||||
|
memory: 1024M
|
||||||
|
devices:
|
||||||
|
disks:
|
||||||
|
- name: containerdisk
|
||||||
|
disk:
|
||||||
|
bus: virtio
|
||||||
|
- name: emptydisk
|
||||||
|
disk:
|
||||||
|
bus: virtio
|
||||||
|
- disk:
|
||||||
|
bus: virtio
|
||||||
|
name: cloudinitdisk
|
||||||
|
volumes:
|
||||||
|
- name: containerdisk
|
||||||
|
containerDisk:
|
||||||
|
image: kubevirt/fedora-cloud-container-disk-demo:latest
|
||||||
|
- name: emptydisk
|
||||||
|
emptyDisk:
|
||||||
|
capacity: "2Gi"
|
||||||
|
- name: cloudinitdisk
|
||||||
|
cloudInitNoCloud:
|
||||||
|
userData: |-
|
||||||
|
#cloud-config
|
||||||
|
password: fedora
|
||||||
|
chpasswd: { expire: False }
|
||||||
5
pulumi/k3s-prod-1/visualization/yaml/download.sh
Normal file
5
pulumi/k3s-prod-1/visualization/yaml/download.sh
Normal file
@@ -0,0 +1,5 @@
|
|||||||
|
export RELEASE=$(curl https://storage.googleapis.com/kubevirt-prow/release/kubevirt/kubevirt/stable.txt)
|
||||||
|
echo "The latest kubevirt's version is $RELEASE"
|
||||||
|
|
||||||
|
curl -Lo kubevirt-operator-${RELEASE}.yaml https://github.com/kubevirt/kubevirt/releases/download/${RELEASE}/kubevirt-operator.yaml
|
||||||
|
curl -Lo kubevirt-cr-${RELEASE}.yaml https://github.com/kubevirt/kubevirt/releases/download/${RELEASE}/kubevirt-cr.yaml
|
||||||
14
pulumi/k3s-prod-1/visualization/yaml/kubevirt-cr-v1.1.1.yaml
Normal file
14
pulumi/k3s-prod-1/visualization/yaml/kubevirt-cr-v1.1.1.yaml
Normal file
@@ -0,0 +1,14 @@
|
|||||||
|
---
|
||||||
|
apiVersion: kubevirt.io/v1
|
||||||
|
kind: KubeVirt
|
||||||
|
metadata:
|
||||||
|
name: kubevirt
|
||||||
|
namespace: kubevirt
|
||||||
|
spec:
|
||||||
|
certificateRotateStrategy: {}
|
||||||
|
configuration:
|
||||||
|
developerConfiguration:
|
||||||
|
featureGates: []
|
||||||
|
customizeComponents: {}
|
||||||
|
imagePullPolicy: IfNotPresent
|
||||||
|
workloadUpdateStrategy: {}
|
||||||
7529
pulumi/k3s-prod-1/visualization/yaml/kubevirt-operator-v1.1.1.yaml
Normal file
7529
pulumi/k3s-prod-1/visualization/yaml/kubevirt-operator-v1.1.1.yaml
Normal file
File diff suppressed because it is too large
Load Diff
@@ -117,6 +117,43 @@ in {
|
|||||||
};
|
};
|
||||||
homelab_tailscale_gw_tags = ["tailscale-gw" "homelab-network"];
|
homelab_tailscale_gw_tags = ["tailscale-gw" "homelab-network"];
|
||||||
|
|
||||||
|
# --- Kubevirt Nodes --- #
|
||||||
|
kubevirt_shoryu_modules = {
|
||||||
|
nixos-modules =
|
||||||
|
[
|
||||||
|
../hosts/k8s/kubevirt_shoryu
|
||||||
|
]
|
||||||
|
++ kube_base_modules.nixos-modules;
|
||||||
|
home-module.imports = [
|
||||||
|
../home/linux/server.nix
|
||||||
|
];
|
||||||
|
};
|
||||||
|
kubevirt_shoryu_tags = ["virt-shoryu"];
|
||||||
|
|
||||||
|
kubevirt_shushou_modules = {
|
||||||
|
nixos-modules =
|
||||||
|
[
|
||||||
|
../hosts/k8s/kubevirt_shushou
|
||||||
|
]
|
||||||
|
++ kube_base_modules.nixos-modules;
|
||||||
|
home-module.imports = [
|
||||||
|
../home/linux/server.nix
|
||||||
|
];
|
||||||
|
};
|
||||||
|
kubevirt_shushou_tags = ["virt-shushou"];
|
||||||
|
|
||||||
|
kubevirt_youko_modules = {
|
||||||
|
nixos-modules =
|
||||||
|
[
|
||||||
|
../hosts/k8s/kubevirt_youko
|
||||||
|
]
|
||||||
|
++ kube_base_modules.nixos-modules;
|
||||||
|
home-module.imports = [
|
||||||
|
../home/linux/server.nix
|
||||||
|
];
|
||||||
|
};
|
||||||
|
kubevirt_youko_tags = ["virt-youko"];
|
||||||
|
|
||||||
# --- Kubernetes Nodes --- #
|
# --- Kubernetes Nodes --- #
|
||||||
|
|
||||||
k3s_prod_1_master_1_modules = {
|
k3s_prod_1_master_1_modules = {
|
||||||
@@ -129,7 +166,7 @@ in {
|
|||||||
../home/linux/server.nix
|
../home/linux/server.nix
|
||||||
];
|
];
|
||||||
};
|
};
|
||||||
k3s_prod_1_master_1_tags = ["k8s-prod-master"];
|
k3s_prod_1_master_1_tags = ["k8s-prod-master-1"];
|
||||||
|
|
||||||
k3s_prod_1_master_2_modules = {
|
k3s_prod_1_master_2_modules = {
|
||||||
nixos-modules =
|
nixos-modules =
|
||||||
@@ -138,7 +175,7 @@ in {
|
|||||||
]
|
]
|
||||||
++ kube_base_modules.nixos-modules;
|
++ kube_base_modules.nixos-modules;
|
||||||
};
|
};
|
||||||
k3s_prod_1_master_2_tags = ["k8s-prod-master"];
|
k3s_prod_1_master_2_tags = ["k8s-prod-master-2"];
|
||||||
|
|
||||||
k3s_prod_1_master_3_modules = {
|
k3s_prod_1_master_3_modules = {
|
||||||
nixos-modules =
|
nixos-modules =
|
||||||
@@ -147,7 +184,7 @@ in {
|
|||||||
]
|
]
|
||||||
++ kube_base_modules.nixos-modules;
|
++ kube_base_modules.nixos-modules;
|
||||||
};
|
};
|
||||||
k3s_prod_1_master_3_tags = ["k8s-prod-master"];
|
k3s_prod_1_master_3_tags = ["k8s-prod-master-3"];
|
||||||
|
|
||||||
k3s_prod_1_worker_1_modules = {
|
k3s_prod_1_worker_1_modules = {
|
||||||
nixos-modules =
|
nixos-modules =
|
||||||
@@ -156,7 +193,7 @@ in {
|
|||||||
]
|
]
|
||||||
++ kube_base_modules.nixos-modules;
|
++ kube_base_modules.nixos-modules;
|
||||||
};
|
};
|
||||||
k3s_prod_1_worker_1_tags = ["k8s-prod-worker"];
|
k3s_prod_1_worker_1_tags = ["k8s-prod-worker-1"];
|
||||||
|
|
||||||
k3s_prod_1_worker_2_modules = {
|
k3s_prod_1_worker_2_modules = {
|
||||||
nixos-modules =
|
nixos-modules =
|
||||||
@@ -165,7 +202,7 @@ in {
|
|||||||
]
|
]
|
||||||
++ kube_base_modules.nixos-modules;
|
++ kube_base_modules.nixos-modules;
|
||||||
};
|
};
|
||||||
k3s_prod_1_worker_2_tags = ["k8s-prod-worker"];
|
k3s_prod_1_worker_2_tags = ["k8s-prod-worker-2"];
|
||||||
|
|
||||||
k3s_prod_1_worker_3_modules = {
|
k3s_prod_1_worker_3_modules = {
|
||||||
nixos-modules =
|
nixos-modules =
|
||||||
@@ -174,7 +211,7 @@ in {
|
|||||||
]
|
]
|
||||||
++ kube_base_modules.nixos-modules;
|
++ kube_base_modules.nixos-modules;
|
||||||
};
|
};
|
||||||
k3s_prod_1_worker_3_tags = ["k8s-prod-worker"];
|
k3s_prod_1_worker_3_tags = ["k8s-prod-worker-3"];
|
||||||
|
|
||||||
# --- RISC-V / AARCH64 Systems --- #
|
# --- RISC-V / AARCH64 Systems --- #
|
||||||
|
|
||||||
|
|||||||
@@ -1,6 +1,6 @@
|
|||||||
{lib, ...}: rec {
|
{lib, ...}: rec {
|
||||||
mainGateway = "192.168.5.1"; # main router
|
mainGateway = "192.168.5.1"; # main router
|
||||||
defaultGateway = "192.168.5.101"; # subrouter with a transparent proxy
|
defaultGateway = "192.168.5.101"; # subrouter with a transparent proxy
|
||||||
nameservers = [
|
nameservers = [
|
||||||
"119.29.29.29" # DNSPod
|
"119.29.29.29" # DNSPod
|
||||||
"223.5.5.5" # AliDNS
|
"223.5.5.5" # AliDNS
|
||||||
@@ -25,6 +25,9 @@
|
|||||||
"k3s-prod-1-worker-1" = "192.168.5.111";
|
"k3s-prod-1-worker-1" = "192.168.5.111";
|
||||||
"k3s-prod-1-worker-2" = "192.168.5.112";
|
"k3s-prod-1-worker-2" = "192.168.5.112";
|
||||||
"k3s-prod-1-worker-3" = "192.168.5.113";
|
"k3s-prod-1-worker-3" = "192.168.5.113";
|
||||||
|
"kubevirt-shoryu" = "192.168.5.176";
|
||||||
|
"kubevirt-shushou" = "192.168.5.177";
|
||||||
|
"kubevirt-youko" = "192.168.5.178";
|
||||||
"tailscale-gw" = "192.168.5.192";
|
"tailscale-gw" = "192.168.5.192";
|
||||||
};
|
};
|
||||||
|
|
||||||
|
|||||||
Reference in New Issue
Block a user