From 8d17a07c9be5cd9d400644c34ea50177535d15f6 Mon Sep 17 00:00:00 2001 From: Alex Auvolat Date: Sat, 24 Dec 2022 22:59:37 +0100 Subject: reorganize some things --- experimental/app/csi-s3/deploy/csi-s3.hcl | 39 ++++ experimental/app/csi-s3/deploy/dummy-volume.hcl | 20 ++ experimental/app/nextcloud/config/litestream.yml | 10 + experimental/app/nextcloud/deploy/nextcloud.hcl | 137 ++++++++++++++ .../app/nextcloud/secrets/nextcloud/admin_pass | 1 + .../app/nextcloud/secrets/nextcloud/admin_user | 1 + .../app/nextcloud/secrets/nextcloud/s3_access_key | 1 + .../app/nextcloud/secrets/nextcloud/s3_secret_key | 1 + experimental/app/ssb/deploy/go-ssb-room.hcl | 83 +++++++++ experimental/app/ssb/deploy/ssb-room.hcl | 59 ++++++ .../app/telemetry-elastic/config/apm-config.yaml | 20 ++ .../app/telemetry-elastic/config/filebeat.yml | 46 +++++ .../config/grafana-litestream.yml | 10 + .../grafana/provisioning/datasources/elastic.yaml | 21 +++ .../app/telemetry-elastic/config/otel-config.yaml | 56 ++++++ .../telemetry-elastic/deploy/telemetry-system.hcl | 182 ++++++++++++++++++ .../app/telemetry-elastic/deploy/telemetry.hcl | 181 ++++++++++++++++++ experimental/app/yugabyte/deploy/yugabyte.hcl | 204 +++++++++++++++++++++ experimental/bad.csi-s3/deploy/csi-s3.hcl | 39 ---- experimental/bad.csi-s3/deploy/dummy-volume.hcl | 20 -- experimental/bad.nextcloud/config/litestream.yml | 10 - experimental/bad.nextcloud/deploy/nextcloud.hcl | 137 -------------- .../bad.nextcloud/secrets/nextcloud/admin_pass | 1 - .../bad.nextcloud/secrets/nextcloud/admin_user | 1 - .../bad.nextcloud/secrets/nextcloud/s3_access_key | 1 - .../bad.nextcloud/secrets/nextcloud/s3_secret_key | 1 - experimental/bad.ssb/deploy/go-ssb-room.hcl | 83 --------- experimental/bad.ssb/deploy/ssb-room.hcl | 59 ------ .../bad.telemetry-elastic/config/apm-config.yaml | 20 -- .../bad.telemetry-elastic/config/filebeat.yml | 46 ----- .../config/grafana-litestream.yml | 10 - .../grafana/provisioning/datasources/elastic.yaml | 21 --- .../bad.telemetry-elastic/config/otel-config.yaml | 56 ------ .../deploy/telemetry-system.hcl | 182 ------------------ .../bad.telemetry-elastic/deploy/telemetry.hcl | 181 ------------------ experimental/bad.yugabyte/deploy/yugabyte.hcl | 204 --------------------- .../luks-fde/example-hardware-configuration.nix | 38 ++++ experimental/luks-fde/nixos-install-luks.md | 182 ++++++++++++++++++ 38 files changed, 1292 insertions(+), 1072 deletions(-) create mode 100644 experimental/app/csi-s3/deploy/csi-s3.hcl create mode 100644 experimental/app/csi-s3/deploy/dummy-volume.hcl create mode 100644 experimental/app/nextcloud/config/litestream.yml create mode 100644 experimental/app/nextcloud/deploy/nextcloud.hcl create mode 100644 experimental/app/nextcloud/secrets/nextcloud/admin_pass create mode 100644 experimental/app/nextcloud/secrets/nextcloud/admin_user create mode 100644 experimental/app/nextcloud/secrets/nextcloud/s3_access_key create mode 100644 experimental/app/nextcloud/secrets/nextcloud/s3_secret_key create mode 100644 experimental/app/ssb/deploy/go-ssb-room.hcl create mode 100644 experimental/app/ssb/deploy/ssb-room.hcl create mode 100644 experimental/app/telemetry-elastic/config/apm-config.yaml create mode 100644 experimental/app/telemetry-elastic/config/filebeat.yml create mode 100644 experimental/app/telemetry-elastic/config/grafana-litestream.yml create mode 100644 experimental/app/telemetry-elastic/config/grafana/provisioning/datasources/elastic.yaml create mode 100644 experimental/app/telemetry-elastic/config/otel-config.yaml create mode 100644 experimental/app/telemetry-elastic/deploy/telemetry-system.hcl create mode 100644 experimental/app/telemetry-elastic/deploy/telemetry.hcl create mode 100644 experimental/app/yugabyte/deploy/yugabyte.hcl delete mode 100644 experimental/bad.csi-s3/deploy/csi-s3.hcl delete mode 100644 experimental/bad.csi-s3/deploy/dummy-volume.hcl delete mode 100644 experimental/bad.nextcloud/config/litestream.yml delete mode 100644 experimental/bad.nextcloud/deploy/nextcloud.hcl delete mode 100644 experimental/bad.nextcloud/secrets/nextcloud/admin_pass delete mode 100644 experimental/bad.nextcloud/secrets/nextcloud/admin_user delete mode 100644 experimental/bad.nextcloud/secrets/nextcloud/s3_access_key delete mode 100644 experimental/bad.nextcloud/secrets/nextcloud/s3_secret_key delete mode 100644 experimental/bad.ssb/deploy/go-ssb-room.hcl delete mode 100644 experimental/bad.ssb/deploy/ssb-room.hcl delete mode 100644 experimental/bad.telemetry-elastic/config/apm-config.yaml delete mode 100644 experimental/bad.telemetry-elastic/config/filebeat.yml delete mode 100644 experimental/bad.telemetry-elastic/config/grafana-litestream.yml delete mode 100644 experimental/bad.telemetry-elastic/config/grafana/provisioning/datasources/elastic.yaml delete mode 100644 experimental/bad.telemetry-elastic/config/otel-config.yaml delete mode 100644 experimental/bad.telemetry-elastic/deploy/telemetry-system.hcl delete mode 100644 experimental/bad.telemetry-elastic/deploy/telemetry.hcl delete mode 100644 experimental/bad.yugabyte/deploy/yugabyte.hcl create mode 100644 experimental/luks-fde/example-hardware-configuration.nix create mode 100644 experimental/luks-fde/nixos-install-luks.md (limited to 'experimental') diff --git a/experimental/app/csi-s3/deploy/csi-s3.hcl b/experimental/app/csi-s3/deploy/csi-s3.hcl new file mode 100644 index 0000000..8e70c6a --- /dev/null +++ b/experimental/app/csi-s3/deploy/csi-s3.hcl @@ -0,0 +1,39 @@ +job "plugin-csi-s3-nodes" { + datacenters = ["neptune", "pluton"] + + # you can run node plugins as service jobs as well, but this ensures + # that all nodes in the DC have a copy. + type = "system" + + group "nodes" { + task "plugin" { + driver = "docker" + + config { + image = "ctrox/csi-s3:v1.2.0-rc.1" + + args = [ + "--endpoint=unix://csi/csi.sock", + "--nodeid=${node.unique.id}", + "--logtostderr", + "--v=5", + ] + + # node plugins must run as privileged jobs because they + # mount disks to the host + privileged = true + } + + csi_plugin { + id = "csi-s3" + type = "node" + mount_dir = "/csi" + } + + resources { + cpu = 500 + memory = 256 + } + } + } +} diff --git a/experimental/app/csi-s3/deploy/dummy-volume.hcl b/experimental/app/csi-s3/deploy/dummy-volume.hcl new file mode 100644 index 0000000..67dfd39 --- /dev/null +++ b/experimental/app/csi-s3/deploy/dummy-volume.hcl @@ -0,0 +1,20 @@ +id = "dummy-volume" +name = "dummy-volume" +type = "csi" +plugin_id = "csi-s3" + +capability { + access_mode = "single-node-writer" + attachment_mode = "file-system" +} + +secrets { + accessKeyId = "GKfd94f06139bb73de5642baf5" + secretAccessKey = "a4fa6c956d847b145a823c4615e4655126c67babf3cce2337b4d73cd381d7f06" + endpoint = "https://garage-staging.home.adnab.me" + region = "garage-staging" +} + +parameters { + mounter = "rclone" +} diff --git a/experimental/app/nextcloud/config/litestream.yml b/experimental/app/nextcloud/config/litestream.yml new file mode 100644 index 0000000..46eca93 --- /dev/null +++ b/experimental/app/nextcloud/config/litestream.yml @@ -0,0 +1,10 @@ +dbs: + - path: /ephemeral/nextcloud.db + replicas: + - url: s3://nextcloud-db/nextcloud.db + region: garage-staging + endpoint: http://{{ env "attr.unique.network.ip-address" }}:3990 + access-key-id: {{ key "secrets/nextcloud/s3_access_key" | trimSpace }} + secret-access-key: {{ key "secrets/nextcloud/s3_secret_key" | trimSpace }} + force-path-style: true + sync-interval: 60s diff --git a/experimental/app/nextcloud/deploy/nextcloud.hcl b/experimental/app/nextcloud/deploy/nextcloud.hcl new file mode 100644 index 0000000..45d1b6e --- /dev/null +++ b/experimental/app/nextcloud/deploy/nextcloud.hcl @@ -0,0 +1,137 @@ +job "nextcloud" { + datacenters = ["neptune"] + type = "service" + + group "nextcloud" { + count = 1 + + network { + port "http" { + to = 80 + } + } + + ephemeral_disk { + size = 10000 + } + + restart { + attempts = 10 + delay = "30s" + } + + task "restore-db" { + lifecycle { + hook = "prestart" + sidecar = false + } + + driver = "docker" + config { + image = "litestream/litestream:0.3.7" + args = [ + "restore", "-config", "/etc/litestream.yml", "/ephemeral/nextcloud.db" + ] + volumes = [ + "../alloc/data:/ephemeral", + "secrets/litestream.yml:/etc/litestream.yml" + ] + } + user = "33" + + template { + data = file("../config/litestream.yml") + destination = "secrets/litestream.yml" + } + + resources { + memory = 200 + cpu = 1000 + } + } + + task "nextcloud" { + driver = "docker" + config { + image = "nextcloud:22.2.3-apache" + ports = [ "http" ] + #entrypoint = [ "/bin/sh", "-c" ] + #command = "apache2-foreground" + + volumes = [ + "../alloc/data:/var/www/html/data", + ] + } + user = "33" + + template { + data = < +``` + +If necessary, assign static IP. E.g. `ip addr add 192.168.1.40/24 dev eno1` or sth (replace ip and device appropriately) + +Remotely: `ssh-copy-id @`. Check SSH access is good. + +## Deploy from this repo + +See the documentation in `/doc` in this repo. The old procedure described here is partly obsolete. + +## Old guide + +It's time! + +**Files in this repo to create/change:** + +- create node `.nix` file and symlink for node `.site.nix` (create site and + cluster `.nix` files if necessary; use existing files of e.g. the staging + cluster as examples/templates) +- make sure values are filled in correctly +- add node to `ssh_config` with it's LAN IP, we don't have VPN at this stage + +**Configuration steps on the node:** + +```bash +# On node being installed +mkdir -p /var/lib/deuxfleurs/remote-unlock +cd /var/lib/deuxfleurs/remote-unlock +ssh-keygen -t ed25519 -N "" -f ./ssh_host_ed25519_key +``` + +**Try to deploy:** + +```bash +# In nixcfg repository from your PC +./deploy.sh +``` + +Reboot. + +Check remote unlocking works: `ssh -p 222 root@` + +### Configure wireguard + +```bash +# On node being installed +mkdir -p /var/lib/deuxfleurs/wireguard-keys +cd /var/lib/deuxfleurs/wireguard-keys +wg genkey | tee private | wg pubkey > public +``` + +Get the public key, make sure it is in `cluster.nix` so that nodes know one +another. Also put it anywhere else like in your local wireguard config for +instance so that you can access the node from your PC by its wireguard address +and not only its LAN address. + +Redo a deploy (`./deploy.sh `) + +Check VPN works. Change IP in `ssh_config` to use VPN IP instead of LAN IP (required for deploy when away from home). + +### Commit changes to `nixcfg` repo + +This is a good point to commit your new/modified `.nix` files. + +### Configure Nomad and Consul TLS + +If you are bootstraping a new cluster, you need to `./genpki.sh ` to +make a TLS PKI for the Nomad+Consul cluster to work. Then redo a deploy. -- cgit v1.2.3