made disko/impermanence config into a module
This commit is contained in:
		
							parent
							
								
									c28731a1a7
								
							
						
					
					
						commit
						30ad3c91b9
					
				
					 10 changed files with 571 additions and 466 deletions
				
			
		| 
						 | 
				
			
			@ -1,11 +1,6 @@
 | 
			
		|||
# server nas
 | 
			
		||||
{
 | 
			
		||||
  inputs,
 | 
			
		||||
  pkgs,
 | 
			
		||||
  ...
 | 
			
		||||
}: {
 | 
			
		||||
{pkgs, ...}: {
 | 
			
		||||
  imports = [
 | 
			
		||||
    inputs.disko.nixosModules.disko
 | 
			
		||||
    # ./services.nix
 | 
			
		||||
  ];
 | 
			
		||||
 | 
			
		||||
| 
						 | 
				
			
			@ -21,6 +16,33 @@
 | 
			
		|||
      ester.isNormalUser = false;
 | 
			
		||||
      eve.isNormalUser = false;
 | 
			
		||||
    };
 | 
			
		||||
    impermanence.enable = true;
 | 
			
		||||
    storage = {
 | 
			
		||||
      enable = true;
 | 
			
		||||
      encryption = true;
 | 
			
		||||
      pool = {
 | 
			
		||||
        drives = [
 | 
			
		||||
          "ata-ST18000NE000-3G6101_ZVTCXVEB"
 | 
			
		||||
          "ata-ST18000NE000-3G6101_ZVTCXWSC"
 | 
			
		||||
          "ata-ST18000NE000-3G6101_ZVTD10EH"
 | 
			
		||||
          "ata-ST18000NT001-3NF101_ZVTE0S3Q"
 | 
			
		||||
          "ata-ST18000NT001-3NF101_ZVTEF27J"
 | 
			
		||||
          "ata-ST18000NT001-3NF101_ZVTEZACV"
 | 
			
		||||
        ];
 | 
			
		||||
        cache = [
 | 
			
		||||
          "nvme-Samsung_SSD_990_PRO_4TB_S7KGNU0X907881F"
 | 
			
		||||
        ];
 | 
			
		||||
        # extraDatasets = {
 | 
			
		||||
        #   "persist/system/var/lib/jellyfin/media" = {
 | 
			
		||||
        #     type = "zfs_fs";
 | 
			
		||||
        #     mountpoint = "/persist/system/var/lib/jellyfin/media";
 | 
			
		||||
        #   };
 | 
			
		||||
        # };
 | 
			
		||||
      };
 | 
			
		||||
    };
 | 
			
		||||
  };
 | 
			
		||||
  networking = {
 | 
			
		||||
    hostId = "c51763d6";
 | 
			
		||||
  };
 | 
			
		||||
 | 
			
		||||
  # apps = {
 | 
			
		||||
| 
						 | 
				
			
			
 | 
			
		|||
| 
						 | 
				
			
			@ -1,9 +1,7 @@
 | 
			
		|||
# server nas
 | 
			
		||||
{...}: {
 | 
			
		||||
  imports = [
 | 
			
		||||
    ./disko-config.nix
 | 
			
		||||
    ./hardware-configuration.nix
 | 
			
		||||
    ./impermanence.nix
 | 
			
		||||
    ./configuration.nix
 | 
			
		||||
  ];
 | 
			
		||||
}
 | 
			
		||||
| 
						 | 
				
			
			
 | 
			
		|||
| 
						 | 
				
			
			@ -1,181 +0,0 @@
 | 
			
		|||
{lib, ...}: let
 | 
			
		||||
  zfsDisk = devicePath: {
 | 
			
		||||
    type = "disk";
 | 
			
		||||
    device = devicePath;
 | 
			
		||||
    content = {
 | 
			
		||||
      type = "gpt";
 | 
			
		||||
      partitions = {
 | 
			
		||||
        zfs = {
 | 
			
		||||
          size = "100%";
 | 
			
		||||
          content = {
 | 
			
		||||
            type = "zfs";
 | 
			
		||||
            pool = "rpool";
 | 
			
		||||
          };
 | 
			
		||||
        };
 | 
			
		||||
      };
 | 
			
		||||
    };
 | 
			
		||||
  };
 | 
			
		||||
  cacheDisk = devicePath: {
 | 
			
		||||
    type = "disk";
 | 
			
		||||
    device = devicePath;
 | 
			
		||||
    content = {
 | 
			
		||||
      type = "gpt";
 | 
			
		||||
      partitions = {
 | 
			
		||||
        # We are having to boot off of the nvm cache drive because I cant figure out how to boot via the HBA
 | 
			
		||||
        ESP = {
 | 
			
		||||
          size = "64M";
 | 
			
		||||
          type = "EF00";
 | 
			
		||||
          content = {
 | 
			
		||||
            type = "filesystem";
 | 
			
		||||
            format = "vfat";
 | 
			
		||||
            mountpoint = "/boot";
 | 
			
		||||
            mountOptions = ["umask=0077"];
 | 
			
		||||
          };
 | 
			
		||||
        };
 | 
			
		||||
        zfs = {
 | 
			
		||||
          size = "100%";
 | 
			
		||||
          content = {
 | 
			
		||||
            type = "zfs";
 | 
			
		||||
            pool = "rpool";
 | 
			
		||||
          };
 | 
			
		||||
        };
 | 
			
		||||
      };
 | 
			
		||||
    };
 | 
			
		||||
  };
 | 
			
		||||
in {
 | 
			
		||||
  disko.devices = {
 | 
			
		||||
    disk = {
 | 
			
		||||
      hd_18_tb_a = zfsDisk "/dev/disk/by-id/ata-ST18000NE000-3G6101_ZVTCXVEB";
 | 
			
		||||
      hd_18_tb_b = zfsDisk "/dev/disk/by-id/ata-ST18000NE000-3G6101_ZVTCXWSC";
 | 
			
		||||
      hd_18_tb_c = zfsDisk "/dev/disk/by-id/ata-ST18000NE000-3G6101_ZVTD10EH";
 | 
			
		||||
      hd_18_tb_d = zfsDisk "/dev/disk/by-id/ata-ST18000NT001-3NF101_ZVTE0S3Q";
 | 
			
		||||
      hd_18_tb_e = zfsDisk "/dev/disk/by-id/ata-ST18000NT001-3NF101_ZVTEF27J";
 | 
			
		||||
      hd_18_tb_f = zfsDisk "/dev/disk/by-id/ata-ST18000NT001-3NF101_ZVTEZACV";
 | 
			
		||||
 | 
			
		||||
      ssd_4_tb_a = cacheDisk "/dev/disk/by-id/nvme-Samsung_SSD_990_PRO_4TB_S7KGNU0X907881F";
 | 
			
		||||
    };
 | 
			
		||||
    zpool = {
 | 
			
		||||
      rpool = {
 | 
			
		||||
        type = "zpool";
 | 
			
		||||
        mode = {
 | 
			
		||||
          topology = {
 | 
			
		||||
            type = "topology";
 | 
			
		||||
            vdev = [
 | 
			
		||||
              {
 | 
			
		||||
                mode = "raidz2";
 | 
			
		||||
                members = [
 | 
			
		||||
                  "hd_18_tb_a"
 | 
			
		||||
                  "hd_18_tb_b"
 | 
			
		||||
                  "hd_18_tb_c"
 | 
			
		||||
                  "hd_18_tb_d"
 | 
			
		||||
                  "hd_18_tb_e"
 | 
			
		||||
                  "hd_18_tb_f"
 | 
			
		||||
                ];
 | 
			
		||||
              }
 | 
			
		||||
            ];
 | 
			
		||||
            cache = ["ssd_4_tb_a"];
 | 
			
		||||
          };
 | 
			
		||||
        };
 | 
			
		||||
 | 
			
		||||
        options = {
 | 
			
		||||
          ashift = "12";
 | 
			
		||||
          autotrim = "on";
 | 
			
		||||
        };
 | 
			
		||||
 | 
			
		||||
        rootFsOptions =
 | 
			
		||||
          {
 | 
			
		||||
            canmount = "off";
 | 
			
		||||
            mountpoint = "none";
 | 
			
		||||
 | 
			
		||||
            xattr = "sa";
 | 
			
		||||
            acltype = "posixacl";
 | 
			
		||||
            relatime = "on";
 | 
			
		||||
 | 
			
		||||
            compression = "lz4";
 | 
			
		||||
 | 
			
		||||
            "com.sun:auto-snapshot" = "false";
 | 
			
		||||
          }
 | 
			
		||||
          # TODO: have an option to enable encryption
 | 
			
		||||
          // lib.attrsets.optionalAttrs false {
 | 
			
		||||
            encryption = "on";
 | 
			
		||||
            keyformat = "hex";
 | 
			
		||||
            keylocation = "prompt";
 | 
			
		||||
          };
 | 
			
		||||
 | 
			
		||||
        datasets = {
 | 
			
		||||
          # local datasets are for data that should be considered ephemeral
 | 
			
		||||
          "local" = {
 | 
			
		||||
            type = "zfs_fs";
 | 
			
		||||
            options.canmount = "off";
 | 
			
		||||
          };
 | 
			
		||||
          # the nix directory is local because its all generable from our configuration
 | 
			
		||||
          "local/system/nix" = {
 | 
			
		||||
            type = "zfs_fs";
 | 
			
		||||
            mountpoint = "/nix";
 | 
			
		||||
            options = {
 | 
			
		||||
              atime = "off";
 | 
			
		||||
              relatime = "off";
 | 
			
		||||
              canmount = "on";
 | 
			
		||||
            };
 | 
			
		||||
          };
 | 
			
		||||
          "local/system/sops" = {
 | 
			
		||||
            type = "zfs_fs";
 | 
			
		||||
            mountpoint = import ../../../const/sops_age_key_directory.nix;
 | 
			
		||||
            options = {
 | 
			
		||||
              atime = "off";
 | 
			
		||||
              relatime = "off";
 | 
			
		||||
              canmount = "on";
 | 
			
		||||
            };
 | 
			
		||||
          };
 | 
			
		||||
          "local/system/root" = {
 | 
			
		||||
            type = "zfs_fs";
 | 
			
		||||
            mountpoint = "/";
 | 
			
		||||
            options = {
 | 
			
		||||
              canmount = "on";
 | 
			
		||||
            };
 | 
			
		||||
            postCreateHook = ''
 | 
			
		||||
              zfs snapshot rpool/local/system/root@blank
 | 
			
		||||
            '';
 | 
			
		||||
          };
 | 
			
		||||
          "local/home/leyla" = {
 | 
			
		||||
            type = "zfs_fs";
 | 
			
		||||
            mountpoint = "/home/leyla";
 | 
			
		||||
            options = {
 | 
			
		||||
              canmount = "on";
 | 
			
		||||
            };
 | 
			
		||||
            postCreateHook = ''
 | 
			
		||||
              zfs snapshot rpool/local/home/leyla@blank
 | 
			
		||||
            '';
 | 
			
		||||
          };
 | 
			
		||||
 | 
			
		||||
          # persist datasets are datasets that contain information that we would like to keep around
 | 
			
		||||
          "persist" = {
 | 
			
		||||
            type = "zfs_fs";
 | 
			
		||||
            options.canmount = "off";
 | 
			
		||||
          };
 | 
			
		||||
          "persist/system/root" = {
 | 
			
		||||
            type = "zfs_fs";
 | 
			
		||||
            mountpoint = "/persist/system/root";
 | 
			
		||||
            options = {
 | 
			
		||||
              "com.sun:auto-snapshot" = "true";
 | 
			
		||||
              mountpoint = "/persist/system/root";
 | 
			
		||||
            };
 | 
			
		||||
          };
 | 
			
		||||
          "persist/home/leyla" = {
 | 
			
		||||
            type = "zfs_fs";
 | 
			
		||||
            mountpoint = "/persist/home/leyla";
 | 
			
		||||
            options = {
 | 
			
		||||
              "com.sun:auto-snapshot" = "true";
 | 
			
		||||
              mountpoint = "/persist/home/leyla";
 | 
			
		||||
            };
 | 
			
		||||
          };
 | 
			
		||||
 | 
			
		||||
          # TODO: separate dataset for logs that wont participate in snapshots and rollbacks with the rest of the system
 | 
			
		||||
        };
 | 
			
		||||
      };
 | 
			
		||||
    };
 | 
			
		||||
  };
 | 
			
		||||
  networking = {
 | 
			
		||||
    hostId = "c51763d6";
 | 
			
		||||
  };
 | 
			
		||||
}
 | 
			
		||||
| 
						 | 
				
			
			@ -1,83 +1,4 @@
 | 
			
		|||
{lib, ...}: {
 | 
			
		||||
  boot.initrd.postResumeCommands = lib.mkAfter ''
 | 
			
		||||
    zfs rollback -r rpool/local/system/root@blank
 | 
			
		||||
    zfs rollback -r rpool/local/home/leyla@blank
 | 
			
		||||
  '';
 | 
			
		||||
 | 
			
		||||
  # systemd.services = {
 | 
			
		||||
  #   # https://github.com/openzfs/zfs/issues/10891
 | 
			
		||||
  #   systemd-udev-settle.enable = false;
 | 
			
		||||
  #   # Snapshots are not accessible on boot for some reason this should fix it
 | 
			
		||||
  #   # https://github.com/NixOS/nixpkgs/issues/257505
 | 
			
		||||
  #   zfs-mount = {
 | 
			
		||||
  #     serviceConfig = {
 | 
			
		||||
  #       ExecStart = ["zfs mount -a -o remount"];
 | 
			
		||||
  #       # ExecStart = [
 | 
			
		||||
  #       #   "${lib.getExe' pkgs.util-linux "mount"} -t zfs rpool/local -o remount"
 | 
			
		||||
  #       #   "${lib.getExe' pkgs.util-linux "mount"} -t zfs rpool/persistent -o remount"
 | 
			
		||||
  #       # ];
 | 
			
		||||
  #     };
 | 
			
		||||
  #   };
 | 
			
		||||
  # };
 | 
			
		||||
 | 
			
		||||
  # boot.initrd.systemd.services.rollback = {
 | 
			
		||||
  #   description = "Rollback filesystem to a pristine state on boot";
 | 
			
		||||
  #   wantedBy = [
 | 
			
		||||
  #     "initrd.target"
 | 
			
		||||
  #   ];
 | 
			
		||||
  #   after = [
 | 
			
		||||
  #     "zfs-import-rpool.service"
 | 
			
		||||
  #   ];
 | 
			
		||||
  #   before = [
 | 
			
		||||
  #     "sysroot.mount"
 | 
			
		||||
  #   ];
 | 
			
		||||
  #   requiredBy = [
 | 
			
		||||
  #     "sysroot.mount"
 | 
			
		||||
  #   ];
 | 
			
		||||
  #   serviceConfig = {
 | 
			
		||||
  #     Type = "oneshot";
 | 
			
		||||
  #     ExecStart = ''
 | 
			
		||||
  #       zfs rollback -r rpool/local/system/root@blank
 | 
			
		||||
  #       zfs rollback -r rpool/local/home@blank
 | 
			
		||||
  #     '';
 | 
			
		||||
  #   };
 | 
			
		||||
  # };
 | 
			
		||||
 | 
			
		||||
  fileSystems."/".neededForBoot = true;
 | 
			
		||||
  fileSystems."/home/leyla".neededForBoot = true;
 | 
			
		||||
  fileSystems."/persist/system/root".neededForBoot = true;
 | 
			
		||||
  fileSystems."/persist/home/leyla".neededForBoot = true;
 | 
			
		||||
  fileSystems.${import ../../../const/sops_age_key_directory.nix}.neededForBoot = true;
 | 
			
		||||
 | 
			
		||||
  environment.persistence."/persist/system/root" = {
 | 
			
		||||
    enable = true;
 | 
			
		||||
    hideMounts = true;
 | 
			
		||||
    directories = [
 | 
			
		||||
      "/run/secrets"
 | 
			
		||||
 | 
			
		||||
      "/etc/ssh"
 | 
			
		||||
 | 
			
		||||
      "/var/log"
 | 
			
		||||
      "/var/lib/nixos"
 | 
			
		||||
      "/var/lib/systemd/coredump"
 | 
			
		||||
 | 
			
		||||
      # config.apps.pihole.directory.root
 | 
			
		||||
 | 
			
		||||
      # config.apps.jellyfin.mediaDirectory
 | 
			
		||||
      # config.services.jellyfin.configDir
 | 
			
		||||
      # config.services.jellyfin.cacheDir
 | 
			
		||||
      # config.services.jellyfin.dataDir
 | 
			
		||||
 | 
			
		||||
      # "/var/hass" # config.users.users.hass.home
 | 
			
		||||
      # "/var/postgresql" # config.users.users.postgresql.home
 | 
			
		||||
      # "/var/forgejo" # config.users.users.forgejo.home
 | 
			
		||||
      # "/var/nextcloud" # config.users.users.nextcloud.home
 | 
			
		||||
      # "/var/headscale" # config.users.users.headscale.home
 | 
			
		||||
    ];
 | 
			
		||||
    files = [
 | 
			
		||||
      "/etc/machine-id"
 | 
			
		||||
    ];
 | 
			
		||||
  };
 | 
			
		||||
 | 
			
		||||
  security.sudo.extraConfig = "Defaults lecture=never";
 | 
			
		||||
{...}: {
 | 
			
		||||
  # fileSystems."/home/leyla".neededForBoot = true;
 | 
			
		||||
  # fileSystems."/persist/home/leyla".neededForBoot = true;
 | 
			
		||||
}
 | 
			
		||||
| 
						 | 
				
			
			
 | 
			
		|||
| 
						 | 
				
			
			@ -8,5 +8,7 @@
 | 
			
		|||
    ./desktop.nix
 | 
			
		||||
    ./ssh.nix
 | 
			
		||||
    ./i18n.nix
 | 
			
		||||
    ./impermanence.nix
 | 
			
		||||
    ./disko.nix
 | 
			
		||||
  ];
 | 
			
		||||
}
 | 
			
		||||
| 
						 | 
				
			
			
 | 
			
		|||
							
								
								
									
										168
									
								
								modules/nixos-modules/disko.nix
									
										
									
									
									
										Normal file
									
								
							
							
						
						
									
										168
									
								
								modules/nixos-modules/disko.nix
									
										
									
									
									
										Normal file
									
								
							| 
						 | 
				
			
			@ -0,0 +1,168 @@
 | 
			
		|||
{
 | 
			
		||||
  lib,
 | 
			
		||||
  config,
 | 
			
		||||
  inputs,
 | 
			
		||||
  ...
 | 
			
		||||
}: let
 | 
			
		||||
  # there currently is a bug with disko that causes long disk names to be generated improperly this hash function should alleviate it when used for disk names instead of what we are defaulting to
 | 
			
		||||
  # max gpt length is 36 and disk adds formats it like disk-xxxx-zfs which means we need to be 9 characters under that
 | 
			
		||||
  hashDisk = drive: (builtins.substring 0 27 (builtins.hashString "sha256" drive));
 | 
			
		||||
 | 
			
		||||
  vdevs =
 | 
			
		||||
    builtins.map (
 | 
			
		||||
      disks:
 | 
			
		||||
        builtins.map (disk: lib.attrsets.nameValuePair (hashDisk disk) disk) disks
 | 
			
		||||
    )
 | 
			
		||||
    config.host.storage.pool.vdevs;
 | 
			
		||||
  cache =
 | 
			
		||||
    builtins.map (
 | 
			
		||||
      disk: lib.attrsets.nameValuePair (hashDisk disk) disk
 | 
			
		||||
    )
 | 
			
		||||
    config.host.storage.pool.cache;
 | 
			
		||||
in {
 | 
			
		||||
  options.host.storage = {
 | 
			
		||||
    enable = lib.mkEnableOption "are we going create zfs disks with disko on this device";
 | 
			
		||||
    encryption = lib.mkEnableOption "is the vdev going to be encrypted";
 | 
			
		||||
    pool = {
 | 
			
		||||
      vdevs = lib.mkOption {
 | 
			
		||||
        type = lib.types.listOf (lib.types.listOf lib.types.str);
 | 
			
		||||
        description = "list of disks that are going to be in";
 | 
			
		||||
        default = [config.host.storage.pool.drives];
 | 
			
		||||
      };
 | 
			
		||||
      drives = lib.mkOption {
 | 
			
		||||
        type = lib.types.listOf lib.types.str;
 | 
			
		||||
        description = "list of drives that are going to be in the vdev";
 | 
			
		||||
        default = [];
 | 
			
		||||
      };
 | 
			
		||||
      cache = lib.mkOption {
 | 
			
		||||
        type = lib.types.listOf lib.types.str;
 | 
			
		||||
        description = "list of drives that are going to be used as cache";
 | 
			
		||||
        default = [];
 | 
			
		||||
      };
 | 
			
		||||
      extraDatasets = lib.mkOption {
 | 
			
		||||
        type = lib.types.attrsOf (inputs.disko.lib.subType {
 | 
			
		||||
          types = {inherit (inputs.disko.lib.types) zfs_fs zfs_volume;};
 | 
			
		||||
        });
 | 
			
		||||
        description = "List of datasets to define";
 | 
			
		||||
        default = {};
 | 
			
		||||
      };
 | 
			
		||||
    };
 | 
			
		||||
  };
 | 
			
		||||
 | 
			
		||||
  config = lib.mkIf config.host.storage.enable {
 | 
			
		||||
    disko.devices = {
 | 
			
		||||
      disk = (
 | 
			
		||||
        builtins.listToAttrs (
 | 
			
		||||
          (
 | 
			
		||||
            builtins.map
 | 
			
		||||
            (drive:
 | 
			
		||||
              lib.attrsets.nameValuePair (drive.name) {
 | 
			
		||||
                type = "disk";
 | 
			
		||||
                device = "/dev/disk/by-id/${drive.value}";
 | 
			
		||||
                content = {
 | 
			
		||||
                  type = "gpt";
 | 
			
		||||
                  partitions = {
 | 
			
		||||
                    zfs = {
 | 
			
		||||
                      size = "100%";
 | 
			
		||||
                      content = {
 | 
			
		||||
                        type = "zfs";
 | 
			
		||||
                        pool = "rpool";
 | 
			
		||||
                      };
 | 
			
		||||
                    };
 | 
			
		||||
                  };
 | 
			
		||||
                };
 | 
			
		||||
              })
 | 
			
		||||
            (lib.lists.flatten vdevs)
 | 
			
		||||
          )
 | 
			
		||||
          ++ (
 | 
			
		||||
            builtins.map
 | 
			
		||||
            (drive:
 | 
			
		||||
              lib.attrsets.nameValuePair (drive.name) {
 | 
			
		||||
                type = "disk";
 | 
			
		||||
                device = "/dev/disk/by-id/${drive.value}";
 | 
			
		||||
                content = {
 | 
			
		||||
                  type = "gpt";
 | 
			
		||||
                  partitions = {
 | 
			
		||||
                    # We are having to boot off of the nvm cache drive because I cant figure out how to boot via the HBA
 | 
			
		||||
                    ESP = {
 | 
			
		||||
                      size = "64M";
 | 
			
		||||
                      type = "EF00";
 | 
			
		||||
                      content = {
 | 
			
		||||
                        type = "filesystem";
 | 
			
		||||
                        format = "vfat";
 | 
			
		||||
                        mountpoint = "/boot";
 | 
			
		||||
                        mountOptions = ["umask=0077"];
 | 
			
		||||
                      };
 | 
			
		||||
                    };
 | 
			
		||||
                    zfs = {
 | 
			
		||||
                      size = "100%";
 | 
			
		||||
                      content = {
 | 
			
		||||
                        type = "zfs";
 | 
			
		||||
                        pool = "rpool";
 | 
			
		||||
                      };
 | 
			
		||||
                    };
 | 
			
		||||
                  };
 | 
			
		||||
                };
 | 
			
		||||
              })
 | 
			
		||||
            cache
 | 
			
		||||
          )
 | 
			
		||||
        )
 | 
			
		||||
      );
 | 
			
		||||
      zpool = {
 | 
			
		||||
        rpool = {
 | 
			
		||||
          type = "zpool";
 | 
			
		||||
          mode = {
 | 
			
		||||
            topology = {
 | 
			
		||||
              type = "topology";
 | 
			
		||||
              vdev = (
 | 
			
		||||
                builtins.map (disks: {
 | 
			
		||||
                  mode = "raidz2";
 | 
			
		||||
                  members =
 | 
			
		||||
                    builtins.map (disk: disk.name) disks;
 | 
			
		||||
                })
 | 
			
		||||
                vdevs
 | 
			
		||||
              );
 | 
			
		||||
              cache = builtins.map (disk: disk.name) cache;
 | 
			
		||||
            };
 | 
			
		||||
          };
 | 
			
		||||
 | 
			
		||||
          options = {
 | 
			
		||||
            ashift = "12";
 | 
			
		||||
            autotrim = "on";
 | 
			
		||||
          };
 | 
			
		||||
 | 
			
		||||
          rootFsOptions =
 | 
			
		||||
            {
 | 
			
		||||
              canmount = "off";
 | 
			
		||||
              mountpoint = "none";
 | 
			
		||||
 | 
			
		||||
              xattr = "sa";
 | 
			
		||||
              acltype = "posixacl";
 | 
			
		||||
              relatime = "on";
 | 
			
		||||
 | 
			
		||||
              compression = "lz4";
 | 
			
		||||
 | 
			
		||||
              "com.sun:auto-snapshot" = "false";
 | 
			
		||||
            }
 | 
			
		||||
            // (
 | 
			
		||||
              lib.attrsets.optionalAttrs config.host.storage.encryption {
 | 
			
		||||
                encryption = "on";
 | 
			
		||||
                keyformat = "hex";
 | 
			
		||||
                keylocation = "prompt";
 | 
			
		||||
              }
 | 
			
		||||
            );
 | 
			
		||||
 | 
			
		||||
          datasets = lib.mkMerge [
 | 
			
		||||
            (lib.attrsets.mapAttrs (name: value: {
 | 
			
		||||
                type = value.type;
 | 
			
		||||
                options = value.options;
 | 
			
		||||
                mountpoint = value.mountpoint;
 | 
			
		||||
                postCreateHook = value.postCreateHook;
 | 
			
		||||
              })
 | 
			
		||||
              config.host.storage.pool.extraDatasets)
 | 
			
		||||
          ];
 | 
			
		||||
        };
 | 
			
		||||
      };
 | 
			
		||||
    };
 | 
			
		||||
  };
 | 
			
		||||
}
 | 
			
		||||
							
								
								
									
										110
									
								
								modules/nixos-modules/impermanence.nix
									
										
									
									
									
										Normal file
									
								
							
							
						
						
									
										110
									
								
								modules/nixos-modules/impermanence.nix
									
										
									
									
									
										Normal file
									
								
							| 
						 | 
				
			
			@ -0,0 +1,110 @@
 | 
			
		|||
{
 | 
			
		||||
  config,
 | 
			
		||||
  lib,
 | 
			
		||||
  ...
 | 
			
		||||
}: {
 | 
			
		||||
  options.host.impermanence.enable = lib.mkEnableOption "are we going to use impermanence on this device";
 | 
			
		||||
 | 
			
		||||
  # TODO: validate that config.host.storage.enable is enabled
 | 
			
		||||
  config = lib.mkMerge [
 | 
			
		||||
    {
 | 
			
		||||
      assertions = [
 | 
			
		||||
        {
 | 
			
		||||
          assertion = !(config.host.impermanence.enable && !config.host.storage.enable);
 | 
			
		||||
          message = ''
 | 
			
		||||
            Disko storage must be enabled to use impermanence.
 | 
			
		||||
          '';
 | 
			
		||||
        }
 | 
			
		||||
      ];
 | 
			
		||||
    }
 | 
			
		||||
    (
 | 
			
		||||
      lib.mkIf config.host.impermanence.enable {
 | 
			
		||||
        boot.initrd.postResumeCommands = lib.mkAfter ''
 | 
			
		||||
                    zfs rollback -r rpool/local/system/root@blank
 | 
			
		||||
          1        '';
 | 
			
		||||
 | 
			
		||||
        fileSystems = {
 | 
			
		||||
          "/".neededForBoot = true;
 | 
			
		||||
          "/persist/system/root".neededForBoot = true;
 | 
			
		||||
        };
 | 
			
		||||
 | 
			
		||||
        host.storage.pool.extraDatasets = {
 | 
			
		||||
          # local datasets are for data that should be considered ephemeral
 | 
			
		||||
          "local" = {
 | 
			
		||||
            type = "zfs_fs";
 | 
			
		||||
            options.canmount = "off";
 | 
			
		||||
          };
 | 
			
		||||
          # nix directory needs to be available pre persist and doesn't need to be snapshotted or backed up
 | 
			
		||||
          "local/system/nix" = {
 | 
			
		||||
            type = "zfs_fs";
 | 
			
		||||
            mountpoint = "/nix";
 | 
			
		||||
            options = {
 | 
			
		||||
              atime = "off";
 | 
			
		||||
              relatime = "off";
 | 
			
		||||
              canmount = "on";
 | 
			
		||||
            };
 | 
			
		||||
          };
 | 
			
		||||
          # dataset for root that gets rolled back on every boot
 | 
			
		||||
          "local/system/root" = {
 | 
			
		||||
            type = "zfs_fs";
 | 
			
		||||
            mountpoint = "/";
 | 
			
		||||
            options = {
 | 
			
		||||
              canmount = "on";
 | 
			
		||||
            };
 | 
			
		||||
            postCreateHook = ''
 | 
			
		||||
              zfs snapshot rpool/local/system/root@blank
 | 
			
		||||
            '';
 | 
			
		||||
          };
 | 
			
		||||
 | 
			
		||||
          # persist datasets are datasets that contain information that we would like to keep around
 | 
			
		||||
          "persist" = {
 | 
			
		||||
            type = "zfs_fs";
 | 
			
		||||
            options.canmount = "off";
 | 
			
		||||
          };
 | 
			
		||||
          # this is where root data actually lives
 | 
			
		||||
          "persist/system/root" = {
 | 
			
		||||
            type = "zfs_fs";
 | 
			
		||||
            mountpoint = "/persist/system/root";
 | 
			
		||||
            options = {
 | 
			
		||||
              "com.sun:auto-snapshot" = "true";
 | 
			
		||||
            };
 | 
			
		||||
          };
 | 
			
		||||
          "persist/system/var/log" = {
 | 
			
		||||
            type = "zfs_fs";
 | 
			
		||||
            mountpoint = "/persist/system/var/log";
 | 
			
		||||
          };
 | 
			
		||||
        };
 | 
			
		||||
 | 
			
		||||
        environment.persistence."/persist/system/root" = {
 | 
			
		||||
          enable = true;
 | 
			
		||||
          hideMounts = true;
 | 
			
		||||
          directories = [
 | 
			
		||||
            "/etc/ssh"
 | 
			
		||||
 | 
			
		||||
            "/var/log"
 | 
			
		||||
            "/var/lib/nixos"
 | 
			
		||||
            "/var/lib/systemd/coredump"
 | 
			
		||||
 | 
			
		||||
            # config.apps.pihole.directory.root
 | 
			
		||||
 | 
			
		||||
            # config.apps.jellyfin.mediaDirectory
 | 
			
		||||
            # config.services.jellyfin.configDir
 | 
			
		||||
            # config.services.jellyfin.cacheDir
 | 
			
		||||
            # config.services.jellyfin.dataDir
 | 
			
		||||
 | 
			
		||||
            # "/var/hass" # config.users.users.hass.home
 | 
			
		||||
            # "/var/postgresql" # config.users.users.postgresql.home
 | 
			
		||||
            # "/var/forgejo" # config.users.users.forgejo.home
 | 
			
		||||
            # "/var/nextcloud" # config.users.users.nextcloud.home
 | 
			
		||||
            # "/var/headscale" # config.users.users.headscale.home
 | 
			
		||||
          ];
 | 
			
		||||
          files = [
 | 
			
		||||
            "/etc/machine-id"
 | 
			
		||||
          ];
 | 
			
		||||
        };
 | 
			
		||||
 | 
			
		||||
        security.sudo.extraConfig = "Defaults lecture=never";
 | 
			
		||||
      }
 | 
			
		||||
    )
 | 
			
		||||
  ];
 | 
			
		||||
}
 | 
			
		||||
| 
						 | 
				
			
			@ -10,7 +10,7 @@
 | 
			
		|||
 | 
			
		||||
  principleUsers = host.principleUsers;
 | 
			
		||||
  terminalUsers = host.terminalUsers;
 | 
			
		||||
  # normalUsers = host.normalUsers;
 | 
			
		||||
  normalUsers = host.normalUsers;
 | 
			
		||||
 | 
			
		||||
  uids = {
 | 
			
		||||
    leyla = 1000;
 | 
			
		||||
| 
						 | 
				
			
			@ -43,213 +43,273 @@
 | 
			
		|||
  ester = users.ester.name;
 | 
			
		||||
  eve = users.eve.name;
 | 
			
		||||
in {
 | 
			
		||||
  config = {
 | 
			
		||||
    # principle users are by definition trusted
 | 
			
		||||
    nix.settings.trusted-users = builtins.map (user: user.name) principleUsers;
 | 
			
		||||
  config = lib.mkMerge [
 | 
			
		||||
    {
 | 
			
		||||
      # principle users are by definition trusted
 | 
			
		||||
      nix.settings.trusted-users = builtins.map (user: user.name) principleUsers;
 | 
			
		||||
 | 
			
		||||
    # we should only be able to ssh into principle users of a computer who are also set up for terminal access
 | 
			
		||||
    services.openssh.settings.AllowUsers = builtins.map (user: user.name) (lib.lists.intersectLists terminalUsers principleUsers);
 | 
			
		||||
      # we should only be able to ssh into principle users of a computer who are also set up for terminal access
 | 
			
		||||
      services.openssh.settings.AllowUsers = builtins.map (user: user.name) (lib.lists.intersectLists terminalUsers principleUsers);
 | 
			
		||||
 | 
			
		||||
    # we need to set up env variables to nix can find keys to decrypt passwords on rebuild
 | 
			
		||||
    environment = {
 | 
			
		||||
      sessionVariables = {
 | 
			
		||||
        SOPS_AGE_KEY_DIRECTORY = SOPS_AGE_KEY_DIRECTORY;
 | 
			
		||||
        SOPS_AGE_KEY_FILE = "${SOPS_AGE_KEY_DIRECTORY}/key.txt";
 | 
			
		||||
      };
 | 
			
		||||
    };
 | 
			
		||||
 | 
			
		||||
    # set up user passwords
 | 
			
		||||
    sops = {
 | 
			
		||||
      defaultSopsFormat = "yaml";
 | 
			
		||||
      gnupg.sshKeyPaths = [];
 | 
			
		||||
 | 
			
		||||
      age = {
 | 
			
		||||
        keyFile = "/var/lib/sops-nix/key.txt";
 | 
			
		||||
        sshKeyPaths = [];
 | 
			
		||||
        # generateKey = true;
 | 
			
		||||
      };
 | 
			
		||||
 | 
			
		||||
      secrets = {
 | 
			
		||||
        "passwords/leyla" = {
 | 
			
		||||
          neededForUsers = true;
 | 
			
		||||
          sopsFile = "${inputs.secrets}/user-passwords.yaml";
 | 
			
		||||
        };
 | 
			
		||||
        "passwords/ester" = {
 | 
			
		||||
          neededForUsers = true;
 | 
			
		||||
          sopsFile = "${inputs.secrets}/user-passwords.yaml";
 | 
			
		||||
        };
 | 
			
		||||
        "passwords/eve" = {
 | 
			
		||||
          neededForUsers = true;
 | 
			
		||||
          sopsFile = "${inputs.secrets}/user-passwords.yaml";
 | 
			
		||||
      # we need to set up env variables to nix can find keys to decrypt passwords on rebuild
 | 
			
		||||
      environment = {
 | 
			
		||||
        sessionVariables = {
 | 
			
		||||
          SOPS_AGE_KEY_DIRECTORY = SOPS_AGE_KEY_DIRECTORY;
 | 
			
		||||
          SOPS_AGE_KEY_FILE = "${SOPS_AGE_KEY_DIRECTORY}/key.txt";
 | 
			
		||||
        };
 | 
			
		||||
      };
 | 
			
		||||
 | 
			
		||||
      # set up user passwords
 | 
			
		||||
      sops = {
 | 
			
		||||
        defaultSopsFormat = "yaml";
 | 
			
		||||
        gnupg.sshKeyPaths = [];
 | 
			
		||||
 | 
			
		||||
        age = {
 | 
			
		||||
          keyFile = "/var/lib/sops-nix/key.txt";
 | 
			
		||||
          sshKeyPaths = [];
 | 
			
		||||
          # generateKey = true;
 | 
			
		||||
        };
 | 
			
		||||
 | 
			
		||||
        secrets = {
 | 
			
		||||
          "passwords/leyla" = {
 | 
			
		||||
            neededForUsers = true;
 | 
			
		||||
            sopsFile = "${inputs.secrets}/user-passwords.yaml";
 | 
			
		||||
          };
 | 
			
		||||
          "passwords/ester" = {
 | 
			
		||||
            neededForUsers = true;
 | 
			
		||||
            sopsFile = "${inputs.secrets}/user-passwords.yaml";
 | 
			
		||||
          };
 | 
			
		||||
          "passwords/eve" = {
 | 
			
		||||
            neededForUsers = true;
 | 
			
		||||
            sopsFile = "${inputs.secrets}/user-passwords.yaml";
 | 
			
		||||
          };
 | 
			
		||||
        };
 | 
			
		||||
      };
 | 
			
		||||
    };
 | 
			
		||||
 | 
			
		||||
    users = {
 | 
			
		||||
      mutableUsers = false;
 | 
			
		||||
      users = {
 | 
			
		||||
        leyla = {
 | 
			
		||||
          uid = lib.mkForce uids.leyla;
 | 
			
		||||
          name = lib.mkForce host.users.leyla.name;
 | 
			
		||||
          description = "Leyla";
 | 
			
		||||
          extraGroups =
 | 
			
		||||
            (lib.lists.optionals host.users.leyla.isNormalUser ["networkmanager"])
 | 
			
		||||
            ++ (lib.lists.optionals host.users.leyla.isPrincipleUser ["wheel" "dialout"])
 | 
			
		||||
            ++ (lib.lists.optionals host.users.leyla.isDesktopUser ["adbusers"]);
 | 
			
		||||
          hashedPasswordFile = config.sops.secrets."passwords/leyla".path;
 | 
			
		||||
          isNormalUser = host.users.leyla.isNormalUser;
 | 
			
		||||
          isSystemUser = !host.users.leyla.isNormalUser;
 | 
			
		||||
          group = config.users.users.leyla.name;
 | 
			
		||||
        };
 | 
			
		||||
 | 
			
		||||
        ester = {
 | 
			
		||||
          uid = lib.mkForce uids.ester;
 | 
			
		||||
          name = lib.mkForce host.users.ester.name;
 | 
			
		||||
          description = "Ester";
 | 
			
		||||
          extraGroups = lib.optionals host.users.ester.isNormalUser ["networkmanager"];
 | 
			
		||||
          hashedPasswordFile = config.sops.secrets."passwords/ester".path;
 | 
			
		||||
          isNormalUser = host.users.ester.isNormalUser;
 | 
			
		||||
          isSystemUser = !host.users.ester.isNormalUser;
 | 
			
		||||
          group = config.users.users.ester.name;
 | 
			
		||||
        };
 | 
			
		||||
 | 
			
		||||
        eve = {
 | 
			
		||||
          uid = lib.mkForce uids.eve;
 | 
			
		||||
          name = lib.mkForce host.users.eve.name;
 | 
			
		||||
          description = "Eve";
 | 
			
		||||
          extraGroups = lib.optionals host.users.eve.isNormalUser ["networkmanager"];
 | 
			
		||||
          hashedPasswordFile = config.sops.secrets."passwords/eve".path;
 | 
			
		||||
          isNormalUser = host.users.eve.isNormalUser;
 | 
			
		||||
          isSystemUser = !host.users.eve.isNormalUser;
 | 
			
		||||
          group = config.users.users.eve.name;
 | 
			
		||||
        };
 | 
			
		||||
 | 
			
		||||
        jellyfin = {
 | 
			
		||||
          uid = lib.mkForce uids.jellyfin;
 | 
			
		||||
          isSystemUser = true;
 | 
			
		||||
          group = config.users.users.jellyfin.name;
 | 
			
		||||
        };
 | 
			
		||||
 | 
			
		||||
        forgejo = {
 | 
			
		||||
          uid = lib.mkForce uids.forgejo;
 | 
			
		||||
          isSystemUser = true;
 | 
			
		||||
          group = config.users.users.forgejo.name;
 | 
			
		||||
        };
 | 
			
		||||
 | 
			
		||||
        pihole = {
 | 
			
		||||
          uid = lib.mkForce uids.pihole;
 | 
			
		||||
          isSystemUser = true;
 | 
			
		||||
          group = config.users.users.pihole.name;
 | 
			
		||||
        };
 | 
			
		||||
 | 
			
		||||
        hass = {
 | 
			
		||||
          uid = lib.mkForce uids.hass;
 | 
			
		||||
          isSystemUser = true;
 | 
			
		||||
          group = config.users.users.hass.name;
 | 
			
		||||
        };
 | 
			
		||||
 | 
			
		||||
        headscale = {
 | 
			
		||||
          uid = lib.mkForce uids.headscale;
 | 
			
		||||
          isSystemUser = true;
 | 
			
		||||
          group = config.users.users.headscale.name;
 | 
			
		||||
        };
 | 
			
		||||
 | 
			
		||||
        nextcloud = {
 | 
			
		||||
          uid = lib.mkForce uids.nextcloud;
 | 
			
		||||
          isSystemUser = true;
 | 
			
		||||
          group = config.users.users.nextcloud.name;
 | 
			
		||||
        };
 | 
			
		||||
      };
 | 
			
		||||
 | 
			
		||||
      groups = {
 | 
			
		||||
        leyla = {
 | 
			
		||||
          gid = lib.mkForce gids.leyla;
 | 
			
		||||
          members = [
 | 
			
		||||
            leyla
 | 
			
		||||
          ];
 | 
			
		||||
        };
 | 
			
		||||
 | 
			
		||||
        ester = {
 | 
			
		||||
          gid = lib.mkForce gids.ester;
 | 
			
		||||
          members = [
 | 
			
		||||
            ester
 | 
			
		||||
          ];
 | 
			
		||||
        };
 | 
			
		||||
 | 
			
		||||
        eve = {
 | 
			
		||||
          gid = lib.mkForce gids.eve;
 | 
			
		||||
          members = [
 | 
			
		||||
            eve
 | 
			
		||||
          ];
 | 
			
		||||
        };
 | 
			
		||||
 | 
			
		||||
        mutableUsers = false;
 | 
			
		||||
        users = {
 | 
			
		||||
          gid = lib.mkForce gids.users;
 | 
			
		||||
          members = [
 | 
			
		||||
            leyla
 | 
			
		||||
            ester
 | 
			
		||||
            eve
 | 
			
		||||
          ];
 | 
			
		||||
          leyla = {
 | 
			
		||||
            uid = lib.mkForce uids.leyla;
 | 
			
		||||
            name = lib.mkForce host.users.leyla.name;
 | 
			
		||||
            description = "Leyla";
 | 
			
		||||
            extraGroups =
 | 
			
		||||
              (lib.lists.optionals host.users.leyla.isNormalUser ["networkmanager"])
 | 
			
		||||
              ++ (lib.lists.optionals host.users.leyla.isPrincipleUser ["wheel" "dialout"])
 | 
			
		||||
              ++ (lib.lists.optionals host.users.leyla.isDesktopUser ["adbusers"]);
 | 
			
		||||
            hashedPasswordFile = config.sops.secrets."passwords/leyla".path;
 | 
			
		||||
            isNormalUser = host.users.leyla.isNormalUser;
 | 
			
		||||
            isSystemUser = !host.users.leyla.isNormalUser;
 | 
			
		||||
            group = config.users.users.leyla.name;
 | 
			
		||||
          };
 | 
			
		||||
 | 
			
		||||
          ester = {
 | 
			
		||||
            uid = lib.mkForce uids.ester;
 | 
			
		||||
            name = lib.mkForce host.users.ester.name;
 | 
			
		||||
            description = "Ester";
 | 
			
		||||
            extraGroups = lib.optionals host.users.ester.isNormalUser ["networkmanager"];
 | 
			
		||||
            hashedPasswordFile = config.sops.secrets."passwords/ester".path;
 | 
			
		||||
            isNormalUser = host.users.ester.isNormalUser;
 | 
			
		||||
            isSystemUser = !host.users.ester.isNormalUser;
 | 
			
		||||
            group = config.users.users.ester.name;
 | 
			
		||||
          };
 | 
			
		||||
 | 
			
		||||
          eve = {
 | 
			
		||||
            uid = lib.mkForce uids.eve;
 | 
			
		||||
            name = lib.mkForce host.users.eve.name;
 | 
			
		||||
            description = "Eve";
 | 
			
		||||
            extraGroups = lib.optionals host.users.eve.isNormalUser ["networkmanager"];
 | 
			
		||||
            hashedPasswordFile = config.sops.secrets."passwords/eve".path;
 | 
			
		||||
            isNormalUser = host.users.eve.isNormalUser;
 | 
			
		||||
            isSystemUser = !host.users.eve.isNormalUser;
 | 
			
		||||
            group = config.users.users.eve.name;
 | 
			
		||||
          };
 | 
			
		||||
 | 
			
		||||
          jellyfin = {
 | 
			
		||||
            uid = lib.mkForce uids.jellyfin;
 | 
			
		||||
            isSystemUser = true;
 | 
			
		||||
            group = config.users.users.jellyfin.name;
 | 
			
		||||
          };
 | 
			
		||||
 | 
			
		||||
          forgejo = {
 | 
			
		||||
            uid = lib.mkForce uids.forgejo;
 | 
			
		||||
            isSystemUser = true;
 | 
			
		||||
            group = config.users.users.forgejo.name;
 | 
			
		||||
          };
 | 
			
		||||
 | 
			
		||||
          pihole = {
 | 
			
		||||
            uid = lib.mkForce uids.pihole;
 | 
			
		||||
            isSystemUser = true;
 | 
			
		||||
            group = config.users.users.pihole.name;
 | 
			
		||||
          };
 | 
			
		||||
 | 
			
		||||
          hass = {
 | 
			
		||||
            uid = lib.mkForce uids.hass;
 | 
			
		||||
            isSystemUser = true;
 | 
			
		||||
            group = config.users.users.hass.name;
 | 
			
		||||
          };
 | 
			
		||||
 | 
			
		||||
          headscale = {
 | 
			
		||||
            uid = lib.mkForce uids.headscale;
 | 
			
		||||
            isSystemUser = true;
 | 
			
		||||
            group = config.users.users.headscale.name;
 | 
			
		||||
          };
 | 
			
		||||
 | 
			
		||||
          nextcloud = {
 | 
			
		||||
            uid = lib.mkForce uids.nextcloud;
 | 
			
		||||
            isSystemUser = true;
 | 
			
		||||
            group = config.users.users.nextcloud.name;
 | 
			
		||||
          };
 | 
			
		||||
        };
 | 
			
		||||
 | 
			
		||||
        jellyfin_media = {
 | 
			
		||||
          gid = lib.mkForce gids.jellyfin_media;
 | 
			
		||||
          members = [
 | 
			
		||||
            users.jellyfin.name
 | 
			
		||||
            leyla
 | 
			
		||||
            ester
 | 
			
		||||
            eve
 | 
			
		||||
          ];
 | 
			
		||||
        };
 | 
			
		||||
        groups = {
 | 
			
		||||
          leyla = {
 | 
			
		||||
            gid = lib.mkForce gids.leyla;
 | 
			
		||||
            members = [
 | 
			
		||||
              leyla
 | 
			
		||||
            ];
 | 
			
		||||
          };
 | 
			
		||||
 | 
			
		||||
        jellyfin = {
 | 
			
		||||
          gid = lib.mkForce gids.jellyfin;
 | 
			
		||||
          members = [
 | 
			
		||||
            users.jellyfin.name
 | 
			
		||||
            # leyla
 | 
			
		||||
          ];
 | 
			
		||||
        };
 | 
			
		||||
          ester = {
 | 
			
		||||
            gid = lib.mkForce gids.ester;
 | 
			
		||||
            members = [
 | 
			
		||||
              ester
 | 
			
		||||
            ];
 | 
			
		||||
          };
 | 
			
		||||
 | 
			
		||||
        forgejo = {
 | 
			
		||||
          gid = lib.mkForce gids.forgejo;
 | 
			
		||||
          members = [
 | 
			
		||||
            users.forgejo.name
 | 
			
		||||
            # leyla
 | 
			
		||||
          ];
 | 
			
		||||
        };
 | 
			
		||||
          eve = {
 | 
			
		||||
            gid = lib.mkForce gids.eve;
 | 
			
		||||
            members = [
 | 
			
		||||
              eve
 | 
			
		||||
            ];
 | 
			
		||||
          };
 | 
			
		||||
 | 
			
		||||
        pihole = {
 | 
			
		||||
          gid = lib.mkForce gids.pihole;
 | 
			
		||||
          members = [
 | 
			
		||||
            users.pihole.name
 | 
			
		||||
            # leyla
 | 
			
		||||
          ];
 | 
			
		||||
        };
 | 
			
		||||
          users = {
 | 
			
		||||
            gid = lib.mkForce gids.users;
 | 
			
		||||
            members = [
 | 
			
		||||
              leyla
 | 
			
		||||
              ester
 | 
			
		||||
              eve
 | 
			
		||||
            ];
 | 
			
		||||
          };
 | 
			
		||||
 | 
			
		||||
        hass = {
 | 
			
		||||
          gid = lib.mkForce gids.hass;
 | 
			
		||||
          members = [
 | 
			
		||||
            users.hass.name
 | 
			
		||||
            # leyla
 | 
			
		||||
          ];
 | 
			
		||||
        };
 | 
			
		||||
          jellyfin_media = {
 | 
			
		||||
            gid = lib.mkForce gids.jellyfin_media;
 | 
			
		||||
            members = [
 | 
			
		||||
              users.jellyfin.name
 | 
			
		||||
              leyla
 | 
			
		||||
              ester
 | 
			
		||||
              eve
 | 
			
		||||
            ];
 | 
			
		||||
          };
 | 
			
		||||
 | 
			
		||||
        headscale = {
 | 
			
		||||
          gid = lib.mkForce gids.headscale;
 | 
			
		||||
          members = [
 | 
			
		||||
            users.headscale.name
 | 
			
		||||
            # leyla
 | 
			
		||||
          ];
 | 
			
		||||
        };
 | 
			
		||||
          jellyfin = {
 | 
			
		||||
            gid = lib.mkForce gids.jellyfin;
 | 
			
		||||
            members = [
 | 
			
		||||
              users.jellyfin.name
 | 
			
		||||
              # leyla
 | 
			
		||||
            ];
 | 
			
		||||
          };
 | 
			
		||||
 | 
			
		||||
        nextcloud = {
 | 
			
		||||
          gid = lib.mkForce gids.nextcloud;
 | 
			
		||||
          members = [
 | 
			
		||||
            users.nextcloud.name
 | 
			
		||||
            # leyla
 | 
			
		||||
          ];
 | 
			
		||||
          forgejo = {
 | 
			
		||||
            gid = lib.mkForce gids.forgejo;
 | 
			
		||||
            members = [
 | 
			
		||||
              users.forgejo.name
 | 
			
		||||
              # leyla
 | 
			
		||||
            ];
 | 
			
		||||
          };
 | 
			
		||||
 | 
			
		||||
          pihole = {
 | 
			
		||||
            gid = lib.mkForce gids.pihole;
 | 
			
		||||
            members = [
 | 
			
		||||
              users.pihole.name
 | 
			
		||||
              # leyla
 | 
			
		||||
            ];
 | 
			
		||||
          };
 | 
			
		||||
 | 
			
		||||
          hass = {
 | 
			
		||||
            gid = lib.mkForce gids.hass;
 | 
			
		||||
            members = [
 | 
			
		||||
              users.hass.name
 | 
			
		||||
              # leyla
 | 
			
		||||
            ];
 | 
			
		||||
          };
 | 
			
		||||
 | 
			
		||||
          headscale = {
 | 
			
		||||
            gid = lib.mkForce gids.headscale;
 | 
			
		||||
            members = [
 | 
			
		||||
              users.headscale.name
 | 
			
		||||
              # leyla
 | 
			
		||||
            ];
 | 
			
		||||
          };
 | 
			
		||||
 | 
			
		||||
          nextcloud = {
 | 
			
		||||
            gid = lib.mkForce gids.nextcloud;
 | 
			
		||||
            members = [
 | 
			
		||||
              users.nextcloud.name
 | 
			
		||||
              # leyla
 | 
			
		||||
            ];
 | 
			
		||||
          };
 | 
			
		||||
        };
 | 
			
		||||
      };
 | 
			
		||||
    };
 | 
			
		||||
  };
 | 
			
		||||
    }
 | 
			
		||||
    (lib.mkIf config.host.impermanence.enable {
 | 
			
		||||
      boot.initrd.postResumeCommands = lib.mkAfter (
 | 
			
		||||
        lib.strings.concatStrings (builtins.map (user: ''
 | 
			
		||||
            zfs rollback -r rpool/local/home/${user.name}@blank
 | 
			
		||||
          '')
 | 
			
		||||
          normalUsers)
 | 
			
		||||
      );
 | 
			
		||||
 | 
			
		||||
      fileSystems.${SOPS_AGE_KEY_DIRECTORY}.neededForBoot = true;
 | 
			
		||||
 | 
			
		||||
      environment.persistence."/persist/system/root" = {
 | 
			
		||||
        enable = true;
 | 
			
		||||
        hideMounts = true;
 | 
			
		||||
        directories = [
 | 
			
		||||
          "/run/secrets"
 | 
			
		||||
        ];
 | 
			
		||||
      };
 | 
			
		||||
 | 
			
		||||
      host.storage.pool.extraDatasets = lib.mkMerge [
 | 
			
		||||
        {
 | 
			
		||||
          # sops age key needs to be available to pre persist for user generation
 | 
			
		||||
          "local/system/sops" = {
 | 
			
		||||
            type = "zfs_fs";
 | 
			
		||||
            mountpoint = SOPS_AGE_KEY_DIRECTORY;
 | 
			
		||||
            options = {
 | 
			
		||||
              atime = "off";
 | 
			
		||||
              relatime = "off";
 | 
			
		||||
              canmount = "on";
 | 
			
		||||
            };
 | 
			
		||||
          };
 | 
			
		||||
        }
 | 
			
		||||
        (
 | 
			
		||||
          lib.mkMerge
 | 
			
		||||
          (
 | 
			
		||||
            builtins.map (user: {
 | 
			
		||||
              "local/home/${user.name}" = {
 | 
			
		||||
                type = "zfs_fs";
 | 
			
		||||
                mountpoint = "/home/${user.name}";
 | 
			
		||||
                options = {
 | 
			
		||||
                  canmount = "on";
 | 
			
		||||
                };
 | 
			
		||||
                postCreateHook = ''
 | 
			
		||||
                  zfs snapshot rpool/local/home/${user.name}@blank
 | 
			
		||||
                '';
 | 
			
		||||
              };
 | 
			
		||||
              "persist/home/${user.name}" = {
 | 
			
		||||
                type = "zfs_fs";
 | 
			
		||||
                mountpoint = "/persist/home/${user.name}";
 | 
			
		||||
                options = {
 | 
			
		||||
                  "com.sun:auto-snapshot" = "true";
 | 
			
		||||
                };
 | 
			
		||||
              };
 | 
			
		||||
            })
 | 
			
		||||
            normalUsers
 | 
			
		||||
          )
 | 
			
		||||
        )
 | 
			
		||||
      ];
 | 
			
		||||
    })
 | 
			
		||||
  ];
 | 
			
		||||
}
 | 
			
		||||
| 
						 | 
				
			
			
 | 
			
		|||
| 
						 | 
				
			
			@ -67,10 +67,13 @@ in {
 | 
			
		|||
      default = lib.lists.filter (user: user.isPrincipleUser) hostUsers;
 | 
			
		||||
    };
 | 
			
		||||
    normalUsers = lib.mkOption {
 | 
			
		||||
      default = lib.lists.filter (user: user.isTerminalUser) hostUsers;
 | 
			
		||||
      default = lib.lists.filter (user: user.isNormalUser) hostUsers;
 | 
			
		||||
    };
 | 
			
		||||
    desktopUsers = lib.mkOption {
 | 
			
		||||
      default = lib.lists.filter (user: user.isDesktopUser) hostUsers;
 | 
			
		||||
    };
 | 
			
		||||
    terminalUsers = lib.mkOption {
 | 
			
		||||
      default = lib.lists.filter (user: user.isNormalUser) hostUsers;
 | 
			
		||||
      default = lib.lists.filter (user: user.isTerminalUser) hostUsers;
 | 
			
		||||
    };
 | 
			
		||||
  };
 | 
			
		||||
 | 
			
		||||
| 
						 | 
				
			
			
 | 
			
		|||
| 
						 | 
				
			
			@ -7,6 +7,7 @@
 | 
			
		|||
  home-manager = inputs.home-manager;
 | 
			
		||||
  nix-darwin = inputs.nix-darwin;
 | 
			
		||||
  sops-nix = inputs.sops-nix;
 | 
			
		||||
  disko = inputs.disko;
 | 
			
		||||
  impermanence = inputs.impermanence;
 | 
			
		||||
 | 
			
		||||
  systems = [
 | 
			
		||||
| 
						 | 
				
			
			@ -74,6 +75,7 @@ in {
 | 
			
		|||
          sops-nix.nixosModules.sops
 | 
			
		||||
          impermanence.nixosModules.impermanence
 | 
			
		||||
          home-manager.nixosModules.home-manager
 | 
			
		||||
          disko.nixosModules.disko
 | 
			
		||||
          ../modules/nixos-modules
 | 
			
		||||
          ../configurations/nixos/${host}
 | 
			
		||||
        ];
 | 
			
		||||
| 
						 | 
				
			
			
 | 
			
		|||
		Loading…
	
	Add table
		Add a link
		
	
		Reference in a new issue