mirror of
https://github.com/NixOS/nixpkgs.git
synced 2025-06-17 15:09:26 +03:00

After final improvements to the official formatter implementation, this commit now performs the first treewide reformat of Nix files using it. This is part of the implementation of RFC 166. Only "inactive" files are reformatted, meaning only files that aren't being touched by any PR with activity in the past 2 months. This is to avoid conflicts for PRs that might soon be merged. Later we can do a full treewide reformat to get the rest, which should not cause as many conflicts. A CI check has already been running for some time to ensure that new and already-formatted files are formatted, so the files being reformatted here should also stay formatted. This commit was automatically created and can be verified using nix-builda08b3a4d19
.tar.gz \ --argstr baseRevb32a094368
result/bin/apply-formatting $NIXPKGS_PATH
237 lines
7.1 KiB
Nix
237 lines
7.1 KiB
Nix
{
|
|
config,
|
|
lib,
|
|
pkgs,
|
|
...
|
|
}:
|
|
let
|
|
cfg = config.services.hadoop;
|
|
|
|
# Config files for hadoop services
|
|
hadoopConf = "${import ./conf.nix { inherit cfg pkgs lib; }}/";
|
|
|
|
# Generator for HDFS service options
|
|
hadoopServiceOption =
|
|
{
|
|
serviceName,
|
|
firewallOption ? true,
|
|
extraOpts ? null,
|
|
}:
|
|
{
|
|
enable = lib.mkEnableOption serviceName;
|
|
restartIfChanged = lib.mkOption {
|
|
type = lib.types.bool;
|
|
description = ''
|
|
Automatically restart the service on config change.
|
|
This can be set to false to defer restarts on clusters running critical applications.
|
|
Please consider the security implications of inadvertently running an older version,
|
|
and the possibility of unexpected behavior caused by inconsistent versions across a cluster when disabling this option.
|
|
'';
|
|
default = false;
|
|
};
|
|
extraFlags = lib.mkOption {
|
|
type = with lib.types; listOf str;
|
|
default = [ ];
|
|
description = "Extra command line flags to pass to ${serviceName}";
|
|
example = [
|
|
"-Dcom.sun.management.jmxremote"
|
|
"-Dcom.sun.management.jmxremote.port=8010"
|
|
];
|
|
};
|
|
extraEnv = lib.mkOption {
|
|
type = with lib.types; attrsOf str;
|
|
default = { };
|
|
description = "Extra environment variables for ${serviceName}";
|
|
};
|
|
}
|
|
// (lib.optionalAttrs firewallOption {
|
|
openFirewall = lib.mkOption {
|
|
type = lib.types.bool;
|
|
default = false;
|
|
description = "Open firewall ports for ${serviceName}.";
|
|
};
|
|
})
|
|
// (lib.optionalAttrs (extraOpts != null) extraOpts);
|
|
|
|
# Generator for HDFS service configs
|
|
hadoopServiceConfig =
|
|
{
|
|
name,
|
|
serviceOptions ? cfg.hdfs."${lib.toLower name}",
|
|
description ? "Hadoop HDFS ${name}",
|
|
User ? "hdfs",
|
|
allowedTCPPorts ? [ ],
|
|
preStart ? "",
|
|
environment ? { },
|
|
extraConfig ? { },
|
|
}:
|
|
(
|
|
|
|
lib.mkIf serviceOptions.enable (
|
|
lib.mkMerge [
|
|
{
|
|
systemd.services."hdfs-${lib.toLower name}" = {
|
|
inherit description preStart;
|
|
environment = environment // serviceOptions.extraEnv;
|
|
wantedBy = [ "multi-user.target" ];
|
|
inherit (serviceOptions) restartIfChanged;
|
|
serviceConfig = {
|
|
inherit User;
|
|
SyslogIdentifier = "hdfs-${lib.toLower name}";
|
|
ExecStart = "${cfg.package}/bin/hdfs --config ${hadoopConf} ${lib.toLower name} ${lib.escapeShellArgs serviceOptions.extraFlags}";
|
|
Restart = "always";
|
|
};
|
|
};
|
|
|
|
services.hadoop.gatewayRole.enable = true;
|
|
|
|
networking.firewall.allowedTCPPorts = lib.mkIf (
|
|
(builtins.hasAttr "openFirewall" serviceOptions) && serviceOptions.openFirewall
|
|
) allowedTCPPorts;
|
|
}
|
|
extraConfig
|
|
]
|
|
)
|
|
);
|
|
|
|
in
|
|
{
|
|
options.services.hadoop.hdfs = {
|
|
|
|
namenode = hadoopServiceOption { serviceName = "HDFS NameNode"; } // {
|
|
formatOnInit = lib.mkOption {
|
|
type = lib.types.bool;
|
|
default = false;
|
|
description = ''
|
|
Format HDFS namenode on first start. This is useful for quickly spinning up
|
|
ephemeral HDFS clusters with a single namenode.
|
|
For HA clusters, initialization involves multiple steps across multiple nodes.
|
|
Follow this guide to initialize an HA cluster manually:
|
|
<https://hadoop.apache.org/docs/stable/hadoop-project-dist/hadoop-hdfs/HDFSHighAvailabilityWithQJM.html>
|
|
'';
|
|
};
|
|
};
|
|
|
|
datanode = hadoopServiceOption { serviceName = "HDFS DataNode"; } // {
|
|
dataDirs = lib.mkOption {
|
|
default = null;
|
|
description = "Tier and path definitions for datanode storage.";
|
|
type =
|
|
with lib.types;
|
|
nullOr (
|
|
listOf (submodule {
|
|
options = {
|
|
type = lib.mkOption {
|
|
type = enum [
|
|
"SSD"
|
|
"DISK"
|
|
"ARCHIVE"
|
|
"RAM_DISK"
|
|
];
|
|
description = ''
|
|
Storage types ([SSD]/[DISK]/[ARCHIVE]/[RAM_DISK]) for HDFS storage policies.
|
|
'';
|
|
};
|
|
path = lib.mkOption {
|
|
type = path;
|
|
example = [ "/var/lib/hadoop/hdfs/dn" ];
|
|
description = "Determines where on the local filesystem a data node should store its blocks.";
|
|
};
|
|
};
|
|
})
|
|
);
|
|
};
|
|
};
|
|
|
|
journalnode = hadoopServiceOption { serviceName = "HDFS JournalNode"; };
|
|
|
|
zkfc = hadoopServiceOption {
|
|
serviceName = "HDFS ZooKeeper failover controller";
|
|
firewallOption = false;
|
|
};
|
|
|
|
httpfs = hadoopServiceOption { serviceName = "HDFS JournalNode"; } // {
|
|
tempPath = lib.mkOption {
|
|
type = lib.types.path;
|
|
default = "/tmp/hadoop/httpfs";
|
|
description = "HTTPFS_TEMP path used by HTTPFS";
|
|
};
|
|
};
|
|
|
|
};
|
|
|
|
config = lib.mkMerge [
|
|
(hadoopServiceConfig {
|
|
name = "NameNode";
|
|
allowedTCPPorts = [
|
|
9870 # namenode.http-address
|
|
8020 # namenode.rpc-address
|
|
8022 # namenode.servicerpc-address
|
|
8019 # dfs.ha.zkfc.port
|
|
];
|
|
preStart = (
|
|
lib.mkIf cfg.hdfs.namenode.formatOnInit "${cfg.package}/bin/hdfs --config ${hadoopConf} namenode -format -nonInteractive || true"
|
|
);
|
|
})
|
|
|
|
(hadoopServiceConfig {
|
|
name = "DataNode";
|
|
# port numbers for datanode changed between hadoop 2 and 3
|
|
allowedTCPPorts =
|
|
if lib.versionAtLeast cfg.package.version "3" then
|
|
[
|
|
9864 # datanode.http.address
|
|
9866 # datanode.address
|
|
9867 # datanode.ipc.address
|
|
]
|
|
else
|
|
[
|
|
50075 # datanode.http.address
|
|
50010 # datanode.address
|
|
50020 # datanode.ipc.address
|
|
];
|
|
extraConfig.services.hadoop.hdfsSiteInternal."dfs.datanode.data.dir" = lib.mkIf (
|
|
cfg.hdfs.datanode.dataDirs != null
|
|
) (lib.concatMapStringsSep "," (x: "[" + x.type + "]file://" + x.path) cfg.hdfs.datanode.dataDirs);
|
|
})
|
|
|
|
(hadoopServiceConfig {
|
|
name = "JournalNode";
|
|
allowedTCPPorts = [
|
|
8480 # dfs.journalnode.http-address
|
|
8485 # dfs.journalnode.rpc-address
|
|
];
|
|
})
|
|
|
|
(hadoopServiceConfig {
|
|
name = "zkfc";
|
|
description = "Hadoop HDFS ZooKeeper failover controller";
|
|
})
|
|
|
|
(hadoopServiceConfig {
|
|
name = "HTTPFS";
|
|
environment.HTTPFS_TEMP = cfg.hdfs.httpfs.tempPath;
|
|
preStart = "mkdir -p $HTTPFS_TEMP";
|
|
User = "httpfs";
|
|
allowedTCPPorts = [
|
|
14000 # httpfs.http.port
|
|
];
|
|
})
|
|
|
|
(lib.mkIf cfg.gatewayRole.enable {
|
|
users.users.hdfs = {
|
|
description = "Hadoop HDFS user";
|
|
group = "hadoop";
|
|
uid = config.ids.uids.hdfs;
|
|
};
|
|
})
|
|
(lib.mkIf cfg.hdfs.httpfs.enable {
|
|
users.users.httpfs = {
|
|
description = "Hadoop HTTPFS user";
|
|
group = "hadoop";
|
|
isSystemUser = true;
|
|
};
|
|
})
|
|
|
|
];
|
|
}
|