mirror of
https://github.com/NixOS/nixpkgs.git
synced 2025-06-10 11:45:45 +03:00

Format all Nix files using the officially approved formatter,
making the CI check introduced in the previous commit succeed:
nix-build ci -A fmt.check
This is the next step of the of the [implementation](https://github.com/NixOS/nixfmt/issues/153)
of the accepted [RFC 166](https://github.com/NixOS/rfcs/pull/166).
This commit will lead to merge conflicts for a number of PRs,
up to an estimated ~1100 (~33%) among the PRs with activity in the past 2
months, but that should be lower than what it would be without the previous
[partial treewide format](https://github.com/NixOS/nixpkgs/pull/322537).
Merge conflicts caused by this commit can now automatically be resolved while rebasing using the
[auto-rebase script](8616af08d9/maintainers/scripts/auto-rebase
).
If you run into any problems regarding any of this, please reach out to the
[formatting team](https://nixos.org/community/teams/formatting/) by
pinging @NixOS/nix-formatting.
198 lines
5.3 KiB
Nix
198 lines
5.3 KiB
Nix
{
|
|
system ? builtins.currentSystem,
|
|
pkgs ? import ../../.. { inherit system; },
|
|
}:
|
|
with import ./base.nix { inherit system; };
|
|
let
|
|
domain = "my.zyx";
|
|
|
|
redisPod = pkgs.writeText "redis-pod.json" (
|
|
builtins.toJSON {
|
|
kind = "Pod";
|
|
apiVersion = "v1";
|
|
metadata.name = "redis";
|
|
metadata.labels.name = "redis";
|
|
spec.containers = [
|
|
{
|
|
name = "redis";
|
|
image = "redis";
|
|
args = [
|
|
"--bind"
|
|
"0.0.0.0"
|
|
];
|
|
imagePullPolicy = "Never";
|
|
ports = [
|
|
{
|
|
name = "redis-server";
|
|
containerPort = 6379;
|
|
}
|
|
];
|
|
}
|
|
];
|
|
}
|
|
);
|
|
|
|
redisService = pkgs.writeText "redis-service.json" (
|
|
builtins.toJSON {
|
|
kind = "Service";
|
|
apiVersion = "v1";
|
|
metadata.name = "redis";
|
|
spec = {
|
|
ports = [
|
|
{
|
|
port = 6379;
|
|
targetPort = 6379;
|
|
}
|
|
];
|
|
selector = {
|
|
name = "redis";
|
|
};
|
|
};
|
|
}
|
|
);
|
|
|
|
redisImage = pkgs.dockerTools.buildImage {
|
|
name = "redis";
|
|
tag = "latest";
|
|
copyToRoot = pkgs.buildEnv {
|
|
name = "image-root";
|
|
pathsToLink = [ "/bin" ];
|
|
paths = [
|
|
pkgs.redis
|
|
pkgs.bind.host
|
|
];
|
|
};
|
|
config.Entrypoint = [ "/bin/redis-server" ];
|
|
};
|
|
|
|
probePod = pkgs.writeText "probe-pod.json" (
|
|
builtins.toJSON {
|
|
kind = "Pod";
|
|
apiVersion = "v1";
|
|
metadata.name = "probe";
|
|
metadata.labels.name = "probe";
|
|
spec.containers = [
|
|
{
|
|
name = "probe";
|
|
image = "probe";
|
|
args = [ "-f" ];
|
|
tty = true;
|
|
imagePullPolicy = "Never";
|
|
}
|
|
];
|
|
}
|
|
);
|
|
|
|
probeImage = pkgs.dockerTools.buildImage {
|
|
name = "probe";
|
|
tag = "latest";
|
|
copyToRoot = pkgs.buildEnv {
|
|
name = "image-root";
|
|
pathsToLink = [ "/bin" ];
|
|
paths = [
|
|
pkgs.bind.host
|
|
pkgs.busybox
|
|
];
|
|
};
|
|
config.Entrypoint = [ "/bin/tail" ];
|
|
};
|
|
|
|
extraConfiguration =
|
|
{
|
|
config,
|
|
pkgs,
|
|
lib,
|
|
...
|
|
}:
|
|
{
|
|
environment.systemPackages = [ pkgs.bind.host ];
|
|
services.dnsmasq.enable = true;
|
|
services.dnsmasq.settings.server = [
|
|
"/cluster.local/${config.services.kubernetes.addons.dns.clusterIp}#53"
|
|
];
|
|
};
|
|
|
|
base = {
|
|
name = "dns";
|
|
inherit domain extraConfiguration;
|
|
};
|
|
|
|
singleNodeTest = {
|
|
test = ''
|
|
# prepare machine1 for test
|
|
machine1.wait_until_succeeds("kubectl get node machine1.${domain} | grep -w Ready")
|
|
machine1.wait_until_succeeds(
|
|
"${pkgs.gzip}/bin/zcat ${redisImage} | ${pkgs.containerd}/bin/ctr -n k8s.io image import -"
|
|
)
|
|
machine1.wait_until_succeeds(
|
|
"kubectl create -f ${redisPod}"
|
|
)
|
|
machine1.wait_until_succeeds(
|
|
"kubectl create -f ${redisService}"
|
|
)
|
|
machine1.wait_until_succeeds(
|
|
"${pkgs.gzip}/bin/zcat ${probeImage} | ${pkgs.containerd}/bin/ctr -n k8s.io image import -"
|
|
)
|
|
machine1.wait_until_succeeds(
|
|
"kubectl create -f ${probePod}"
|
|
)
|
|
|
|
# check if pods are running
|
|
machine1.wait_until_succeeds("kubectl get pod redis | grep Running")
|
|
machine1.wait_until_succeeds("kubectl get pod probe | grep Running")
|
|
machine1.wait_until_succeeds("kubectl get pods -n kube-system | grep 'coredns.*1/1'")
|
|
|
|
# check dns on host (dnsmasq)
|
|
machine1.succeed("host redis.default.svc.cluster.local")
|
|
|
|
# check dns inside the container
|
|
machine1.succeed("kubectl exec probe -- /bin/host redis.default.svc.cluster.local")
|
|
'';
|
|
};
|
|
|
|
multiNodeTest = {
|
|
test = ''
|
|
# Node token exchange
|
|
machine1.wait_until_succeeds(
|
|
"cp -f /var/lib/cfssl/apitoken.secret /tmp/shared/apitoken.secret"
|
|
)
|
|
machine2.wait_until_succeeds(
|
|
"cat /tmp/shared/apitoken.secret | nixos-kubernetes-node-join"
|
|
)
|
|
|
|
# prepare machines for test
|
|
machine1.wait_until_succeeds("kubectl get node machine2.${domain} | grep -w Ready")
|
|
machine2.wait_until_succeeds(
|
|
"${pkgs.gzip}/bin/zcat ${redisImage} | ${pkgs.containerd}/bin/ctr -n k8s.io image import -"
|
|
)
|
|
machine1.wait_until_succeeds(
|
|
"kubectl create -f ${redisPod}"
|
|
)
|
|
machine1.wait_until_succeeds(
|
|
"kubectl create -f ${redisService}"
|
|
)
|
|
machine2.wait_until_succeeds(
|
|
"${pkgs.gzip}/bin/zcat ${probeImage} | ${pkgs.containerd}/bin/ctr -n k8s.io image import -"
|
|
)
|
|
machine1.wait_until_succeeds(
|
|
"kubectl create -f ${probePod}"
|
|
)
|
|
|
|
# check if pods are running
|
|
machine1.wait_until_succeeds("kubectl get pod redis | grep Running")
|
|
machine1.wait_until_succeeds("kubectl get pod probe | grep Running")
|
|
machine1.wait_until_succeeds("kubectl get pods -n kube-system | grep 'coredns.*1/1'")
|
|
|
|
# check dns on hosts (dnsmasq)
|
|
machine1.succeed("host redis.default.svc.cluster.local")
|
|
machine2.succeed("host redis.default.svc.cluster.local")
|
|
|
|
# check dns inside the container
|
|
machine1.succeed("kubectl exec probe -- /bin/host redis.default.svc.cluster.local")
|
|
'';
|
|
};
|
|
in
|
|
{
|
|
singlenode = mkKubernetesSingleNodeTest (base // singleNodeTest);
|
|
multinode = mkKubernetesMultiNodeTest (base // multiNodeTest);
|
|
}
|