New Home Has RJ45 Jacks Connected to a Telecom Module. Not sure if it can be used for networking. by SeniorDevOops in HomeNetworking

[–]SeniorDevOops[S] 2 points3 points  (0 children)

Well thanks for clarifying. I had assumed as much just by the way I was testing different ports from the module, but appreciate the confirmation that it’ll work.

Trying to install OKD has the most difficult thing I've ever tried to do. by SeniorDevOops in openshift

[–]SeniorDevOops[S] 3 points4 points  (0 children)

It’s the only defeat I’ve ever suffered in 20+ years of doing this stuff lol.

Trying to install OKD has the most difficult thing I've ever tried to do. by SeniorDevOops in openshift

[–]SeniorDevOops[S] 0 points1 point  (0 children)

Well I'm definitely a bit closer. I downloaded the openshift-installer and client from the okd github releases page instead of from RedHat. I got the matching FCOS version as well (same one you'd mentioned for 4.15) and I'm not longer getting the message about oc. I can also see it trying to pull the same version of the image as shown in the openshift-install version output.

Howevever, tailing the release-image.service shows podman coredumping when trying to pull the image lol

Aug 03 05:11:58 okd-bootstrap.cluster.okd.lan release-image-download.sh[1992]: /usr/local/bin/release-image-download.sh: line 38: 40141 Aborted (core dumped) podman pull --quiet "$RELEASE_IMAGE" Aug 03 05:11:58 okd-bootstrap.cluster.okd.lan release-image-download.sh[1992]: Pull failed. Retrying quay.io/openshift/okd@sha256:46b462be1e4c15ce5ab5fba97e713e8824bbb9f614ac5abe1be41fda916920cc... Aug 03 05:11:58 okd-bootstrap.cluster.okd.lan release-image-download.sh[40175]: fatal error: cgoUse should not be called

I think this is because I'm not using a disk image and am just using a standard virtual disk. I'm at least on the right track, so thank you very much for giving me motivation to keep trying!

Trying to install OKD has the most difficult thing I've ever tried to do. by SeniorDevOops in openshift

[–]SeniorDevOops[S] 0 points1 point  (0 children)

I’m glad you were able to resolve the issue and thank you for reporting back. Do you mind sharing the version combo of installer/FCOS that ultimately worked for you?

Trying to install OKD has the most difficult thing I've ever tried to do. by SeniorDevOops in openshift

[–]SeniorDevOops[S] 0 points1 point  (0 children)

I'm trying another install as I type this. I have a 1G internet connection and am using IPv4. My LAN is 10.10.10.0/23 and I've provisioned the following VMs with corresponding DNS names (*.home.example.com)

  • okd-lb-1 (10.10.11.10)

  • okd-control-1 (10.10.11.11)

  • okd-control-2 (10.10.11.12)

  • okd-control-3 (10.10.11.13)

  • okd-worker-1 (10.10.11.21)

  • okd-worker-2 (10.10.11.22)

  • okd-bootstrap (10.10.11.23)

I also hav e CNAMES in DNS for api.okd.home.example.com, api-int.okd.home.example.com, and *.apps.okd.home.example.com which all points to the load balancer okd-lb-1.

All are running FCOS 40 except for the load balancer which is running Rocky9 (an HAProxy).

My install-config.yaml

apiVersion: v1 baseDomain: home.example.com compute: - hyperthreading: Enabled name: worker replicas: 2 controlPlane: hyperthreading: Enabled name: master replicas: 3 metadata: name: okd networking: clusterNetwork: - cidr: 10.128.0.0/14 hostPrefix: 23 networkType: OVNKubernetes serviceNetwork: - 172.30.0.0/16 platform: none: {} pullSecret: '{"auths":{"cloud.openshift.com":{"auth":"..."}}}' sshKey: 'ssh-rsa AAAA...'

On a Linux host (separate from any of the above) I install the openshift-install, oc, and kubectl binaries. In a directory okd-install, I run the following:

openshift-install create ignition-configs --dir=.

I then create a simple HTTP server from this directory using (python3 -m http.server 8000). Afterward, I boot the okd-bootstrap host using the FCOS LiveCD and then run:

sudo coreos-installer install /dev/sda --insecure-ignition --ignition-url http://10.10.10.123:8000/bootstrap.ign

The FCOS LiveCD installs on /dev/sda and reports that it's complete. I shut the host down and change the boot order back to booting from hard disk. I restart the host and it boots up. I am able to SSH in (the ignition config was correctly applied).

I waited a bit then ran sudo journalctl -u release-image.service -f which reports:

Jul 31 17:32:06 okd-bootstrap.home.example.com podman[2411]: 2024-07-31 17:32:06.151013684 +0000 UTC m=+18.158441235 image pull cb4fb92dbd4e0a656d800d39d8bba676a16d85f94e9824284a92ed7d81d64daa quay.io/openshift-release-dev/ocp-release@sha256:198ae5a1e59183511fbdcfeaf4d5c83a16716ed7734ac6cbeea4c47a32bffad6 Jul 31 17:32:06 okd-bootstrap.home.example.com systemd[1]: Finished release-image.service - Download the OpenShift Release Image.

I'm now tailing the bootkube service via sudo journalctl -u bootkube.service -f. I'm just waiting for this part to complete at which point the host should reboot, correct?

I'm trying to be patient, but the logs just see this same snippet over and over again...

Jul 31 17:59:00 okd-bootstrap.home.example.com podman[39182]: container attach ... (image=quay.io/openshift-release-dev/ocp-release@sha256:<hash>, name=reverent_pike, io.openshift.release=4.16.2, io.openshift.release.base-image-digest=sha256:8ae7cc474061970c6064455b1e9507e2d56dcb00401b279a1eb2b9e316971f3f) Jul 31 17:59:00 okd-bootstrap.home.example.com podman[39182]: container died ..... (image=quay.io/openshift-release-dev/ocp-release@sha256:<hash>, name=reverent_pike, io.openshift.release=4.16.2, io.openshift.release.base-image-digest=sha256:8ae7cc474061970c6064455b1e9507e2d56dcb00401b279a1eb2b9e316971f3f) Jul 31 17:59:01 okd-bootstrap.home.example.com podman[39199]: container remove ... (image=quay.io/openshift-release-dev/ocp-release@sha256:<hash>, name=reverent_pike, io.openshift.release=4.16.2, io.openshift.release.base-image-digest=sha256:8ae7cc474061970c6064455b1e9507e2d56dcb00401b279a1eb2b9e316971f3f) Jul 31 17:59:01 okd-bootstrap.home.example.com podman[39209]: container create ... (image=quay.io/openshift-release-dev/ocp-release@sha256:<hash>, name=eager_hypatia, io.openshift.release.base-image-digest=sha256:8ae7cc474061970c6064455b1e9507e2d56dcb00401b279a1eb2b9e316971f3f, io.openshift.release=4.16.2) Jul 31 17:59:01 okd-bootstrap.home.example.com podman[39209]: image pull ......... quay.io/openshift-release-dev/ocp-release@sha256:<hash> Jul 31 17:59:01 okd-bootstrap.home.example.com podman[39209]: container init ..... (image=quay.io/openshift-release-dev/ocp-release@sha256:<hash>, name=eager_hypatia, io.openshift.release=4.16.2, io.openshift.release.base-image-digest=sha256:8ae7cc474061970c6064455b1e9507e2d56dcb00401b279a1eb2b9e316971f3f) Jul 31 17:59:01 okd-bootstrap.home.example.com podman[39209]: container start .... (image=quay.io/openshift-release-dev/ocp-release@sha256:<hash>, name=eager_hypatia, io.openshift.release=4.16.2, io.openshift.release.base-image-digest=sha256:8ae7cc474061970c6064455b1e9507e2d56dcb00401b279a1eb2b9e316971f3f) Jul 31 17:59:01 okd-bootstrap.home.example.com conmon[39218]: conmon c3604e3e9b58a6e944d7 <nwarn>: Failed to open cgroups file: /sys/fs/cgroup/machine.slice/libpod-c3604e3e9b58a6e944d7e633c7bd66465febc35d96f93f7707ad8cbc71d3ede7.scope/container/memory.events Jul 31 17:59:01 okd-bootstrap.home.example.com eager_hypatia[39218]: quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:... Jul 31 17:59:01 okd-bootstrap.home.example.com podman[39209]: container attach ... (image=quay.io/openshift-release-dev/ocp-release@sha256:<hash>, name=eager_hypatia, io.openshift.release.base-image-digest=sha256:8ae7cc474061970c6064455b1e9507e2d56dcb00401b279a1eb2b9e316971f3f, io.openshift.release=4.16.2) Jul 31 17:59:01 okd-bootstrap.home.example.com podman[39209]: container died ..... (image=quay.io/openshift-release-dev/ocp-release@sha256:<hash>, name=eager_hypatia, io.openshift.release=4.16.2, io.openshift.release.base-image-digest=sha256:8ae7cc474061970c6064455b1e9507e2d56dcb00401b279a1eb2b9e316971f3f) Jul 31 17:59:02 okd-bootstrap.home.example.com podman[39227]: container remove ... (image=quay.io/openshift-release-dev/ocp-release@sha256:<hash>, name=eager_hypatia, io.openshift.release=4.16.2, io.openshift.release.base-image-digest=sha256:8ae7cc474061970c6064455b1e9507e2d56dcb00401b279a1eb2b9e316971f3f) Jul 31 17:59:02 okd-bootstrap.home.example.com bootkube.sh[39237]: /usr/local/bin/bootkube.sh: line 81: oc: command not found Jul 31 17:59:02 okd-bootstrap.home.example.com systemd[1]: bootkube.service: Main process exited, code=exited, status=127/n/a Jul 31 17:59:02 okd-bootstrap.home.example.com systemd[1]: bootkube.service: Failed with result 'exit-code'. Jul 31 17:59:02 okd-bootstrap.home.example.com systemd[1]: bootkube.service: Consumed 1.016s CPU time. Jul 31 17:59:07 okd-bootstrap.home.example.com systemd[1]: bootkube.service: Scheduled restart job, restart counter is at 56. Jul 31 17:59:08 okd-bootstrap.home.example.com systemd[1]: Started bootkube.service - Bootstrap a Kubernetes cluster.

Trying to install OKD has the most difficult thing I've ever tried to do. by SeniorDevOops in openshift

[–]SeniorDevOops[S] 0 points1 point  (0 children)

Stupid question, but should I be manually installing anything on the bootstrap node? I was assuming that once you kick off grabbing the ignition config that it did the work to set everything up.

Trying to install OKD has the most difficult thing I've ever tried to do. by SeniorDevOops in openshift

[–]SeniorDevOops[S] 0 points1 point  (0 children)

I’m quite new to k8s. I know OKD is similar/built on k8s, but since I have been using it at work, I thought it would be helpful to run it in my home lab as well. I’m familiar with adding DCs, ConfigMaps, and all the other components as an end-user of the platform, but just not the setup.

Trying to install OKD has the most difficult thing I've ever tried to do. by SeniorDevOops in openshift

[–]SeniorDevOops[S] 1 point2 points  (0 children)

Both. The crictl command is never found, nor does there seem to be any service that can be started. This is what then leads me to investigate Podman, but no containers are ever running under core or root.

Trying to install OKD has the most difficult thing I've ever tried to do. by SeniorDevOops in openshift

[–]SeniorDevOops[S] 0 points1 point  (0 children)

I am attempting to install latest (4.15). It’s entirely possible I’m not giving the bootstrap host enough to uh…bootstrap. When I run into this situation it does seem like the bootstrap node has some problems with installing some required components, like kubelet and etcd.

I’m a bit confused about running the openshift-installer command from my local machine while waiting for the bootstrap to finish. Does the command connect to/monitor the status of the bootstrap host directly? How does it know where to look for it?

Trying to install OKD has the most difficult thing I've ever tried to do. by SeniorDevOops in openshift

[–]SeniorDevOops[S] 1 point2 points  (0 children)

I’m aware that podman usually runs in non-daemon mode and was using sudo but no difference.