Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Olares installation failure on debian bookworm #662

Closed
piyushaswani55 opened this issue Nov 24, 2024 · 11 comments
Closed

Olares installation failure on debian bookworm #662

piyushaswani55 opened this issue Nov 24, 2024 · 11 comments

Comments

@piyushaswani55
Copy link

I have debian installed on my raspberry pi and am trying to install olares by following the instructions but unable to install due to below error.

preparing installation environment...

current: root
2024-11-24T19:25:18.457+0530	[Job] [Prepare the System Environment] start ...
2024-11-24T19:25:18.458+0530	[Module] PreCheckOs
2024-11-24T19:25:18.459+0530	[A] LocalHost: PreCheckSupport success (195.815µs)
2024-11-24T19:25:18.460+0530	[A] linux: PatchAppArmor skipped (128.722µs)
2024-11-24T19:25:18.461+0530	[A] linux: RaspbianCheck success (245.037µs)
2024-11-24T19:25:18.462+0530	[A] linux: CorrectHostname success (253.388µs)
Failed to stop systemd-resolved.service: Unit systemd-resolved.service not loaded.
Failed to disable unit: Unit file systemd-resolved.service does not exist.
nameserver
nameserver
nameserver
nameserver
2024-11-24T19:25:28.869+0530	[A] linux: DisableLocalDNS success (10.407280845s)
2024-11-24T19:25:28.870+0530	[INFO] installing and configuring OS dependencies ...
2024-11-24T19:25:28.870+0530	[Module] InstallDeps
Unable to handle repository shortcut 'universe'
2024-11-24T19:25:30.385+0530	add os repo error Failed to exec command: /bin/sh -c sudo -E /bin/bash -c "add-apt-repository universe -y"
Unable to handle repository shortcut 'universe': exit status 1
2024-11-24T19:25:30.387+0530	linux Failed to exec command: /bin/sh -c sudo -E /bin/bash -c "add-apt-repository universe -y"
Unable to handle repository shortcut 'universe': exit status 1
Unable to handle repository shortcut 'universe'
2024-11-24T19:25:36.924+0530	add os repo error Failed to exec command: /bin/sh -c sudo -E /bin/bash -c "add-apt-repository universe -y"
Unable to handle repository shortcut 'universe': exit status 1
2024-11-24T19:25:36.924+0530	linux Failed to exec command: /bin/sh -c sudo -E /bin/bash -c "add-apt-repository universe -y"
Unable to handle repository shortcut 'universe': exit status 1
Unable to handle repository shortcut 'universe'
2024-11-24T19:25:43.455+0530	add os repo error Failed to exec command: /bin/sh -c sudo -E /bin/bash -c "add-apt-repository universe -y"
Unable to handle repository shortcut 'universe': exit status 1
2024-11-24T19:25:43.455+0530	linux Failed to exec command: /bin/sh -c sudo -E /bin/bash -c "add-apt-repository universe -y"
Unable to handle repository shortcut 'universe': exit status 1
2024-11-24T19:25:43.456+0530	[A] linux: PatchOs failed (14.585621472s)
2024-11-24T19:25:43.456+0530	[Job] [Prepare the System Environment] execute failed <nil>
2024-11-24T19:25:43.456+0530	[FATAL] prepare system error: Module[InstallDeps] exec failed:
failed - linux: [A] PatchOs: PatchOs exec failed after 3 retires: Failed to exec command: /bin/sh -c sudo -E /bin/bash -c "add-apt-repository universe -y"
Unable to handle repository shortcut 'universe': exit status 1
bytetrade.io/web3os/installer/cmd/ctl/os.NewCmdPrepare.func1
	/home/runner/work/Installer/Installer/cmd/ctl/os/prepare.go:28
github.com/spf13/cobra.(*Command).execute
	/home/runner/go/pkg/mod/github.com/spf13/cobra@v1.8.0/command.go:987
github.com/spf13/cobra.(*Command).ExecuteC
	/home/runner/go/pkg/mod/github.com/spf13/cobra@v1.8.0/command.go:1115
github.com/spf13/cobra.(*Command).Execute
	/home/runner/go/pkg/mod/github.com/spf13/cobra@v1.8.0/command.go:1039
main.main
	/home/runner/work/Installer/Installer/cmd/main.go:15
runtime.main
	/opt/hostedtoolcache/go/1.22.4/x64/src/runtime/proc.go:271

I even tried raspberrypi os but there was no luck, some hwclock related issues were coming.

I am starting to give up on the installation of olares now. The installation is not straight forward even for someone with good technical knowledge.

@pengpeng
Copy link
Contributor

pengpeng commented Nov 26, 2024

We have fixed the Debian issue in the daily build.

The hwclock issue did not occur on our Raspberry Pi, we are trying to reproduce the issue. Could you provide your Raspberry Pi model and the specific version of the OS?

@piyushaswani55
Copy link
Author

Thank you for the update @pengpeng and thank you @aby913 & @eball for fixing the Debian issue (#668) really quick, really appreciate it.
I confirm that this issue has been resolved with build 1.11.0-20241125 however the source which is added using add-apt-repository is already present in my sources.list file and it's getting duplicated every time I run the Olares installation but that's not a major concern or blocker anymore with the installation.

But the hwclock issue still persists with Debian as well as official Raspberry Pi OS.

I am using Raspberry Pi 4, 8GB RAM model and I am running DietPi on it right now which is based on Debian but even with official Raspberry Pi OS I faced the same issue.
Mine doesn't have any external RTC hardware connected to it, maybe yours has a hardware clock attached to it or maybe you have Raspberry Pi 5 which has an inbuilt RTC hardware to it.

So I think that's why whenever hwclock -w command is executed it fails due to below error.

No usable clock interface found
hwclock: Cannot access the Hardware Clock via any known method

I also tried to use fake-hwclock but there was still no luck. I couldn't figure out anything which would fix the issue or bypass this specific error.

Also, for most of the devices, the time is already in sync with NTP, I don't understand the use or need of a hardware clock here. I could be wrong or unaware of the use case of hwclock with Olares.

@eball
Copy link
Collaborator

eball commented Nov 27, 2024

@piyushaswani55 We'll be running tests on Raspberry Pi 4B soon later. In our case of hwclock is to make sure the time will be accurate even if the machine is shut down for a long time and then restarted. And we are also going to modify this step, let it be not a critical step, then make the NTP service more robust.

@piyushaswani55
Copy link
Author

Well, I guess until then I have no other choice but to wait or arrange a hardware clock and try installation again hoping that some other issue does not arise.

@piyushaswani55
Copy link
Author

But wait do I see a PR (#677) for bypassing this issue already? I guess it won't be a long wait then 😅
appreciation++ @aby913

@piyushaswani55
Copy link
Author

piyushaswani55 commented Nov 27, 2024

Update : I confirm that the above fix is working because I tried the installation again with a custom install.sh script with latest stable Olares version as 1.10.4 and CLI version as mentioned in the fix #677.
I see below logs and this time the installation didn't fail due to hwclock issue and instead used NTP.

CLOCK: time stepped by -0.076406
2024-11-27 20:04:35.262324 (+0530) -0.076406 +/- 0.012974 pool.ntp.org 192.46.210.39 s2 no-leap
2024-11-27T20:04:36.528+0530    [A] nebulae: UpdateNtpDate success (23.026861041s)
hwclock: Cannot access the Hardware Clock via any known method.
hwclock: Use the --verbose option to see the details of our search for an access method.
CLOCK: time stepped by 0.000233

But after hours of installation I got yet another error now, trying to re-install or do something about it.

2024-11-27T21:34:07.883+0530    linux enable k3s failed: Failed to exec command: /bin/sh -c sudo -E /bin/bash -c "systemctl daemon-reload && systemctl enable --now k3s" 
Created symlink /etc/systemd/system/multi-user.target.wants/k3s.service → /etc/systemd/system/k3s.service.
Job for k3s.service failed because the control process exited with error code.
See "systemctl status k3s.service" and "journalctl -xeu k3s.service" for details.: exit status 1
2024-11-27T21:34:24.269+0530    linux enable k3s failed: Failed to exec command: /bin/sh -c sudo -E /bin/bash -c "systemctl daemon-reload && systemctl enable --now k3s" 
Job for k3s.service failed because the control process exited with error code.
See "systemctl status k3s.service" and "journalctl -xeu k3s.service" for details.: exit status 1
2024-11-27T21:34:39.980+0530    linux enable k3s failed: Failed to exec command: /bin/sh -c sudo -E /bin/bash -c "systemctl daemon-reload && systemctl enable --now k3s" 
Job for k3s.service failed because the control process exited with error code.
See "systemctl status k3s.service" and "journalctl -xeu k3s.service" for details.: exit status 1
2024-11-27T21:34:39.980+0530    [A] linux: EnableK3sService failed (1m10.479664329s)
2024-11-27T21:34:39.980+0530    [Job] [Install the System] execute failed <nil>
2024-11-27T21:34:39.981+0530    [FATAL] install Olares error: Module[K3sInitClusterModule] exec failed: 
failed - linux: [A] EnableK3sService: EnableK3sService exec failed after 3 retires: enable k3s failed: Failed to exec command: /bin/sh -c sudo -E /bin/bash -c "systemctl daemon-reload && systemctl enable --now k3s" 
Job for k3s.service failed because the control process exited with error code.
See "systemctl status k3s.service" and "journalctl -xeu k3s.service" for details.: exit status 1
bytetrade.io/web3os/installer/cmd/ctl/os.NewCmdInstallOs.func1
        /home/runner/work/Installer/Installer/cmd/ctl/os/install.go:27
github.com/spf13/cobra.(*Command).execute
        /home/runner/go/pkg/mod/github.com/spf13/cobra@v1.8.0/command.go:987
github.com/spf13/cobra.(*Command).ExecuteC
        /home/runner/go/pkg/mod/github.com/spf13/cobra@v1.8.0/command.go:1115
github.com/spf13/cobra.(*Command).Execute
        /home/runner/go/pkg/mod/github.com/spf13/cobra@v1.8.0/command.go:1039
main.main
        /home/runner/work/Installer/Installer/cmd/main.go:15
runtime.main
        /opt/hostedtoolcache/go/1.22.4/x64/src/runtime/proc.go:271

On manually running the above command

sudo systemctl enable --now k3s
Job for k3s.service failed because the control process exited with error code.
See "systemctl status k3s.service" and "journalctl -xeu k3s.service" for details.
root@linux:~$ sudo systemctl status k3s.service
● k3s.service - Lightweight Kubernetes
     Loaded: loaded (/etc/systemd/system/k3s.service; enabled; preset: enabled)
     Active: activating (auto-restart) (Result: exit-code) since Wed 2024-11-27 22:57:35 IST; 571ms ago
       Docs: https://k3s.io
    Process: 37268 ExecStartPre=/sbin/modprobe br_netfilter (code=exited, status=0/SUCCESS)
    Process: 37269 ExecStartPre=/sbin/modprobe overlay (code=exited, status=0/SUCCESS)
    Process: 37270 ExecStartPre=/usr/local/bin/juicefs summary /olares/rootfs (code=exited, status=0/SUCCESS)
    Process: 37279 ExecStart=/usr/local/bin/k3s $K3S_ROLE $K3S_ARGS $K3S_EXTRA_ARGS $K3S_SERVER_ARGS (code=exited, status=1/FAILURE)
   Main PID: 37279 (code=exited, status=1/FAILURE)
        CPU: 7.704s

@piyushaswani55
Copy link
Author

On manually running the above command

sudo systemctl enable --now k3s
Job for k3s.service failed because the control process exited with error code.
See "systemctl status k3s.service" and "journalctl -xeu k3s.service" for details.
root@linux:~$ sudo systemctl status k3s.service
● k3s.service - Lightweight Kubernetes
     Loaded: loaded (/etc/systemd/system/k3s.service; enabled; preset: enabled)
     Active: activating (auto-restart) (Result: exit-code) since Wed 2024-11-27 22:57:35 IST; 571ms ago
       Docs: https://k3s.io
    Process: 37268 ExecStartPre=/sbin/modprobe br_netfilter (code=exited, status=0/SUCCESS)
    Process: 37269 ExecStartPre=/sbin/modprobe overlay (code=exited, status=0/SUCCESS)
    Process: 37270 ExecStartPre=/usr/local/bin/juicefs summary /olares/rootfs (code=exited, status=0/SUCCESS)
    Process: 37279 ExecStart=/usr/local/bin/k3s $K3S_ROLE $K3S_ARGS $K3S_EXTRA_ARGS $K3S_SERVER_ARGS (code=exited, status=1/FAILURE)
   Main PID: 37279 (code=exited, status=1/FAILURE)
        CPU: 7.704s

Logs from journalctl

Nov 27 23:03:23 linux systemd[1]: Starting k3s.service - Lightweight Kubernetes...
░░ Subject: A start job for unit k3s.service has begun execution
░░ Defined-By: systemd
░░ Support: https://www.debian.org/support
░░ 
░░ A start job for unit k3s.service has begun execution.
░░ 
░░ The job identifier is 38668.
Nov 27 23:03:23 linux juicefs[44289]: +------+---------+------+-------+
Nov 27 23:03:23 linux juicefs[44289]: | PATH |   SIZE  | DIRS | FILES |
Nov 27 23:03:23 linux juicefs[44289]: +------+---------+------+-------+
Nov 27 23:03:23 linux juicefs[44289]: | /    | 4.0 KiB |    1 |     0 |
Nov 27 23:03:23 linux juicefs[44289]: +------+---------+------+-------+
Nov 27 23:03:24 linux k3s[44298]: time="2024-11-27T23:03:24.828251671+05:30" level=info msg="Starting k3s v1.21.5+k3s1 (acad8ef8)"
Nov 27 23:03:24 linux k3s[44298]: time="2024-11-27T23:03:24.829005430+05:30" level=info msg="Cluster bootstrap already complete"
Nov 27 23:03:24 linux k3s[44298]: time="2024-11-27T23:03:24.889868060+05:30" level=info msg="Running kube-apiserver --advertise-address=192.168.1.4 --advertise-port=6443 --allow-privileged=true --anonymous-auth=false --api-audiences=https://kubernetes.default.svc.cluster.local,k3s --authorization-mode=Node,RBAC --bind-address=127.0.0.1 --cert-dir=/var/lib/rancher/k3s/server/tls/temporary-certs --client-ca-file=/var/lib/rancher/k3s/server/tls/client-ca.crt --enable-admission-plugins=NodeRestriction --etcd-cafile=/etc/ssl/etcd/ssl/ca.pem --etcd-certfile=/etc/ssl/etcd/ssl/node-linux.pem --etcd-keyfile=/etc/ssl/etcd/ssl/node-linux-key.pem --etcd-servers=https://192.168.1.4:2379 --insecure-port=0 --kubelet-certificate-authority=/var/lib/rancher/k3s/server/tls/server-ca.crt --kubelet-client-certificate=/var/lib/rancher/k3s/server/tls/client-kube-apiserver.crt --kubelet-client-key=/var/lib/rancher/k3s/server/tls/client-kube-apiserver.key --profiling=false --proxy-client-cert-file=/var/lib/rancher/k3s/server/tls/client-auth-proxy.crt --proxy-client-key-file=/var/lib/rancher/k3s/server/tls/client-auth-proxy.key --requestheader-allowed-names=system:auth-proxy --requestheader-client-ca-file=/var/lib/rancher/k3s/server/tls/request-header-ca.crt --requestheader-extra-headers-prefix=X-Remote-Extra- --requestheader-group-headers=X-Remote-Group --requestheader-username-headers=X-Remote-User --secure-port=6444 --service-account-issuer=https://kubernetes.default.svc.cluster.local --service-account-key-file=/var/lib/rancher/k3s/server/tls/service.key --service-account-signing-key-file=/var/lib/rancher/k3s/server/tls/service.key --service-cluster-ip-range=10.233.0.0/18 --service-node-port-range=30000-32767 --storage-backend=etcd3 --tls-cert-file=/var/lib/rancher/k3s/server/tls/serving-kube-apiserver.crt --tls-private-key-file=/var/lib/rancher/k3s/server/tls/serving-kube-apiserver.key"
Nov 27 23:03:24 linux k3s[44298]: Flag --insecure-port has been deprecated, This flag has no effect now and will be removed in v1.24.
Nov 27 23:03:24 linux k3s[44298]: I1127 23:03:24.892512   44298 server.go:656] external host was not specified, using 192.168.1.4
Nov 27 23:03:24 linux k3s[44298]: I1127 23:03:24.893305   44298 server.go:195] Version: v1.21.5+k3s1
Nov 27 23:03:24 linux k3s[44298]: I1127 23:03:24.905563   44298 shared_informer.go:240] Waiting for caches to sync for node_authorizer
Nov 27 23:03:24 linux k3s[44298]: I1127 23:03:24.910990   44298 plugins.go:158] Loaded 12 mutating admission controller(s) successfully in the following order: NamespaceLifecycle,LimitRanger,ServiceAccount,NodeRestriction,TaintNodesByCondition,Priority,DefaultTolerationSeconds,DefaultStorageClass,StorageObjectInUseProtection,RuntimeClass,DefaultIngressClass,MutatingAdmissionWebhook.
Nov 27 23:03:24 linux k3s[44298]: I1127 23:03:24.911069   44298 plugins.go:161] Loaded 10 validating admission controller(s) successfully in the following order: LimitRanger,ServiceAccount,Priority,PersistentVolumeClaimResize,RuntimeClass,CertificateApproval,CertificateSigning,CertificateSubjectRestriction,ValidatingAdmissionWebhook,ResourceQuota.
Nov 27 23:03:24 linux k3s[44298]: I1127 23:03:24.915893   44298 plugins.go:158] Loaded 12 mutating admission controller(s) successfully in the following order: NamespaceLifecycle,LimitRanger,ServiceAccount,NodeRestriction,TaintNodesByCondition,Priority,DefaultTolerationSeconds,DefaultStorageClass,StorageObjectInUseProtection,RuntimeClass,DefaultIngressClass,MutatingAdmissionWebhook.
Nov 27 23:03:24 linux k3s[44298]: I1127 23:03:24.915967   44298 plugins.go:161] Loaded 10 validating admission controller(s) successfully in the following order: LimitRanger,ServiceAccount,Priority,PersistentVolumeClaimResize,RuntimeClass,CertificateApproval,CertificateSigning,CertificateSubjectRestriction,ValidatingAdmissionWebhook,ResourceQuota.
Nov 27 23:03:25 linux k3s[44298]: I1127 23:03:25.146951   44298 instance.go:283] Using reconciler: lease
Nov 27 23:03:25 linux k3s[44298]: I1127 23:03:25.833831   44298 rest.go:130] the default service ipfamily for this cluster is: IPv4
Nov 27 23:03:28 linux k3s[44298]: W1127 23:03:28.864613   44298 genericapiserver.go:425] Skipping API node.k8s.io/v1alpha1 because it has no resources.
Nov 27 23:03:28 linux k3s[44298]: W1127 23:03:28.926530   44298 genericapiserver.go:425] Skipping API rbac.authorization.k8s.io/v1alpha1 because it has no resources.
Nov 27 23:03:28 linux k3s[44298]: W1127 23:03:28.949028   44298 genericapiserver.go:425] Skipping API scheduling.k8s.io/v1alpha1 because it has no resources.
Nov 27 23:03:28 linux k3s[44298]: W1127 23:03:28.987828   44298 genericapiserver.go:425] Skipping API storage.k8s.io/v1alpha1 because it has no resources.
Nov 27 23:03:29 linux k3s[44298]: W1127 23:03:29.007871   44298 genericapiserver.go:425] Skipping API flowcontrol.apiserver.k8s.io/v1alpha1 because it has no resources.
Nov 27 23:03:29 linux k3s[44298]: W1127 23:03:29.052046   44298 genericapiserver.go:425] Skipping API apps/v1beta2 because it has no resources.
Nov 27 23:03:29 linux k3s[44298]: W1127 23:03:29.052571   44298 genericapiserver.go:425] Skipping API apps/v1beta1 because it has no resources.
Nov 27 23:03:29 linux k3s[44298]: I1127 23:03:29.130984   44298 plugins.go:158] Loaded 12 mutating admission controller(s) successfully in the following order: NamespaceLifecycle,LimitRanger,ServiceAccount,NodeRestriction,TaintNodesByCondition,Priority,DefaultTolerationSeconds,DefaultStorageClass,StorageObjectInUseProtection,RuntimeClass,DefaultIngressClass,MutatingAdmissionWebhook.
Nov 27 23:03:29 linux k3s[44298]: I1127 23:03:29.131595   44298 plugins.go:161] Loaded 10 validating admission controller(s) successfully in the following order: LimitRanger,ServiceAccount,Priority,PersistentVolumeClaimResize,RuntimeClass,CertificateApproval,CertificateSigning,CertificateSubjectRestriction,ValidatingAdmissionWebhook,ResourceQuota.
Nov 27 23:03:29 linux k3s[44298]: time="2024-11-27T23:03:29.255077816+05:30" level=info msg="Running kube-scheduler --address=127.0.0.1 --bind-address=127.0.0.1 --kubeconfig=/var/lib/rancher/k3s/server/cred/scheduler.kubeconfig --leader-elect-lease-duration=60s --leader-elect-renew-deadline=30s --port=10251 --profiling=false --secure-port=0"
Nov 27 23:03:29 linux k3s[44298]: time="2024-11-27T23:03:29.255100002+05:30" level=info msg="Waiting for API server to become available"
Nov 27 23:03:29 linux k3s[44298]: time="2024-11-27T23:03:29.258322353+05:30" level=info msg="Running kube-controller-manager --address=127.0.0.1 --allocate-node-cidrs=true --bind-address=127.0.0.1 --cluster-cidr=10.233.64.0/18 --cluster-signing-kube-apiserver-client-cert-file=/var/lib/rancher/k3s/server/tls/client-ca.crt --cluster-signing-kube-apiserver-client-key-file=/var/lib/rancher/k3s/server/tls/client-ca.key --cluster-signing-kubelet-client-cert-file=/var/lib/rancher/k3s/server/tls/client-ca.crt --cluster-signing-kubelet-client-key-file=/var/lib/rancher/k3s/server/tls/client-ca.key --cluster-signing-kubelet-serving-cert-file=/var/lib/rancher/k3s/server/tls/server-ca.crt --cluster-signing-kubelet-serving-key-file=/var/lib/rancher/k3s/server/tls/server-ca.key --cluster-signing-legacy-unknown-cert-file=/var/lib/rancher/k3s/server/tls/client-ca.crt --cluster-signing-legacy-unknown-key-file=/var/lib/rancher/k3s/server/tls/client-ca.key --kubeconfig=/var/lib/rancher/k3s/server/cred/controller.kubeconfig --leader-elect-lease-duration=60s --leader-elect-renew-deadline=30s --pod-eviction-timeout=3m0s --port=10252 --profiling=false --root-ca-file=/var/lib/rancher/k3s/server/tls/server-ca.crt --secure-port=0 --service-account-private-key-file=/var/lib/rancher/k3s/server/tls/service.key --terminated-pod-gc-threshold=5 --use-service-account-credentials=true"
Nov 27 23:03:29 linux k3s[44298]: time="2024-11-27T23:03:29.265983057+05:30" level=info msg="Node token is available at /var/lib/rancher/k3s/server/token"
Nov 27 23:03:29 linux k3s[44298]: time="2024-11-27T23:03:29.266135057+05:30" level=info msg="To join node to cluster: k3s agent -s https://192.168.1.4:6443 -t ${NODE_TOKEN}"
Nov 27 23:03:29 linux k3s[44298]: time="2024-11-27T23:03:29.274130076+05:30" level=info msg="Wrote kubeconfig /etc/rancher/k3s/k3s.yaml"
Nov 27 23:03:29 linux k3s[44298]: time="2024-11-27T23:03:29.274358705+05:30" level=info msg="Run: k3s kubectl"
Nov 27 23:03:29 linux k3s[44298]: time="2024-11-27T23:03:29.408396761+05:30" level=info msg="Cluster-Http-Server 2024/11/27 23:03:29 http: TLS handshake error from 127.0.0.1:58698: remote error: tls: bad certificate"
Nov 27 23:03:29 linux k3s[44298]: time="2024-11-27T23:03:29.434787168+05:30" level=info msg="Cluster-Http-Server 2024/11/27 23:03:29 http: TLS handshake error from 127.0.0.1:58732: remote error: tls: bad certificate"
Nov 27 23:03:29 linux k3s[44298]: time="2024-11-27T23:03:29.528061261+05:30" level=info msg="certificate CN=linux signed by CN=k3s-server-ca@1732723434: notBefore=2024-11-27 16:03:54 +0000 UTC notAfter=2025-11-27 17:33:29 +0000 UTC"
Nov 27 23:03:29 linux k3s[44298]: time="2024-11-27T23:03:29.556735557+05:30" level=info msg="certificate CN=system:node:linux,O=system:nodes signed by CN=k3s-client-ca@1732723434: notBefore=2024-11-27 16:03:54 +0000 UTC notAfter=2025-11-27 17:33:29 +0000 UTC"
Nov 27 23:03:29 linux k3s[44298]: time="2024-11-27T23:03:29.584173224+05:30" level=info msg="Module overlay was already loaded"
Nov 27 23:03:29 linux k3s[44298]: time="2024-11-27T23:03:29.584278038+05:30" level=info msg="Module nf_conntrack was already loaded"
Nov 27 23:03:29 linux k3s[44298]: time="2024-11-27T23:03:29.584327483+05:30" level=info msg="Module br_netfilter was already loaded"
Nov 27 23:03:29 linux k3s[44298]: time="2024-11-27T23:03:29.584372094+05:30" level=info msg="Module iptable_nat was already loaded"
Nov 27 23:03:29 linux k3s[44298]: W1127 23:03:29.584479   44298 sysinfo.go:203] Nodes topology is not available, providing CPU topology
Nov 27 23:03:29 linux k3s[44298]: time="2024-11-27T23:03:29.654190224+05:30" level=info msg="Connecting to proxy" url="wss://127.0.0.1:6443/v1-k3s/connect"
Nov 27 23:03:29 linux k3s[44298]: time="2024-11-27T23:03:29.665973501+05:30" level=info msg="Handling backend connection request [linux]"
Nov 27 23:03:29 linux k3s[44298]: time="2024-11-27T23:03:29.670420724+05:30" level=info msg="Running kubelet --address=0.0.0.0 --anonymous-auth=false --authentication-token-webhook=true --authorization-mode=Webhook --cgroup-driver=systemd --client-ca-file=/var/lib/rancher/k3s/agent/client-ca.crt --cluster-dns=10.233.0.3 --cluster-domain=cluster.local --cni-bin-dir=/opt/cni/bin --cni-conf-dir=/etc/cni/net.d --config=/etc/rancher/k3s/kubelet.config --container-runtime-endpoint=unix://unix:///run/containerd/containerd.sock --container-runtime=remote --containerd=/run/containerd/containerd.sock --eviction-hard=memory.available<5%,nodefs.available<10% --eviction-minimum-reclaim=imagefs.available=10%,nodefs.available=10% --fail-swap-on=false --healthz-bind-address=127.0.0.1 --hostname-override=linux --kube-reserved=cpu=200m,memory=250Mi,ephemeral-storage=1Gi --kubeconfig=/var/lib/rancher/k3s/agent/kubelet.kubeconfig --network-plugin=cni --node-labels= --pod-manifest-path=/var/lib/rancher/k3s/agent/pod-manifests --read-only-port=0 --resolv-conf=/etc/resolv.conf --serialize-image-pulls=false --system-reserved=cpu=200m,memory=250Mi,ephemeral-storage=1Gi --tls-cert-file=/var/lib/rancher/k3s/agent/serving-kubelet.crt --tls-private-key-file=/var/lib/rancher/k3s/agent/serving-kubelet.key"
Nov 27 23:03:29 linux k3s[44298]: time="2024-11-27T23:03:29.672831446+05:30" level=info msg="Running kube-proxy --cluster-cidr=10.233.64.0/18 --conntrack-max-per-core=0 --conntrack-tcp-timeout-close-wait=0s --conntrack-tcp-timeout-established=0s --healthz-bind-address=127.0.0.1 --hostname-override=linux --kubeconfig=/var/lib/rancher/k3s/agent/kubeproxy.kubeconfig --proxy-mode=ipvs"
Nov 27 23:03:29 linux k3s[44298]: Flag --cni-bin-dir has been deprecated, will be removed along with dockershim.
Nov 27 23:03:29 linux k3s[44298]: Flag --cni-conf-dir has been deprecated, will be removed along with dockershim.
Nov 27 23:03:29 linux k3s[44298]: Flag --containerd has been deprecated, This is a cadvisor flag that was mistakenly registered with the Kubelet. Due to legacy concerns, it will follow the standard CLI deprecation timeline before being removed.
Nov 27 23:03:29 linux k3s[44298]: Flag --network-plugin has been deprecated, will be removed along with dockershim.
Nov 27 23:03:29 linux k3s[44298]: W1127 23:03:29.673824   44298 server.go:220] WARNING: all flags other than --config, --write-config-to, and --cleanup are deprecated. Please begin using a config file ASAP.
Nov 27 23:03:29 linux k3s[44298]: Flag --cni-bin-dir has been deprecated, will be removed along with dockershim.
Nov 27 23:03:29 linux k3s[44298]: Flag --cni-conf-dir has been deprecated, will be removed along with dockershim.
Nov 27 23:03:29 linux k3s[44298]: Flag --network-plugin has been deprecated, will be removed along with dockershim.
Nov 27 23:03:29 linux k3s[44298]: I1127 23:03:29.781041   44298 server.go:436] "Kubelet version" kubeletVersion="v1.21.5+k3s1"
Nov 27 23:03:29 linux k3s[44298]: E1127 23:03:29.807531   44298 node.go:161] Failed to retrieve node info: nodes "linux" is forbidden: User "system:kube-proxy" cannot get resource "nodes" in API group "" at the cluster scope
Nov 27 23:03:29 linux k3s[44298]: I1127 23:03:29.922518   44298 dynamic_cafile_content.go:167] Starting client-ca-bundle::/var/lib/rancher/k3s/agent/client-ca.crt
Nov 27 23:03:29 linux k3s[44298]: W1127 23:03:29.922542   44298 manager.go:159] Cannot detect current cgroup on cgroup v2
Nov 27 23:03:30 linux k3s[44298]: W1127 23:03:30.168629   44298 sysinfo.go:203] Nodes topology is not available, providing CPU topology
Nov 27 23:03:30 linux k3s[44298]: I1127 23:03:30.281139   44298 server.go:660] "--cgroups-per-qos enabled, but --cgroup-root was not specified.  defaulting to /"
Nov 27 23:03:30 linux k3s[44298]: I1127 23:03:30.282515   44298 container_manager_linux.go:291] "Container manager verified user specified cgroup-root exists" cgroupRoot=[]
Nov 27 23:03:30 linux k3s[44298]: I1127 23:03:30.283290   44298 container_manager_linux.go:296] "Creating Container Manager object based on Node Config" nodeConfig={RuntimeCgroupsName: SystemCgroupsName: KubeletCgroupsName: ContainerRuntime:remote CgroupsPerQOS:true CgroupRoot:/ CgroupDriver:systemd KubeletRootDir:/var/lib/kubelet ProtectKernelDefaults:false NodeAllocatableConfig:{KubeReservedCgroupName: SystemReservedCgroupName: ReservedSystemCPUs: EnforceNodeAllocatable:map[pods:{}] KubeReserved:map[cpu:{i:{value:200 scale:-3} d:{Dec:<nil>} s:200m Format:DecimalSI} ephemeral-storage:{i:{value:1073741824 scale:0} d:{Dec:<nil>} s:1Gi Format:BinarySI} memory:{i:{value:262144000 scale:0} d:{Dec:<nil>} s:250Mi Format:BinarySI}] SystemReserved:map[cpu:{i:{value:200 scale:-3} d:{Dec:<nil>} s:200m Format:DecimalSI} ephemeral-storage:{i:{value:1073741824 scale:0} d:{Dec:<nil>} s:1Gi Format:BinarySI} memory:{i:{value:262144000 scale:0} d:{Dec:<nil>} s:250Mi Format:BinarySI}] HardEvictionThresholds:[{Signal:memory.available Operator:LessThan Value:{Quantity:<nil> Percentage:0.05} GracePeriod:0s MinReclaim:<nil>} {Signal:nodefs.available Operator:LessThan Value:{Quantity:<nil> Percentage:0.1} GracePeriod:0s MinReclaim:<nil>}]} QOSReserved:map[] ExperimentalCPUManagerPolicy:none ExperimentalTopologyManagerScope:container ExperimentalCPUManagerReconcilePeriod:10s ExperimentalMemoryManagerPolicy:None ExperimentalMemoryManagerReservedMemory:[] ExperimentalPodPidsLimit:-1 EnforceCPULimits:true CPUCFSQuotaPeriod:100ms ExperimentalTopologyManagerPolicy:none Rootless:false}
Nov 27 23:03:30 linux k3s[44298]: I1127 23:03:30.284032   44298 topology_manager.go:120] "Creating topology manager with policy per scope" topologyPolicyName="none" topologyScopeName="container"
Nov 27 23:03:30 linux k3s[44298]: I1127 23:03:30.284474   44298 container_manager_linux.go:327] "Initializing Topology Manager" policy="none" scope="container"
Nov 27 23:03:30 linux k3s[44298]: I1127 23:03:30.284830   44298 container_manager_linux.go:332] "Creating device plugin manager" devicePluginEnabled=true
Nov 27 23:03:30 linux k3s[44298]: I1127 23:03:30.286139   44298 kubelet.go:404] "Attempting to sync node with API server"
Nov 27 23:03:30 linux k3s[44298]: I1127 23:03:30.286641   44298 kubelet.go:272] "Adding static pod path" path="/var/lib/rancher/k3s/agent/pod-manifests"
Nov 27 23:03:30 linux k3s[44298]: I1127 23:03:30.287203   44298 kubelet.go:283] "Adding apiserver pod source"
Nov 27 23:03:30 linux k3s[44298]: I1127 23:03:30.287618   44298 apiserver.go:42] "Waiting for node sync before watching apiserver pods"
Nov 27 23:03:30 linux k3s[44298]: E1127 23:03:30.294734   44298 remote_runtime.go:86] "Version from runtime service failed" err="rpc error: code = Unimplemented desc = unknown service runtime.v1alpha2.RuntimeService"
Nov 27 23:03:30 linux k3s[44298]: E1127 23:03:30.294919   44298 kuberuntime_manager.go:208] "Get runtime version failed" err="get remote runtime typed version failed: rpc error: code = Unimplemented desc = unknown service runtime.v1alpha2.RuntimeService"
Nov 27 23:03:30 linux k3s[44298]: E1127 23:03:30.294990   44298 server.go:288] "Failed to run kubelet" err="failed to run Kubelet: failed to create kubelet: get remote runtime typed version failed: rpc error: code = Unimplemented desc = unknown service runtime.v1alpha2.RuntimeService"
Nov 27 23:03:30 linux systemd[1]: k3s.service: Main process exited, code=exited, status=1/FAILURE
░░ Subject: Unit process exited
░░ Defined-By: systemd
░░ Support: https://www.debian.org/support
░░ 
░░ An ExecStart= process belonging to unit k3s.service has exited.
░░ 
░░ The process' exit code is 'exited' and its exit status is 1.
Nov 27 23:03:30 linux systemd[1]: k3s.service: Failed with result 'exit-code'.
░░ Subject: Unit failed
░░ Defined-By: systemd
░░ Support: https://www.debian.org/support
░░ 
░░ The unit k3s.service has entered the 'failed' state with result 'exit-code'.
Nov 27 23:03:30 linux systemd[1]: Failed to start k3s.service - Lightweight Kubernetes.
░░ Subject: A start job for unit k3s.service has failed
░░ Defined-By: systemd
░░ Support: https://www.debian.org/support
░░ 
░░ A start job for unit k3s.service has finished with a failure.
░░ 
░░ The job identifier is 38668 and the job result is failed.
Nov 27 23:03:30 linux systemd[1]: k3s.service: Consumed 6.991s CPU time.
░░ Subject: Resources consumed by unit runtime
░░ Defined-By: systemd
░░ Support: https://www.debian.org/support
░░ 
░░ The unit k3s.service completed and consumed the indicated resources.

@piyushaswani55
Copy link
Author

I would close this as the original issue is resolved but before that I just want to leave my notes about Olares as I won't be pursuing the installation any further.

  • The offering looks very promising from a high-level, I hope it does deliver whatever it claims and if & when it does it would be a huge win for self-hosting as well as open-source communities.
  • For now, the project is in very early stages with a complex installation method having multiple failure points and re-installation is a drag. Also, I observed that it comes with a lot of packages which it tries to install and beefs up to 64GB which goes beyond my SD card attached to my Raspberry Pi.
  • So, I wish the installation was easier and only the required packages were added on need basis only. I think this could be offered as a separate OS itself or a containerised system so that the existing system isn't affected and is easy to lift and shift.
  • Also, I think there should be some sort of demo or videos about the usage of Olares so that people can get a hand-on experience. At-least that's what I was looking for when I wanted to install Olares, I was so excited to get my hands on it from the moment I got to know about it but am a little disappointed after spending so much time and energy for nothing.

In the end, I don't want to discourage anyone. I really appreciate all the hard work you guys are putting into it and the active participation to quickly fix all the incoming issues is really commendable. I know it takes a lot of time to build things specially something like this, it takes much-much more time that the time we put into installing or trying it as a user/consumer.

So, I think all of it was expected since the project is in its initial phase, but I am sure it will grow to become as stable as possible someday soon and hopefully I would give it another shot.

Until then, thank you and happy coding.

@dkeven
Copy link
Contributor

dkeven commented Nov 28, 2024

@piyushaswani55

Thank you for your interest in our project, we deeply appreciate the time and effort you've spent trying to install Olares, and sharing your experience and thoughts with us, we're truly sorry for the frustration the failure points have caused. Your patience and understanding means a lot to us, we're committed to making the installation process smoother and more reliable, and your feedback helps us identify the areas for improvement.

As for the failure point of your last installation, we believe it's most likely caused by an already installed Docker on your environment, which disables the cri plugin of the containerd by default, a scenario we hadn't taken into consideration, can you help us confirm that?

Thank you again for your support!

@piyushaswani55
Copy link
Author

piyushaswani55 commented Nov 28, 2024

Thank you @dkeven, I will be eagerly waiting for the installation and the application as a whole to get stable which I am sure would be very soon looking at Olares teams' dedication.

Yes, you are right, my machine already has docker and docker compose installed so your hypothesis could be correct here.
And now that I know what the root cause could be maybe I'll give it another shot over the weekend if I get some spare time.

@dkeven
Copy link
Contributor

dkeven commented Dec 2, 2024

@piyushaswani55 very glad to hear that, looking forward to your next shot to the installation program, and more importantly, to Olares OS itself😊

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

4 participants