You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Kepler fails on all Ubuntu 20 hosts in my K8s cluster, producing the following logs:
$ kubectl logs -n monitoring kepler-6hrms
WARNING: failed to read int from file: open /sys/devices/system/cpu/cpu0/online: no such file or directory
I1128 17:15:12.843579 1 exporter.go:103] Kepler running on version: v0.7.12-dirty
I1128 17:15:12.844340 1 config.go:293] using gCgroup ID in the BPF program: true
I1128 17:15:12.844406 1 config.go:295] kernel version: 5.4
I1128 17:15:12.844693 1 power.go:78] Unable to obtain power, use estimate method
I1128 17:15:12.844720 1 redfish.go:169] failed to get redfish credential file path
I1128 17:15:12.853436 1 acpi.go:71] Could not find any ACPI power meter path. Is it a VM?
I1128 17:15:12.853459 1 power.go:79] using none to obtain power
E1128 17:15:12.853478 1 accelerator.go:154] [DUMMY] doesn't contain GPU
E1128 17:15:12.853507 1 exporter.go:154] failed to init GPU accelerators: no devices found
WARNING: failed to read int from file: open /sys/devices/system/cpu/cpu0/online: no such file or directory
I1128 17:15:12.854860 1 exporter.go:84] Number of CPUs: 2
panic: runtime error: invalid memory address or nil pointer dereference
[signal SIGSEGV: segmentation violation code=0x1 addr=0x0 pc=0x87b273]
goroutine 1 [running]:
github.com/sustainable-computing-io/kepler/pkg/bpf.(*hardwarePerfEvents).close(0x0)
/workspace/pkg/bpf/exporter.go:274 +0x13
github.com/sustainable-computing-io/kepler/pkg/bpf.(*exporter).Detach(0xc0001a4000)
/workspace/pkg/bpf/exporter.go:195 +0x15a
github.com/sustainable-computing-io/kepler/pkg/bpf.NewExporter()
/workspace/pkg/bpf/exporter.go:58 +0x13e
main.main()
/workspace/cmd/exporter/exporter.go:159 +0x86b
Pods running as expected against U22 hosts in the same cluster
What did you expect to happen?
Kepler runs on Ubuntu 20 hosts
How can we reproduce it (as minimally and precisely as possible)?
Install via helm at latest version against a cluster with virtualised Ubuntu 20 nodes
What happened?
Kepler fails on all Ubuntu 20 hosts in my K8s cluster, producing the following logs:
Pods running as expected against U22 hosts in the same cluster
What did you expect to happen?
Kepler runs on Ubuntu 20 hosts
How can we reproduce it (as minimally and precisely as possible)?
Install via helm at latest version against a cluster with virtualised Ubuntu 20 nodes
Anything else we need to know?
Virtualised hosts running on Hyper-V
Kepler image tag
Kubernetes version
Server Version: v1.31.2
Cloud provider or bare metal
OS version
The text was updated successfully, but these errors were encountered: