rook: Edgefs: rook-edgefs-target CrashLoopBackOff
Is this a bug report or feature request?
- Bug Report
Deviation from expected behavior:
[root@k8s-master-1 edgefs]# kubectl get po -n rook-edgefs
NAME READY STATUS RESTARTS AGE IP NODE NOMINATED NODE
rook-edgefs-mgr-7555f9fb8d-4pzw6 1/1 Running 0 17m 192.168.1.239 k8s-node-3 <none>
rook-edgefs-target-0 2/3 CrashLoopBackOff 8 17m 192.168.1.239 k8s-node-3 <none>
rook-edgefs-target-1 2/3 CrashLoopBackOff 8 17m 192.168.1.218 centos73-218-lb2 <none>
rook-edgefs-target-2 2/3 CrashLoopBackOff 8 17m 192.168.1.219 centos73-219-lb3 <none>
rook-edgefs-target-3 2/3 CrashLoopBackOff 8 17m 192.168.1.217 centos73-217-lb1 <none>
rook-edgefs-target-4 2/3 CrashLoopBackOff 8 17m 192.168.1.233 k8s-node-2 <none>
rook-edgefs-target-5 2/3 CrashLoopBackOff 8 17m 192.168.1.234 k8s-node-1 <none>
name: daemon
ready: false
restartCount: 8
state:
waiting:
message: Back-off 5m0s restarting failed container=daemon pod=rook-edgefs-target-0_rook-edgefs(3444b277-09cf-11e9-b890-005056852a45)
reason: CrashLoopBackOff
[root@k8s-node-3 ~]# docker logs -f ba04d2ce523e
Running in container
Starting ccowd in foreground.
Dec 27 12:05:59.507 k8s-node-3 ccowd[1] error [13] reptrans-rd.c:9463 : Couldn't find device name at index 0
Dec 27 12:05:59.508 k8s-node-3 ccowd[1] error [13] reptrans-rd.c:9693 : Couldn't fetch VDEVs configuration
Dec 27 12:05:59.508 k8s-node-3 ccowd[1] error [13] reptrans-rd.c:10963 : Configure file format, RD version or flags error
Dec 27 12:05:59.508 k8s-node-3 ccowd[1] error [13] ccowd.c:1605 : FATAL: reptrans failed to init -2
How to reproduce it (minimal and precise):
Step 1
kubectl create -f edgefs/operator.yaml
Step 2
kubectl create -f edgefs/cluster.yaml
apiVersion: edgefs.rook.io/v1alpha1
kind: Cluster
metadata:
name: rook-edgefs
namespace: rook-edgefs
spec:
edgefsImageName: 192.168.1.52/edgefs/edgefs:1.0.0
serviceAccount: rook-edgefs-cluster
dataDirHostPath: /var/lib/edgefs
network:
serverIfName: "ens160"
brokerIfName: "ens160"
storage:
useAllNodes: true
useAllDevices: true
config:
rtVerifyChid: "0"
lmdbPageSize: "32768"
useMetadataOffload: "false"
Environment:
- OS (e.g. from /etc/os-release): centos 7.3
- Kernel (e.g.
uname -a): 3.10 - Cloud provider or hardware configuration:
- Rook version (use
rook versioninside of a Rook Pod): v0.9.0 - Kubernetes version (use
kubectl version): v1.12.3 - Kubernetes cluster type (e.g. Tectonic, GKE, OpenShift):
- Storage backend status (e.g. for Ceph use
ceph healthin the Rook Ceph toolbox):
About this issue
- Original URL
- State: closed
- Created 6 years ago
- Comments: 15 (6 by maintainers)
Hi all,
We ran into this same issue running on a VMWare cluster. We arent the administrators of the VMWare cluster itself, so we had to improvise.
In the end we got this fixed by simply defining the disks like this:
apiVersion: edgefs.rook.io/v1alpha1 kind: Cluster metadata: name: rook-edgefs namespace: rook-edgefs spec: edgefsImageName: edgefs/edgefs:latest # specify version here, i.e. edgefs/edgefs:1.0.0 etc serviceAccount: rook-edgefs-cluster dataDirHostPath: /var/lib/edgefs dashboard: storage: # cluster level storage configuration and selection useAllNodes: true useAllDevices: false config: sync: "2" nodes: - name: "<ip 1>" devices: - name: "sdc" - name: "<ip 2>" devices: - name: "sdc" - name: "<ip 3>" devices: - name: "sdc" - name: "<ip 4>" devices: - name: "sdc" resources: limits: cpu: "2" memory: "4096Mi"