rook: Edgefs: rook-edgefs-target CrashLoopBackOff

Is this a bug report or feature request?

  • Bug Report

Deviation from expected behavior:

[root@k8s-master-1 edgefs]# kubectl get po -n rook-edgefs
NAME                               READY   STATUS             RESTARTS   AGE   IP              NODE               NOMINATED NODE
rook-edgefs-mgr-7555f9fb8d-4pzw6   1/1     Running            0          17m   192.168.1.239   k8s-node-3         <none>
rook-edgefs-target-0               2/3     CrashLoopBackOff   8          17m   192.168.1.239   k8s-node-3         <none>
rook-edgefs-target-1               2/3     CrashLoopBackOff   8          17m   192.168.1.218   centos73-218-lb2   <none>
rook-edgefs-target-2               2/3     CrashLoopBackOff   8          17m   192.168.1.219   centos73-219-lb3   <none>
rook-edgefs-target-3               2/3     CrashLoopBackOff   8          17m   192.168.1.217   centos73-217-lb1   <none>
rook-edgefs-target-4               2/3     CrashLoopBackOff   8          17m   192.168.1.233   k8s-node-2         <none>
rook-edgefs-target-5               2/3     CrashLoopBackOff   8          17m   192.168.1.234   k8s-node-1         <none>

name: daemon
  ready: false
  restartCount: 8
  state:
    waiting:
      message: Back-off 5m0s restarting failed container=daemon pod=rook-edgefs-target-0_rook-edgefs(3444b277-09cf-11e9-b890-005056852a45)
      reason: CrashLoopBackOff
[root@k8s-node-3 ~]# docker logs -f ba04d2ce523e
Running in container
Starting ccowd in foreground.
Dec 27 12:05:59.507 k8s-node-3 ccowd[1] error   [13]  reptrans-rd.c:9463 : Couldn't find device name at index 0
Dec 27 12:05:59.508 k8s-node-3 ccowd[1] error   [13]  reptrans-rd.c:9693 : Couldn't fetch VDEVs configuration
Dec 27 12:05:59.508 k8s-node-3 ccowd[1] error   [13]  reptrans-rd.c:10963 : Configure file format, RD version or flags error
Dec 27 12:05:59.508 k8s-node-3 ccowd[1] error   [13]        ccowd.c:1605 : FATAL: reptrans failed to init -2

How to reproduce it (minimal and precise):

Step 1

 kubectl create -f  edgefs/operator.yaml 

Step 2

 kubectl create -f  edgefs/cluster.yaml 
apiVersion: edgefs.rook.io/v1alpha1
kind: Cluster
metadata:
  name: rook-edgefs
  namespace: rook-edgefs
spec:
  edgefsImageName: 192.168.1.52/edgefs/edgefs:1.0.0   
  serviceAccount: rook-edgefs-cluster
  dataDirHostPath: /var/lib/edgefs
  network: 
    serverIfName: "ens160"
    brokerIfName: "ens160"
  storage: 
    useAllNodes: true
    useAllDevices: true
    config:
      rtVerifyChid: "0"           
      lmdbPageSize: "32768"       
      useMetadataOffload: "false"

Environment:

  • OS (e.g. from /etc/os-release): centos 7.3
  • Kernel (e.g. uname -a): 3.10
  • Cloud provider or hardware configuration:
  • Rook version (use rook version inside of a Rook Pod): v0.9.0
  • Kubernetes version (use kubectl version): v1.12.3
  • Kubernetes cluster type (e.g. Tectonic, GKE, OpenShift):
  • Storage backend status (e.g. for Ceph use ceph health in the Rook Ceph toolbox):

About this issue

  • Original URL
  • State: closed
  • Created 6 years ago
  • Comments: 15 (6 by maintainers)

Most upvoted comments

Hi all,

We ran into this same issue running on a VMWare cluster. We arent the administrators of the VMWare cluster itself, so we had to improvise.

In the end we got this fixed by simply defining the disks like this: apiVersion: edgefs.rook.io/v1alpha1 kind: Cluster metadata: name: rook-edgefs namespace: rook-edgefs spec: edgefsImageName: edgefs/edgefs:latest # specify version here, i.e. edgefs/edgefs:1.0.0 etc serviceAccount: rook-edgefs-cluster dataDirHostPath: /var/lib/edgefs dashboard: storage: # cluster level storage configuration and selection useAllNodes: true useAllDevices: false config: sync: "2" nodes: - name: "<ip 1>" devices: - name: "sdc" - name: "<ip 2>" devices: - name: "sdc" - name: "<ip 3>" devices: - name: "sdc" - name: "<ip 4>" devices: - name: "sdc" resources: limits: cpu: "2" memory: "4096Mi"