actions-runner-controller: failed calling webhook "mutate.runnerdeployment.actions.summerwind.dev ... lscontext deadline exceeded

Controller Version

actions-runner-controller-0.19.0

Helm Chart Version

0.24.0

Deployment Method

Helm

Checks

  • This isn’t a question or user support case (For Q&A and community support, go to Discussions. It might also be a good idea to contract with any of contributors and maintainers if your business is so critical and therefore you need priority support
  • I’ve read releasenotes before submitting this issue and I’m sure it’s not due to any recently-introduced backward-incompatible changes
  • My actions-runner-controller version (v0.x.y) does support the feature
  • I’ve already upgraded ARC to the latest and it didn’t fix the issue

Resource Definitions

apiVersion: actions.summerwind.dev/v1alpha1
kind: Runner
metadata:
  name: example-runner
spec:
  repository: my-repo
  env: []

To Reproduce

1. I followed the steps to install runner actions from: https://github.com/actions-runner-controller/actions-runner-controller#installation. (I used a personal token access)
2. kubectl --namespace self-hosted-runners apply -f runner.yaml

got: Error from server (InternalError): error when creating "self-hosted-runner.yaml": Internal error occurred: failed calling webhook "mutate.runnerdeployment.actions.summerwind.dev": Post "https://actions-runner-controller-webhook.actions-runner-system.svc:443/mutate-actions-summerwind-dev-v1alpha1-runnerdeployment?timeout=10s": context deadline exceeded

Describe the bug

Im trying to register a self-hoster runner on k8s and got a failling error when it tries to call mutate.runnerdeployment.actions.summerwind.dev. My nodes are bare metal so we are not using cloud. Can someone help to know what am I missing? or how to fix this? The certs controller looks good:

# kubectl --namespace actions-runner-system get all
NAME                                            READY   STATUS    RESTARTS   AGE
pod/actions-runner-controller-bc58c5b87-2dm4q   2/2     Running   0          4m21s

NAME                                                TYPE        CLUSTER-IP      EXTERNAL-IP   PORT(S)    AGE
service/actions-runner-controller-metrics-service   ClusterIP  <.....>   <none>        8443/TCP   4m22s
service/actions-runner-controller-webhook           ClusterIP   <.....>    <none>        443/TCP    4m22s
service/controller-manager-metrics-service          ClusterIP   <.....>    <none>        8443/TCP   17m
service/webhook-service                             ClusterIP   <.....>    <none>        443/TCP    17m

NAME                                        READY   UP-TO-DATE   AVAILABLE   AGE
deployment.apps/actions-runner-controller   1/1     1            1           4m22s

NAME                                                  DESIRED   CURRENT   READY   AGE
replicaset.apps/actions-runner-controller-bc58c5b87   1         1         1       4m22s```

### Describe the expected behavior

Regsiter a self-hosted runner

### Controller Logs

```shell
.

Runner Pod Logs

.

Additional Context

No response

About this issue

  • Original URL
  • State: closed
  • Created 2 years ago
  • Comments: 37 (2 by maintainers)

Most upvoted comments

WORKED after deleting:

kubectl delete validatingwebhookconfiguration validating-webhook-configuration
kubectl delete mutatingwebhookconfiguration mutating-webhook-configuration
kubectl delete mutatingwebhookconfiguration actions-runner-controller-mutating-webhook-configuration
kubectl delete validatingwebhookconfiguration actions-runner-controller-validating-webhook-configuration

@0xF4D3C0D3 Hey! The issue happens when you had conflicting/dangling ARC-related resources from its previous installation. Note that at least one validating and mutating webhook configuration are needed for ARC to function correctly.

For my case, I have to create firewall rule to allow communication from control plane to nodes as I use private kubernetes node.

k8s runner for github is working now! thanks so much for the help

@mikejoseph-ef Hey. I’d appreciate it very much if you could do so!

yes please, it will be nice 😃