datadog-agent: Tagger error on Kubernetes
Describe what happened: I am trying to set up log collection for Docker containers on a Kubernetes cluster.
Snippet from my config:
logs:
- type: docker
image: "datadog/agent"
servie: datadog
source: datadog
Logs are not being sent to Datadog, and I see a lot of these errors in the DD container log:
2018-01-10 16:47:50 UTC | WARN | (tagger.go:248 in Tag) | error collecting from kubelet: container docker://3cacb7a1688be922adcb89f39950d4526cafb1c0e92ee050f630fe0e90132b0b not found in podlist
Describe what you expected: Working log collection.
Steps to reproduce the issue: Deploy DD agent on Kuberentes with Daemonset.
Additional environment details (Operating System, Cloud provider, etc):
Kubernetes 1.8.6 Docker 0.13
agent status
===================== Agent (v6.0.0-beta.7) =====================Status date: 2018-01-10 16:49:09.760419 UTC Pid: 7 Python Version: 2.7.14 Logs: Check Runners: 10 Log Level: info
Paths
Config File: /etc/datadog-agent/datadog.yaml
conf.d: /etc/datadog-agent/conf.d
checks.d: /etc/datadog-agent/checks.d
Clocks
NTP offset: 0.00326513 s
System UTC time: 2018-01-10 16:49:09.760419 UTC
Host Info
bootTime: 2018-01-09 20:04:23.000000 UTC
kernelVersion: 4.4.65-k8s
os: linux
platform: debian
platformFamily: debian
platformVersion: 9.3
procs: 63
uptime: 74247
virtualizationRole: guest
virtualizationSystem: xen
Hostnames
ec2-hostname: ip-172-20-174-85.ec2.internal
hostname: i-0d84be79a0113f36b
instance-id: i-0d84be79a0113f36b
socket-fqdn: dd-agent-j6b4h
socket-hostname: dd-agent-j6b4h
========= Collector
Running Checks
cpu
---
Total Runs: 29
Metrics: 6, Total Metrics: 168
Events: 0, Total Events: 0
Service Checks: 0, Total Service Checks: 0
disk
----
Total Runs: 29
Metrics: 160, Total Metrics: 4640
Events: 0, Total Events: 0
Service Checks: 0, Total Service Checks: 0
docker
------
Total Runs: 29
Metrics: 266, Total Metrics: 7342
Events: 0, Total Events: 5
Service Checks: 1, Total Service Checks: 29
file_handle
-----------
Total Runs: 29
Metrics: 1, Total Metrics: 29
Events: 0, Total Events: 0
Service Checks: 0, Total Service Checks: 0
io
--
Total Runs: 29
Metrics: 52, Total Metrics: 1472
Events: 0, Total Events: 0
Service Checks: 0, Total Service Checks: 0
kube_dns
--------
Total Runs: 29
Metrics: 41, Total Metrics: 1189
Events: 0, Total Events: 0
Service Checks: 0, Total Service Checks: 0
load
----
Total Runs: 29
Metrics: 6, Total Metrics: 174
Events: 0, Total Events: 0
Service Checks: 0, Total Service Checks: 0
memory
------
Total Runs: 29
Metrics: 14, Total Metrics: 406
Events: 0, Total Events: 0
Service Checks: 0, Total Service Checks: 0
network
-------
Total Runs: 29
Metrics: 20, Total Metrics: 580
Events: 0, Total Events: 0
Service Checks: 0, Total Service Checks: 0
ntp
---
Total Runs: 29
Metrics: 1, Total Metrics: 28
Events: 0, Total Events: 0
Service Checks: 1, Total Service Checks: 29
uptime
------
Total Runs: 29
Metrics: 1, Total Metrics: 29
Events: 0, Total Events: 0
Service Checks: 0, Total Service Checks: 0
Loading Errors
docker_daemon
-------------
Core Check Loader:
Check docker_daemon not found in Catalog
JMX Check Loader:
check is not a jmx check, or unable to determine if it's so
Python Check Loader:
No module named docker_daemon
======== JMXFetch
Initialized checks
no checks
Failed checks
no checks
========= Forwarder
CheckRunsV1: 29 IntakeV1: 6 RetryQueueSize: 0 Success: 64 TimeseriesV1: 29
API Keys status
https://6-0-0-app.agent.datadoghq.com,*************************4aa08: API Key valid
========= DogStatsD
Checks Metric Sample: 16694 Event: 6 Events Flushed: 6 Number Of Flushes: 29 Series Flushed: 12074 Service Check: 377 Service Checks Flushed: 393 Dogstatsd Metric Sample: 697
About this issue
- Original URL
- State: closed
- Created 6 years ago
- Reactions: 3
- Comments: 17 (6 by maintainers)
I got the same error.
stable/datadog)values.yaml
@mfpierre when will the 6.1.1 release be available as a Helm chart? Why doesn’t every new release result in a new version of the Helm chart? Related to #1447
Hi everyone, we have a fix https://github.com/DataDog/datadog-agent/pull/1345 that should resolve the logging issues and will be included in the next 6.1 release.
We’re aware of an issue on kubernetes side where static pods are not correctly updated in the kubelet podlist (#1447) but we’ll keep an eye on this to see if there could be other issues.
Hi everyone,
A little message just to signal you that my DataDog agent is reporting the same kind of errors. I have an opened ticket that talk about that here: https://help.datadoghq.com/hc/en-us/requests/130699 (with logs from the agent)
I have a k8s 1.9.2 cluster. My Datadog agent is the v6 one and it is deployed thanks to the stable chart you provide.
I deploy my agent thanks to this command:
with the following
values.yml:I’m facing the same issue on beta9, except I’m not getting logs even with the same config as from 2 comments above, i.e. lots of the following in the agent logs:Pod metrics/tagging seem to work fine though.Edit: disregard everything, I wasn’t paying attention close enough, and I was mounting the checks configs in
/opt/datadog-agent/conf.dinstead of just/conf.d.When I mount them in the correct location, I get some different errors though:
They’re under these weird paths because that’s how k8s mounts configmaps as files. Despite these errors it seems that logs-agent does attempt to read container logs though:
It seems like our Docker (1.12.6, CoreOS) is too old? @macat if you don’t mind me asking, what Docker version are you running? In your initial comment you said 0.13, but that seems wrong, I’m guessing it’s 1.13.x?
OK, thank you. Actually, I realized that error does not prevent my logs to be shipped. After removing the image filter, all logs are being transferred to DataDog, which is great. The agent even picked up the Kubernetes attributes from the containers, so I’m very happy.