hi all
Am trying to setup Kubernetes cluster (i have done offline install of K8) on RHEL 7.3. I see my docker images are disappearing randomly, any help around this ?
docker info
Containers: 21
Running: 12
Paused: 0
Stopped: 9
Images: 9
Server Version: 1.12.6
Storage Driver: devicemapper
Pool Name: docker-253:8-12648726-pool
Pool Blocksize: 65.54 kB
Base Device Size: 10.74 GB
Backing Filesystem: xfs
Data file: /dev/loop0
Metadata file: /dev/loop1
Data Space Used: 1.226 GB
Data Space Total: 107.4 GB
Data Space Available: 878.9 MB
Metadata Space Used: 1.991 MB
Metadata Space Total: 2.147 GB
Metadata Space Available: 878.9 MB
Thin Pool Minimum Free Space: 10.74 GB
Udev Sync Supported: true
Deferred Removal Enabled: false
Deferred Deletion Enabled: false
Deferred Deleted Device Count: 0
Data loop file: /var/lib/docker/devicemapper/devicemapper/data
WARNING: Usage of loopback devices is strongly discouraged for production use. Use `--storage-opt dm.thinpooldev` to specify a custom block storage device.
Metadata loop file: /var/lib/docker/devicemapper/devicemapper/metadata
Library Version: 1.02.135-RHEL7 (2016-11-16)
Logging Driver: journald
Cgroup Driver: systemd
Plugins:
Volume: local
Network: bridge host null overlay
Authorization: rhel-push-plugin
Swarm: inactive
Runtimes: docker-runc runc
Default Runtime: docker-runc
Security Options: seccomp
Kernel Version: 3.10.0-514.26.1.el7.x86_64
Operating System: Red Hat Enterprise Linux
OSType: linux
Architecture: x86_64
Number of Docker Hooks: 2
CPUs: 4
Total Memory: 7.64 GiB
Name: vstjenmast01
ID: LNLB:ZA3K:FVHH:NAR6:CBPI:CY46:J563:X5EV:YP2U:37UZ:DKEO:QLDU
Docker Root Dir: /var/lib/docker
Debug Mode (client): false
Debug Mode (server): false
Registry: https://registry.access.redhat.com/v1/
Insecure Registries:
127.0.0.0/8
Registries: registry.access.redhat.com (secure), docker.io (secure)
these are the images i had earlier..
REPOSITORY TAG IMAGE ID CREATED SIZE
quay.io/calico/node v2.6.12 401cc3e56a1a 3 months ago 281.2 MB
quay.io/calico/kube-controllers v1.0.5 b647fdbe067d 3 months ago 52.44 MB
quay.io/calico/cni v1.11.8 5cf513c7d5ed 3 months ago 70.81 MB
k8s.gcr.io/kube-proxy-amd64 v1.10.3 4261d315109d 9 months ago 97.06 MB
k8s.gcr.io/kube-controller-manager-amd64 v1.10.3 40c8d10b2d11 9 months ago 147.9 MB
k8s.gcr.io/kube-apiserver-amd64 v1.10.3 e03746fe22c3 9 months ago 225.1 MB
k8s.gcr.io/kube-scheduler-amd64 v1.10.3 353b8f1d102e 9 months ago 50.41 MB
k8s.gcr.io/etcd-amd64 3.1.12 52920ad46f5b 11 months ago 193.2 MB
registry latest d1fd7d86a825 13 months ago 33.26 MB
k8s.gcr.io/pause-amd64 3.1 da86e6ba6ca1 14 months ago 742.5 kB
quay.io/coreos/etcd v3.1.1 47bb9dd99916 19 months ago 34.56 MB
at the moment i see
docker images
REPOSITORY TAG IMAGE ID CREATED SIZE
quay.io/calico/node v2.6.12 401cc3e56a1a 3 months ago 281.2 MB
quay.io/calico/kube-controllers v1.0.5 b647fdbe067d 3 months ago 52.44 MB
quay.io/calico/cni v1.11.8 5cf513c7d5ed 3 months ago 70.81 MB
k8s.gcr.io/kube-proxy-amd64 v1.10.3 4261d315109d 9 months ago 97.06 MB
k8s.gcr.io/kube-controller-manager-amd64 v1.10.3 40c8d10b2d11 9 months ago 147.9 MB
k8s.gcr.io/kube-scheduler-amd64 v1.10.3 353b8f1d102e 9 months ago 50.41 MB
k8s.gcr.io/etcd-amd64 3.1.12 52920ad46f5b 11 months ago 193.2 MB
registry latest d1fd7d86a825 13 months ago 33.26 MB
k8s.gcr.io/pause-amd64 3.1 da86e6ba6ca1 14 months ago 742.5 kB
k8s.gcr.io/kube-apiserver-amd64 has disappeared. I have tried redeploying but same issues...
journalctl docker
Feb 28 13:50:18 masterhost dockerd-current[7365]: 2019-02-28T13:50:18Z masterhost confd[89]: ERROR client: etcd cluster is unavailable or misconfigured; error #0: dial tcp 10.96.232.136:6666: getsockopt: connection refused
Feb 28 13:50:18 masterhost dockerd-current[7365]:
Feb 28 13:50:18 masterhost dockerd-current[7365]: 2019-02-28T13:50:18Z masterhost confd[89]: ERROR client: etcd cluster is unavailable or misconfigured; error #0: dial tcp 10.96.232.136:6666: getsockopt: connection refused
Feb 28 13:50:18 masterhost dockerd-current[7365]:
Feb 28 13:50:18 masterhost dockerd-current[7365]: 2019-02-28T13:50:18Z masterhost confd[89]: ERROR client: etcd cluster is unavailable or misconfigured; error #0: dial tcp 10.96.232.136:6666: getsockopt: connection refused
Feb 28 13:50:18 masterhost dockerd-current[7365]:
Feb 28 13:50:18 masterhost dockerd-current[7365]: 2019-02-28T13:50:18Z masterhost confd[89]: ERROR client: etcd cluster is unavailable or misconfigured; error #0: dial tcp 10.96.232.136:6666: getsockopt: connection refused
Feb 28 13:50:18 masterhost dockerd-current[7365]:
Feb 28 13:50:19 masterhost dockerd-current[7365]: E0228 13:50:19.123528 1 leaderelection.go:224] error retrieving resource lock kube-system/kube-controller-manager: Get https://11.159.183.96:6443/api/v1/namespaces/kube-system/end
Feb 28 13:50:19 masterhost dockerd-current[7365]: 2019-02-28 13:50:19.146 [WARNING][88] syncer.go 353: Net error from etcd error=dial tcp 10.96.232.136:6666: getsockopt: connection refused
Feb 28 13:50:19 masterhost dockerd-current[7365]: 2019-02-28 13:50:19.147 [WARNING][88] syncer.go 284: May be tight looping, throttling retries.
Feb 28 13:50:19 masterhost dockerd-current[7365]: 2019-02-28 13:50:19.276 [WARNING][88] syncer.go 372: Failed to poll etcd server cluster ID error=client: etcd cluster is unavailable or misconfigured; error #0: dial tcp 10.96.232.136:6
Feb 28 13:50:19 masterhost dockerd-current[7365]:
Feb 28 13:50:19 masterhost dockerd-current[7365]: E0228 13:50:19.388335 1 reflector.go:205] k8s.io/kubernetes/cmd/kube-scheduler/app/server.go:594: Failed to list *v1.Pod: Get https://11.159.183.96:6443/api/v1/pods?fieldSelector=
Feb 28 13:50:19 masterhost dockerd-current[7365]: E0228 13:50:19.389070 1 reflector.go:205] k8s.io/kubernetes/vendor/k8s.io/client-go/informers/factory.go:87: Failed to list *v1.Node: Get https://11.159.183.96:6443/api/v1/nodes?l
Feb 28 13:50:19 masterhost dockerd-current[7365]: E0228 13:50:19.390082 1 reflector.go:205] k8s.io/kubernetes/vendor/k8s.io/client-go/informers/factory.go:87: Failed to list *v1.PersistentVolume: Get https://11.159.183.96:6443/ap
Feb 28 13:50:19 masterhost dockerd-current[7365]: E0228 13:50:19.391249 1 reflector.go:205] k8s.io/kubernetes/vendor/k8s.io/client-go/informers/factory.go:87: Failed to list *v1beta1.ReplicaSet: Get https://11.159.183.96:6443/api
Feb 28 13:50:19 masterhost dockerd-current[7365]: E0228 13:50:19.392294 1 reflector.go:205] k8s.io/kubernetes/vendor/k8s.io/client-go/informers/factory.go:87: Failed to list *v1.Service: Get https://11.159.183.96:6443/api/v1/serv
Feb 28 13:50:19 masterhost dockerd-current[7365]: E0228 13:50:19.393350 1 reflector.go:205] k8s.io/kubernetes/vendor/k8s.io/client-go/informers/factory.go:87: Failed to list *v1.ReplicationController: Get https://11.159.183.96:64
Feb 28 13:50:19 masterhost dockerd-current[7365]: E0228 13:50:19.394442 1 reflector.go:205] k8s.io/kubernetes/vendor/k8s.io/client-go/informers/factory.go:87: Failed to list *v1.PersistentVolumeClaim: Get https://11.159.183.96:64
Feb 28 13:50:19 masterhost dockerd-current[7365]: E0228 13:50:19.395526 1 reflector.go:205] k8s.io/kubernetes/vendor/k8s.io/client-go/informers/factory.go:87: Failed to list *v1.StorageClass: Get https://11.159.183.96:6443/apis/s
Feb 28 13:50:19 masterhost dockerd-current[7365]: E0228 13:50:19.396597 1 reflector.go:205] k8s.io/kubernetes/vendor/k8s.io/client-go/informers/factory.go:87: Failed to list *v1beta1.PodDisruptionBudget: Get https://11.159.183.96
Feb 28 13:50:19 masterhost dockerd-current[7365]: E0228 13:50:19.397606 1 reflector.go:205] k8s.io/kubernetes/vendor/k8s.io/client-go/informers/factory.go:87: Failed to list *v1beta1.StatefulSet: Get https://11.159.183.96:6443/ap
Feb 28 13:50:19 masterhost dockerd-current[7365]: E0228 13:50:19.676174 1 reflector.go:205] k8s.io/kubernetes/pkg/client/informers/informers_generated/internalversion/factory.go:86: Failed to list *core.Service: Get https://11.15
Feb 28 13:50:19 masterhost dockerd-current[7365]: E0228 13:50:19.677083 1 reflector.go:205] k8s.io/kubernetes/pkg/client/informers/informers_generated/internalversion/factory.go:86: Failed to list *core.Endpoints: Get https://11.
Feb 28 13:50:20 masterhost dockerd-current[7365]: 2019-02-28 13:50:20.147 [WARNING][88] syncer.go 353: Net error from etcd error=dial tcp 10.96.232.136:6666: getsockopt: connection refused
Feb 28 13:50:20 masterhost dockerd-current[7365]: 2019-02-28 13:50:20.147 [WARNING][88] syncer.go 284: May be tight looping, throttling retries.
Feb 28 13:50:20 masterhost dockerd-current[7365]: E0228 13:50:20.388887 1 reflector.go:205] k8s.io/kubernetes/cmd/kube-scheduler/app/server.go:594: Failed to list *v1.Pod: Get https://11.159.183.96:6443/api/v1/pods?fieldSelector=
Feb 28 13:50:20 masterhost dockerd-current[7365]: E0228 13:50:20.389727 1 reflector.go:205] k8s.io/kubernetes/vendor/k8s.io/client-go/informers/factory.go:87: Failed to list *v1.Node: Get https://11.159.183.96:6443/api/v1/nodes?l
Feb 28 13:50:20 masterhost dockerd-current[7365]: E0228 13:50:20.390765 1 reflector.go:205] k8s.io/kubernetes/vendor/k8s.io/client-go/informers/factory.go:87: Failed to list *v1.PersistentVolume: Get https://11.159.183.96:6443/ap
Feb 28 13:50:20 masterhost dockerd-current[7365]: E0228 13:50:20.391803 1 reflector.go:205] k8s.io/kubernetes/vendor/k8s.io/client-go/informers/factory.go:87: Failed to list *v1beta1.ReplicaSet: Get https://11.159.183.96:6443/api
Feb 28 13:50:20 masterhost dockerd-current[7365]: E0228 13:50:20.392855 1 reflector.go:205] k8s.io/kubernetes/vendor/k8s.io/client-go/informers/factory.go:87: Failed to list *v1.Service: Get https://11.159.183.96:6443/api/v1/serv
Feb 28 13:50:20 masterhost dockerd-current[7365]: E0228 13:50:20.393913 1 reflector.go:205] k8s.io/kubernetes/vendor/k8s.io/client-go/informers/factory.go:87: Failed to list *v1.ReplicationController: Get https://11.159.183.96:64
Feb 28 13:50:20 masterhost dockerd-current[7365]: E0228 13:50:20.395067 1 reflector.go:205] k8s.io/kubernetes/vendor/k8s.io/client-go/informers/factory.go:87: Failed to list *v1.PersistentVolumeClaim: Get https://11.159.183.96:64
Feb 28 13:50:20 masterhost dockerd-current[7365]: E0228 13:50:20.396132 1 reflector.go:205] k8s.io/kubernetes/vendor/k8s.io/client-go/informers/factory.go:87: Failed to list *v1.StorageClass: Get https://11.159.183.96:6443/apis/s
Feb 28 13:50:20 masterhost dockerd-current[7365]: E0228 13:50:20.397212 1 reflector.go:205] k8s.io/kubernetes/vendor/k8s.io/client-go/informers/factory.go:87: Failed to list *v1beta1.PodDisruptionBudget: Get https://11.159.183.96
Feb 28 13:50:20 masterhost dockerd-current[7365]: E0228 13:50:20.398272 1 reflector.go:205] k8s.io/kubernetes/vendor/k8s.io/client-go/informers/factory.go:87: Failed to list *v1beta1.StatefulSet: Get https://11.159.183.96:6443/ap
Feb 28 13:50:20 masterhost dockerd-current[7365]: 2019-02-28 13:50:20.588 [INFO][88] health.go 150: Overall health summary=&health.HealthReport{Live:true, Ready:true}
Feb 28 13:50:20 masterhost dockerd-current[7365]: E0228 13:50:20.676705 1 reflector.go:205] k8s.io/kubernetes/pkg/client/informers/informers_generated/internalversion/factory.go:86: Failed to list *core.Service: Get https://11.15
Feb 28 13:50:20 masterhost dockerd-current[7365]: E0228 13:50:20.677651 1 reflector.go:205] k8s.io/kubernetes/pkg/client/informers/informers_generated/internalversion/factory.go:86: Failed to list *core.Endpoints: Get https://11.
Feb 28 13:50:20 masterhost dockerd-current[7365]: 2019-02-28T13:50:20Z masterhost confd[89]: ERROR client: etcd cluster is unavailable or misconfigured; error #0: dial tcp 10.96.232.136:6666: getsockopt: connection refused
Feb 28 13:50:20 masterhost dockerd-current[7365]:
Feb 28 13:50:20 masterhost dockerd-current[7365]: 2019-02-28T13:50:20Z masterhost confd[89]: ERROR client: etcd cluster is unavailable or misconfigured; error #0: dial tcp 10.96.232.136:6666: getsockopt: connection refused
Feb 28 13:50:20 masterhost dockerd-current[7365]:
Feb 28 13:50:20 masterhost dockerd-current[7365]: 2019-02-28T13:50:20Z masterhost confd[89]: ERROR client: etcd cluster is unavailable or misconfigured; error #0: dial tcp 10.96.232.136:6666: getsockopt: connection refused
Feb 28 13:50:20 masterhost dockerd-current[7365]:
Feb 28 13:50:20 masterhost dockerd-current[7365]: 2019-02-28T13:50:20Z masterhost confd[89]: ERROR client: etcd cluster is unavailable or misconfigured; error #0: dial tcp 10.96.232.136:6666: getsockopt: connection refused
Feb 28 13:50:20 masterhost dockerd-current[7365]:
Feb 28 13:50:20 masterhost dockerd-current[7365]: 2019-02-28T13:50:20Z masterhost confd[89]: ERROR client: etcd cluster is unavailable or misconfigured; error #0: dial tcp 10.96.232.136:6666: getsockopt: connection refused
Feb 28 13:50:20 masterhost dockerd-current[7365]:
Feb 28 13:50:20 masterhost dockerd-current[7365]: 2019-02-28T13:50:20Z masterhost confd[89]: ERROR client: etcd cluster is unavailable or misconfigured; error #0: dial tcp 10.96.232.136:6666: getsockopt: connection refused
Feb 28 13:50:20 masterhost dockerd-current[7365]:
Feb 28 13:50:20 masterhost dockerd-current[7365]: 2019-02-28T13:50:20Z masterhost confd[89]: ERROR client: etcd cluster is unavailable or misconfigured; error #0: dial tcp 10.96.232.136:6666: getsockopt: connection refused
Feb 28 13:50:20 masterhost dockerd-current[7365]:
Feb 28 13:50:20 masterhost dockerd-current[7365]: 2019-02-28T13:50:20Z masterhost confd[89]: ERROR client: etcd cluster is unavailable or misconfigured; error #0: dial tcp 10.96.232.136:6666: getsockopt: connection refused
Feb 28 13:50:20 masterhost dockerd-current[7365]:
Feb 28 13:50:20 masterhost dockerd-current[7365]: 2019-02-28T13:50:20Z masterhost confd[89]: ERROR client: etcd cluster is unavailable or misconfigured; error #0: dial tcp 10.96.232.136:6666: getsockopt: connection refused
Feb 28 13:50:20 masterhost dockerd-current[7365]:
Feb 28 13:50:20 masterhost dockerd-current[7365]: 2019-02-28T13:50:20Z masterhost confd[89]: ERROR client: etcd cluster is unavailable or misconfigured; error #0: dial tcp 10.96.232.136:6666: getsockopt: connection refused
Feb 28 13:50:20 masterhost dockerd-current[7365]:
Feb 28 13:50:20 masterhost dockerd-current[7365]: 2019-02-28T13:50:20Z masterhost confd[89]: ERROR client: etcd cluster is unavailable or misconfigured; error #0: dial tcp 10.96.232.136:6666: getsockopt: connection refused
It looks like you do not have enough disk space left. Kubelet automatically deletes unused images when disk usage exceeds 85%.
https://kubernetes.io/docs/concepts/cluster-administration/kubelet-garbage-collection/#user-configuration
1 Like