I'm new on kubernetes.
I couldn't get deployments using kubectl but I can see all deployments on kubernetes dashboard. How can i fix this problem?
user@master:~$ kubectl get deployments
error: the server doesn't have a resource type "deployments"
kubernetes version: 1.12
kubectl version: 1.13
kubectl api-versions:
apiregistration.k8s.io/v1
apiregistration.k8s.io/v1beta1
v1
api-resources:
user@master:~$ kubectl api-resources
NAME SHORTNAMES APIGROUP NAMESPACED
KIND
bindings true
Binding
componentstatuses cs false
ComponentStatus
configmaps cm true
ConfigMap
endpoints ep true
Endpoints
events ev true
Event
limitranges limits true
LimitRange
namespaces ns false
Namespace
nodes no false
Node
persistentvolumeclaims pvc true
PersistentVolumeClaim
persistentvolumes pv false
PersistentVolume
pods po true
Pod
podtemplates true
PodTemplate
replicationcontrollers rc true
ReplicationController
resourcequotas quota true
ResourceQuota
secrets true
Secret
serviceaccounts sa true
ServiceAccount
services svc true
Service
apiservices apiregistration.k8s.io false
APIService
Thanks for your helps.
-----------Edit 1-----------
Hello @EduardoBaitello, Thank you for quicly reply. The problem is not related to permission.
user@master:~$ kubectl auth can-i get deployments
Warning: the server doesn't have a resource type 'deployments'
yes
user@master:~$ kubectl auth can-i get deployment
Warning: the server doesn't have a resource type 'deployment'
yes
user@master:~$ kubectl auth can-i get namespaces
yes
user@master:~$ kubectl auth can-i get pods
yes
So I think this is not a duplicated question.
user@master:~$ kubectl get po --namespace=kube-system
NAME READY STATUS RESTARTS AGE
calico-kube-controllers-7c6b876df8-nk7nm 1/1 Running 2 118d
calico-node-8lt9f 1/1 Running 3 118d
calico-node-d9r9l 1/1 Running 2 118d
calico-node-ffqlj 1/1 Running 2 118d
dns-autoscaler-57ff59dd4c-c9tjv 1/1 Running 2 118d
kube-apiserver-node1 1/1 Running 3 118d
kube-controller-manager-node1 1/1 Running 6 118d
kube-dns-84467597f5-hf2fn 3/3 Running 6 118d
kube-dns-84467597f5-sttgx 3/3 Running 9 118d
kube-proxy-node1 1/1 Running 3 118d
kube-proxy-node2 1/1 Running 2 118d
kube-proxy-node3 1/1 Running 2 118d
kube-scheduler-node1 1/1 Running 6 118d
kubernetes-dashboard-5db4d9f45f-gkl6w 1/1 Running 3 118d
nginx-proxy-node2 1/1 Running 2 118d
nginx-proxy-node3 1/1 Running 2 118d
tiller-deploy-6f6fd74b68-27fqc 1/1 Running 0 16d
user@master:~$ kubectl get componentstatus
NAME STATUS MESSAGE
scheduler Healthy ok
controller-manager Healthy ok
etcd-2 Healthy {"health": "true"}
etcd-1 Healthy {"health": "true"}
etcd-0 Healthy {"health": "true"}
I was experiencing exactly the same behavior as in "Edit 1" above with Kubernetes 1.13.5 (client and server). Removing the
~/.kube/http-cache
directory on the client worked for me.@AliCan
I suspect this is because of Kubectl binary version. I noticed your kubectl version(1.13) is higher than the Kubernetes version(1.12).
can you try downgrading your kubectl version to 1.12 or lower?