Skip to main content

Coredns in pending state in Kubernetes cluster

 in order to add nodes to K8S cluster you must deploy a Container Network Interface (CNI) 

NAMESPACE     NAME                            READY   STATUS    RESTARTS   AGE

kube-system   coredns-64897985d-4npkg          0/1     Pending   0          5h33m
kube-system   coredns-64897985d-qm8gt          0/1     Pending   0          5h33m
kube-system   etcd-KubeMngr                    1/1     Running   1          5h33m
kube-system   kube-apiserver-KubeMngr          1/1     Running   1          5h33m
kube-system   kube-controller-manager-KubeMngr 1/1     Running   1          5h33m
kube-system   kube-proxy-w6mmf                 1/1     Running   0          5h33m
kube-system   kube-scheduler-KubeMngr          1/1     Running   1          5h33m

here is an example of how to install calico CNI:

download YAML file:

curl https://docs.projectcalico.org/manifests/calico.yaml -O


apply it using kubectl cli:

kubectl apply -f calico.yaml


and:
[root@KubeMngr kubernetes]# kubectl get pods --all-namespaces
NAMESPACE     NAME                                READY   STATUS    RESTARTS   AGE
kube-system   calico-kube-controllers-647d84984b   1/1     Running   0   110s
kube-system   calico-node-zn98r                    1/1     Running   0         112s
kube-system   coredns-64897985d-4npkg              1/1     Running   0         5h38m
kube-system   coredns-64897985d-qm8gt              1/1     Running   0         5h38m
kube-system   etcd-KubeMngr                        1/1     Running   1         5h38m
kube-system   kube-apiserver-KubeMngr              1/1     Running   1         5h38m
kube-system   kube-controller-manager-KubeMngr     1/1     Running   1       5h38m
kube-system   kube-proxy-w6mmf                     1/1     Running   0         5h38m
kube-system   kube-scheduler-KubeMngr              1/1     Running   1         5h38m

Comments

Popular posts from this blog

ORA-27104: system-defined limits for shared memory was misconfigured

I faced this error while trying to restore & recover of a PDB (pluggable database) part of the log file and the solution is described here below: log: initialization parameters used for automatic instance: db_name=CDB db_unique_name=gbux_pitr_PDB1_CDB compatible=12.1.0.2.0 db_block_size=8192 db_files=200 diagnostic_dest=/oracle/app/oracle _system_trig_enabled=FALSE sga_target=1888M processes=200 db_create_file_dest=/oracle/auxilary log_archive_dest_1='location=/oracle/auxilary' enable_pluggable_database=true _clone_one_pdb_recovery=true #No auxiliary parameter file used starting up automatic instance CDB RMAN-00571: =========================================================== RMAN-00569: =============== ERROR MESSAGE STACK FOLLOWS =============== RMAN-00571: =========================================================== RMAN-03002: failure of recover command at 07/31/2016 16:22:20 RMAN-04014: startup failed: ORA-27104: system-defined limits for shared

mount.nfs: backgrounding

if you face this kind of error with your remote NFS: [root@Vertica000 ~]# mount /files/application/Rremote3 mount.nfs: backgrounding "10.0.0.2:/files/application/remoteFiles" mount.nfs: mount options: "bg,hard,nointr,rsize=65536,wsize=65536,tcp,actimeo=0,vers=3,timeo=600,addr=10.0.0.2" look for the problem in the log file: cat /var/log/messages | grep mount  mount to NFS server '10.0.0.2' failed: timed out, retrying Solution: In most of the cases, you have a problem with your iptables in the destination server login as root to dest server (10.0.0.2) in my case and type this command: iptables --flush  the go back to your origin server to try remount the problematic NFS file system of course this is in case nfs server was installed and functioning properly. Good luck.

ora-65035 while create pluggable database from another one

or : ORA-21000: error number argument to raise_application_error of -65035 is out of range Cause: this error caused because of an unrecoverable transaction in the source DB. Solution: 1. either stop and start your pluggable source DB 2. identify the open transaction and kill it if you can. use this query to do so: select * from v$transaction t,v$session  s where t.ses_addr = s.saddr;