0

In my GKE cluster, I have a node pool with 2 nodes.
Also, I have an external dedicated server with a database.
During the google maintenance, one of the nodes in this node pool was replaced with a new one.
After this, my pods on the new node can't connect to my external server with the error 'no route to host'.
But pods that are located on the old node can connect to my external server without any problems.

So the problem is with the new node only.
The settings of the network and firewall are default in the cluster.

As a result, I have 2 nodes in the node pool, but correctly works only one of them.

The hotfix that works it's a replace the problemed node with a new one.
But a new node would work correctly with a probability of 50% (yes or no).
If the new one also doesn't work, then repeat this step until you get a node that works correctly.
I think it's a very bad solution.

I add a connectivity test from this problemed node to my external dedicated server and database port.
This test says that everything is ok and the destination is reached.

But, when I connect to the node by ssh and try to connect to an external dedicated server with telnet, I get the same 'no route to host' as in pods on that node.

How is it possible?

Also didn't help:

  • Additional firewall rule to allow all traffic to my dedicated server
  • Add ip-masq-agent

0 Answers0