I'll have to resubmit this issue as it's still a problem for my installed cluster.
My environment consists of 2 masters and one load-balancer that has the default HAProxy installed by the automated install procedure via the openshift-ansible project. So it's basically the "default" version for a cluster with 2 masters.
What I find is that the config map for the web-console gets created by default with one of the masters as consolePublicURL and masterPublicURL and not the load-balancer entry-point, as I would have expected.
So if I try to simulate a master fail for the host that's configured in the configMap, this effectively means that I cannot reach the webconsole, even though I do have the second master available.
I have tried editing the configMap for the webconsole and the oauthclient openshift-web-console, but this results in invalid request errors when trying to access the web-console via a web-browser.
The important edits are:
- in configmap webconsole-config:
- in oauthclient openshift-web-console:
I've already gone through some mail exchanges with Sam Padgett who has pointed out that it might be load-balancer config related, but, as I stated, the load-balancer is the default one configured by the installer. The HAProxy config file seems to be the default one provided in the openshift-ansible project.
I'd appreciate if you have any ideas about where to look for this problem.
Thanks in advance!