actually i have issue when i tried to add new node in cluster i faced the below issue during starting the nodes one of old give me error
Error: unable to start all nodes
nodeb.example.com: Error connecting to nodeb.example.com - (HTTP error: 400)
and when try to make auth from node a to nodeb i faced the below error
pcs host auth nodeb.example.com
and in pcs status
* Node nodeb.example.com: UNCLEAN (offline)
* Online: [ nodea.example.com nodec.example.com ]
and after checking /var/log/messages on nodeb give the below error
Jul 11 06:53:26 nodeb corosync: [MAIN ] Corosync Cluster Engine 3.1.5 starting up
Jul 11 06:53:26 nodeb corosync: [MAIN ] Corosync built-in features: dbus systemd xmlconf vqsim nozzle snmp pie relro bindn ow
Jul 11 06:53:27 nodeb corosync: [MAIN ] failed to parse node address 'nodec.example.com'
Jul 11 06:53:27 nodeb corosync: [MAIN ] Corosync Cluster Engine exiting with status 8 at main.c:1417.
Jul 11 06:53:27 nodeb systemd: corosync.service: Main process exited, code=exited, status=8/n/a
Jul 11 06:53:27 nodeb systemd: corosync.service: Failed with result 'exit-code'.
Jul 11 06:53:27 nodeb systemd: Failed to start Corosync Cluster Engine.
so i wanna what must
Maybe there is a local node resolving to a different IP internally. When that happens, commands return HTTP code 400. Please check your IP address/hostname to all the nodes to be sure and you might use the following command: