Bartas1979
Mission Specialist
Mission Specialist
  • 1,844 Views

Global lock failed: check that global lockspace is started

Hi all,

I'm preparing for EX436 exam on RH labs and testing many scenarios. One of these scenarios is prepare cluster with assigned two IP addresses to each node and then use iscsi target with dlm, multipath and LVM's.

Everything goes somoothly till... I starting with LVM configuration. I getting error "Global lock failed: check that global lockspace is started".

Below steps which I using for reproduce that error:

1. pcs cluster setup prod1 \
node1.mydomain.com addr=172.25.250.10 addr=172.25.250.50\
node2.mydomain.com addr=172.25.250.11 addr=172.25.250.51\
node3.mydomain.com addr=172.25.250.12 addr=172.25.250.52\
 
2. pcs stonith create fence_node1 fence_ipmilan pcmk_host_list=node1.mydomain.com  ip=192.168.0.101 username=myadmin password=secret_password lanplus=1 power_timeout=180
pcs stonith create fence_node2 fence_ipmilan pcmk_host_list=node2.mydomain.com  ip=192.168.0.101 username=myadmin password=secret_password lanplus=1 power_timeout=180
pcs stonith create fence_node3 fence_ipmilan pcmk_host_list=node2.mydomain.com  ip=192.168.0.101 username=myadmin password=secret_password lanplus=1 power_timeout=180
 
3. cat /etc/corosync/corosync.conf
nodelist {
node {
ring0_addr 172.25.250.10
ring1_addr 172.25.250.50
name: node1.mydomain.com
nodeid :1
}
node {
ring0_addr 172.25.250.11
ring1_addr 172.25.250.51
name: node2.mydomain.com
nodeid :2
}
node {
ring0_addr 172.25.250.12
ring1_addr 172.25.250.52
name: node3.mydomain.com
nodeid :3
}
}
 
4. dnf install -y dlm iscsi-initiator-utils lvm2-lockd device-mapper-multipath gfs2-utils [on all nodes]
 
5. Edit /etc/iscsi/initiatorname.iscsi file and set the IQN for the client initiator. 
(iqn.2023-11.com.mydomain:<short_hostname>)[on all nodes]
 
6. systemctl enable --now iscsid [on all nodes]
 
7. iscsiadm -m discovery -t st -p 192.168.1.15 192.168.1.15
iscsiadm -m node -T iqn.2023-11.com.mydomain:store-prod -p 192.168.1.15 -l
iscsiadm -m discovery -t st -p 192.168.1.15 192.168.2.15
iscsiadm -m node -T iqn.2023-11.com.mydomain:store-prod -p 192.168.2.15 -l
[on all nodes]
 
8. mpathconf --enable
 
9. iscsiadm -m session -P 3
 
10. udevadm info /dev/sdb | grep ID_SERIAL=
 
11. Edit the /etc/multipath.conf
multipaths {
multipath {
wwid 3600140562aeac25dc4c4eb5842574c7a
alias diska
}
}
 
12. Copy /etc/multipath.conf to two nodes in cluster node2 and node3:
scp /etc/multipath.conf root@node2:/etc/
scp /etc/multipath.conf root@node3:/etc/
 
13. systemctl enable --now multipathd [on all nodes]
 
14. pcs resource create dlm ocf:pacemaker:controld --group=locking
 
15. pcs resource create lvmlockd ocf:heartbeat:lvmlockd --group=locking
 
16. pcs resource clone locking interleave=true
 
17. pvcreate /dev/mapper/diska
 
Global lock failed: check that global lockspace is started
 
Could I ask you for help to find me where I made mistake? Thank you.
Labels (2)
10 Replies
Chaitanya83
Flight Engineer
Flight Engineer
  • 1,632 Views

Hello

Are u able to list the disks on all the nodes when u type lsblk on machines
?
Bartas1979
Mission Specialist
Mission Specialist
  • 1,604 Views

Yes. Without any problem. 

Chetan_Tiwary_
Moderator
Moderator
  • 1,622 Views

Hello @Bartas1979 !

Thanks for reaching out !

Could you please quickly let me know the guided exercise or the chapter -section which you are practising for this issue in RH436 course ?

0 Kudos
Bartas1979
Mission Specialist
Mission Specialist
  • 1,591 Views

Hello,

Well... there is no section exactly like this. I'm basing it on "compreview-gfs2" with a few small changes (as described above). 

Chetan_Tiwary_
Moderator
Moderator
  • 1,565 Views

Ok @Bartas1979 ! What about use_lvmlockd = 1 in the global section of /etc/lvm/lvm.conf ? if it is currently set to 1 then change it to 0 and then recheck. 

or Ensure that only one heartbeat address is configured per cluster node in /etc/corosync/corosync.conf

( check /var/log/messages for the dlm error )

0 Kudos
Bartas1979
Mission Specialist
Mission Specialist
  • 1,525 Views

Hi @Chetan_Tiwary_ 

Thank you for your suggestions. I tried some time ago change use_lvmlockd parameter on 0. Result is pvcreate /dev/mapper/diska command works but next step is failed again:

vgcreate --shared vg1 /dev/mapper/diska

Using a shared lock type requires lvmlockd.

About second part of your suggestion - "Ensure that only one heartbeat address is..." I guess you mean to remove one address from section:

ring0_addr 172.25.250.10
ring1_addr 172.25.250.50
 
Unfortunately it is not option beacuse in my case, one of the goals I want to achieve is two addresses assigned to node.
Bartas1979
Mission Specialist
Mission Specialist
  • 1,488 Views

Thank you @Chetan_Tiwary_  for links with documentation. Really appreciate that but... it still not cover my issue.  

0 Kudos
mohamed42
Cadet
Cadet
  • 893 Views

are you able to fix this issue ? because i am facing the same senario also and I don't understand where is the issue really, the shared volume group will require "use_lvmlockd = 1 " which will activate automatic i shouldn't touch the file /etc/lvm/lvm.conf 

 

0 Kudos
Join the discussion
You must log in to join this conversation.