Lines Matching +full:cluster +full:- +full:mode
1 // SPDX-License-Identifier: GPL-2.0-only
5 ** Copyright (C) Sistina Software, Inc. 1997-2003 All rights reserved.
6 ** Copyright (C) 2004-2011 Red Hat, Inc. All rights reserved.
43 ls = dlm_find_lockspace_local(ls->ls_local_handle); in dlm_control_store()
45 return -EINVAL; in dlm_control_store()
55 ret = -EINVAL; in dlm_control_store()
63 int rc = kstrtoint(buf, 0, &ls->ls_uevent_result); in dlm_event_store()
67 set_bit(LSFL_UEVENT_WAIT, &ls->ls_flags); in dlm_event_store()
68 wake_up(&ls->ls_uevent_wait); in dlm_event_store()
74 return snprintf(buf, PAGE_SIZE, "%u\n", ls->ls_global_id); in dlm_id_show()
79 int rc = kstrtouint(buf, 0, &ls->ls_global_id); in dlm_id_store()
99 set_bit(LSFL_NODIR, &ls->ls_flags); in dlm_nodir_store()
111 return snprintf(buf, PAGE_SIZE, "%d\n", ls->ls_recover_nodeid); in dlm_recover_nodeid_show()
121 .attr = {.name = "control", .mode = S_IWUSR},
126 .attr = {.name = "event_done", .mode = S_IWUSR},
131 .attr = {.name = "id", .mode = S_IRUGO | S_IWUSR},
137 .attr = {.name = "nodir", .mode = S_IRUGO | S_IWUSR},
143 .attr = {.name = "recover_status", .mode = S_IRUGO},
148 .attr = {.name = "recover_nodeid", .mode = S_IRUGO},
168 return a->show ? a->show(ls, buf) : 0; in dlm_attr_show()
176 return a->store ? a->store(ls, buf, len) : len; in dlm_attr_store()
201 kobject_uevent(&ls->ls_kobj, KOBJ_ONLINE); in do_uevent()
203 kobject_uevent(&ls->ls_kobj, KOBJ_OFFLINE); in do_uevent()
210 wait_event(ls->ls_uevent_wait, in do_uevent()
211 test_and_clear_bit(LSFL_UEVENT_WAIT, &ls->ls_flags)); in do_uevent()
213 log_rinfo(ls, "group event done %d", ls->ls_uevent_result); in do_uevent()
215 return ls->ls_uevent_result; in do_uevent()
222 add_uevent_var(env, "LOCKSPACE=%s", ls->ls_name); in dlm_uevent()
240 return -ENOMEM; in dlm_lockspace_init()
256 if (time_after_eq(jiffies, ls->ls_scan_time + in find_ls_to_scan()
274 ls->ls_scan_time = jiffies; in dlm_scand()
278 ls->ls_scan_time += HZ; in dlm_scand()
312 if (ls->ls_global_id == id) { in dlm_find_lockspace_global()
313 atomic_inc(&ls->ls_count); in dlm_find_lockspace_global()
329 if (ls->ls_local_handle == lockspace) { in dlm_find_lockspace_local()
330 atomic_inc(&ls->ls_count); in dlm_find_lockspace_local()
346 if (ls->ls_device.minor == minor) { in dlm_find_lockspace_device()
347 atomic_inc(&ls->ls_count); in dlm_find_lockspace_device()
359 if (atomic_dec_and_test(&ls->ls_count)) in dlm_put_lockspace()
360 wake_up(&ls->ls_count_wait); in dlm_put_lockspace()
366 wait_event(ls->ls_count_wait, atomic_read(&ls->ls_count) == 0); in remove_lockspace()
369 if (atomic_read(&ls->ls_count) != 0) { in remove_lockspace()
374 WARN_ON(ls->ls_create_count != 0); in remove_lockspace()
375 list_del(&ls->ls_list); in remove_lockspace()
404 static int new_lockspace(const char *name, const char *cluster, in new_lockspace() argument
415 return -EINVAL; in new_lockspace()
418 return -EINVAL; in new_lockspace()
421 return -EINVAL; in new_lockspace()
425 error = -EUNATCH; in new_lockspace()
431 *ops_result = -EOPNOTSUPP; in new_lockspace()
436 if (!cluster) in new_lockspace()
437 log_print("dlm cluster name '%s' is being used without an application provided cluster name", in new_lockspace()
440 if (dlm_config.ci_recover_callbacks && cluster && in new_lockspace()
441 strncmp(cluster, dlm_config.ci_cluster_name, DLM_LOCKSPACE_LEN)) { in new_lockspace()
442 log_print("dlm cluster name '%s' does not match " in new_lockspace()
443 "the application cluster name '%s'", in new_lockspace()
444 dlm_config.ci_cluster_name, cluster); in new_lockspace()
445 error = -EBADR; in new_lockspace()
453 WARN_ON(ls->ls_create_count <= 0); in new_lockspace()
454 if (ls->ls_namelen != namelen) in new_lockspace()
456 if (memcmp(ls->ls_name, name, namelen)) in new_lockspace()
459 error = -EEXIST; in new_lockspace()
462 ls->ls_create_count++; in new_lockspace()
472 error = -ENOMEM; in new_lockspace()
477 memcpy(ls->ls_name, name, namelen); in new_lockspace()
478 ls->ls_namelen = namelen; in new_lockspace()
479 ls->ls_lvblen = lvblen; in new_lockspace()
480 atomic_set(&ls->ls_count, 0); in new_lockspace()
481 init_waitqueue_head(&ls->ls_count_wait); in new_lockspace()
482 ls->ls_flags = 0; in new_lockspace()
483 ls->ls_scan_time = jiffies; in new_lockspace()
486 ls->ls_ops = ops; in new_lockspace()
487 ls->ls_ops_arg = ops_arg; in new_lockspace()
493 ls->ls_exflags = (flags & ~(DLM_LSFL_FS | DLM_LSFL_NEWEXCL)); in new_lockspace()
496 ls->ls_rsbtbl_size = size; in new_lockspace()
498 ls->ls_rsbtbl = vmalloc(array_size(size, sizeof(struct dlm_rsbtable))); in new_lockspace()
499 if (!ls->ls_rsbtbl) in new_lockspace()
502 ls->ls_rsbtbl[i].keep.rb_node = NULL; in new_lockspace()
503 ls->ls_rsbtbl[i].toss.rb_node = NULL; in new_lockspace()
504 spin_lock_init(&ls->ls_rsbtbl[i].lock); in new_lockspace()
508 ls->ls_remove_names[i] = kzalloc(DLM_RESNAME_MAXLEN+1, in new_lockspace()
510 if (!ls->ls_remove_names[i]) in new_lockspace()
514 idr_init(&ls->ls_lkbidr); in new_lockspace()
515 spin_lock_init(&ls->ls_lkbidr_spin); in new_lockspace()
517 INIT_LIST_HEAD(&ls->ls_waiters); in new_lockspace()
518 mutex_init(&ls->ls_waiters_mutex); in new_lockspace()
519 INIT_LIST_HEAD(&ls->ls_orphans); in new_lockspace()
520 mutex_init(&ls->ls_orphans_mutex); in new_lockspace()
522 INIT_LIST_HEAD(&ls->ls_new_rsb); in new_lockspace()
523 spin_lock_init(&ls->ls_new_rsb_spin); in new_lockspace()
525 INIT_LIST_HEAD(&ls->ls_nodes); in new_lockspace()
526 INIT_LIST_HEAD(&ls->ls_nodes_gone); in new_lockspace()
527 ls->ls_num_nodes = 0; in new_lockspace()
528 ls->ls_low_nodeid = 0; in new_lockspace()
529 ls->ls_total_weight = 0; in new_lockspace()
530 ls->ls_node_array = NULL; in new_lockspace()
532 memset(&ls->ls_local_rsb, 0, sizeof(struct dlm_rsb)); in new_lockspace()
533 ls->ls_local_rsb.res_ls = ls; in new_lockspace()
535 ls->ls_debug_rsb_dentry = NULL; in new_lockspace()
536 ls->ls_debug_waiters_dentry = NULL; in new_lockspace()
538 init_waitqueue_head(&ls->ls_uevent_wait); in new_lockspace()
539 ls->ls_uevent_result = 0; in new_lockspace()
540 init_completion(&ls->ls_recovery_done); in new_lockspace()
541 ls->ls_recovery_result = -1; in new_lockspace()
543 spin_lock_init(&ls->ls_cb_lock); in new_lockspace()
544 INIT_LIST_HEAD(&ls->ls_cb_delay); in new_lockspace()
546 ls->ls_recoverd_task = NULL; in new_lockspace()
547 mutex_init(&ls->ls_recoverd_active); in new_lockspace()
548 spin_lock_init(&ls->ls_recover_lock); in new_lockspace()
549 spin_lock_init(&ls->ls_rcom_spin); in new_lockspace()
550 get_random_bytes(&ls->ls_rcom_seq, sizeof(uint64_t)); in new_lockspace()
551 ls->ls_recover_status = 0; in new_lockspace()
552 ls->ls_recover_seq = get_random_u64(); in new_lockspace()
553 ls->ls_recover_args = NULL; in new_lockspace()
554 init_rwsem(&ls->ls_in_recovery); in new_lockspace()
555 init_rwsem(&ls->ls_recv_active); in new_lockspace()
556 INIT_LIST_HEAD(&ls->ls_requestqueue); in new_lockspace()
557 atomic_set(&ls->ls_requestqueue_cnt, 0); in new_lockspace()
558 init_waitqueue_head(&ls->ls_requestqueue_wait); in new_lockspace()
559 mutex_init(&ls->ls_requestqueue_mutex); in new_lockspace()
560 spin_lock_init(&ls->ls_clear_proc_locks); in new_lockspace()
567 ls->ls_recover_buf = kmalloc(DLM_MAX_SOCKET_BUFSIZE, GFP_NOFS); in new_lockspace()
568 if (!ls->ls_recover_buf) in new_lockspace()
571 ls->ls_slot = 0; in new_lockspace()
572 ls->ls_num_slots = 0; in new_lockspace()
573 ls->ls_slots_size = 0; in new_lockspace()
574 ls->ls_slots = NULL; in new_lockspace()
576 INIT_LIST_HEAD(&ls->ls_recover_list); in new_lockspace()
577 spin_lock_init(&ls->ls_recover_list_lock); in new_lockspace()
578 idr_init(&ls->ls_recover_idr); in new_lockspace()
579 spin_lock_init(&ls->ls_recover_idr_lock); in new_lockspace()
580 ls->ls_recover_list_count = 0; in new_lockspace()
581 ls->ls_local_handle = ls; in new_lockspace()
582 init_waitqueue_head(&ls->ls_wait_general); in new_lockspace()
583 INIT_LIST_HEAD(&ls->ls_root_list); in new_lockspace()
584 init_rwsem(&ls->ls_root_sem); in new_lockspace()
587 ls->ls_create_count = 1; in new_lockspace()
588 list_add(&ls->ls_list, &lslist); in new_lockspace()
599 init_waitqueue_head(&ls->ls_recover_lock_wait); in new_lockspace()
603 * initializes ls_in_recovery as locked in "down" mode. We need in new_lockspace()
605 * has to start out in down mode. in new_lockspace()
614 wait_event(ls->ls_recover_lock_wait, in new_lockspace()
615 test_bit(LSFL_RECOVER_LOCK, &ls->ls_flags)); in new_lockspace()
620 ls->ls_kobj.kset = dlm_kset; in new_lockspace()
621 error = kobject_init_and_add(&ls->ls_kobj, &dlm_ktype, NULL, in new_lockspace()
622 "%s", ls->ls_name); in new_lockspace()
625 kobject_uevent(&ls->ls_kobj, KOBJ_ADD); in new_lockspace()
629 cluster infrastructure.) Once it's done that, it tells us who the in new_lockspace()
638 wait_for_completion(&ls->ls_recovery_done); in new_lockspace()
639 error = ls->ls_recovery_result; in new_lockspace()
652 kfree(ls->ls_node_array); in new_lockspace()
659 list_del(&ls->ls_list); in new_lockspace()
661 idr_destroy(&ls->ls_recover_idr); in new_lockspace()
662 kfree(ls->ls_recover_buf); in new_lockspace()
664 idr_destroy(&ls->ls_lkbidr); in new_lockspace()
667 kfree(ls->ls_remove_names[i]); in new_lockspace()
668 vfree(ls->ls_rsbtbl); in new_lockspace()
671 kobject_put(&ls->ls_kobj); in new_lockspace()
679 static int __dlm_new_lockspace(const char *name, const char *cluster, in __dlm_new_lockspace() argument
693 error = new_lockspace(name, cluster, flags, lvblen, ops, ops_arg, in __dlm_new_lockspace()
709 int dlm_new_lockspace(const char *name, const char *cluster, uint32_t flags, in dlm_new_lockspace() argument
714 return __dlm_new_lockspace(name, cluster, flags | DLM_LSFL_FS, lvblen, in dlm_new_lockspace()
718 int dlm_new_user_lockspace(const char *name, const char *cluster, in dlm_new_user_lockspace() argument
724 return __dlm_new_lockspace(name, cluster, flags, lvblen, ops, in dlm_new_user_lockspace()
732 return lkb->lkb_nodeid == 0 && lkb->lkb_grmode != DLM_LOCK_IV; in lkb_idr_is_local()
744 if (lkb->lkb_lvbptr && test_bit(DLM_IFL_MSTCPY_BIT, &lkb->lkb_iflags)) in lkb_idr_free()
745 dlm_free_lvb(lkb->lkb_lvbptr); in lkb_idr_free()
759 spin_lock(&ls->ls_lkbidr_spin); in lockspace_busy()
761 rv = idr_for_each(&ls->ls_lkbidr, lkb_idr_is_any, ls); in lockspace_busy()
763 rv = idr_for_each(&ls->ls_lkbidr, lkb_idr_is_local, ls); in lockspace_busy()
767 spin_unlock(&ls->ls_lkbidr_spin); in lockspace_busy()
780 if (ls->ls_create_count == 1) { in release_lockspace()
782 rv = -EBUSY; in release_lockspace()
785 ls->ls_create_count = 0; in release_lockspace()
788 } else if (ls->ls_create_count > 1) { in release_lockspace()
789 rv = --ls->ls_create_count; in release_lockspace()
791 rv = -EINVAL; in release_lockspace()
822 idr_destroy(&ls->ls_recover_idr); in release_lockspace()
823 kfree(ls->ls_recover_buf); in release_lockspace()
829 idr_for_each(&ls->ls_lkbidr, lkb_idr_free, ls); in release_lockspace()
830 idr_destroy(&ls->ls_lkbidr); in release_lockspace()
836 for (i = 0; i < ls->ls_rsbtbl_size; i++) { in release_lockspace()
837 while ((n = rb_first(&ls->ls_rsbtbl[i].keep))) { in release_lockspace()
839 rb_erase(n, &ls->ls_rsbtbl[i].keep); in release_lockspace()
843 while ((n = rb_first(&ls->ls_rsbtbl[i].toss))) { in release_lockspace()
845 rb_erase(n, &ls->ls_rsbtbl[i].toss); in release_lockspace()
850 vfree(ls->ls_rsbtbl); in release_lockspace()
853 kfree(ls->ls_remove_names[i]); in release_lockspace()
855 while (!list_empty(&ls->ls_new_rsb)) { in release_lockspace()
856 rsb = list_first_entry(&ls->ls_new_rsb, struct dlm_rsb, in release_lockspace()
858 list_del(&rsb->res_hashchain); in release_lockspace()
867 kfree(ls->ls_recover_args); in release_lockspace()
870 kfree(ls->ls_node_array); in release_lockspace()
872 kobject_put(&ls->ls_kobj); in release_lockspace()
887 * 0 - don't destroy lockspace if it has any LKBs
888 * 1 - destroy lockspace if it has remote LKBs but not if it has local LKBs
889 * 2 - destroy lockspace regardless of LKBs
890 * 3 - destroy lockspace as part of a forced shutdown
900 return -EINVAL; in dlm_release_lockspace()
906 ls_count--; in dlm_release_lockspace()
923 if (!test_bit(LSFL_RUNNING, &ls->ls_flags)) { in dlm_stop_lockspaces()