| /kernel/linux/linux-5.10/kernel/sched/ |
| D | core_ctl.c | 60 struct cluster_data *cluster; member 69 #define for_each_cluster(cluster, idx) \ argument 70 for (; (idx) < num_clusters && ((cluster) = &cluster_state[idx]);\ 81 static unsigned int get_active_cpu_count(const struct cluster_data *cluster); 161 struct cluster_data *cluster; in show_global_state() local 168 cluster = c->cluster; in show_global_state() 169 if (!cluster || !cluster->inited) in show_global_state() 184 cluster->first_cpu); in show_global_state() 193 "\tNr running: %u\n", cluster->nrrun); in show_global_state() 195 "\tActive CPUs: %u\n", get_active_cpu_count(cluster)); in show_global_state() [all …]
|
| /kernel/linux/linux-6.6/kernel/sched/ |
| D | core_ctl.c | 60 struct cluster_data *cluster; member 69 #define for_each_cluster(cluster, idx) \ argument 70 for (; (idx) < num_clusters && ((cluster) = &cluster_state[idx]);\ 81 static unsigned int get_active_cpu_count(const struct cluster_data *cluster); 161 struct cluster_data *cluster; in show_global_state() local 168 cluster = c->cluster; in show_global_state() 169 if (!cluster || !cluster->inited) in show_global_state() 184 cluster->first_cpu); in show_global_state() 193 "\tNr running: %u\n", cluster->nrrun); in show_global_state() 195 "\tActive CPUs: %u\n", get_active_cpu_count(cluster)); in show_global_state() [all …]
|
| /kernel/linux/linux-5.10/arch/arm/common/ |
| D | mcpm_entry.c | 3 * arch/arm/common/mcpm_entry.c -- entry point for multi-cluster PM 24 * see Documentation/arm/cluster-pm-race-avoidance.rst. 34 static void __mcpm_cpu_going_down(unsigned int cpu, unsigned int cluster) in __mcpm_cpu_going_down() argument 36 mcpm_sync.clusters[cluster].cpus[cpu].cpu = CPU_GOING_DOWN; in __mcpm_cpu_going_down() 37 sync_cache_w(&mcpm_sync.clusters[cluster].cpus[cpu].cpu); in __mcpm_cpu_going_down() 42 * cluster can be torn down without disrupting this CPU. 47 static void __mcpm_cpu_down(unsigned int cpu, unsigned int cluster) in __mcpm_cpu_down() argument 50 mcpm_sync.clusters[cluster].cpus[cpu].cpu = CPU_DOWN; in __mcpm_cpu_down() 51 sync_cache_w(&mcpm_sync.clusters[cluster].cpus[cpu].cpu); in __mcpm_cpu_down() 56 * __mcpm_outbound_leave_critical: Leave the cluster teardown critical section. [all …]
|
| /kernel/linux/linux-6.6/arch/arm/common/ |
| D | mcpm_entry.c | 3 * arch/arm/common/mcpm_entry.c -- entry point for multi-cluster PM 24 * see Documentation/arch/arm/cluster-pm-race-avoidance.rst. 34 static void __mcpm_cpu_going_down(unsigned int cpu, unsigned int cluster) in __mcpm_cpu_going_down() argument 36 mcpm_sync.clusters[cluster].cpus[cpu].cpu = CPU_GOING_DOWN; in __mcpm_cpu_going_down() 37 sync_cache_w(&mcpm_sync.clusters[cluster].cpus[cpu].cpu); in __mcpm_cpu_going_down() 42 * cluster can be torn down without disrupting this CPU. 47 static void __mcpm_cpu_down(unsigned int cpu, unsigned int cluster) in __mcpm_cpu_down() argument 50 mcpm_sync.clusters[cluster].cpus[cpu].cpu = CPU_DOWN; in __mcpm_cpu_down() 51 sync_cache_w(&mcpm_sync.clusters[cluster].cpus[cpu].cpu); in __mcpm_cpu_down() 56 * __mcpm_outbound_leave_critical: Leave the cluster teardown critical section. [all …]
|
| /kernel/linux/linux-5.10/arch/arm/mach-sunxi/ |
| D | mc_smp.c | 11 * Cluster cache enable trampoline code adapted from MCPM framework 88 static bool sunxi_core_is_cortex_a15(unsigned int core, unsigned int cluster) in sunxi_core_is_cortex_a15() argument 91 int cpu = cluster * SUNXI_CPUS_PER_CLUSTER + core; in sunxi_core_is_cortex_a15() 103 * would be mid way in a core or cluster power sequence. in sunxi_core_is_cortex_a15() 105 pr_err("%s: Couldn't get CPU cluster %u core %u device node\n", in sunxi_core_is_cortex_a15() 106 __func__, cluster, core); in sunxi_core_is_cortex_a15() 116 static int sunxi_cpu_power_switch_set(unsigned int cpu, unsigned int cluster, in sunxi_cpu_power_switch_set() argument 122 reg = readl(prcm_base + PRCM_PWR_SWITCH_REG(cluster, cpu)); in sunxi_cpu_power_switch_set() 125 pr_debug("power clamp for cluster %u cpu %u already open\n", in sunxi_cpu_power_switch_set() 126 cluster, cpu); in sunxi_cpu_power_switch_set() [all …]
|
| /kernel/linux/linux-6.6/arch/arm/mach-sunxi/ |
| D | mc_smp.c | 11 * Cluster cache enable trampoline code adapted from MCPM framework 87 static bool sunxi_core_is_cortex_a15(unsigned int core, unsigned int cluster) in sunxi_core_is_cortex_a15() argument 90 int cpu = cluster * SUNXI_CPUS_PER_CLUSTER + core; in sunxi_core_is_cortex_a15() 102 * would be mid way in a core or cluster power sequence. in sunxi_core_is_cortex_a15() 104 pr_err("%s: Couldn't get CPU cluster %u core %u device node\n", in sunxi_core_is_cortex_a15() 105 __func__, cluster, core); in sunxi_core_is_cortex_a15() 115 static int sunxi_cpu_power_switch_set(unsigned int cpu, unsigned int cluster, in sunxi_cpu_power_switch_set() argument 121 reg = readl(prcm_base + PRCM_PWR_SWITCH_REG(cluster, cpu)); in sunxi_cpu_power_switch_set() 124 pr_debug("power clamp for cluster %u cpu %u already open\n", in sunxi_cpu_power_switch_set() 125 cluster, cpu); in sunxi_cpu_power_switch_set() [all …]
|
| /kernel/linux/linux-5.10/arch/arm/mach-vexpress/ |
| D | tc2_pm.c | 48 static int tc2_pm_cpu_powerup(unsigned int cpu, unsigned int cluster) in tc2_pm_cpu_powerup() argument 50 pr_debug("%s: cpu %u cluster %u\n", __func__, cpu, cluster); in tc2_pm_cpu_powerup() 51 if (cluster >= TC2_CLUSTERS || cpu >= tc2_nr_cpus[cluster]) in tc2_pm_cpu_powerup() 53 ve_spc_set_resume_addr(cluster, cpu, in tc2_pm_cpu_powerup() 55 ve_spc_cpu_wakeup_irq(cluster, cpu, true); in tc2_pm_cpu_powerup() 59 static int tc2_pm_cluster_powerup(unsigned int cluster) in tc2_pm_cluster_powerup() argument 61 pr_debug("%s: cluster %u\n", __func__, cluster); in tc2_pm_cluster_powerup() 62 if (cluster >= TC2_CLUSTERS) in tc2_pm_cluster_powerup() 64 ve_spc_powerdown(cluster, false); in tc2_pm_cluster_powerup() 68 static void tc2_pm_cpu_powerdown_prepare(unsigned int cpu, unsigned int cluster) in tc2_pm_cpu_powerdown_prepare() argument [all …]
|
| D | dcscb.c | 3 * arch/arm/mach-vexpress/dcscb.c - Dual Cluster System Configuration Block 39 static int dcscb_cpu_powerup(unsigned int cpu, unsigned int cluster) in dcscb_cpu_powerup() argument 43 pr_debug("%s: cpu %u cluster %u\n", __func__, cpu, cluster); in dcscb_cpu_powerup() 44 if (cluster >= 2 || !(cpumask & dcscb_allcpus_mask[cluster])) in dcscb_cpu_powerup() 47 rst_hold = readl_relaxed(dcscb_base + RST_HOLD0 + cluster * 4); in dcscb_cpu_powerup() 49 writel_relaxed(rst_hold, dcscb_base + RST_HOLD0 + cluster * 4); in dcscb_cpu_powerup() 53 static int dcscb_cluster_powerup(unsigned int cluster) in dcscb_cluster_powerup() argument 57 pr_debug("%s: cluster %u\n", __func__, cluster); in dcscb_cluster_powerup() 58 if (cluster >= 2) in dcscb_cluster_powerup() 61 /* remove cluster reset and add individual CPU's reset */ in dcscb_cluster_powerup() [all …]
|
| D | spc.c | 58 /* SPC CPU/cluster reset statue */ 79 /* TC2 static dual-cluster configuration */ 105 * A15s cluster identifier 119 static inline bool cluster_is_a15(u32 cluster) in cluster_is_a15() argument 121 return cluster == info->a15_clusid; in cluster_is_a15() 154 * @cluster: mpidr[15:8] bitfield describing cluster affinity level 158 void ve_spc_cpu_wakeup_irq(u32 cluster, u32 cpu, bool set) in ve_spc_cpu_wakeup_irq() argument 162 if (cluster >= MAX_CLUSTERS) in ve_spc_cpu_wakeup_irq() 167 if (!cluster_is_a15(cluster)) in ve_spc_cpu_wakeup_irq() 183 * @cluster: mpidr[15:8] bitfield describing cluster affinity level [all …]
|
| /kernel/linux/linux-6.6/arch/arm/mach-versatile/ |
| D | tc2_pm.c | 46 static int tc2_pm_cpu_powerup(unsigned int cpu, unsigned int cluster) in tc2_pm_cpu_powerup() argument 48 pr_debug("%s: cpu %u cluster %u\n", __func__, cpu, cluster); in tc2_pm_cpu_powerup() 49 if (cluster >= TC2_CLUSTERS || cpu >= tc2_nr_cpus[cluster]) in tc2_pm_cpu_powerup() 51 ve_spc_set_resume_addr(cluster, cpu, in tc2_pm_cpu_powerup() 53 ve_spc_cpu_wakeup_irq(cluster, cpu, true); in tc2_pm_cpu_powerup() 57 static int tc2_pm_cluster_powerup(unsigned int cluster) in tc2_pm_cluster_powerup() argument 59 pr_debug("%s: cluster %u\n", __func__, cluster); in tc2_pm_cluster_powerup() 60 if (cluster >= TC2_CLUSTERS) in tc2_pm_cluster_powerup() 62 ve_spc_powerdown(cluster, false); in tc2_pm_cluster_powerup() 66 static void tc2_pm_cpu_powerdown_prepare(unsigned int cpu, unsigned int cluster) in tc2_pm_cpu_powerdown_prepare() argument [all …]
|
| D | dcscb.c | 3 * dcscb.c - Dual Cluster System Configuration Block 39 static int dcscb_cpu_powerup(unsigned int cpu, unsigned int cluster) in dcscb_cpu_powerup() argument 43 pr_debug("%s: cpu %u cluster %u\n", __func__, cpu, cluster); in dcscb_cpu_powerup() 44 if (cluster >= 2 || !(cpumask & dcscb_allcpus_mask[cluster])) in dcscb_cpu_powerup() 47 rst_hold = readl_relaxed(dcscb_base + RST_HOLD0 + cluster * 4); in dcscb_cpu_powerup() 49 writel_relaxed(rst_hold, dcscb_base + RST_HOLD0 + cluster * 4); in dcscb_cpu_powerup() 53 static int dcscb_cluster_powerup(unsigned int cluster) in dcscb_cluster_powerup() argument 57 pr_debug("%s: cluster %u\n", __func__, cluster); in dcscb_cluster_powerup() 58 if (cluster >= 2) in dcscb_cluster_powerup() 61 /* remove cluster reset and add individual CPU's reset */ in dcscb_cluster_powerup() [all …]
|
| D | spc.c | 50 /* SPC CPU/cluster reset statue */ 71 /* TC2 static dual-cluster configuration */ 97 * A15s cluster identifier 111 static inline bool cluster_is_a15(u32 cluster) in cluster_is_a15() argument 113 return cluster == info->a15_clusid; in cluster_is_a15() 142 * @cluster: mpidr[15:8] bitfield describing cluster affinity level 150 void ve_spc_cpu_wakeup_irq(u32 cluster, u32 cpu, bool set) in ve_spc_cpu_wakeup_irq() argument 154 if (cluster >= MAX_CLUSTERS) in ve_spc_cpu_wakeup_irq() 159 if (!cluster_is_a15(cluster)) in ve_spc_cpu_wakeup_irq() 175 * @cluster: mpidr[15:8] bitfield describing cluster affinity level [all …]
|
| /kernel/linux/linux-5.10/Documentation/arm/ |
| D | cluster-pm-race-avoidance.rst | 2 Cluster-wide Power-up/power-down race avoidance algorithm 6 cluster setup and teardown operations and to manage hardware coherency 29 cluster-level operations are only performed when it is truly safe to do 34 are not immediately enabled when a cluster powers up. Since enabling or 38 power-down and power-up at the cluster level. 48 Each cluster and CPU is assigned a state, as follows: 67 The CPU or cluster is not coherent, and is either powered off or 71 The CPU or cluster has committed to moving to the UP state. 76 The CPU or cluster is active and coherent at the hardware 81 The CPU or cluster has committed to moving to the DOWN [all …]
|
| /kernel/linux/linux-6.6/Documentation/arch/arm/ |
| D | cluster-pm-race-avoidance.rst | 2 Cluster-wide Power-up/power-down race avoidance algorithm 6 cluster setup and teardown operations and to manage hardware coherency 29 cluster-level operations are only performed when it is truly safe to do 34 are not immediately enabled when a cluster powers up. Since enabling or 38 power-down and power-up at the cluster level. 48 Each cluster and CPU is assigned a state, as follows: 67 The CPU or cluster is not coherent, and is either powered off or 71 The CPU or cluster has committed to moving to the UP state. 76 The CPU or cluster is active and coherent at the hardware 81 The CPU or cluster has committed to moving to the DOWN [all …]
|
| /kernel/linux/linux-5.10/fs/ocfs2/cluster/ |
| D | nodemanager.c | 20 * cluster active at a time. Changing this will require trickling 21 * cluster references throughout where nodes are looked up */ 51 struct o2nm_cluster *cluster = o2nm_single_cluster; in o2nm_configured_node_map() local 53 BUG_ON(bytes < (sizeof(cluster->cl_nodes_bitmap))); in o2nm_configured_node_map() 55 if (cluster == NULL) in o2nm_configured_node_map() 58 read_lock(&cluster->cl_nodes_lock); in o2nm_configured_node_map() 59 memcpy(map, cluster->cl_nodes_bitmap, sizeof(cluster->cl_nodes_bitmap)); in o2nm_configured_node_map() 60 read_unlock(&cluster->cl_nodes_lock); in o2nm_configured_node_map() 66 static struct o2nm_node *o2nm_node_ip_tree_lookup(struct o2nm_cluster *cluster, in o2nm_node_ip_tree_lookup() argument 71 struct rb_node **p = &cluster->cl_node_ip_tree.rb_node; in o2nm_node_ip_tree_lookup() [all …]
|
| /kernel/linux/linux-6.6/fs/ocfs2/cluster/ |
| D | nodemanager.c | 18 * cluster active at a time. Changing this will require trickling 19 * cluster references throughout where nodes are looked up */ 49 struct o2nm_cluster *cluster = o2nm_single_cluster; in o2nm_configured_node_map() local 51 BUG_ON(bytes < (sizeof(cluster->cl_nodes_bitmap))); in o2nm_configured_node_map() 53 if (cluster == NULL) in o2nm_configured_node_map() 56 read_lock(&cluster->cl_nodes_lock); in o2nm_configured_node_map() 57 bitmap_copy(map, cluster->cl_nodes_bitmap, O2NM_MAX_NODES); in o2nm_configured_node_map() 58 read_unlock(&cluster->cl_nodes_lock); in o2nm_configured_node_map() 64 static struct o2nm_node *o2nm_node_ip_tree_lookup(struct o2nm_cluster *cluster, in o2nm_node_ip_tree_lookup() argument 69 struct rb_node **p = &cluster->cl_node_ip_tree.rb_node; in o2nm_node_ip_tree_lookup() [all …]
|
| /kernel/linux/linux-6.6/drivers/perf/ |
| D | qcom_l2_pmu.c | 121 * The cache is made up of one or more clusters, each cluster has its own PMU. 122 * Each cluster is associated with one or more CPUs. 143 /* The CPU that is used for collecting events on this cluster */ 145 /* All the CPUs associated with this cluster */ 243 static void cluster_pmu_set_resr(struct cluster_pmu *cluster, in cluster_pmu_set_resr() argument 254 spin_lock_irqsave(&cluster->pmu_lock, flags); in cluster_pmu_set_resr() 262 spin_unlock_irqrestore(&cluster->pmu_lock, flags); in cluster_pmu_set_resr() 268 * all CPUS, subunits and ID independent events in this cluster. 319 static void l2_cache_cluster_set_period(struct cluster_pmu *cluster, in l2_cache_cluster_set_period() argument 339 static int l2_cache_get_event_idx(struct cluster_pmu *cluster, in l2_cache_get_event_idx() argument [all …]
|
| /kernel/linux/linux-5.10/drivers/perf/ |
| D | qcom_l2_pmu.c | 121 * The cache is made up of one or more clusters, each cluster has its own PMU. 122 * Each cluster is associated with one or more CPUs. 143 /* The CPU that is used for collecting events on this cluster */ 145 /* All the CPUs associated with this cluster */ 243 static void cluster_pmu_set_resr(struct cluster_pmu *cluster, in cluster_pmu_set_resr() argument 254 spin_lock_irqsave(&cluster->pmu_lock, flags); in cluster_pmu_set_resr() 262 spin_unlock_irqrestore(&cluster->pmu_lock, flags); in cluster_pmu_set_resr() 268 * all CPUS, subunits and ID independent events in this cluster. 319 static void l2_cache_cluster_set_period(struct cluster_pmu *cluster, in l2_cache_cluster_set_period() argument 339 static int l2_cache_get_event_idx(struct cluster_pmu *cluster, in l2_cache_get_event_idx() argument [all …]
|
| /kernel/linux/linux-6.6/arch/arm/mach-exynos/ |
| D | mcpm-exynos.c | 57 static int exynos_cpu_powerup(unsigned int cpu, unsigned int cluster) in exynos_cpu_powerup() argument 59 unsigned int cpunr = cpu + (cluster * EXYNOS5420_CPUS_PER_CLUSTER); in exynos_cpu_powerup() 62 pr_debug("%s: cpu %u cluster %u\n", __func__, cpu, cluster); in exynos_cpu_powerup() 64 cluster >= EXYNOS5420_NR_CLUSTERS) in exynos_cpu_powerup() 71 * This assumes the cluster number of the big cores(Cortex A15) in exynos_cpu_powerup() 76 if (cluster && in exynos_cpu_powerup() 77 cluster == MPIDR_AFFINITY_LEVEL(cpu_logical_map(0), 1)) { in exynos_cpu_powerup() 92 pr_err("cpu %u cluster %u powerup failed\n", in exynos_cpu_powerup() 93 cpu, cluster); in exynos_cpu_powerup() 106 static int exynos_cluster_powerup(unsigned int cluster) in exynos_cluster_powerup() argument [all …]
|
| /kernel/linux/linux-5.10/arch/arm/mach-exynos/ |
| D | mcpm-exynos.c | 59 static int exynos_cpu_powerup(unsigned int cpu, unsigned int cluster) in exynos_cpu_powerup() argument 61 unsigned int cpunr = cpu + (cluster * EXYNOS5420_CPUS_PER_CLUSTER); in exynos_cpu_powerup() 64 pr_debug("%s: cpu %u cluster %u\n", __func__, cpu, cluster); in exynos_cpu_powerup() 66 cluster >= EXYNOS5420_NR_CLUSTERS) in exynos_cpu_powerup() 73 * This assumes the cluster number of the big cores(Cortex A15) in exynos_cpu_powerup() 78 if (cluster && in exynos_cpu_powerup() 79 cluster == MPIDR_AFFINITY_LEVEL(cpu_logical_map(0), 1)) { in exynos_cpu_powerup() 94 pr_err("cpu %u cluster %u powerup failed\n", in exynos_cpu_powerup() 95 cpu, cluster); in exynos_cpu_powerup() 108 static int exynos_cluster_powerup(unsigned int cluster) in exynos_cluster_powerup() argument [all …]
|
| /kernel/linux/linux-6.6/arch/arm/include/asm/ |
| D | mcpm.h | 13 * Maximum number of possible clusters / CPUs per cluster. 39 * This is used to indicate where the given CPU from given cluster should 44 void mcpm_set_entry_vector(unsigned cpu, unsigned cluster, void *ptr); 51 void mcpm_set_early_poke(unsigned cpu, unsigned cluster, 55 * CPU/cluster power operations API for higher subsystems to use. 66 * mcpm_cpu_power_up - make given CPU in given cluster runable 68 * @cpu: CPU number within given cluster 69 * @cluster: cluster number for the CPU 71 * The identified CPU is brought out of reset. If the cluster was powered 73 * in the cluster run, and ensuring appropriate cluster setup. [all …]
|
| /kernel/linux/linux-5.10/arch/arm/include/asm/ |
| D | mcpm.h | 13 * Maximum number of possible clusters / CPUs per cluster. 39 * This is used to indicate where the given CPU from given cluster should 44 void mcpm_set_entry_vector(unsigned cpu, unsigned cluster, void *ptr); 51 void mcpm_set_early_poke(unsigned cpu, unsigned cluster, 55 * CPU/cluster power operations API for higher subsystems to use. 66 * mcpm_cpu_power_up - make given CPU in given cluster runable 68 * @cpu: CPU number within given cluster 69 * @cluster: cluster number for the CPU 71 * The identified CPU is brought out of reset. If the cluster was powered 73 * in the cluster run, and ensuring appropriate cluster setup. [all …]
|
| /kernel/linux/linux-6.6/arch/arm/mach-hisi/ |
| D | platmcpm.c | 71 static bool hip04_cluster_is_down(unsigned int cluster) in hip04_cluster_is_down() argument 76 if (hip04_cpu_table[cluster][i]) in hip04_cluster_is_down() 81 static void hip04_set_snoop_filter(unsigned int cluster, unsigned int on) in hip04_set_snoop_filter() argument 89 data |= 1 << cluster; in hip04_set_snoop_filter() 91 data &= ~(1 << cluster); in hip04_set_snoop_filter() 100 unsigned int mpidr, cpu, cluster; in hip04_boot_secondary() local 106 cluster = MPIDR_AFFINITY_LEVEL(mpidr, 1); in hip04_boot_secondary() 110 if (cluster >= HIP04_MAX_CLUSTERS || cpu >= HIP04_MAX_CPUS_PER_CLUSTER) in hip04_boot_secondary() 115 if (hip04_cpu_table[cluster][cpu]) in hip04_boot_secondary() 118 sys_dreq = sysctrl + SC_CPU_RESET_DREQ(cluster); in hip04_boot_secondary() [all …]
|
| /kernel/linux/linux-5.10/arch/arm/mach-hisi/ |
| D | platmcpm.c | 71 static bool hip04_cluster_is_down(unsigned int cluster) in hip04_cluster_is_down() argument 76 if (hip04_cpu_table[cluster][i]) in hip04_cluster_is_down() 81 static void hip04_set_snoop_filter(unsigned int cluster, unsigned int on) in hip04_set_snoop_filter() argument 89 data |= 1 << cluster; in hip04_set_snoop_filter() 91 data &= ~(1 << cluster); in hip04_set_snoop_filter() 100 unsigned int mpidr, cpu, cluster; in hip04_boot_secondary() local 106 cluster = MPIDR_AFFINITY_LEVEL(mpidr, 1); in hip04_boot_secondary() 110 if (cluster >= HIP04_MAX_CLUSTERS || cpu >= HIP04_MAX_CPUS_PER_CLUSTER) in hip04_boot_secondary() 115 if (hip04_cpu_table[cluster][cpu]) in hip04_boot_secondary() 118 sys_dreq = sysctrl + SC_CPU_RESET_DREQ(cluster); in hip04_boot_secondary() [all …]
|
| /kernel/linux/linux-5.10/Documentation/ABI/testing/ |
| D | sysfs-ocfs2 | 14 covers how ocfs2 uses distributed locking between cluster 18 cluster nodes can interoperate if they have an identical 34 the available plugins to support ocfs2 cluster operation. 35 A cluster plugin is required to use ocfs2 in a cluster. 38 * 'o2cb' - The classic o2cb cluster stack that ocfs2 has 40 * 'user' - A plugin supporting userspace cluster software 54 cluster plugin is currently in use by the filesystem. 62 the cluster stack in use. The contents may change 63 when all filesystems are unmounted and the cluster stack 71 of current ocfs2 cluster stack. This value is set by [all …]
|