Home
last modified time | relevance | path

Searched refs:mn (Results 1 – 25 of 53) sorted by relevance

123

/kernel/linux/linux-5.10/drivers/gpu/drm/i915/gem/
Di915_gem_userptr.c21 struct i915_mmu_notifier *mn; member
33 struct mmu_notifier mn; member
39 struct i915_mmu_notifier *mn; member
47 interval_tree_insert(&mo->it, &mo->mn->objects); in add_object()
55 interval_tree_remove(&mo->it, &mo->mn->objects); in del_object()
78 spin_lock(&mo->mn->lock); in __i915_gem_userptr_set_active()
83 spin_unlock(&mo->mn->lock); in __i915_gem_userptr_set_active()
90 struct i915_mmu_notifier *mn = in userptr_mn_invalidate_range_start() local
91 container_of(_mn, struct i915_mmu_notifier, mn); in userptr_mn_invalidate_range_start()
96 if (RB_EMPTY_ROOT(&mn->objects.rb_root)) in userptr_mn_invalidate_range_start()
[all …]
Di915_gem_object.c109 struct i915_mmap_offset *mmo, *mn; in i915_gem_close_object() local
133 rbtree_postorder_for_each_entry_safe(mmo, mn, &obj->mmo.offsets, offset) in i915_gem_close_object()
182 struct i915_mmap_offset *mmo, *mn; in __i915_gem_object_free_mmaps() local
186 rbtree_postorder_for_each_entry_safe(mmo, mn, in __i915_gem_object_free_mmaps()
/kernel/linux/linux-5.10/drivers/clk/qcom/
Dclk-rcg.c106 static u32 md_to_m(struct mn *mn, u32 md) in md_to_m() argument
108 md >>= mn->m_val_shift; in md_to_m()
109 md &= BIT(mn->width) - 1; in md_to_m()
132 static u32 mn_to_md(struct mn *mn, u32 m, u32 n, u32 md) in mn_to_md() argument
136 mask_w = BIT(mn->width) - 1; in mn_to_md()
137 mask = (mask_w << mn->m_val_shift) | mask_w; in mn_to_md()
141 m <<= mn->m_val_shift; in mn_to_md()
149 static u32 ns_m_to_n(struct mn *mn, u32 ns, u32 m) in ns_m_to_n() argument
151 ns = ~ns >> mn->n_val_shift; in ns_m_to_n()
152 ns &= BIT(mn->width) - 1; in ns_m_to_n()
[all …]
Dclk-rcg.h30 struct mn { struct
79 struct mn mn; member
117 struct mn mn[2]; member
Dgcc-msm8660.c103 .mn = {
154 .mn = {
205 .mn = {
256 .mn = {
307 .mn = {
358 .mn = {
409 .mn = {
460 .mn = {
509 .mn = {
558 .mn = {
[all …]
Dgcc-mdm9615.c187 .mn = {
238 .mn = {
289 .mn = {
340 .mn = {
391 .mn = {
454 .mn = {
503 .mn = {
552 .mn = {
601 .mn = {
650 .mn = {
[all …]
Dgcc-ipq806x.c343 .mn = {
394 .mn = {
445 .mn = {
496 .mn = {
547 .mn = {
598 .mn = {
662 .mn = {
711 .mn = {
760 .mn = {
809 .mn = {
[all …]
Dgcc-msm8960.c329 .mn = {
380 .mn = {
431 .mn = {
482 .mn = {
533 .mn = {
584 .mn = {
635 .mn = {
686 .mn = {
735 .mn = {
784 .mn = {
[all …]
Dmmcc-msm8960.c172 .mn = {
221 .mn = {
270 .mn = {
326 .mn = {
390 .mn = {
454 .mn = {
705 .mn = {
807 .mn[0] = {
815 .mn[1] = {
867 .mn[0] = {
[all …]
Dlcc-ipq806x.c111 .mn = {
225 .mn = {
305 .mn = {
364 .mn = {
Dlcc-msm8960.c92 .mn = {
198 .mn = { \
342 .mn = {
410 .mn = {
Dlcc-mdm9615.c94 .mn = {
200 .mn = { \
344 .mn = {
412 .mn = {
/kernel/linux/linux-5.10/drivers/misc/sgi-gru/
Dgrutlbpurge.c209 static int gru_invalidate_range_start(struct mmu_notifier *mn, in gru_invalidate_range_start() argument
212 struct gru_mm_struct *gms = container_of(mn, struct gru_mm_struct, in gru_invalidate_range_start()
224 static void gru_invalidate_range_end(struct mmu_notifier *mn, in gru_invalidate_range_end() argument
227 struct gru_mm_struct *gms = container_of(mn, struct gru_mm_struct, in gru_invalidate_range_end()
252 static void gru_free_notifier(struct mmu_notifier *mn) in gru_free_notifier() argument
254 kfree(container_of(mn, struct gru_mm_struct, ms_notifier)); in gru_free_notifier()
267 struct mmu_notifier *mn; in gru_register_mmu_notifier() local
269 mn = mmu_notifier_get_locked(&gru_mmuops, current->mm); in gru_register_mmu_notifier()
270 if (IS_ERR(mn)) in gru_register_mmu_notifier()
271 return ERR_CAST(mn); in gru_register_mmu_notifier()
[all …]
/kernel/linux/linux-5.10/drivers/net/ethernet/mellanox/mlx5/core/diag/
Den_rep_tracepoint.h24 TP_fast_assign(const struct mlx5e_neigh *mn = &nhe->m_neigh;
28 __assign_str(devname, mn->dev->name);
34 if (mn->family == AF_INET) {
35 *p32 = mn->dst_ip.v4;
37 } else if (mn->family == AF_INET6) {
38 *pin6 = mn->dst_ip.v6;
Den_tc_tracepoint.h85 TP_fast_assign(const struct mlx5e_neigh *mn = &nhe->m_neigh;
89 __assign_str(devname, mn->dev->name);
94 if (mn->family == AF_INET) {
95 *p32 = mn->dst_ip.v4;
97 } else if (mn->family == AF_INET6) {
98 *pin6 = mn->dst_ip.v6;
/kernel/linux/linux-5.10/drivers/gpu/drm/nouveau/nvkm/core/
Dmemory.c38 nvkm_mm_free(&fb->tags, &tags->mn); in nvkm_memory_tags_put()
61 if (tags->mn && tags->mn->length != nr) { in nvkm_memory_tags_get()
77 if (!nvkm_mm_head(&fb->tags, 0, 1, nr, nr, 1, &tags->mn)) { in nvkm_memory_tags_get()
79 clr(device, tags->mn->offset, tags->mn->length); in nvkm_memory_tags_get()
90 tags->mn = NULL; in nvkm_memory_tags_get()
/kernel/linux/linux-5.10/drivers/infiniband/hw/hfi1/
Dmmu_rb.c99 INIT_HLIST_NODE(&h->mn.hlist); in hfi1_mmu_rb_register()
101 h->mn.ops = &mn_opts; in hfi1_mmu_rb_register()
107 ret = mmu_notifier_register(&h->mn, current->mm); in hfi1_mmu_rb_register()
125 mmu_notifier_unregister(&handler->mn, handler->mn.mm); in hfi1_mmu_rb_unregister()
158 if (current->mm != handler->mn.mm) in hfi1_mmu_rb_insert()
213 if (current->mm != handler->mn.mm) in hfi1_mmu_rb_remove_unless_exact()
238 if (current->mm != handler->mn.mm) in hfi1_mmu_rb_evict()
274 if (current->mm != handler->mn.mm) in hfi1_mmu_rb_remove()
287 static int mmu_notifier_range_start(struct mmu_notifier *mn, in mmu_notifier_range_start() argument
291 container_of(mn, struct mmu_rb_handler, mn); in mmu_notifier_range_start()
/kernel/linux/linux-5.10/mm/
Dpage_ext.c341 struct memory_notify *mn = arg; in page_ext_callback() local
346 ret = online_page_ext(mn->start_pfn, in page_ext_callback()
347 mn->nr_pages, mn->status_change_nid); in page_ext_callback()
350 offline_page_ext(mn->start_pfn, in page_ext_callback()
351 mn->nr_pages, mn->status_change_nid); in page_ext_callback()
354 offline_page_ext(mn->start_pfn, in page_ext_callback()
355 mn->nr_pages, mn->status_change_nid); in page_ext_callback()
/kernel/linux/linux-5.10/drivers/gpu/drm/nouveau/nvkm/subdev/fb/
Dram.c34 struct nvkm_mm_node *mn; member
45 .mem = vram->mn, in nvkm_vram_map()
54 return (u64)nvkm_mm_size(nvkm_vram(memory)->mn) << NVKM_RAM_MM_SHIFT; in nvkm_vram_size()
61 if (!nvkm_mm_contiguous(vram->mn)) in nvkm_vram_addr()
63 return (u64)nvkm_mm_addr(vram->mn) << NVKM_RAM_MM_SHIFT; in nvkm_vram_addr()
82 struct nvkm_mm_node *next = vram->mn; in nvkm_vram_dtor()
130 node = &vram->mn; in nvkm_ram_get()
/kernel/linux/linux-5.10/drivers/iommu/amd/
Diommu_v2.c40 struct mmu_notifier mn; /* mmu_notifier handle */ member
335 mmu_notifier_unregister(&pasid_state->mn, pasid_state->mm); in free_pasid_states()
354 static struct pasid_state *mn_to_state(struct mmu_notifier *mn) in mn_to_state() argument
356 return container_of(mn, struct pasid_state, mn); in mn_to_state()
359 static void mn_invalidate_range(struct mmu_notifier *mn, in mn_invalidate_range() argument
366 pasid_state = mn_to_state(mn); in mn_invalidate_range()
376 static void mn_release(struct mmu_notifier *mn, struct mm_struct *mm) in mn_release() argument
384 pasid_state = mn_to_state(mn); in mn_release()
637 pasid_state->mn.ops = &iommu_mn; in amd_iommu_bind_pasid()
642 mmu_notifier_register(&pasid_state->mn, mm); in amd_iommu_bind_pasid()
[all …]
/kernel/linux/linux-5.10/drivers/gpu/drm/nouveau/nvkm/subdev/instmem/
Dgk20a.c53 struct nvkm_mm_node *mn; member
130 return (u64)gk20a_instobj(memory)->mn->offset << 12; in gk20a_instobj_addr()
136 return (u64)gk20a_instobj(memory)->mn->length << 12; in gk20a_instobj_size()
288 .mem = node->mn, in gk20a_instobj_map()
304 dma_free_attrs(dev, (u64)node->base.mn->length << PAGE_SHIFT, in gk20a_instobj_dtor_dma()
317 struct nvkm_mm_node *r = node->base.mn; in gk20a_instobj_dtor_iommu()
335 for (i = 0; i < node->base.mn->length; i++) { in gk20a_instobj_dtor_iommu()
416 node->base.mn = &node->r; in gk20a_instobj_ctor_dma()
493 node->base.mn = r; in gk20a_instobj_ctor_iommu()
542 size, align, (u64)node->mn->offset << 12); in gk20a_instobj_new()
/kernel/linux/linux-5.10/drivers/scsi/qla2xxx/
Dqla_bsg.c1002 struct verify_chip_entry_84xx *mn = NULL; in qla84xx_updatefw() local
1048 mn = dma_pool_zalloc(ha->s_dma_pool, GFP_KERNEL, &mn_dma); in qla84xx_updatefw()
1049 if (!mn) { in qla84xx_updatefw()
1059 mn->entry_type = VERIFY_CHIP_IOCB_TYPE; in qla84xx_updatefw()
1060 mn->entry_count = 1; in qla84xx_updatefw()
1066 mn->options = cpu_to_le16(options); in qla84xx_updatefw()
1067 mn->fw_ver = cpu_to_le32(fw_ver); in qla84xx_updatefw()
1068 mn->fw_size = cpu_to_le32(data_len); in qla84xx_updatefw()
1069 mn->fw_seq_size = cpu_to_le32(data_len); in qla84xx_updatefw()
1070 put_unaligned_le64(fw_dma, &mn->dsd.address); in qla84xx_updatefw()
[all …]
/kernel/linux/linux-5.10/drivers/gpu/drm/radeon/
Dradeon_mn.c48 static bool radeon_mn_invalidate(struct mmu_interval_notifier *mn, in radeon_mn_invalidate() argument
52 struct radeon_bo *bo = container_of(mn, struct radeon_bo, notifier); in radeon_mn_invalidate()
/kernel/linux/linux-5.10/arch/x86/kernel/apic/
Dx2apic_uv_x.c1236 struct mn { struct
1244 static void get_mn(struct mn *mnp) in get_mn() argument
1266 struct mn mn; in uv_init_hub_info() local
1268 get_mn(&mn); in uv_init_hub_info()
1269 hi->gpa_mask = mn.m_val ? in uv_init_hub_info()
1270 (1UL << (mn.m_val + mn.n_val)) - 1 : in uv_init_hub_info()
1273 hi->m_val = mn.m_val; in uv_init_hub_info()
1274 hi->n_val = mn.n_val; in uv_init_hub_info()
1275 hi->m_shift = mn.m_shift; in uv_init_hub_info()
1276 hi->n_lshift = mn.n_lshift ? mn.n_lshift : 0; in uv_init_hub_info()
[all …]
/kernel/linux/linux-5.10/net/sched/
Dsch_teql.c231 struct neighbour *mn; in __teql_resolve() local
233 mn = __neigh_lookup_errno(n->tbl, n->primary_key, dev); in __teql_resolve()
235 if (IS_ERR(mn)) in __teql_resolve()
236 return PTR_ERR(mn); in __teql_resolve()
237 n = mn; in __teql_resolve()

123