Home
last modified time | relevance | path

Searched full:virt (Results 1 – 25 of 1558) sorted by relevance

12345678910>>...63

/kernel/linux/linux-5.10/drivers/gpu/drm/amd/amdgpu/
Damdgpu_virt.c114 struct amdgpu_virt *virt = &adev->virt; in amdgpu_virt_request_full_gpu() local
117 if (virt->ops && virt->ops->req_full_gpu) { in amdgpu_virt_request_full_gpu()
118 r = virt->ops->req_full_gpu(adev, init); in amdgpu_virt_request_full_gpu()
122 adev->virt.caps &= ~AMDGPU_SRIOV_CAPS_RUNTIME; in amdgpu_virt_request_full_gpu()
137 struct amdgpu_virt *virt = &adev->virt; in amdgpu_virt_release_full_gpu() local
140 if (virt->ops && virt->ops->rel_full_gpu) { in amdgpu_virt_release_full_gpu()
141 r = virt->ops->rel_full_gpu(adev, init); in amdgpu_virt_release_full_gpu()
145 adev->virt.caps |= AMDGPU_SRIOV_CAPS_RUNTIME; in amdgpu_virt_release_full_gpu()
158 struct amdgpu_virt *virt = &adev->virt; in amdgpu_virt_reset_gpu() local
161 if (virt->ops && virt->ops->reset_gpu) { in amdgpu_virt_reset_gpu()
[all …]
Damdgpu_vf_error.c41 mutex_lock(&adev->virt.vf_errors.lock); in amdgpu_vf_error_put()
42 index = adev->virt.vf_errors.write_count % AMDGPU_VF_ERROR_ENTRY_SIZE; in amdgpu_vf_error_put()
43 adev->virt.vf_errors.code [index] = error_code; in amdgpu_vf_error_put()
44 adev->virt.vf_errors.flags [index] = error_flags; in amdgpu_vf_error_put()
45 adev->virt.vf_errors.data [index] = error_data; in amdgpu_vf_error_put()
46 adev->virt.vf_errors.write_count ++; in amdgpu_vf_error_put()
47 mutex_unlock(&adev->virt.vf_errors.lock); in amdgpu_vf_error_put()
58 (!adev->virt.ops) || (!adev->virt.ops->trans_msg)) { in amdgpu_vf_error_trans_all()
69 mutex_lock(&adev->virt.vf_errors.lock); in amdgpu_vf_error_trans_all()
71 …if (adev->virt.vf_errors.write_count - adev->virt.vf_errors.read_count > AMDGPU_VF_ERROR_ENTRY_SIZ… in amdgpu_vf_error_trans_all()
[all …]
Dmxgpu_nv.c178 adev->virt.req_init_data_ver = 0; in xgpu_nv_send_access_requests()
182 adev->virt.req_init_data_ver = in xgpu_nv_send_access_requests()
186 if (adev->virt.req_init_data_ver < 1) in xgpu_nv_send_access_requests()
187 adev->virt.req_init_data_ver = 1; in xgpu_nv_send_access_requests()
193 adev->virt.fw_reserve.checksum_key = in xgpu_nv_send_access_requests()
259 struct amdgpu_virt *virt = container_of(work, struct amdgpu_virt, flr_work); in xgpu_nv_mailbox_flr_work() local
260 struct amdgpu_device *adev = container_of(virt, struct amdgpu_device, virt); in xgpu_nv_mailbox_flr_work()
320 schedule_work(&adev->virt.flr_work); in xgpu_nv_mailbox_rcv_irq()
348 adev->virt.ack_irq.num_types = 1; in xgpu_nv_mailbox_set_irq_funcs()
349 adev->virt.ack_irq.funcs = &xgpu_nv_mailbox_ack_irq_funcs; in xgpu_nv_mailbox_set_irq_funcs()
[all …]
Dmxgpu_ai.c179 adev->virt.fw_reserve.checksum_key = in xgpu_ai_send_access_requests()
238 struct amdgpu_virt *virt = container_of(work, struct amdgpu_virt, flr_work); in xgpu_ai_mailbox_flr_work() local
239 struct amdgpu_device *adev = container_of(virt, struct amdgpu_device, virt); in xgpu_ai_mailbox_flr_work()
293 schedule_work(&adev->virt.flr_work); in xgpu_ai_mailbox_rcv_irq()
324 adev->virt.ack_irq.num_types = 1; in xgpu_ai_mailbox_set_irq_funcs()
325 adev->virt.ack_irq.funcs = &xgpu_ai_mailbox_ack_irq_funcs; in xgpu_ai_mailbox_set_irq_funcs()
326 adev->virt.rcv_irq.num_types = 1; in xgpu_ai_mailbox_set_irq_funcs()
327 adev->virt.rcv_irq.funcs = &xgpu_ai_mailbox_rcv_irq_funcs; in xgpu_ai_mailbox_set_irq_funcs()
334 r = amdgpu_irq_add_id(adev, SOC15_IH_CLIENTID_BIF, 135, &adev->virt.rcv_irq); in xgpu_ai_mailbox_add_irq_id()
338 r = amdgpu_irq_add_id(adev, SOC15_IH_CLIENTID_BIF, 138, &adev->virt.ack_irq); in xgpu_ai_mailbox_add_irq_id()
[all …]
/kernel/linux/linux-6.6/drivers/gpu/drm/amd/amdgpu/
Damdgpu_virt.c136 struct amdgpu_virt *virt = &adev->virt; in amdgpu_virt_request_full_gpu() local
139 if (virt->ops && virt->ops->req_full_gpu) { in amdgpu_virt_request_full_gpu()
140 r = virt->ops->req_full_gpu(adev, init); in amdgpu_virt_request_full_gpu()
146 adev->virt.caps &= ~AMDGPU_SRIOV_CAPS_RUNTIME; in amdgpu_virt_request_full_gpu()
161 struct amdgpu_virt *virt = &adev->virt; in amdgpu_virt_release_full_gpu() local
164 if (virt->ops && virt->ops->rel_full_gpu) { in amdgpu_virt_release_full_gpu()
165 r = virt->ops->rel_full_gpu(adev, init); in amdgpu_virt_release_full_gpu()
169 adev->virt.caps |= AMDGPU_SRIOV_CAPS_RUNTIME; in amdgpu_virt_release_full_gpu()
182 struct amdgpu_virt *virt = &adev->virt; in amdgpu_virt_reset_gpu() local
185 if (virt->ops && virt->ops->reset_gpu) { in amdgpu_virt_reset_gpu()
[all …]
Damdgpu_vf_error.c41 mutex_lock(&adev->virt.vf_errors.lock); in amdgpu_vf_error_put()
42 index = adev->virt.vf_errors.write_count % AMDGPU_VF_ERROR_ENTRY_SIZE; in amdgpu_vf_error_put()
43 adev->virt.vf_errors.code [index] = error_code; in amdgpu_vf_error_put()
44 adev->virt.vf_errors.flags [index] = error_flags; in amdgpu_vf_error_put()
45 adev->virt.vf_errors.data [index] = error_data; in amdgpu_vf_error_put()
46 adev->virt.vf_errors.write_count ++; in amdgpu_vf_error_put()
47 mutex_unlock(&adev->virt.vf_errors.lock); in amdgpu_vf_error_put()
58 (!adev->virt.ops) || (!adev->virt.ops->trans_msg)) { in amdgpu_vf_error_trans_all()
69 mutex_lock(&adev->virt.vf_errors.lock); in amdgpu_vf_error_trans_all()
71 …if (adev->virt.vf_errors.write_count - adev->virt.vf_errors.read_count > AMDGPU_VF_ERROR_ENTRY_SIZ… in amdgpu_vf_error_trans_all()
[all …]
Dmxgpu_nv.c187 adev->virt.req_init_data_ver = 0; in xgpu_nv_send_access_requests()
190 adev->virt.req_init_data_ver = in xgpu_nv_send_access_requests()
194 if (adev->virt.req_init_data_ver < 1) in xgpu_nv_send_access_requests()
195 adev->virt.req_init_data_ver = 1; in xgpu_nv_send_access_requests()
201 adev->virt.fw_reserve.checksum_key = in xgpu_nv_send_access_requests()
276 struct amdgpu_virt *virt = container_of(work, struct amdgpu_virt, flr_work); in xgpu_nv_mailbox_flr_work() local
277 struct amdgpu_device *adev = container_of(virt, struct amdgpu_device, virt); in xgpu_nv_mailbox_flr_work()
352 &adev->virt.flr_work), in xgpu_nv_mailbox_rcv_irq()
382 adev->virt.ack_irq.num_types = 1; in xgpu_nv_mailbox_set_irq_funcs()
383 adev->virt.ack_irq.funcs = &xgpu_nv_mailbox_ack_irq_funcs; in xgpu_nv_mailbox_set_irq_funcs()
[all …]
Dmxgpu_ai.c181 adev->virt.fw_reserve.checksum_key = in xgpu_ai_send_access_requests()
189 adev->virt.req_init_data_ver = 0; in xgpu_ai_send_access_requests()
254 struct amdgpu_virt *virt = container_of(work, struct amdgpu_virt, flr_work); in xgpu_ai_mailbox_flr_work() local
255 struct amdgpu_device *adev = container_of(virt, struct amdgpu_device, virt); in xgpu_ai_mailbox_flr_work()
324 &adev->virt.flr_work), in xgpu_ai_mailbox_rcv_irq()
357 adev->virt.ack_irq.num_types = 1; in xgpu_ai_mailbox_set_irq_funcs()
358 adev->virt.ack_irq.funcs = &xgpu_ai_mailbox_ack_irq_funcs; in xgpu_ai_mailbox_set_irq_funcs()
359 adev->virt.rcv_irq.num_types = 1; in xgpu_ai_mailbox_set_irq_funcs()
360 adev->virt.rcv_irq.funcs = &xgpu_ai_mailbox_rcv_irq_funcs; in xgpu_ai_mailbox_set_irq_funcs()
367 r = amdgpu_irq_add_id(adev, SOC15_IH_CLIENTID_BIF, 135, &adev->virt.rcv_irq); in xgpu_ai_mailbox_add_irq_id()
[all …]
/kernel/linux/linux-6.6/drivers/staging/media/atomisp/pci/hmm/
Dhmm.c230 void hmm_free(ia_css_ptr virt) in hmm_free() argument
234 dev_dbg(atomisp_dev, "%s: free 0x%08x\n", __func__, virt); in hmm_free()
236 if (WARN_ON(virt == mmgr_EXCEPTION)) in hmm_free()
239 bo = hmm_bo_device_search_start(&bo_device, (unsigned int)virt); in hmm_free()
244 (unsigned int)virt); in hmm_free()
278 static int load_and_flush_by_kmap(ia_css_ptr virt, void *data, in load_and_flush_by_kmap() argument
286 bo = hmm_bo_device_search_in_range(&bo_device, virt); in load_and_flush_by_kmap()
287 ret = hmm_check_bo(bo, virt); in load_and_flush_by_kmap()
293 idx = (virt - bo->start) >> PAGE_SHIFT; in load_and_flush_by_kmap()
294 offset = (virt - bo->start) - (idx << PAGE_SHIFT); in load_and_flush_by_kmap()
[all …]
/kernel/linux/linux-5.10/drivers/staging/media/atomisp/pci/hmm/
Dhmm.c285 void hmm_free(ia_css_ptr virt) in hmm_free() argument
289 dev_dbg(atomisp_dev, "%s: free 0x%08x\n", __func__, virt); in hmm_free()
291 if (WARN_ON(virt == mmgr_EXCEPTION)) in hmm_free()
294 bo = hmm_bo_device_search_start(&bo_device, (unsigned int)virt); in hmm_free()
299 (unsigned int)virt); in hmm_free()
335 static int load_and_flush_by_kmap(ia_css_ptr virt, void *data, in load_and_flush_by_kmap() argument
343 bo = hmm_bo_device_search_in_range(&bo_device, virt); in load_and_flush_by_kmap()
344 ret = hmm_check_bo(bo, virt); in load_and_flush_by_kmap()
350 idx = (virt - bo->start) >> PAGE_SHIFT; in load_and_flush_by_kmap()
351 offset = (virt - bo->start) - (idx << PAGE_SHIFT); in load_and_flush_by_kmap()
[all …]
/kernel/linux/linux-6.6/drivers/staging/media/atomisp/include/hmm/
Dhmm.h43 int hmm_load(ia_css_ptr virt, void *data, unsigned int bytes);
44 int hmm_store(ia_css_ptr virt, const void *data, unsigned int bytes);
45 int hmm_set(ia_css_ptr virt, int c, unsigned int bytes);
46 int hmm_flush(ia_css_ptr virt, unsigned int bytes);
51 phys_addr_t hmm_virt_to_phys(ia_css_ptr virt);
54 * map ISP memory starts with virt to kernel virtual address
57 * virt must be the start address of ISP memory (return by hmm_alloc),
60 void *hmm_vmap(ia_css_ptr virt, bool cached);
61 void hmm_vunmap(ia_css_ptr virt);
67 void hmm_flush_vmap(ia_css_ptr virt);
[all …]
/kernel/linux/linux-5.10/include/asm-generic/
Dsections.h85 * with virt == phys kernel mapping, for code that wants to check if an address
101 * @virt: virtual address of the memory object
104 * Returns: true if the object specified by @virt and @size is entirely
108 static inline bool memory_contains(void *begin, void *end, void *virt, in memory_contains() argument
111 return virt >= begin && virt + size <= end; in memory_contains()
119 * @virt: virtual address of the memory object
122 * Returns: true if an object's memory region, specified by @virt and @size,
125 static inline bool memory_intersects(void *begin, void *end, void *virt, in memory_intersects() argument
128 void *vend = virt + size; in memory_intersects()
130 if (virt < end && vend > begin) in memory_intersects()
[all …]
/kernel/linux/linux-5.10/arch/m68k/mm/
Dsun3kmap.c29 static inline void do_page_mapin(unsigned long phys, unsigned long virt, in do_page_mapin() argument
39 sun3_put_pte(virt, pte); in do_page_mapin()
43 print_pte_vaddr(virt); in do_page_mapin()
48 static inline void do_pmeg_mapin(unsigned long phys, unsigned long virt, in do_pmeg_mapin() argument
52 if(sun3_get_segmap(virt & ~SUN3_PMEG_MASK) == SUN3_INVALID_PMEG) in do_pmeg_mapin()
53 mmu_emu_map_pmeg(sun3_get_context(), virt); in do_pmeg_mapin()
56 do_page_mapin(phys, virt, type); in do_pmeg_mapin()
58 virt += PAGE_SIZE; in do_pmeg_mapin()
67 unsigned long offset, virt, ret; in sun3_ioremap() local
83 pr_info("ioremap: got virt %p size %lx(%lx)\n", area->addr, size, in sun3_ioremap()
[all …]
/kernel/linux/linux-6.6/arch/m68k/mm/
Dsun3kmap.c29 static inline void do_page_mapin(unsigned long phys, unsigned long virt, in do_page_mapin() argument
39 sun3_put_pte(virt, pte); in do_page_mapin()
43 print_pte_vaddr(virt); in do_page_mapin()
48 static inline void do_pmeg_mapin(unsigned long phys, unsigned long virt, in do_pmeg_mapin() argument
52 if(sun3_get_segmap(virt & ~SUN3_PMEG_MASK) == SUN3_INVALID_PMEG) in do_pmeg_mapin()
53 mmu_emu_map_pmeg(sun3_get_context(), virt); in do_pmeg_mapin()
56 do_page_mapin(phys, virt, type); in do_pmeg_mapin()
58 virt += PAGE_SIZE; in do_pmeg_mapin()
67 unsigned long offset, virt, ret; in sun3_ioremap() local
83 pr_info("ioremap: got virt %p size %lx(%lx)\n", area->addr, size, in sun3_ioremap()
[all …]
/kernel/linux/linux-5.10/drivers/staging/media/atomisp/include/hmm/
Dhmm.h45 int hmm_load(ia_css_ptr virt, void *data, unsigned int bytes);
46 int hmm_store(ia_css_ptr virt, const void *data, unsigned int bytes);
47 int hmm_set(ia_css_ptr virt, int c, unsigned int bytes);
48 int hmm_flush(ia_css_ptr virt, unsigned int bytes);
53 phys_addr_t hmm_virt_to_phys(ia_css_ptr virt);
56 * map ISP memory starts with virt to kernel virtual address
59 * virt must be the start address of ISP memory (return by hmm_alloc),
62 void *hmm_vmap(ia_css_ptr virt, bool cached);
63 void hmm_vunmap(ia_css_ptr virt);
69 void hmm_flush_vmap(ia_css_ptr virt);
[all …]
/kernel/linux/linux-6.6/drivers/net/ethernet/qlogic/qed/
Dqed_chain.c73 void *virt, *virt_next; in qed_chain_free_next_ptr() local
77 virt = chain->p_virt_addr; in qed_chain_free_next_ptr()
81 if (!virt) in qed_chain_free_next_ptr()
84 next = virt + size; in qed_chain_free_next_ptr()
88 dma_free_coherent(dev, chain->page_size, virt, phys); in qed_chain_free_next_ptr()
90 virt = virt_next; in qed_chain_free_next_ptr()
202 void *virt, *virt_prev = NULL; in qed_chain_alloc_next_ptr() local
207 virt = dma_alloc_coherent(dev, chain->page_size, &phys, in qed_chain_alloc_next_ptr()
209 if (!virt) in qed_chain_alloc_next_ptr()
213 qed_chain_init_mem(chain, virt, phys); in qed_chain_alloc_next_ptr()
[all …]
/kernel/linux/linux-5.10/drivers/net/ethernet/qlogic/qed/
Dqed_chain.c73 void *virt, *virt_next; in qed_chain_free_next_ptr() local
77 virt = chain->p_virt_addr; in qed_chain_free_next_ptr()
81 if (!virt) in qed_chain_free_next_ptr()
84 next = virt + size; in qed_chain_free_next_ptr()
88 dma_free_coherent(dev, chain->page_size, virt, phys); in qed_chain_free_next_ptr()
90 virt = virt_next; in qed_chain_free_next_ptr()
202 void *virt, *virt_prev = NULL; in qed_chain_alloc_next_ptr() local
207 virt = dma_alloc_coherent(dev, chain->page_size, &phys, in qed_chain_alloc_next_ptr()
209 if (!virt) in qed_chain_alloc_next_ptr()
213 qed_chain_init_mem(chain, virt, phys); in qed_chain_alloc_next_ptr()
[all …]
/kernel/linux/linux-6.6/include/asm-generic/
Dsections.h84 * @virt: virtual address of the memory object
87 * Returns: true if the object specified by @virt and @size is entirely
91 static inline bool memory_contains(void *begin, void *end, void *virt, in memory_contains() argument
94 return virt >= begin && virt + size <= end; in memory_contains()
102 * @virt: virtual address of the memory object
105 * Returns: true if an object's memory region, specified by @virt and @size,
108 static inline bool memory_intersects(void *begin, void *end, void *virt, in memory_intersects() argument
111 void *vend = virt + size; in memory_intersects()
113 if (virt < end && vend > begin) in memory_intersects()
122 * @virt: virtual address of the memory object
[all …]
/kernel/linux/linux-6.6/Documentation/devicetree/bindings/interconnect/
Dqcom,rpmh.yaml29 - qcom,sc7180-camnoc-virt
34 - qcom,sc7180-mc-virt
37 - qcom,sc7180-qup-virt
41 - qcom,sc8180x-camnoc-virt
46 - qcom,sc8180x-mc-virt
48 - qcom,sc8180x-qup-virt
66 - qcom,sdx55-mc-virt
69 - qcom,sdx65-mc-virt
79 - qcom,sm8150-mc-virt
88 - qcom,sm8250-mc-virt
[all …]
/kernel/linux/linux-6.6/drivers/media/platform/samsung/s5p-mfc/
Ds5p_mfc_opr.c54 b->virt = dev->mem_virt + offset; in s5p_mfc_alloc_priv_buf()
61 b->virt = dma_alloc_coherent(mem_dev, b->size, &b->dma, GFP_KERNEL); in s5p_mfc_alloc_priv_buf()
62 if (!b->virt) in s5p_mfc_alloc_priv_buf()
67 dma_free_coherent(mem_dev, b->size, b->virt, b->dma); in s5p_mfc_alloc_priv_buf()
72 mfc_debug(3, "Allocated addr %p %pad\n", b->virt, &b->dma); in s5p_mfc_alloc_priv_buf()
87 b->virt = dma_alloc_coherent(mem_dev, b->size, &b->dma, GFP_KERNEL); in s5p_mfc_alloc_generic_buf()
88 if (!b->virt) in s5p_mfc_alloc_generic_buf()
91 mfc_debug(3, "Allocated addr %p %pad\n", b->virt, &b->dma); in s5p_mfc_alloc_generic_buf()
109 dma_free_coherent(mem_dev, b->size, b->virt, b->dma); in s5p_mfc_release_priv_buf()
111 b->virt = NULL; in s5p_mfc_release_priv_buf()
[all …]
/kernel/linux/linux-5.10/drivers/media/platform/s5p-mfc/
Ds5p_mfc_opr.c54 b->virt = dev->mem_virt + offset; in s5p_mfc_alloc_priv_buf()
61 b->virt = dma_alloc_coherent(mem_dev, b->size, &b->dma, GFP_KERNEL); in s5p_mfc_alloc_priv_buf()
62 if (!b->virt) in s5p_mfc_alloc_priv_buf()
67 dma_free_coherent(mem_dev, b->size, b->virt, b->dma); in s5p_mfc_alloc_priv_buf()
72 mfc_debug(3, "Allocated addr %p %pad\n", b->virt, &b->dma); in s5p_mfc_alloc_priv_buf()
87 b->virt = dma_alloc_coherent(mem_dev, b->size, &b->dma, GFP_KERNEL); in s5p_mfc_alloc_generic_buf()
88 if (!b->virt) in s5p_mfc_alloc_generic_buf()
91 mfc_debug(3, "Allocated addr %p %pad\n", b->virt, &b->dma); in s5p_mfc_alloc_generic_buf()
109 dma_free_coherent(mem_dev, b->size, b->virt, b->dma); in s5p_mfc_release_priv_buf()
111 b->virt = NULL; in s5p_mfc_release_priv_buf()
[all …]
/kernel/linux/linux-6.6/drivers/scsi/lpfc/
Dlpfc_mem.c107 pool->elements[i].virt = dma_pool_alloc(phba->lpfc_mbuf_pool, in lpfc_mem_alloc()
109 if (!pool->elements[i].virt) in lpfc_mem_alloc()
180 dma_pool_free(phba->lpfc_mbuf_pool, pool->elements[i].virt, in lpfc_mem_alloc()
252 dma_pool_free(phba->lpfc_mbuf_pool, pool->elements[i].virt, in lpfc_mem_free()
297 lpfc_mbuf_free(phba, mp->virt, mp->phys); in lpfc_mem_free_all()
307 lpfc_mbuf_free(phba, mp->virt, mp->phys); in lpfc_mem_free_all()
321 lpfc_mbuf_free(phba, mp->virt, mp->phys); in lpfc_mem_free_all()
342 phba->cgn_i->virt, phba->cgn_i->phys); in lpfc_mem_free_all()
391 ret = pool->elements[pool->current_count].virt; in lpfc_mbuf_alloc()
401 * @virt: mbuf to free
[all …]
/kernel/linux/linux-6.6/drivers/gpio/
Dgpio-stp-xway.c82 void __iomem *virt; member
105 return (xway_stp_r32(chip->virt, XWAY_STP_CPU0) & BIT(gpio)); in xway_stp_get()
124 xway_stp_w32(chip->virt, chip->shadow, XWAY_STP_CPU0); in xway_stp_set()
126 xway_stp_w32_mask(chip->virt, 0, XWAY_STP_CON_SWU, XWAY_STP_CON0); in xway_stp_set()
170 xway_stp_w32(chip->virt, 0, XWAY_STP_AR); in xway_stp_hw_init()
171 xway_stp_w32(chip->virt, 0, XWAY_STP_CPU0); in xway_stp_hw_init()
172 xway_stp_w32(chip->virt, 0, XWAY_STP_CPU1); in xway_stp_hw_init()
173 xway_stp_w32(chip->virt, XWAY_STP_CON_SWU, XWAY_STP_CON0); in xway_stp_hw_init()
174 xway_stp_w32(chip->virt, 0, XWAY_STP_CON1); in xway_stp_hw_init()
177 xway_stp_w32_mask(chip->virt, XWAY_STP_EDGE_MASK, in xway_stp_hw_init()
[all …]
/kernel/linux/linux-5.10/drivers/gpio/
Dgpio-stp-xway.c81 void __iomem *virt; member
104 return (xway_stp_r32(chip->virt, XWAY_STP_CPU0) & BIT(gpio)); in xway_stp_get()
123 xway_stp_w32(chip->virt, chip->shadow, XWAY_STP_CPU0); in xway_stp_set()
125 xway_stp_w32_mask(chip->virt, 0, XWAY_STP_CON_SWU, XWAY_STP_CON0); in xway_stp_set()
169 xway_stp_w32(chip->virt, 0, XWAY_STP_AR); in xway_stp_hw_init()
170 xway_stp_w32(chip->virt, 0, XWAY_STP_CPU0); in xway_stp_hw_init()
171 xway_stp_w32(chip->virt, 0, XWAY_STP_CPU1); in xway_stp_hw_init()
172 xway_stp_w32(chip->virt, XWAY_STP_CON_SWU, XWAY_STP_CON0); in xway_stp_hw_init()
173 xway_stp_w32(chip->virt, 0, XWAY_STP_CON1); in xway_stp_hw_init()
176 xway_stp_w32_mask(chip->virt, XWAY_STP_EDGE_MASK, in xway_stp_hw_init()
[all …]
/kernel/linux/linux-5.10/drivers/scsi/lpfc/
Dlpfc_mem.c107 pool->elements[i].virt = dma_pool_alloc(phba->lpfc_mbuf_pool, in lpfc_mem_alloc()
109 if (!pool->elements[i].virt) in lpfc_mem_alloc()
180 dma_pool_free(phba->lpfc_mbuf_pool, pool->elements[i].virt, in lpfc_mem_alloc()
252 dma_pool_free(phba->lpfc_mbuf_pool, pool->elements[i].virt, in lpfc_mem_free()
297 lpfc_mbuf_free(phba, mp->virt, mp->phys); in lpfc_mem_free_all()
307 lpfc_mbuf_free(phba, mp->virt, mp->phys); in lpfc_mem_free_all()
321 lpfc_mbuf_free(phba, mp->virt, mp->phys); in lpfc_mem_free_all()
375 ret = pool->elements[pool->current_count].virt; in lpfc_mbuf_alloc()
385 * @virt: mbuf to free
397 __lpfc_mbuf_free(struct lpfc_hba * phba, void *virt, dma_addr_t dma) in __lpfc_mbuf_free() argument
[all …]

12345678910>>...63