Lines Matching +full:dma +full:- +full:pool
4 * Copyright (C) 2017-2018 Broadcom. All Rights Reserved. The term *
6 * Copyright (C) 2004-2014 Emulex. All rights reserved. *
9 * Portions Copyright (C) 2004-2005 Christoph Hellwig *
17 * FITNESS FOR A PARTICULAR PURPOSE, OR NON-INFRINGEMENT, ARE *
45 #define LPFC_MBUF_POOL_SIZE 64 /* max elements in MBUF safety pool */
46 #define LPFC_MEM_POOL_SIZE 64 /* max elem in non-DMA safety pool */
47 #define LPFC_DEVICE_DATA_POOL_SIZE 64 /* max elements in device data pool */
48 #define LPFC_RRQ_POOL_SIZE 256 /* max elements in non-DMA pool */
49 #define LPFC_MBX_POOL_SIZE 256 /* max elements in MBX non-DMA pool */
54 int max_xri = phba->sli4_hba.max_cfg_param.max_xri; in lpfc_mem_alloc_active_rrq_pool_s4()
57 return -ENOMEM; in lpfc_mem_alloc_active_rrq_pool_s4()
58 bytes = ((BITS_PER_LONG - 1 + max_xri) / BITS_PER_LONG) * in lpfc_mem_alloc_active_rrq_pool_s4()
60 phba->cfg_rrq_xri_bitmap_sz = bytes; in lpfc_mem_alloc_active_rrq_pool_s4()
61 phba->active_rrq_pool = mempool_create_kmalloc_pool(LPFC_MEM_POOL_SIZE, in lpfc_mem_alloc_active_rrq_pool_s4()
63 if (!phba->active_rrq_pool) in lpfc_mem_alloc_active_rrq_pool_s4()
64 return -ENOMEM; in lpfc_mem_alloc_active_rrq_pool_s4()
70 * lpfc_mem_alloc - create and allocate all PCI and memory pools
75 * lpfc_hrb_pool. Creates and allocates kmalloc-backed mempools
78 * Notes: Not interrupt-safe. Must be called with no locks held. If any
83 * -ENOMEM on failure (if any memory allocations fail)
88 struct lpfc_dma_pool *pool = &phba->lpfc_mbuf_safety_pool; in lpfc_mem_alloc() local
92 phba->lpfc_mbuf_pool = dma_pool_create("lpfc_mbuf_pool", &phba->pcidev->dev, in lpfc_mem_alloc()
95 if (!phba->lpfc_mbuf_pool) in lpfc_mem_alloc()
98 pool->elements = kmalloc_array(LPFC_MBUF_POOL_SIZE, in lpfc_mem_alloc()
101 if (!pool->elements) in lpfc_mem_alloc()
104 pool->max_count = 0; in lpfc_mem_alloc()
105 pool->current_count = 0; in lpfc_mem_alloc()
107 pool->elements[i].virt = dma_pool_alloc(phba->lpfc_mbuf_pool, in lpfc_mem_alloc()
108 GFP_KERNEL, &pool->elements[i].phys); in lpfc_mem_alloc()
109 if (!pool->elements[i].virt) in lpfc_mem_alloc()
111 pool->max_count++; in lpfc_mem_alloc()
112 pool->current_count++; in lpfc_mem_alloc()
115 phba->mbox_mem_pool = mempool_create_kmalloc_pool(LPFC_MBX_POOL_SIZE, in lpfc_mem_alloc()
117 if (!phba->mbox_mem_pool) in lpfc_mem_alloc()
120 phba->nlp_mem_pool = mempool_create_kmalloc_pool(LPFC_MEM_POOL_SIZE, in lpfc_mem_alloc()
122 if (!phba->nlp_mem_pool) in lpfc_mem_alloc()
125 if (phba->sli_rev == LPFC_SLI_REV4) { in lpfc_mem_alloc()
126 phba->rrq_pool = in lpfc_mem_alloc()
129 if (!phba->rrq_pool) in lpfc_mem_alloc()
131 phba->lpfc_hrb_pool = dma_pool_create("lpfc_hrb_pool", in lpfc_mem_alloc()
132 &phba->pcidev->dev, in lpfc_mem_alloc()
134 if (!phba->lpfc_hrb_pool) in lpfc_mem_alloc()
137 phba->lpfc_drb_pool = dma_pool_create("lpfc_drb_pool", in lpfc_mem_alloc()
138 &phba->pcidev->dev, in lpfc_mem_alloc()
140 if (!phba->lpfc_drb_pool) in lpfc_mem_alloc()
142 phba->lpfc_hbq_pool = NULL; in lpfc_mem_alloc()
144 phba->lpfc_hbq_pool = dma_pool_create("lpfc_hbq_pool", in lpfc_mem_alloc()
145 &phba->pcidev->dev, LPFC_BPL_SIZE, align, 0); in lpfc_mem_alloc()
146 if (!phba->lpfc_hbq_pool) in lpfc_mem_alloc()
148 phba->lpfc_hrb_pool = NULL; in lpfc_mem_alloc()
149 phba->lpfc_drb_pool = NULL; in lpfc_mem_alloc()
152 if (phba->cfg_EnableXLane) { in lpfc_mem_alloc()
153 phba->device_data_mem_pool = mempool_create_kmalloc_pool( in lpfc_mem_alloc()
156 if (!phba->device_data_mem_pool) in lpfc_mem_alloc()
159 phba->device_data_mem_pool = NULL; in lpfc_mem_alloc()
164 dma_pool_destroy(phba->lpfc_drb_pool); in lpfc_mem_alloc()
165 phba->lpfc_drb_pool = NULL; in lpfc_mem_alloc()
167 dma_pool_destroy(phba->lpfc_hrb_pool); in lpfc_mem_alloc()
168 phba->lpfc_hrb_pool = NULL; in lpfc_mem_alloc()
170 mempool_destroy(phba->rrq_pool); in lpfc_mem_alloc()
171 phba->rrq_pool = NULL; in lpfc_mem_alloc()
173 mempool_destroy(phba->nlp_mem_pool); in lpfc_mem_alloc()
174 phba->nlp_mem_pool = NULL; in lpfc_mem_alloc()
176 mempool_destroy(phba->mbox_mem_pool); in lpfc_mem_alloc()
177 phba->mbox_mem_pool = NULL; in lpfc_mem_alloc()
179 while (i--) in lpfc_mem_alloc()
180 dma_pool_free(phba->lpfc_mbuf_pool, pool->elements[i].virt, in lpfc_mem_alloc()
181 pool->elements[i].phys); in lpfc_mem_alloc()
182 kfree(pool->elements); in lpfc_mem_alloc()
184 dma_pool_destroy(phba->lpfc_mbuf_pool); in lpfc_mem_alloc()
185 phba->lpfc_mbuf_pool = NULL; in lpfc_mem_alloc()
187 return -ENOMEM; in lpfc_mem_alloc()
193 phba->lpfc_nvmet_drb_pool = in lpfc_nvmet_mem_alloc()
195 &phba->pcidev->dev, LPFC_NVMET_DATA_BUF_SIZE, in lpfc_nvmet_mem_alloc()
197 if (!phba->lpfc_nvmet_drb_pool) { in lpfc_nvmet_mem_alloc()
199 "6024 Can't enable NVME Target - no memory\n"); in lpfc_nvmet_mem_alloc()
200 return -ENOMEM; in lpfc_nvmet_mem_alloc()
206 * lpfc_mem_free - Frees memory allocated by lpfc_mem_alloc
218 struct lpfc_dma_pool *pool = &phba->lpfc_mbuf_safety_pool; in lpfc_mem_free() local
223 dma_pool_destroy(phba->lpfc_nvmet_drb_pool); in lpfc_mem_free()
224 phba->lpfc_nvmet_drb_pool = NULL; in lpfc_mem_free()
226 dma_pool_destroy(phba->lpfc_drb_pool); in lpfc_mem_free()
227 phba->lpfc_drb_pool = NULL; in lpfc_mem_free()
229 dma_pool_destroy(phba->lpfc_hrb_pool); in lpfc_mem_free()
230 phba->lpfc_hrb_pool = NULL; in lpfc_mem_free()
232 dma_pool_destroy(phba->lpfc_hbq_pool); in lpfc_mem_free()
233 phba->lpfc_hbq_pool = NULL; in lpfc_mem_free()
235 mempool_destroy(phba->rrq_pool); in lpfc_mem_free()
236 phba->rrq_pool = NULL; in lpfc_mem_free()
238 /* Free NLP memory pool */ in lpfc_mem_free()
239 mempool_destroy(phba->nlp_mem_pool); in lpfc_mem_free()
240 phba->nlp_mem_pool = NULL; in lpfc_mem_free()
241 if (phba->sli_rev == LPFC_SLI_REV4 && phba->active_rrq_pool) { in lpfc_mem_free()
242 mempool_destroy(phba->active_rrq_pool); in lpfc_mem_free()
243 phba->active_rrq_pool = NULL; in lpfc_mem_free()
246 /* Free mbox memory pool */ in lpfc_mem_free()
247 mempool_destroy(phba->mbox_mem_pool); in lpfc_mem_free()
248 phba->mbox_mem_pool = NULL; in lpfc_mem_free()
250 /* Free MBUF memory pool */ in lpfc_mem_free()
251 for (i = 0; i < pool->current_count; i++) in lpfc_mem_free()
252 dma_pool_free(phba->lpfc_mbuf_pool, pool->elements[i].virt, in lpfc_mem_free()
253 pool->elements[i].phys); in lpfc_mem_free()
254 kfree(pool->elements); in lpfc_mem_free()
256 dma_pool_destroy(phba->lpfc_mbuf_pool); in lpfc_mem_free()
257 phba->lpfc_mbuf_pool = NULL; in lpfc_mem_free()
259 /* Free Device Data memory pool */ in lpfc_mem_free()
260 if (phba->device_data_mem_pool) { in lpfc_mem_free()
261 /* Ensure all objects have been returned to the pool */ in lpfc_mem_free()
262 while (!list_empty(&phba->luns)) { in lpfc_mem_free()
263 device_data = list_first_entry(&phba->luns, in lpfc_mem_free()
266 list_del(&device_data->listentry); in lpfc_mem_free()
267 mempool_free(device_data, phba->device_data_mem_pool); in lpfc_mem_free()
269 mempool_destroy(phba->device_data_mem_pool); in lpfc_mem_free()
271 phba->device_data_mem_pool = NULL; in lpfc_mem_free()
276 * lpfc_mem_free_all - Frees all PCI and driver memory
281 * kmalloc-backed mempools for LPFC_MBOXQ_t and lpfc_nodelist. Also frees
289 struct lpfc_sli *psli = &phba->sli; in lpfc_mem_free_all()
293 /* Free memory used in mailbox queue back to mailbox memory pool */ in lpfc_mem_free_all()
294 list_for_each_entry_safe(mbox, next_mbox, &psli->mboxq, list) { in lpfc_mem_free_all()
295 mp = (struct lpfc_dmabuf *)(mbox->ctx_buf); in lpfc_mem_free_all()
297 lpfc_mbuf_free(phba, mp->virt, mp->phys); in lpfc_mem_free_all()
300 list_del(&mbox->list); in lpfc_mem_free_all()
301 mempool_free(mbox, phba->mbox_mem_pool); in lpfc_mem_free_all()
303 /* Free memory used in mailbox cmpl list back to mailbox memory pool */ in lpfc_mem_free_all()
304 list_for_each_entry_safe(mbox, next_mbox, &psli->mboxq_cmpl, list) { in lpfc_mem_free_all()
305 mp = (struct lpfc_dmabuf *)(mbox->ctx_buf); in lpfc_mem_free_all()
307 lpfc_mbuf_free(phba, mp->virt, mp->phys); in lpfc_mem_free_all()
310 list_del(&mbox->list); in lpfc_mem_free_all()
311 mempool_free(mbox, phba->mbox_mem_pool); in lpfc_mem_free_all()
313 /* Free the active mailbox command back to the mailbox memory pool */ in lpfc_mem_free_all()
314 spin_lock_irq(&phba->hbalock); in lpfc_mem_free_all()
315 psli->sli_flag &= ~LPFC_SLI_MBOX_ACTIVE; in lpfc_mem_free_all()
316 spin_unlock_irq(&phba->hbalock); in lpfc_mem_free_all()
317 if (psli->mbox_active) { in lpfc_mem_free_all()
318 mbox = psli->mbox_active; in lpfc_mem_free_all()
319 mp = (struct lpfc_dmabuf *)(mbox->ctx_buf); in lpfc_mem_free_all()
321 lpfc_mbuf_free(phba, mp->virt, mp->phys); in lpfc_mem_free_all()
324 mempool_free(mbox, phba->mbox_mem_pool); in lpfc_mem_free_all()
325 psli->mbox_active = NULL; in lpfc_mem_free_all()
331 /* Free DMA buffer memory pool */ in lpfc_mem_free_all()
332 dma_pool_destroy(phba->lpfc_sg_dma_buf_pool); in lpfc_mem_free_all()
333 phba->lpfc_sg_dma_buf_pool = NULL; in lpfc_mem_free_all()
335 dma_pool_destroy(phba->lpfc_cmd_rsp_buf_pool); in lpfc_mem_free_all()
336 phba->lpfc_cmd_rsp_buf_pool = NULL; in lpfc_mem_free_all()
339 kfree(psli->iocbq_lookup); in lpfc_mem_free_all()
340 psli->iocbq_lookup = NULL; in lpfc_mem_free_all()
346 * lpfc_mbuf_alloc - Allocate an mbuf from the lpfc_mbuf_pool PCI pool
347 * @phba: HBA which owns the pool to allocate from
349 * @handle: used to return the DMA-mapped address of the mbuf
351 * Description: Allocates a DMA-mapped buffer from the lpfc_mbuf_pool PCI pool.
354 * HBA's pool.
356 * Notes: Not interrupt-safe. Must be called with no locks held. Takes
357 * phba->hbalock.
366 struct lpfc_dma_pool *pool = &phba->lpfc_mbuf_safety_pool; in lpfc_mbuf_alloc() local
370 ret = dma_pool_alloc(phba->lpfc_mbuf_pool, GFP_KERNEL, handle); in lpfc_mbuf_alloc()
372 spin_lock_irqsave(&phba->hbalock, iflags); in lpfc_mbuf_alloc()
373 if (!ret && (mem_flags & MEM_PRI) && pool->current_count) { in lpfc_mbuf_alloc()
374 pool->current_count--; in lpfc_mbuf_alloc()
375 ret = pool->elements[pool->current_count].virt; in lpfc_mbuf_alloc()
376 *handle = pool->elements[pool->current_count].phys; in lpfc_mbuf_alloc()
378 spin_unlock_irqrestore(&phba->hbalock, iflags); in lpfc_mbuf_alloc()
383 * __lpfc_mbuf_free - Free an mbuf from the lpfc_mbuf_pool PCI pool (locked)
384 * @phba: HBA which owns the pool to return to
386 * @dma: the DMA-mapped address of the lpfc_mbuf_pool to be freed
391 * Notes: Must be called with phba->hbalock held to synchronize access to
397 __lpfc_mbuf_free(struct lpfc_hba * phba, void *virt, dma_addr_t dma) in __lpfc_mbuf_free() argument
399 struct lpfc_dma_pool *pool = &phba->lpfc_mbuf_safety_pool; in __lpfc_mbuf_free() local
401 if (pool->current_count < pool->max_count) { in __lpfc_mbuf_free()
402 pool->elements[pool->current_count].virt = virt; in __lpfc_mbuf_free()
403 pool->elements[pool->current_count].phys = dma; in __lpfc_mbuf_free()
404 pool->current_count++; in __lpfc_mbuf_free()
406 dma_pool_free(phba->lpfc_mbuf_pool, virt, dma); in __lpfc_mbuf_free()
412 * lpfc_mbuf_free - Free an mbuf from the lpfc_mbuf_pool PCI pool (unlocked)
413 * @phba: HBA which owns the pool to return to
415 * @dma: the DMA-mapped address of the lpfc_mbuf_pool to be freed
420 * Notes: Takes phba->hbalock. Can be called with or without other locks held.
425 lpfc_mbuf_free(struct lpfc_hba * phba, void *virt, dma_addr_t dma) in lpfc_mbuf_free() argument
429 spin_lock_irqsave(&phba->hbalock, iflags); in lpfc_mbuf_free()
430 __lpfc_mbuf_free(phba, virt, dma); in lpfc_mbuf_free()
431 spin_unlock_irqrestore(&phba->hbalock, iflags); in lpfc_mbuf_free()
436 * lpfc_nvmet_buf_alloc - Allocate an nvmet_buf from the
437 * lpfc_sg_dma_buf_pool PCI pool
438 * @phba: HBA which owns the pool to allocate from
440 * @handle: used to return the DMA-mapped address of the nvmet_buf
442 * Description: Allocates a DMA-mapped buffer from the lpfc_sg_dma_buf_pool
443 * PCI pool. Allocates from generic dma_pool_alloc function.
454 ret = dma_pool_alloc(phba->lpfc_sg_dma_buf_pool, GFP_KERNEL, handle); in lpfc_nvmet_buf_alloc()
459 * lpfc_nvmet_buf_free - Free an nvmet_buf from the lpfc_sg_dma_buf_pool
460 * PCI pool
461 * @phba: HBA which owns the pool to return to
463 * @dma: the DMA-mapped address of the lpfc_sg_dma_buf_pool to be freed
468 lpfc_nvmet_buf_free(struct lpfc_hba *phba, void *virt, dma_addr_t dma) in lpfc_nvmet_buf_free() argument
470 dma_pool_free(phba->lpfc_sg_dma_buf_pool, virt, dma); in lpfc_nvmet_buf_free()
474 * lpfc_els_hbq_alloc - Allocate an HBQ buffer
477 * Description: Allocates a DMA-mapped HBQ buffer from the lpfc_hrb_pool PCI
478 * pool along a non-DMA-mapped container for it.
480 * Notes: Not interrupt-safe. Must be called with no locks held.
495 hbqbp->dbuf.virt = dma_pool_alloc(phba->lpfc_hbq_pool, GFP_KERNEL, in lpfc_els_hbq_alloc()
496 &hbqbp->dbuf.phys); in lpfc_els_hbq_alloc()
497 if (!hbqbp->dbuf.virt) { in lpfc_els_hbq_alloc()
501 hbqbp->total_size = LPFC_BPL_SIZE; in lpfc_els_hbq_alloc()
506 * lpfc_els_hbq_free - Frees an HBQ buffer allocated with lpfc_els_hbq_alloc
510 * Description: Frees both the container and the DMA-mapped buffer returned by
520 dma_pool_free(phba->lpfc_hbq_pool, hbqbp->dbuf.virt, hbqbp->dbuf.phys); in lpfc_els_hbq_free()
526 * lpfc_sli4_rb_alloc - Allocate an SLI4 Receive buffer
529 * Description: Allocates a DMA-mapped receive buffer from the lpfc_hrb_pool PCI
530 * pool along a non-DMA-mapped container for it.
532 * Notes: Not interrupt-safe. Must be called with no locks held.
547 dma_buf->hbuf.virt = dma_pool_alloc(phba->lpfc_hrb_pool, GFP_KERNEL, in lpfc_sli4_rb_alloc()
548 &dma_buf->hbuf.phys); in lpfc_sli4_rb_alloc()
549 if (!dma_buf->hbuf.virt) { in lpfc_sli4_rb_alloc()
553 dma_buf->dbuf.virt = dma_pool_alloc(phba->lpfc_drb_pool, GFP_KERNEL, in lpfc_sli4_rb_alloc()
554 &dma_buf->dbuf.phys); in lpfc_sli4_rb_alloc()
555 if (!dma_buf->dbuf.virt) { in lpfc_sli4_rb_alloc()
556 dma_pool_free(phba->lpfc_hrb_pool, dma_buf->hbuf.virt, in lpfc_sli4_rb_alloc()
557 dma_buf->hbuf.phys); in lpfc_sli4_rb_alloc()
561 dma_buf->total_size = LPFC_DATA_BUF_SIZE; in lpfc_sli4_rb_alloc()
566 * lpfc_sli4_rb_free - Frees a receive buffer
568 * @dmab: DMA Buffer container returned by lpfc_sli4_hbq_alloc
570 * Description: Frees both the container and the DMA-mapped buffers returned by
580 dma_pool_free(phba->lpfc_hrb_pool, dmab->hbuf.virt, dmab->hbuf.phys); in lpfc_sli4_rb_free()
581 dma_pool_free(phba->lpfc_drb_pool, dmab->dbuf.virt, dmab->dbuf.phys); in lpfc_sli4_rb_free()
586 * lpfc_sli4_nvmet_alloc - Allocate an SLI4 Receive buffer
589 * Description: Allocates a DMA-mapped receive buffer from the lpfc_hrb_pool PCI
590 * pool along a non-DMA-mapped container for it.
605 dma_buf->hbuf.virt = dma_pool_alloc(phba->lpfc_hrb_pool, GFP_KERNEL, in lpfc_sli4_nvmet_alloc()
606 &dma_buf->hbuf.phys); in lpfc_sli4_nvmet_alloc()
607 if (!dma_buf->hbuf.virt) { in lpfc_sli4_nvmet_alloc()
611 dma_buf->dbuf.virt = dma_pool_alloc(phba->lpfc_nvmet_drb_pool, in lpfc_sli4_nvmet_alloc()
612 GFP_KERNEL, &dma_buf->dbuf.phys); in lpfc_sli4_nvmet_alloc()
613 if (!dma_buf->dbuf.virt) { in lpfc_sli4_nvmet_alloc()
614 dma_pool_free(phba->lpfc_hrb_pool, dma_buf->hbuf.virt, in lpfc_sli4_nvmet_alloc()
615 dma_buf->hbuf.phys); in lpfc_sli4_nvmet_alloc()
619 dma_buf->total_size = LPFC_NVMET_DATA_BUF_SIZE; in lpfc_sli4_nvmet_alloc()
624 * lpfc_sli4_nvmet_free - Frees a receive buffer
626 * @dmab: DMA Buffer container returned by lpfc_sli4_rbq_alloc
628 * Description: Frees both the container and the DMA-mapped buffers returned by
638 dma_pool_free(phba->lpfc_hrb_pool, dmab->hbuf.virt, dmab->hbuf.phys); in lpfc_sli4_nvmet_free()
639 dma_pool_free(phba->lpfc_nvmet_drb_pool, in lpfc_sli4_nvmet_free()
640 dmab->dbuf.virt, dmab->dbuf.phys); in lpfc_sli4_nvmet_free()
645 * lpfc_in_buf_free - Free a DMA buffer
649 * Description: Frees the given DMA buffer in the appropriate way given if the
652 * Notes: Takes phba->hbalock. Can be called with or without other locks held.
665 if (phba->sli3_options & LPFC_SLI3_HBQ_ENABLED) { in lpfc_in_buf_free()
668 spin_lock_irqsave(&phba->hbalock, flags); in lpfc_in_buf_free()
669 if (!phba->hbq_in_use) { in lpfc_in_buf_free()
670 spin_unlock_irqrestore(&phba->hbalock, flags); in lpfc_in_buf_free()
673 list_del(&hbq_entry->dbuf.list); in lpfc_in_buf_free()
674 if (hbq_entry->tag == -1) { in lpfc_in_buf_free()
675 (phba->hbqs[LPFC_ELS_HBQ].hbq_free_buffer) in lpfc_in_buf_free()
680 spin_unlock_irqrestore(&phba->hbalock, flags); in lpfc_in_buf_free()
682 lpfc_mbuf_free(phba, mp->virt, mp->phys); in lpfc_in_buf_free()
689 * lpfc_rq_buf_free - Free a RQ DMA buffer
693 * Description: Frees the given DMA buffer in the appropriate way given by
696 * Notes: Takes phba->hbalock. Can be called with or without other locks held.
714 rqbp = rqb_entry->hrq->rqbp; in lpfc_rq_buf_free()
716 spin_lock_irqsave(&phba->hbalock, flags); in lpfc_rq_buf_free()
717 list_del(&rqb_entry->hbuf.list); in lpfc_rq_buf_free()
718 hrqe.address_lo = putPaddrLow(rqb_entry->hbuf.phys); in lpfc_rq_buf_free()
719 hrqe.address_hi = putPaddrHigh(rqb_entry->hbuf.phys); in lpfc_rq_buf_free()
720 drqe.address_lo = putPaddrLow(rqb_entry->dbuf.phys); in lpfc_rq_buf_free()
721 drqe.address_hi = putPaddrHigh(rqb_entry->dbuf.phys); in lpfc_rq_buf_free()
722 rc = lpfc_sli4_rq_put(rqb_entry->hrq, rqb_entry->drq, &hrqe, &drqe); in lpfc_rq_buf_free()
727 rqb_entry->hrq->queue_id, in lpfc_rq_buf_free()
728 rqb_entry->hrq->host_index, in lpfc_rq_buf_free()
729 rqb_entry->hrq->hba_index, in lpfc_rq_buf_free()
730 rqb_entry->hrq->entry_count, in lpfc_rq_buf_free()
731 rqb_entry->drq->host_index, in lpfc_rq_buf_free()
732 rqb_entry->drq->hba_index); in lpfc_rq_buf_free()
733 (rqbp->rqb_free_buffer)(phba, rqb_entry); in lpfc_rq_buf_free()
735 list_add_tail(&rqb_entry->hbuf.list, &rqbp->rqb_buffer_list); in lpfc_rq_buf_free()
736 rqbp->buffer_count++; in lpfc_rq_buf_free()
739 spin_unlock_irqrestore(&phba->hbalock, flags); in lpfc_rq_buf_free()