/drivers/net/ethernet/qlogic/qlcnic/ |
D | qlcnic_hw.c | 370 req.qhdr = cpu_to_le64(QLCNIC_REQUEST << 23); in qlcnic_sre_macaddr_change() 373 req.req_hdr = cpu_to_le64(word); in qlcnic_sre_macaddr_change() 466 req.qhdr = cpu_to_le64(QLCNIC_HOST_REQUEST << 23); in qlcnic_nic_set_promisc() 470 req.req_hdr = cpu_to_le64(word); in qlcnic_nic_set_promisc() 472 req.words[0] = cpu_to_le64(mode); in qlcnic_nic_set_promisc() 550 req.qhdr = cpu_to_le64(QLCNIC_HOST_REQUEST << 23); in qlcnic_set_fw_loopback() 551 req.req_hdr = cpu_to_le64(QLCNIC_H2C_OPCODE_CONFIG_LOOPBACK | in qlcnic_set_fw_loopback() 554 req.words[0] = cpu_to_le64(flag); in qlcnic_set_fw_loopback() 603 req.qhdr = cpu_to_le64(QLCNIC_HOST_REQUEST << 23); in qlcnic_config_intr_coalesce() 605 req.req_hdr = cpu_to_le64(QLCNIC_CONFIG_INTR_COALESCE | in qlcnic_config_intr_coalesce() [all …]
|
D | qlcnic_ctx.c | 234 prq->host_rsp_dma_addr = cpu_to_le64(cardrsp_phys_addr); in qlcnic_fw_cmd_create_rx_ctx() 266 prq_rds[i].host_phys_addr = cpu_to_le64(rds_ring->phys_addr); in qlcnic_fw_cmd_create_rx_ctx() 269 prq_rds[i].buff_size = cpu_to_le64(rds_ring->dma_size); in qlcnic_fw_cmd_create_rx_ctx() 281 prq_sds[i].host_phys_addr = cpu_to_le64(sds_ring->phys_addr); in qlcnic_fw_cmd_create_rx_ctx() 395 prq->host_rsp_dma_addr = cpu_to_le64(rsp_phys_addr); in qlcnic_fw_cmd_create_tx_ctx() 406 prq->cmd_cons_dma_addr = cpu_to_le64(tx_ring->hw_cons_phys_addr); in qlcnic_fw_cmd_create_tx_ctx() 410 prq_cds->host_phys_addr = cpu_to_le64(tx_ring->phys_addr); in qlcnic_fw_cmd_create_tx_ctx()
|
/drivers/net/vmxnet3/ |
D | upt1_defs.h | 91 UPT1_F_RXCSUM = cpu_to_le64(0x0001), /* rx csum verification */ 92 UPT1_F_RSS = cpu_to_le64(0x0002), 93 UPT1_F_RXVLAN = cpu_to_le64(0x0004), /* VLAN tag stripping */ 94 UPT1_F_LRO = cpu_to_le64(0x0008),
|
D | vmxnet3_drv.c | 619 gd->rxd.addr = cpu_to_le64(rbi->dma_addr); in vmxnet3_rq_alloc_rx_buf() 685 ctx->sop_txd->txd.addr = cpu_to_le64(tq->data_ring.basePA + in vmxnet3_map_pkt() 730 gdesc->txd.addr = cpu_to_le64(tbi->dma_addr); in vmxnet3_map_pkt() 759 gdesc->txd.addr = cpu_to_le64(tbi->dma_addr); in vmxnet3_map_pkt() 1223 rxd->addr = cpu_to_le64(rbi->dma_addr); in vmxnet3_rq_rx_complete() 1266 rxd->addr = cpu_to_le64(rbi->dma_addr); in vmxnet3_rq_rx_complete() 2017 rxConf->mfTablePA = cpu_to_le64(virt_to_phys( in vmxnet3_set_mc() 2085 devRead->misc.ddPA = cpu_to_le64(virt_to_phys(adapter)); in vmxnet3_setup_driver_shared() 2100 devRead->misc.queueDescPA = cpu_to_le64(adapter->queue_desc_pa); in vmxnet3_setup_driver_shared() 2111 tqc->txRingBasePA = cpu_to_le64(tq->tx_ring.basePA); in vmxnet3_setup_driver_shared() [all …]
|
/drivers/net/ethernet/qlogic/netxen/ |
D | netxen_nic_ctx.c | 326 prq->host_rsp_dma_addr = cpu_to_le64(cardrsp_phys_addr); in nx_fw_cmd_create_rx_ctx() 352 prq_rds[i].host_phys_addr = cpu_to_le64(rds_ring->phys_addr); in nx_fw_cmd_create_rx_ctx() 355 prq_rds[i].buff_size = cpu_to_le64(rds_ring->dma_size); in nx_fw_cmd_create_rx_ctx() 365 prq_sds[i].host_phys_addr = cpu_to_le64(sds_ring->phys_addr); in nx_fw_cmd_create_rx_ctx() 476 prq->host_rsp_dma_addr = cpu_to_le64(rsp_phys_addr); in nx_fw_cmd_create_tx_ctx() 487 prq->dummy_dma_addr = cpu_to_le64(adapter->dummy_dma.phys_addr); in nx_fw_cmd_create_tx_ctx() 490 prq->cmd_cons_dma_addr = cpu_to_le64(offset); in nx_fw_cmd_create_tx_ctx() 494 prq_cds->host_phys_addr = cpu_to_le64(tx_ring->phys_addr); in nx_fw_cmd_create_tx_ctx() 720 hwctx->cmd_ring_addr = cpu_to_le64(tx_ring->phys_addr); in netxen_init_old_ctx() 728 cpu_to_le64(rds_ring->phys_addr); in netxen_init_old_ctx() [all …]
|
D | netxen_nic_hw.c | 644 req.qhdr = cpu_to_le64(NX_NIC_REQUEST << 23); in nx_p3_sre_macaddr_change() 647 req.req_hdr = cpu_to_le64(word); in nx_p3_sre_macaddr_change() 740 req.qhdr = cpu_to_le64(NX_HOST_REQUEST << 23); in netxen_p3_nic_set_promisc() 744 req.req_hdr = cpu_to_le64(word); in netxen_p3_nic_set_promisc() 746 req.words[0] = cpu_to_le64(mode); in netxen_p3_nic_set_promisc() 787 req.qhdr = cpu_to_le64(NX_HOST_REQUEST << 23); in netxen_config_intr_coalesce() 790 req.req_hdr = cpu_to_le64(word[0]); in netxen_config_intr_coalesce() 794 req.words[i] = cpu_to_le64(word[i]); in netxen_config_intr_coalesce() 816 req.qhdr = cpu_to_le64(NX_HOST_REQUEST << 23); in netxen_config_hw_lro() 819 req.req_hdr = cpu_to_le64(word); in netxen_config_hw_lro() [all …]
|
/drivers/net/ethernet/icplus/ |
D | ipg.c | 750 rxfd->frag_info = cpu_to_le64(pci_map_single(sp->pdev, skb->data, in ipg_get_rxbuff() 755 rxfd->frag_info |= cpu_to_le64((rxfragsize << 48) & IPG_RFI_FRAGLEN); in ipg_get_rxbuff() 798 rxfd->next_desc = cpu_to_le64(sp->rxd_map + in init_rfdlist() 801 sp->rxd[i - 1].next_desc = cpu_to_le64(sp->rxd_map); in init_rfdlist() 824 txfd->tfc = cpu_to_le64(IPG_TFC_TFDDONE); in init_tfdlist() 831 txfd->next_desc = cpu_to_le64(sp->txd_map + in init_tfdlist() 834 sp->txd[i - 1].next_desc = cpu_to_le64(sp->txd_map); in init_tfdlist() 873 if (!(txfd->tfc & cpu_to_le64(IPG_TFC_TFDDONE))) in ipg_nic_txfree() 1335 if (!(rxfd->rfs & cpu_to_le64(IPG_RFS_RFDDONE))) in ipg_nic_rx_jumbo() 1382 cpu_to_le64(IPG_RFS_RFDDONE | IPG_RFS_FRAMESTART | IPG_RFS_FRAMEEND) in ipg_nic_rx() [all …]
|
/drivers/md/ |
D | dm-thin-metadata.c | 209 disk_super->blocknr = cpu_to_le64(dm_block_location(b)); in sb_prepare_for_write() 346 __le64 *data_le, zero = cpu_to_le64(0); in superblock_all_zeroes() 559 details.mapped_blocks = cpu_to_le64(td->mapped_blocks); in __write_changed_details() 560 details.transaction_id = cpu_to_le64(td->transaction_id); in __write_changed_details() 628 disk_super->data_mapping_root = cpu_to_le64(pmd->root); in __commit_transaction() 629 disk_super->device_details_root = cpu_to_le64(pmd->details_root); in __commit_transaction() 630 disk_super->trans_id = cpu_to_le64(pmd->trans_id); in __commit_transaction() 720 disk_super->magic = cpu_to_le64(THIN_SUPERBLOCK_MAGIC); in dm_pool_metadata_open() 724 disk_super->metadata_nr_blocks = cpu_to_le64(bdev_size >> SECTOR_TO_BLOCK_SHIFT); in dm_pool_metadata_open() 840 details_le.transaction_id = cpu_to_le64(pmd->trans_id); in __open_device() [all …]
|
/drivers/net/ethernet/dlink/ |
D | dl2k.c | 519 cpu_to_le64 (pci_map_single in rio_timer() 524 cpu_to_le64((u64)np->rx_buf_sz << 48); in rio_timer() 559 np->tx_ring[i].status = cpu_to_le64 (TFDDone); in alloc_list() 560 np->tx_ring[i].next_desc = cpu_to_le64 (np->tx_ring_dma + in alloc_list() 567 np->rx_ring[i].next_desc = cpu_to_le64 (np->rx_ring_dma + in alloc_list() 590 cpu_to_le64 ( pci_map_single ( in alloc_list() 593 np->rx_ring[i].fraginfo |= cpu_to_le64((u64)np->rx_buf_sz << 48); in alloc_list() 622 cpu_to_le64 (TCPChecksumEnable | UDPChecksumEnable | in start_xmit() 631 txdesc->fraginfo = cpu_to_le64 (pci_map_single (np->pdev, skb->data, in start_xmit() 634 txdesc->fraginfo |= cpu_to_le64((u64)skb->len << 48); in start_xmit() [all …]
|
/drivers/md/persistent-data/ |
D | dm-space-map-disk.c | 224 root_le.nr_blocks = cpu_to_le64(smd->ll.nr_blocks); in sm_disk_copy_root() 225 root_le.nr_allocated = cpu_to_le64(smd->ll.nr_allocated); in sm_disk_copy_root() 226 root_le.bitmap_root = cpu_to_le64(smd->ll.bitmap_root); in sm_disk_copy_root() 227 root_le.ref_count_root = cpu_to_le64(smd->ll.ref_count_root); in sm_disk_copy_root()
|
D | dm-btree.c | 85 __le64 key_le = cpu_to_le64(key); in insert_at() 443 location = cpu_to_le64(dm_block_location(left)); in btree_split_sibling() 448 location = cpu_to_le64(dm_block_location(right)); in btree_split_sibling() 544 val = cpu_to_le64(dm_block_location(left)); in btree_split_beneath() 549 val = cpu_to_le64(dm_block_location(right)); in btree_split_beneath() 594 __le64 location = cpu_to_le64(dm_block_location(shadow_current(s))); in btree_insert_raw() 622 node->keys[0] = cpu_to_le64(key); in btree_insert_raw() 674 new_le = cpu_to_le64(new_tree); in insert()
|
D | dm-space-map-metadata.c | 372 root_le.nr_blocks = cpu_to_le64(smm->ll.nr_blocks); in sm_metadata_copy_root() 373 root_le.nr_allocated = cpu_to_le64(smm->ll.nr_allocated); in sm_metadata_copy_root() 374 root_le.bitmap_root = cpu_to_le64(smm->ll.bitmap_root); in sm_metadata_copy_root() 375 root_le.ref_count_root = cpu_to_le64(smm->ll.ref_count_root); in sm_metadata_copy_root()
|
D | dm-space-map-common.c | 28 mi_le->blocknr = cpu_to_le64(dm_block_location(b)); in index_prepare_for_write() 78 disk_header->blocknr = cpu_to_le64(dm_block_location(b)); in bitmap_prepare_for_write() 254 idx.blocknr = cpu_to_le64(dm_block_location(b)); in sm_ll_extend() 396 ie_disk.blocknr = cpu_to_le64(dm_block_location(nb)); in sm_ll_insert()
|
/drivers/scsi/aic94xx/ |
D | aic94xx_task.c | 71 sg_arr[0].bus_addr = cpu_to_le64((u64)dma); in asd_map_scatterlist() 100 sg->bus_addr = cpu_to_le64((u64)sg_dma_address(sc)); in asd_map_scatterlist() 108 cpu_to_le64((u64)sg_dma_address(sc)); in asd_map_scatterlist() 115 sg_arr[2].bus_addr=cpu_to_le64((u64)ascb->sg_arr->dma_handle); in asd_map_scatterlist() 120 cpu_to_le64((u64)sg_dma_address(sc)); in asd_map_scatterlist() 451 cpu_to_le64((u64)sg_dma_address(&task->smp_task.smp_req)); in asd_build_smp_ascb() 456 cpu_to_le64((u64)sg_dma_address(&task->smp_task.smp_resp)); in asd_build_smp_ascb()
|
/drivers/infiniband/hw/qib/ |
D | qib_sdma.c | 567 *descqp++ = cpu_to_le64(sdmadesc[0]); in qib_sdma_verbs_send() 568 *descqp++ = cpu_to_le64(sdmadesc[1]); in qib_sdma_verbs_send() 601 *descqp++ = cpu_to_le64(sdmadesc[0]); in qib_sdma_verbs_send() 602 *descqp++ = cpu_to_le64(sdmadesc[1]); in qib_sdma_verbs_send() 635 descqp[0] |= cpu_to_le64(SDMA_DESC_LAST); in qib_sdma_verbs_send() 637 descqp[0] |= cpu_to_le64(SDMA_DESC_DMA_HEAD); in qib_sdma_verbs_send() 639 descqp[0] |= cpu_to_le64(SDMA_DESC_INTR); in qib_sdma_verbs_send()
|
D | qib_user_sdma.c | 666 return cpu_to_le64(/* SDmaPhyAddr[31:0] */ in qib_sdma_make_desc0() 678 return descq | cpu_to_le64(1ULL << 12); in qib_sdma_make_first_desc0() 684 return descq | cpu_to_le64(1ULL << 11 | 1ULL << 13); in qib_sdma_make_last_desc0() 690 return cpu_to_le64(addr >> 32); in qib_sdma_make_desc1() 779 cpu_to_le64(1ULL << 14); in qib_user_sdma_push_pkts()
|
/drivers/block/ |
D | nvme.c | 405 cmd->prp1 = cpu_to_le64(dma_addr); in nvme_setup_prps() 420 cmd->prp2 = cpu_to_le64(dma_addr); in nvme_setup_prps() 435 cmd->prp2 = cpu_to_le64(dma_addr); in nvme_setup_prps() 441 cmd->prp2 = cpu_to_le64(prp_dma); in nvme_setup_prps() 451 old_prp_list[i - 1] = cpu_to_le64(prp_dma); in nvme_setup_prps() 454 prp_list[i++] = cpu_to_le64(dma_addr); in nvme_setup_prps() 598 cmnd->rw.slba = cpu_to_le64(bio->bi_sector >> (ns->lba_shift - 9)); in nvme_submit_bio_queue() 780 c.create_cq.prp1 = cpu_to_le64(nvmeq->cq_dma_addr); in adapter_alloc_cq() 801 c.create_sq.prp1 = cpu_to_le64(nvmeq->sq_dma_addr); in adapter_alloc_sq() 831 c.identify.prp1 = cpu_to_le64(dma_addr); in nvme_identify() [all …]
|
/drivers/infiniband/hw/ipath/ |
D | ipath_sdma.c | 716 *descqp++ = cpu_to_le64(sdmadesc[0]); in ipath_sdma_verbs_send() 717 *descqp++ = cpu_to_le64(sdmadesc[1]); in ipath_sdma_verbs_send() 750 *descqp++ = cpu_to_le64(sdmadesc[0]); in ipath_sdma_verbs_send() 751 *descqp++ = cpu_to_le64(sdmadesc[1]); in ipath_sdma_verbs_send() 785 descqp[0] |= cpu_to_le64(1ULL << 11); in ipath_sdma_verbs_send() 788 descqp[0] |= cpu_to_le64(1ULL << 15); in ipath_sdma_verbs_send()
|
D | ipath_user_sdma.c | 659 return cpu_to_le64(/* SDmaPhyAddr[31:0] */ in ipath_sdma_make_desc0() 671 return descq | cpu_to_le64(1ULL << 12); in ipath_sdma_make_first_desc0() 677 return descq | cpu_to_le64(1ULL << 11 | 1ULL << 13); in ipath_sdma_make_last_desc0() 683 return cpu_to_le64(addr >> 32); in ipath_sdma_make_desc1() 767 cpu_to_le64(1ULL << 14); in ipath_user_sdma_push_pkts()
|
/drivers/net/ethernet/cisco/enic/ |
D | rq_enet_desc.h | 46 desc->address = cpu_to_le64(address); in rq_enet_desc_enc()
|
/drivers/scsi/fnic/ |
D | rq_enet_desc.h | 44 desc->address = cpu_to_le64(address); in rq_enet_desc_enc()
|
/drivers/scsi/mvsas/ |
D | mv_sas.c | 363 hdr->cmd_tbl = cpu_to_le64(sg_dma_address(sg_req)); in mvs_task_prep_smp() 367 hdr->open_frame = cpu_to_le64(buf_tmp_dma); in mvs_task_prep_smp() 375 hdr->prd_tbl = cpu_to_le64(buf_tmp_dma); in mvs_task_prep_smp() 385 hdr->status_buf = cpu_to_le64(buf_tmp_dma); in mvs_task_prep_smp() 500 hdr->cmd_tbl = cpu_to_le64(buf_tmp_dma); in mvs_task_prep_ata() 508 hdr->open_frame = cpu_to_le64(buf_tmp_dma); in mvs_task_prep_ata() 517 hdr->prd_tbl = cpu_to_le64(buf_tmp_dma); in mvs_task_prep_ata() 527 hdr->status_buf = cpu_to_le64(buf_tmp_dma); in mvs_task_prep_ata() 617 hdr->cmd_tbl = cpu_to_le64(buf_tmp_dma); in mvs_task_prep_ssp() 624 hdr->open_frame = cpu_to_le64(buf_tmp_dma); in mvs_task_prep_ssp() [all …]
|
/drivers/staging/rtl8712/ |
D | generic.h | 111 #define cpu_to_le64 __cpu_to_le64 macro
|
/drivers/usb/host/whci/ |
D | qset.c | 127 qset->qh.link = cpu_to_le64(QH_LINK_NTDS(8) | QH_LINK_T); in qset_clear() 234 qtd->page_list_ptr = cpu_to_le64(std->dma_addr); in qset_add_qtds() 382 std->pl_virt[p].buf_ptr = cpu_to_le64(dma_addr); in qset_fill_page_list() 517 std->pl_virt[p].buf_ptr = cpu_to_le64(dma_addr); in qset_add_urb_sg()
|
/drivers/usb/host/ |
D | xhci-mem.c | 106 cpu_to_le64(next->dma); in xhci_link_segments() 664 cpu_to_le64(addr); in xhci_alloc_stream_info() 733 ep_ctx->deq = cpu_to_le64(stream_info->ctx_array_dma); in xhci_setup_streams_ep_input_ctx() 748 ep_ctx->deq = cpu_to_le64(addr | ep->ring->cycle_state); in xhci_setup_no_streams_ep_input_ctx() 986 xhci->dcbaa->dev_context_ptrs[slot_id] = cpu_to_le64(dev->out_ctx->dma); in xhci_alloc_virt_device() 1015 ep0_ctx->deq = cpu_to_le64(xhci_trb_virt_to_dma(ep_ring->enq_seg, in xhci_copy_ep0_dequeue_into_input_ctx() 1194 ep0_ctx->deq = cpu_to_le64(dev->eps[0].ring->first_seg->dma | in xhci_setup_addressable_virt_dev() 1432 ep_ctx->deq = cpu_to_le64(ep_ring->first_seg->dma | ep_ring->cycle_state); in xhci_endpoint_init() 1666 xhci->dcbaa->dev_context_ptrs[0] = cpu_to_le64(xhci->scratchpad->sp_dma); in scratchpad_alloc() 2420 entry->seg_addr = cpu_to_le64(seg->dma); in xhci_mem_init()
|