/drivers/net/ethernet/netronome/nfp/bpf/ |
D | jit.c | 367 u32 dst, u8 imm, u8 byte, u16 addr, u8 defer) in emit_br_byte_neq() argument 372 err = swreg_to_restricted(reg_none(), dst, reg_imm(imm), ®, true); in emit_br_byte_neq() 402 emit_immed(struct nfp_prog *nfp_prog, u32 dst, u16 imm, in emit_immed() argument 413 err = swreg_to_unrestricted(dst, dst, reg_imm(imm & 0xff), ®); in emit_immed() 419 __emit_immed(nfp_prog, reg.areg, reg.breg, imm >> 8, width, in emit_immed() 550 static bool pack_immed(u32 imm, u16 *val, enum immed_shift *shift) in pack_immed() argument 552 if (!(imm & 0xffff0000)) { in pack_immed() 553 *val = imm; in pack_immed() 555 } else if (!(imm & 0xff0000ff)) { in pack_immed() 556 *val = imm >> 8; in pack_immed() [all …]
|
D | verifier.c | 82 u64 imm; in nfp_bpf_check_exit() local 96 imm = reg0->var_off.value; in nfp_bpf_check_exit() 97 if (nfp_prog->act != NN_ACT_DIRECT && imm != 0 && (imm & ~0U) != ~0U) { in nfp_bpf_check_exit() 99 reg0->type, imm); in nfp_bpf_check_exit() 103 if (nfp_prog->act == NN_ACT_DIRECT && imm <= TC_ACT_REDIRECT && in nfp_bpf_check_exit() 104 imm != TC_ACT_SHOT && imm != TC_ACT_STOLEN && in nfp_bpf_check_exit() 105 imm != TC_ACT_QUEUED) { in nfp_bpf_check_exit() 107 reg0->type, imm); in nfp_bpf_check_exit()
|
/drivers/net/ethernet/qlogic/qed/ |
D | qed_debug.c | 85 static u32 cond5(const u32 *r, const u32 *imm) in cond5() argument 87 return ((r[0] & imm[0]) != imm[1]) && ((r[1] & imm[2]) != imm[3]); in cond5() 90 static u32 cond7(const u32 *r, const u32 *imm) in cond7() argument 92 return ((r[0] >> imm[0]) & imm[1]) != imm[2]; in cond7() 95 static u32 cond14(const u32 *r, const u32 *imm) in cond14() argument 97 return (r[0] != imm[0]) && (((r[1] >> imm[1]) & imm[2]) == imm[3]); in cond14() 100 static u32 cond6(const u32 *r, const u32 *imm) in cond6() argument 102 return (r[0] & imm[0]) != imm[1]; in cond6() 105 static u32 cond9(const u32 *r, const u32 *imm) in cond9() argument 107 return ((r[0] & imm[0]) >> imm[1]) != in cond9() [all …]
|
/drivers/infiniband/sw/rxe/ |
D | rxe_hdr.h | 869 __be32 imm; member 876 return immdt->imm; in __immdt_imm() 879 static inline void __immdt_set_imm(void *arg, __be32 imm) in __immdt_set_imm() argument 883 immdt->imm = imm; in __immdt_set_imm() 892 static inline void immdt_set_imm(struct rxe_pkt_info *pkt, __be32 imm) in immdt_set_imm() argument 895 + rxe_opcode[pkt->opcode].offset[RXE_IMMDT], imm); in immdt_set_imm()
|
/drivers/scsi/csiostor/ |
D | csio_scsi.c | 207 uint8_t imm = csio_hw_to_scsim(hw)->proto_cmd_len; in csio_scsi_init_cmd_wr() local 210 FW_SCSI_CMD_WR_IMMDLEN(imm)); in csio_scsi_init_cmd_wr() 367 uint8_t imm = csio_hw_to_scsim(hw)->proto_cmd_len; in csio_scsi_init_read_wr() local 371 FW_SCSI_READ_WR_IMMDLEN(imm)); in csio_scsi_init_read_wr() 398 sizeof(struct fw_scsi_read_wr) + ALIGN(imm, 16)); in csio_scsi_init_read_wr() 420 uint8_t imm = csio_hw_to_scsim(hw)->proto_cmd_len; in csio_scsi_init_write_wr() local 424 FW_SCSI_WRITE_WR_IMMDLEN(imm)); in csio_scsi_init_write_wr() 451 sizeof(struct fw_scsi_write_wr) + ALIGN(imm, 16)); in csio_scsi_init_write_wr() 458 #define CSIO_SCSI_DATA_WRSZ(req, oper, sz, imm) \ argument 461 ALIGN((imm), 16) + /* Immed data */ \
|
/drivers/infiniband/hw/mthca/ |
D | mthca_wqe.h | 61 __be32 imm; /* immediate data */ member
|
D | mthca_cmd.c | 1964 u64 imm; in mthca_MGID_HASH() local 1967 err = mthca_cmd_imm(dev, mailbox->dma, &imm, 0, 0, CMD_MGID_HASH, in mthca_MGID_HASH() 1970 *hash = imm; in mthca_MGID_HASH()
|
D | mthca_srq.c | 92 return (int *) (wqe + offsetof(struct mthca_next_seg, imm)); in wqe_to_link()
|
D | mthca_qp.c | 1669 ((struct mthca_next_seg *) wqe)->imm = wr->ex.imm_data; in mthca_tavor_post_send() 2010 ((struct mthca_next_seg *) wqe)->imm = wr->ex.imm_data; in mthca_arbel_post_send()
|
/drivers/net/ethernet/mellanox/mlx4/ |
D | mcg.c | 58 u64 imm; in mlx4_QP_FLOW_STEERING_ATTACH() local 61 err = mlx4_cmd_imm(dev, mailbox->dma, &imm, size, 0, in mlx4_QP_FLOW_STEERING_ATTACH() 66 *reg_id = imm; in mlx4_QP_FLOW_STEERING_ATTACH() 110 u64 imm; in mlx4_GID_HASH() local 113 err = mlx4_cmd_imm(dev, mailbox->dma, &imm, 0, op_mod, in mlx4_GID_HASH() 118 *hash = imm; in mlx4_GID_HASH()
|
D | en_tx.c | 974 tx_desc->ctrl.imm = get_unaligned((__be32 *)(ethh->h_dest + 2)); in mlx4_en_xmit()
|
/drivers/net/ethernet/cavium/thunder/ |
D | nicvf_queues.c | 1414 struct sq_imm_subdesc *imm; in nicvf_sq_add_cqe_subdesc() local 1432 imm = (struct sq_imm_subdesc *)GET_SQ_DESC(sq, qentry); in nicvf_sq_add_cqe_subdesc() 1433 memset(imm, 0, SND_QUEUE_DESC_SIZE); in nicvf_sq_add_cqe_subdesc() 1434 imm->subdesc_type = SQ_DESC_TYPE_IMMEDIATE; in nicvf_sq_add_cqe_subdesc() 1435 imm->len = 1; in nicvf_sq_add_cqe_subdesc()
|
/drivers/scsi/ |
D | Makefile | 118 obj-$(CONFIG_SCSI_IMM) += imm.o
|
D | Kconfig | 855 then you should say N here and Y to "IOMEGA parallel port (imm - 870 tristate "IOMEGA parallel port (imm - newer drives)" 894 module will be called imm. 897 bool "ppa/imm option - Use slow (but safe) EPP-16" 912 bool "ppa/imm option - Assume slow parport control register"
|
/drivers/net/ethernet/chelsio/cxgb4/ |
D | cxgb4_uld.c | 129 rxq->stats.imm++; in uldrx_handler()
|
D | cxgb4.h | 645 unsigned long imm; /* # of immediate-data packets */ member
|
/drivers/infiniband/hw/mlx5/ |
D | qp.c | 3856 (*ctrl)->imm = send_ieth(wr); in begin_wqe() 3996 ctrl->imm = cpu_to_be32(wr->ex.invalidate_rkey); in mlx5_ib_post_send() 4003 ctrl->imm = cpu_to_be32(reg_wr(wr)->key); in mlx5_ib_post_send() 4016 ctrl->imm = cpu_to_be32(mr->ibmr.rkey); in mlx5_ib_post_send() 4140 ctrl->imm = cpu_to_be32(umr_wr(wr)->mkey); in mlx5_ib_post_send()
|
/drivers/scsi/qla2xxx/ |
D | qla_target.c | 135 struct imm_ntfy_from_isp *imm, int ha_locked); 3437 struct imm_ntfy_from_isp *imm, int ha_locked) in qlt_send_term_imm_notif() argument 3443 rc = __qlt_send_term_imm_notif(vha, imm); in qlt_send_term_imm_notif() 3447 qlt_alloc_qfull_cmd(vha, imm, 0, 0); in qlt_send_term_imm_notif() 3456 rc = __qlt_send_term_imm_notif(vha, imm); in qlt_send_term_imm_notif() 3460 qlt_alloc_qfull_cmd(vha, imm, 0, 0); in qlt_send_term_imm_notif()
|
/drivers/scsi/cxgbi/cxgb4i/ |
D | cxgb4i.c | 645 bool imm = is_ofld_imm(skb); in make_tx_data_wr() local 649 if (imm) { in make_tx_data_wr()
|
/drivers/net/ethernet/mellanox/mlx5/core/fpga/ |
D | conn.c | 164 ctrl->imm = 0; in mlx5_fpga_conn_post_send()
|
/drivers/infiniband/hw/mlx4/ |
D | qp.c | 3219 memcpy(&ctrl->imm, ah->av.eth.mac + 2, 4); in build_mlx_header() 3631 ctrl->imm = send_ieth(wr); in mlx4_ib_post_send()
|
/drivers/net/ethernet/mellanox/mlx5/core/ |
D | en_main.c | 431 cseg->imm = rq->mkey_be; in mlx5e_build_umr_wqe()
|