Home
last modified time | relevance | path

Searched refs:smp (Results 1 – 25 of 27) sorted by relevance

12

/drivers/infiniband/core/
Dsmi.c129 enum smi_action smi_handle_dr_smp_send(struct ib_smp *smp, in smi_handle_dr_smp_send() argument
133 &smp->hop_ptr, smp->hop_cnt, in smi_handle_dr_smp_send()
134 smp->initial_path, in smi_handle_dr_smp_send()
135 smp->return_path, in smi_handle_dr_smp_send()
136 ib_get_smp_direction(smp), in smi_handle_dr_smp_send()
137 smp->dr_dlid == IB_LID_PERMISSIVE, in smi_handle_dr_smp_send()
138 smp->dr_slid == IB_LID_PERMISSIVE); in smi_handle_dr_smp_send()
141 enum smi_action opa_smi_handle_dr_smp_send(struct opa_smp *smp, in opa_smi_handle_dr_smp_send() argument
145 &smp->hop_ptr, smp->hop_cnt, in opa_smi_handle_dr_smp_send()
146 smp->route.dr.initial_path, in opa_smi_handle_dr_smp_send()
[all …]
Dopa_smi.h42 enum smi_action opa_smi_handle_dr_smp_recv(struct opa_smp *smp, bool is_switch,
44 int opa_smi_get_fwd_port(struct opa_smp *smp);
45 extern enum smi_forward_action opa_smi_check_forward_dr_smp(struct opa_smp *smp);
46 extern enum smi_action opa_smi_handle_dr_smp_send(struct opa_smp *smp,
53 static inline enum smi_action opa_smi_check_local_smp(struct opa_smp *smp, in opa_smi_check_local_smp() argument
59 !opa_get_smp_direction(smp) && in opa_smi_check_local_smp()
60 (smp->hop_ptr == smp->hop_cnt + 1)) ? in opa_smi_check_local_smp()
68 static inline enum smi_action opa_smi_check_local_returning_smp(struct opa_smp *smp, in opa_smi_check_local_returning_smp() argument
74 opa_get_smp_direction(smp) && in opa_smi_check_local_returning_smp()
75 !smp->hop_ptr) ? IB_SMI_HANDLE : IB_SMI_DISCARD; in opa_smi_check_local_returning_smp()
Dsmi.h54 enum smi_action smi_handle_dr_smp_recv(struct ib_smp *smp, bool is_switch,
56 int smi_get_fwd_port(struct ib_smp *smp);
57 extern enum smi_forward_action smi_check_forward_dr_smp(struct ib_smp *smp);
58 extern enum smi_action smi_handle_dr_smp_send(struct ib_smp *smp,
65 static inline enum smi_action smi_check_local_smp(struct ib_smp *smp, in smi_check_local_smp() argument
71 !ib_get_smp_direction(smp) && in smi_check_local_smp()
72 (smp->hop_ptr == smp->hop_cnt + 1)) ? in smi_check_local_smp()
80 static inline enum smi_action smi_check_local_returning_smp(struct ib_smp *smp, in smi_check_local_returning_smp() argument
86 ib_get_smp_direction(smp) && in smi_check_local_returning_smp()
87 !smp->hop_ptr) ? IB_SMI_HANDLE : IB_SMI_DISCARD); in smi_check_local_returning_smp()
Dmad.c745 struct ib_smp *smp = mad_send_wr->send_buf.mad; in handle_outgoing_dr_smp() local
746 struct opa_smp *opa_smp = (struct opa_smp *)smp; in handle_outgoing_dr_smp()
763 smp->mgmt_class == IB_MGMT_CLASS_SUBN_DIRECTED_ROUTE) in handle_outgoing_dr_smp()
774 if (opa && smp->class_version == OPA_SMP_CLASS_VERSION) { in handle_outgoing_dr_smp()
802 if ((ib_get_smp_direction(smp) ? smp->dr_dlid : smp->dr_slid) == in handle_outgoing_dr_smp()
804 smi_handle_dr_smp_send(smp, rdma_cap_ib_switch(device), port_num) == in handle_outgoing_dr_smp()
810 drslid = be16_to_cpu(smp->dr_slid); in handle_outgoing_dr_smp()
813 if (smi_check_local_smp(smp, device) == IB_SMI_DISCARD && in handle_outgoing_dr_smp()
814 smi_check_local_returning_smp(smp, device) == IB_SMI_DISCARD) in handle_outgoing_dr_smp()
839 if (opa && smp->base_version == OPA_MGMT_BASE_VERSION) { in handle_outgoing_dr_smp()
[all …]
/drivers/gpu/drm/msm/mdp/mdp5/
Dmdp5_smp.c104 static void update_smp_state(struct mdp5_smp *smp,
108 struct mdp5_kms *get_kms(struct mdp5_smp *smp) in get_kms() argument
110 struct msm_drm_private *priv = smp->dev->dev_private; in get_kms()
134 return mdp5_cfg->smp.clients[pipe] + plane; in pipe2client()
138 static int smp_request_block(struct mdp5_smp *smp, in smp_request_block() argument
141 struct mdp5_kms *mdp5_kms = get_kms(smp); in smp_request_block()
142 struct mdp5_client_smp_state *ps = &smp->client_state[cid]; in smp_request_block()
143 int i, ret, avail, cur_nblks, cnt = smp->blk_cnt; in smp_request_block()
147 reserved = smp->reserved[cid]; in smp_request_block()
149 spin_lock_irqsave(&smp->state_lock, flags); in smp_request_block()
[all …]
Dmdp5_smp.h40 void mdp5_smp_destroy(struct mdp5_smp *smp);
42 int mdp5_smp_request(struct mdp5_smp *smp, enum mdp5_pipe pipe,
44 void mdp5_smp_configure(struct mdp5_smp *smp, enum mdp5_pipe pipe);
45 void mdp5_smp_commit(struct mdp5_smp *smp, enum mdp5_pipe pipe);
46 void mdp5_smp_release(struct mdp5_smp *smp, enum mdp5_pipe pipe);
Dmdp5_cfg.c33 .smp = {
104 .smp = {
175 .smp = {
254 .smp = {
315 .smp = {
Dmdp5_kms.c144 if (mdp5_kms->smp) in mdp5_destroy()
145 mdp5_smp_destroy(mdp5_kms->smp); in mdp5_destroy()
568 mdp5_kms->smp = mdp5_smp_init(mdp5_kms->dev, &config->hw->smp); in mdp5_kms_init()
569 if (IS_ERR(mdp5_kms->smp)) { in mdp5_kms_init()
570 ret = PTR_ERR(mdp5_kms->smp); in mdp5_kms_init()
571 mdp5_kms->smp = NULL; in mdp5_kms_init()
Dmdp5_plane.c708 if (mdp5_kms->smp) { in mdp5_plane_mode_set()
709 ret = mdp5_smp_request(mdp5_kms->smp, in mdp5_plane_mode_set()
721 if (mdp5_kms->smp) in mdp5_plane_mode_set()
722 mdp5_smp_configure(mdp5_kms->smp, pipe); in mdp5_plane_mode_set()
841 if (mdp5_kms->smp) in mdp5_plane_complete_flip()
842 mdp5_smp_commit(mdp5_kms->smp, pipe); in mdp5_plane_complete_flip()
868 if (!plane_enabled(plane->state) && mdp5_kms->smp) { in mdp5_plane_complete_commit()
870 mdp5_smp_release(mdp5_kms->smp, pipe); in mdp5_plane_complete_commit()
Dmdp5_cfg.h83 struct mdp5_smp_block smp; member
Dmdp5_kms.h42 struct mdp5_smp *smp; member
/drivers/staging/rdma/hfi1/
Dmad.c63 static int reply(struct ib_mad_hdr *smp) in reply() argument
69 smp->method = IB_MGMT_METHOD_GET_RESP; in reply()
70 if (smp->mgmt_class == IB_MGMT_CLASS_SUBN_DIRECTED_ROUTE) in reply()
71 smp->status |= IB_SMP_DIRECTION; in reply()
75 static inline void clear_opa_smp_data(struct opa_smp *smp) in clear_opa_smp_data() argument
77 void *data = opa_get_smp_data(smp); in clear_opa_smp_data()
78 size_t size = opa_get_smp_data_size(smp); in clear_opa_smp_data()
87 struct ib_smp *smp; in send_trap() local
119 smp = send_buf->mad; in send_trap()
120 smp->base_version = IB_MGMT_BASE_VERSION; in send_trap()
[all …]
Dud.c553 struct hfi1_qp *qp, u16 slid, struct opa_smp *smp) in opa_smp_check() argument
572 if (smp->mgmt_class != IB_MGMT_CLASS_SUBN_DIRECTED_ROUTE && in opa_smp_check()
573 smp->mgmt_class != IB_MGMT_CLASS_SUBN_LID_ROUTED) { in opa_smp_check()
596 switch (smp->method) { in opa_smp_check()
613 smp->status |= IB_SMP_UNSUP_METHOD; in opa_smp_check()
768 struct opa_smp *smp = (struct opa_smp *)data; in hfi1_ud_rcv() local
775 if (opa_smp_check(ibp, pkey, sc5, qp, slid, smp)) in hfi1_ud_rcv()
782 smp->mgmt_class != IB_MGMT_CLASS_SUBN_DIRECTED_ROUTE) in hfi1_ud_rcv()
Ddiag.c1289 struct ib_smp *smp = NULL; in hfi1_filter_mad_mgmt_class() local
1313 smp = (struct ib_smp *)packet_data; in hfi1_filter_mad_mgmt_class()
1314 if (*((u8 *)value) == smp->mgmt_class) in hfi1_filter_mad_mgmt_class()
/drivers/infiniband/hw/qib/
Dqib_mad.c40 static int reply(struct ib_smp *smp) in reply() argument
46 smp->method = IB_MGMT_METHOD_GET_RESP; in reply()
47 if (smp->mgmt_class == IB_MGMT_CLASS_SUBN_DIRECTED_ROUTE) in reply()
48 smp->status |= IB_SMP_DIRECTION; in reply()
52 static int reply_failure(struct ib_smp *smp) in reply_failure() argument
58 smp->method = IB_MGMT_METHOD_GET_RESP; in reply_failure()
59 if (smp->mgmt_class == IB_MGMT_CLASS_SUBN_DIRECTED_ROUTE) in reply_failure()
60 smp->status |= IB_SMP_DIRECTION; in reply_failure()
68 struct ib_smp *smp; in qib_send_trap() local
91 smp = send_buf->mad; in qib_send_trap()
[all …]
Dqib_ud.c497 struct ib_smp *smp; in qib_ud_rcv() local
502 smp = (struct ib_smp *) data; in qib_ud_rcv()
505 smp->mgmt_class != IB_MGMT_CLASS_SUBN_DIRECTED_ROUTE) in qib_ud_rcv()
Dqib_iba7322.c5496 struct ib_smp *smp; in try_7322_ipg() local
5526 smp = send_buf->mad; in try_7322_ipg()
5527 smp->base_version = IB_MGMT_BASE_VERSION; in try_7322_ipg()
5528 smp->mgmt_class = IB_MGMT_CLASS_SUBN_DIRECTED_ROUTE; in try_7322_ipg()
5529 smp->class_version = 1; in try_7322_ipg()
5530 smp->method = IB_MGMT_METHOD_SEND; in try_7322_ipg()
5531 smp->hop_cnt = 1; in try_7322_ipg()
5532 smp->attr_id = QIB_VENDOR_IPG; in try_7322_ipg()
5533 smp->attr_mod = 0; in try_7322_ipg()
/drivers/staging/rdma/ipath/
Dipath_mad.c46 static int reply(struct ib_smp *smp) in reply() argument
52 smp->method = IB_MGMT_METHOD_GET_RESP; in reply()
53 if (smp->mgmt_class == IB_MGMT_CLASS_SUBN_DIRECTED_ROUTE) in reply()
54 smp->status |= IB_SMP_DIRECTION; in reply()
58 static int recv_subn_get_nodedescription(struct ib_smp *smp, in recv_subn_get_nodedescription() argument
61 if (smp->attr_mod) in recv_subn_get_nodedescription()
62 smp->status |= IB_SMP_INVALID_FIELD; in recv_subn_get_nodedescription()
64 memcpy(smp->data, ibdev->node_desc, sizeof(smp->data)); in recv_subn_get_nodedescription()
66 return reply(smp); in recv_subn_get_nodedescription()
84 static int recv_subn_get_nodeinfo(struct ib_smp *smp, in recv_subn_get_nodeinfo() argument
[all …]
Dipath_ud.c466 struct ib_smp *smp = (struct ib_smp *) data; in ipath_ud_rcv() local
468 if (smp->mgmt_class != IB_MGMT_CLASS_SUBN_DIRECTED_ROUTE) { in ipath_ud_rcv()
/drivers/scsi/csiostor/
Dcsio_defs.h84 csio_set_state(void *smp, void *state) in csio_set_state() argument
86 ((struct csio_sm *)smp)->sm_state = (csio_sm_state_t)state; in csio_set_state()
90 csio_init_state(struct csio_sm *smp, void *state) in csio_init_state() argument
92 csio_set_state(smp, state); in csio_init_state()
96 csio_post_event(void *smp, uint32_t evt) in csio_post_event() argument
98 ((struct csio_sm *)smp)->sm_state(smp, evt); in csio_post_event()
102 csio_get_state(void *smp) in csio_get_state() argument
104 return ((struct csio_sm *)smp)->sm_state; in csio_get_state()
108 csio_match_state(void *smp, void *state) in csio_match_state() argument
110 return (csio_get_state(smp) == (csio_sm_state_t)state); in csio_match_state()
/drivers/sh/intc/
Dcore.c167 unsigned int smp) in save_reg() argument
174 d->smp[cnt] = smp; in save_reg()
184 unsigned int i, k, smp; in register_intc_controller() local
238 d->smp = kzalloc(d->nr_reg * sizeof(*d->smp), GFP_NOWAIT); in register_intc_controller()
239 if (!d->smp) in register_intc_controller()
246 smp = IS_SMP(hw->mask_regs[i]); in register_intc_controller()
247 k += save_reg(d, k, hw->mask_regs[i].set_reg, smp); in register_intc_controller()
248 k += save_reg(d, k, hw->mask_regs[i].clr_reg, smp); in register_intc_controller()
262 smp = IS_SMP(hw->prio_regs[i]); in register_intc_controller()
263 k += save_reg(d, k, hw->prio_regs[i].set_reg, smp); in register_intc_controller()
[all …]
Dinternals.h22 #define IS_SMP(x) (x.smp)
23 #define INTC_REG(d, x, c) (d->reg[(x)] + ((d->smp[(x)] & 0xff) * c))
24 #define SMP_NR(d, x) ((d->smp[(x)] >> 8) ? (d->smp[(x)] >> 8) : 1)
61 unsigned long *smp; member
/drivers/net/ethernet/mellanox/mlx4/
Dcmd.c897 struct ib_smp *smp = inbox->buf; in mlx4_MAD_IFC_wrapper() local
918 if (smp->base_version == 1 && in mlx4_MAD_IFC_wrapper()
919 smp->mgmt_class == IB_MGMT_CLASS_SUBN_LID_ROUTED && in mlx4_MAD_IFC_wrapper()
920 smp->class_version == 1) { in mlx4_MAD_IFC_wrapper()
922 if (!network_view && smp->method == IB_MGMT_METHOD_GET) { in mlx4_MAD_IFC_wrapper()
923 if (smp->attr_id == IB_SMP_ATTR_PKEY_TABLE) { in mlx4_MAD_IFC_wrapper()
924 index = be32_to_cpu(smp->attr_mod); in mlx4_MAD_IFC_wrapper()
945 if (smp->attr_id == IB_SMP_ATTR_PORT_INFO) { in mlx4_MAD_IFC_wrapper()
948 smp->attr_mod = cpu_to_be32(port); in mlx4_MAD_IFC_wrapper()
962 if (smp->attr_id == IB_SMP_ATTR_GUID_INFO) { in mlx4_MAD_IFC_wrapper()
[all …]
/drivers/scsi/isci/
Dscu_task_context.h429 struct smp_task_context smp; member
/drivers/scsi/esas2r/
Datioctl.h1162 struct atto_ioctl_vda_smp_cmd smp; member

12