Home
last modified time | relevance | path

Searched refs:cl (Results 1 – 25 of 331) sorted by relevance

12345678910>>...14

/kernel/linux/linux-5.10/drivers/hid/intel-ish-hid/ishtp/
Dclient.c16 int ishtp_cl_get_tx_free_buffer_size(struct ishtp_cl *cl) in ishtp_cl_get_tx_free_buffer_size() argument
21 spin_lock_irqsave(&cl->tx_free_list_spinlock, tx_free_flags); in ishtp_cl_get_tx_free_buffer_size()
22 size = cl->tx_ring_free_size * cl->device->fw_client->props.max_msg_length; in ishtp_cl_get_tx_free_buffer_size()
23 spin_unlock_irqrestore(&cl->tx_free_list_spinlock, tx_free_flags); in ishtp_cl_get_tx_free_buffer_size()
29 int ishtp_cl_get_tx_free_rings(struct ishtp_cl *cl) in ishtp_cl_get_tx_free_rings() argument
31 return cl->tx_ring_free_size; in ishtp_cl_get_tx_free_rings()
41 static void ishtp_read_list_flush(struct ishtp_cl *cl) in ishtp_read_list_flush() argument
47 spin_lock_irqsave(&cl->dev->read_list_spinlock, flags); in ishtp_read_list_flush()
48 list_for_each_entry_safe(rb, next, &cl->dev->read_list.list, list) in ishtp_read_list_flush()
49 if (rb->cl && ishtp_cl_cmp_id(cl, rb->cl)) { in ishtp_read_list_flush()
[all …]
Dclient-buffers.c19 int ishtp_cl_alloc_rx_ring(struct ishtp_cl *cl) in ishtp_cl_alloc_rx_ring() argument
21 size_t len = cl->device->fw_client->props.max_msg_length; in ishtp_cl_alloc_rx_ring()
27 for (j = 0; j < cl->rx_ring_size; ++j) { in ishtp_cl_alloc_rx_ring()
28 rb = ishtp_io_rb_init(cl); in ishtp_cl_alloc_rx_ring()
36 spin_lock_irqsave(&cl->free_list_spinlock, flags); in ishtp_cl_alloc_rx_ring()
37 list_add_tail(&rb->list, &cl->free_rb_list.list); in ishtp_cl_alloc_rx_ring()
38 spin_unlock_irqrestore(&cl->free_list_spinlock, flags); in ishtp_cl_alloc_rx_ring()
44 dev_err(&cl->device->dev, "error in allocating Rx buffers\n"); in ishtp_cl_alloc_rx_ring()
45 ishtp_cl_free_rx_ring(cl); in ishtp_cl_alloc_rx_ring()
57 int ishtp_cl_alloc_tx_ring(struct ishtp_cl *cl) in ishtp_cl_alloc_tx_ring() argument
[all …]
Dhbm.c55 static inline void ishtp_hbm_cl_hdr(struct ishtp_cl *cl, uint8_t hbm_cmd, in ishtp_hbm_cl_hdr() argument
63 cmd->host_addr = cl->host_client_id; in ishtp_hbm_cl_hdr()
64 cmd->fw_addr = cl->fw_client_id; in ishtp_hbm_cl_hdr()
76 static inline bool ishtp_hbm_cl_addr_equal(struct ishtp_cl *cl, void *buf) in ishtp_hbm_cl_addr_equal() argument
80 return cl->host_client_id == cmd->host_addr && in ishtp_hbm_cl_addr_equal()
81 cl->fw_client_id == cmd->fw_addr; in ishtp_hbm_cl_addr_equal()
262 struct ishtp_cl *cl) in ishtp_hbm_cl_flow_control_req() argument
270 spin_lock_irqsave(&cl->fc_spinlock, flags); in ishtp_hbm_cl_flow_control_req()
273 ishtp_hbm_cl_hdr(cl, ISHTP_FLOW_CONTROL_CMD, &flow_ctrl, len); in ishtp_hbm_cl_flow_control_req()
279 if (cl->out_flow_ctrl_creds) { in ishtp_hbm_cl_flow_control_req()
[all …]
/kernel/linux/linux-5.10/net/sched/
Dsch_hfsc.c186 eltree_insert(struct hfsc_class *cl) in eltree_insert() argument
188 struct rb_node **p = &cl->sched->eligible.rb_node; in eltree_insert()
195 if (cl->cl_e >= cl1->cl_e) in eltree_insert()
200 rb_link_node(&cl->el_node, parent, p); in eltree_insert()
201 rb_insert_color(&cl->el_node, &cl->sched->eligible); in eltree_insert()
205 eltree_remove(struct hfsc_class *cl) in eltree_remove() argument
207 rb_erase(&cl->el_node, &cl->sched->eligible); in eltree_remove()
211 eltree_update(struct hfsc_class *cl) in eltree_update() argument
213 eltree_remove(cl); in eltree_update()
214 eltree_insert(cl); in eltree_update()
[all …]
Dsch_cbq.c163 #define L2T(cl, len) qdisc_l2t((cl)->R_tab, len) argument
181 struct cbq_class *cl; in cbq_reclassify() local
183 for (cl = this->tparent; cl; cl = cl->tparent) { in cbq_reclassify()
184 struct cbq_class *new = cl->defaults[TC_PRIO_BESTEFFORT]; in cbq_reclassify()
210 struct cbq_class *cl = NULL; in cbq_classify() local
219 (cl = cbq_class_lookup(q, prio)) != NULL) in cbq_classify()
220 return cl; in cbq_classify()
237 cl = (void *)res.class; in cbq_classify()
238 if (!cl) { in cbq_classify()
240 cl = cbq_class_lookup(q, res.classid); in cbq_classify()
[all …]
Dsch_htb.c213 struct htb_class *cl; in htb_classify() local
224 cl = htb_find(skb->priority, sch); in htb_classify()
225 if (cl) { in htb_classify()
226 if (cl->level == 0) in htb_classify()
227 return cl; in htb_classify()
229 tcf = rcu_dereference_bh(cl->filter_list); in htb_classify()
247 cl = (void *)res.class; in htb_classify()
248 if (!cl) { in htb_classify()
251 cl = htb_find(res.classid, sch); in htb_classify()
252 if (!cl) in htb_classify()
[all …]
Dsch_drr.c59 struct drr_class *cl = (struct drr_class *)*arg; in drr_change_class() local
84 if (cl != NULL) { in drr_change_class()
86 err = gen_replace_estimator(&cl->bstats, NULL, in drr_change_class()
87 &cl->rate_est, in drr_change_class()
99 cl->quantum = quantum; in drr_change_class()
105 cl = kzalloc(sizeof(struct drr_class), GFP_KERNEL); in drr_change_class()
106 if (cl == NULL) in drr_change_class()
109 cl->common.classid = classid; in drr_change_class()
110 cl->quantum = quantum; in drr_change_class()
111 cl->qdisc = qdisc_create_dflt(sch->dev_queue, in drr_change_class()
[all …]
Dsch_qfq.c313 struct qfq_class *cl) in qfq_add_to_agg() argument
315 cl->agg = agg; in qfq_add_to_agg()
318 if (cl->qdisc->q.qlen > 0) { /* adding an active class */ in qfq_add_to_agg()
319 list_add_tail(&cl->alist, &agg->active); in qfq_add_to_agg()
321 cl && q->in_serv_agg != agg) /* agg was inactive */ in qfq_add_to_agg()
341 static void qfq_deactivate_class(struct qfq_sched *q, struct qfq_class *cl) in qfq_deactivate_class() argument
343 struct qfq_aggregate *agg = cl->agg; in qfq_deactivate_class()
346 list_del(&cl->alist); /* remove from RR queue of the aggregate */ in qfq_deactivate_class()
352 static void qfq_rm_from_agg(struct qfq_sched *q, struct qfq_class *cl) in qfq_rm_from_agg() argument
354 struct qfq_aggregate *agg = cl->agg; in qfq_rm_from_agg()
[all …]
/kernel/linux/linux-5.10/drivers/misc/mei/
Dclient.c339 cb->cl->tx_cb_queued++; in mei_tx_cb_enqueue()
351 if (!WARN_ON(cb->cl->tx_cb_queued == 0)) in mei_tx_cb_dequeue()
352 cb->cl->tx_cb_queued--; in mei_tx_cb_dequeue()
365 static void mei_cl_set_read_by_fp(const struct mei_cl *cl, in mei_cl_set_read_by_fp() argument
370 list_for_each_entry(cl_vtag, &cl->vtag_map, list) { in mei_cl_set_read_by_fp()
387 static struct mei_cl_cb *mei_io_cb_init(struct mei_cl *cl, in mei_io_cb_init() argument
399 cb->cl = cl; in mei_io_cb_init()
414 const struct mei_cl *cl) in mei_io_list_flush_cl() argument
419 if (cl == cb->cl) { in mei_io_list_flush_cl()
435 const struct mei_cl *cl, in mei_io_tx_list_free_cl() argument
[all …]
Dclient.h131 int mei_cl_link(struct mei_cl *cl);
132 int mei_cl_unlink(struct mei_cl *cl);
136 struct mei_cl_cb *mei_cl_read_cb(struct mei_cl *cl, const struct file *fp);
138 void mei_cl_add_rd_completed(struct mei_cl *cl, struct mei_cl_cb *cb);
139 void mei_cl_del_rd_completed(struct mei_cl *cl, struct mei_cl_cb *cb);
141 struct mei_cl_cb *mei_cl_alloc_cb(struct mei_cl *cl, size_t length,
144 struct mei_cl_cb *mei_cl_enqueue_ctrl_wr_cb(struct mei_cl *cl, size_t length,
147 int mei_cl_flush_queues(struct mei_cl *cl, const struct file *fp);
150 const struct file *mei_cl_fp_by_vtag(const struct mei_cl *cl, u8 vtag);
151 int mei_cl_vt_support_check(const struct mei_cl *cl);
[all …]
Dmain.c48 struct mei_cl *cl; in mei_open() local
65 cl = mei_cl_alloc_linked(dev); in mei_open()
66 if (IS_ERR(cl)) { in mei_open()
67 err = PTR_ERR(cl); in mei_open()
71 cl->fp = file; in mei_open()
72 file->private_data = cl; in mei_open()
90 static void mei_cl_vtag_remove_by_fp(const struct mei_cl *cl, in mei_cl_vtag_remove_by_fp() argument
95 list_for_each_entry_safe(vtag_l, next, &cl->vtag_map, list) { in mei_cl_vtag_remove_by_fp()
114 struct mei_cl *cl = file->private_data; in mei_release() local
118 if (WARN_ON(!cl || !cl->dev)) in mei_release()
[all …]
Dinterrupt.c32 struct mei_cl *cl; in mei_irq_compl_handler() local
35 cl = cb->cl; in mei_irq_compl_handler()
39 mei_cl_complete(cl, cb); in mei_irq_compl_handler()
52 static inline int mei_cl_hbm_equal(struct mei_cl *cl, in mei_cl_hbm_equal() argument
55 return mei_cl_host_addr(cl) == mei_hdr->host_addr && in mei_cl_hbm_equal()
56 mei_cl_me_id(cl) == mei_hdr->me_addr; in mei_cl_hbm_equal()
93 static int mei_cl_irq_read_msg(struct mei_cl *cl, in mei_cl_irq_read_msg() argument
98 struct mei_device *dev = cl->dev; in mei_cl_irq_read_msg()
112 cb = list_first_entry_or_null(&cl->rd_pending, struct mei_cl_cb, list); in mei_cl_irq_read_msg()
114 if (!mei_cl_is_fixed_address(cl)) { in mei_cl_irq_read_msg()
[all …]
Dbus.c33 ssize_t __mei_cl_send(struct mei_cl *cl, u8 *buf, size_t length, in __mei_cl_send() argument
40 if (WARN_ON(!cl || !cl->dev)) in __mei_cl_send()
43 bus = cl->dev; in __mei_cl_send()
51 if (!mei_cl_is_connected(cl)) { in __mei_cl_send()
57 if (!mei_me_cl_is_active(cl->me_cl)) { in __mei_cl_send()
62 if (length > mei_cl_mtu(cl)) { in __mei_cl_send()
67 while (cl->tx_cb_queued >= bus->tx_queue_limit) { in __mei_cl_send()
69 rets = wait_event_interruptible(cl->tx_wait, in __mei_cl_send()
70 cl->writing_state == MEI_WRITE_COMPLETE || in __mei_cl_send()
71 (!mei_cl_is_connected(cl))); in __mei_cl_send()
[all …]
/kernel/linux/linux-5.10/drivers/md/bcache/
Dclosure.c16 static inline void closure_put_after_sub(struct closure *cl, int flags) in closure_put_after_sub() argument
24 if (cl->fn && !(flags & CLOSURE_DESTRUCTOR)) { in closure_put_after_sub()
25 atomic_set(&cl->remaining, in closure_put_after_sub()
27 closure_queue(cl); in closure_put_after_sub()
29 struct closure *parent = cl->parent; in closure_put_after_sub()
30 closure_fn *destructor = cl->fn; in closure_put_after_sub()
32 closure_debug_destroy(cl); in closure_put_after_sub()
35 destructor(cl); in closure_put_after_sub()
44 void closure_sub(struct closure *cl, int v) in closure_sub() argument
46 closure_put_after_sub(cl, atomic_sub_return(v, &cl->remaining)); in closure_sub()
[all …]
Dclosure.h169 void closure_sub(struct closure *cl, int v);
170 void closure_put(struct closure *cl);
172 bool closure_wait(struct closure_waitlist *list, struct closure *cl);
173 void __closure_sync(struct closure *cl);
181 static inline void closure_sync(struct closure *cl) in closure_sync() argument
183 if ((atomic_read(&cl->remaining) & CLOSURE_REMAINING_MASK) != 1) in closure_sync()
184 __closure_sync(cl); in closure_sync()
190 void closure_debug_create(struct closure *cl);
191 void closure_debug_destroy(struct closure *cl);
196 static inline void closure_debug_create(struct closure *cl) {} in closure_debug_create() argument
[all …]
Drequest.c28 static void bch_data_insert_start(struct closure *cl);
58 static void bch_data_insert_keys(struct closure *cl) in bch_data_insert_keys() argument
60 struct data_insert_op *op = container_of(cl, struct data_insert_op, cl); in bch_data_insert_keys()
67 op->flush_journal ? cl : NULL); in bch_data_insert_keys()
82 continue_at(cl, bch_data_insert_start, op->wq); in bch_data_insert_keys()
87 closure_return(cl); in bch_data_insert_keys()
108 static void bch_data_invalidate(struct closure *cl) in bch_data_invalidate() argument
110 struct data_insert_op *op = container_of(cl, struct data_insert_op, cl); in bch_data_invalidate()
136 continue_at(cl, bch_data_insert_keys, op->wq); in bch_data_invalidate()
139 static void bch_data_insert_error(struct closure *cl) in bch_data_insert_error() argument
[all …]
/kernel/linux/linux-5.10/drivers/clk/
Dclkdev.c38 struct clk_lookup *p, *cl = NULL; in clk_find() local
62 cl = p; in clk_find()
69 return cl; in clk_find()
74 struct clk_lookup *cl; in clk_find_hw() local
78 cl = clk_find(dev_id, con_id); in clk_find_hw()
79 if (cl) in clk_find_hw()
80 hw = cl->clk_hw; in clk_find_hw()
121 static void __clkdev_add(struct clk_lookup *cl) in __clkdev_add() argument
124 list_add_tail(&cl->node, &clocks); in __clkdev_add()
128 void clkdev_add(struct clk_lookup *cl) in clkdev_add() argument
[all …]
/kernel/linux/linux-5.10/drivers/hsi/
Dhsi_core.c60 struct hsi_client *cl = to_hsi_client(dev); in hsi_client_release() local
62 kfree(cl->tx_cfg.channels); in hsi_client_release()
63 kfree(cl->rx_cfg.channels); in hsi_client_release()
64 kfree(cl); in hsi_client_release()
70 struct hsi_client *cl; in hsi_new_client() local
73 cl = kzalloc(sizeof(*cl), GFP_KERNEL); in hsi_new_client()
74 if (!cl) in hsi_new_client()
77 cl->tx_cfg = info->tx_cfg; in hsi_new_client()
78 if (cl->tx_cfg.channels) { in hsi_new_client()
79 size = cl->tx_cfg.num_channels * sizeof(*cl->tx_cfg.channels); in hsi_new_client()
[all …]
/kernel/linux/linux-5.10/include/linux/hsi/
Dhsi.h138 static inline void hsi_client_set_drvdata(struct hsi_client *cl, void *data) in hsi_client_set_drvdata() argument
140 dev_set_drvdata(&cl->device, data); in hsi_client_set_drvdata()
143 static inline void *hsi_client_drvdata(struct hsi_client *cl) in hsi_client_drvdata() argument
145 return dev_get_drvdata(&cl->device); in hsi_client_drvdata()
148 int hsi_register_port_event(struct hsi_client *cl,
150 int hsi_unregister_port_event(struct hsi_client *cl);
187 struct hsi_client *cl; member
230 int (*setup)(struct hsi_client *cl);
231 int (*flush)(struct hsi_client *cl);
232 int (*start_tx)(struct hsi_client *cl);
[all …]
/kernel/linux/linux-5.10/drivers/hsi/clients/
Dssi_protocol.c35 void ssi_waketest(struct hsi_client *cl, unsigned int enable);
141 struct hsi_client *cl; member
219 struct ssi_protocol *ssi = hsi_client_drvdata(msg->cl); in ssip_release_cmd()
221 dev_dbg(&msg->cl->device, "Release cmd 0x%08x\n", ssip_get_cmd(msg)); in ssip_release_cmd()
333 if (slave->device.parent == ssi->cl->device.parent) { in ssip_slave_get_master()
334 master = ssi->cl; in ssip_slave_get_master()
387 static void ssip_reset(struct hsi_client *cl) in ssip_reset() argument
389 struct ssi_protocol *ssi = hsi_client_drvdata(cl); in ssip_reset()
395 hsi_flush(cl); in ssip_reset()
398 hsi_stop_tx(cl); in ssip_reset()
[all …]
Dhsi_char.c85 struct hsi_client *cl; member
105 struct hsi_client *cl; member
227 struct hsc_client_data *cl_data = hsi_client_drvdata(msg->cl); in hsc_rx_completed()
247 struct hsc_client_data *cl_data = hsi_client_drvdata(msg->cl); in hsc_tx_completed()
267 struct hsc_client_data *cl_data = hsi_client_drvdata(msg->cl); in hsc_break_req_destructor()
275 struct hsc_client_data *cl_data = hsi_client_drvdata(msg->cl); in hsc_break_received()
294 hsi_flush(msg->cl); in hsc_break_received()
295 ret = hsi_async_read(msg->cl, msg); in hsc_break_received()
300 static int hsc_break_request(struct hsi_client *cl) in hsc_break_request() argument
302 struct hsc_client_data *cl_data = hsi_client_drvdata(cl); in hsc_break_request()
[all …]
/kernel/linux/linux-5.10/drivers/gpu/drm/i915/gt/
Dintel_timeline.c126 struct intel_timeline_cacheline *cl = in __rcu_cacheline_free() local
127 container_of(rcu, typeof(*cl), rcu); in __rcu_cacheline_free()
130 i915_gem_object_unpin_map(cl->hwsp->vma->obj); in __rcu_cacheline_free()
131 __idle_hwsp_free(cl->hwsp, ptr_unmask_bits(cl->vaddr, CACHELINE_BITS)); in __rcu_cacheline_free()
133 i915_active_fini(&cl->active); in __rcu_cacheline_free()
134 kfree(cl); in __rcu_cacheline_free()
137 static void __idle_cacheline_free(struct intel_timeline_cacheline *cl) in __idle_cacheline_free() argument
139 GEM_BUG_ON(!i915_active_is_idle(&cl->active)); in __idle_cacheline_free()
140 call_rcu(&cl->rcu, __rcu_cacheline_free); in __idle_cacheline_free()
146 struct intel_timeline_cacheline *cl = in __cacheline_retire() local
[all …]
/kernel/linux/linux-5.10/drivers/gpu/drm/bridge/
Dparade-ps8622.c87 struct i2c_client *cl = ps8622->client; in ps8622_send_config() local
91 err = ps8622_set(cl, 0x02, 0xa1, 0x01); in ps8622_send_config()
96 err = ps8622_set(cl, 0x04, 0x14, 0x01); in ps8622_send_config()
101 err = ps8622_set(cl, 0x04, 0xe3, 0x20); in ps8622_send_config()
106 err = ps8622_set(cl, 0x04, 0xe2, 0x80); in ps8622_send_config()
114 err = ps8622_set(cl, 0x04, 0x8a, 0x0c); in ps8622_send_config()
119 err = ps8622_set(cl, 0x04, 0x89, 0x08); in ps8622_send_config()
124 err = ps8622_set(cl, 0x04, 0x71, 0x2d); in ps8622_send_config()
129 err = ps8622_set(cl, 0x04, 0x7d, 0x07); in ps8622_send_config()
134 err = ps8622_set(cl, 0x04, 0x7b, 0x00); in ps8622_send_config()
[all …]
/kernel/linux/linux-5.10/include/linux/
Dintel-ish-client-if.h64 struct ishtp_cl *cl; member
84 void ishtp_cl_free(struct ishtp_cl *cl);
85 int ishtp_cl_link(struct ishtp_cl *cl);
86 void ishtp_cl_unlink(struct ishtp_cl *cl);
87 int ishtp_cl_disconnect(struct ishtp_cl *cl);
88 int ishtp_cl_connect(struct ishtp_cl *cl);
89 int ishtp_cl_send(struct ishtp_cl *cl, uint8_t *buf, size_t length);
90 int ishtp_cl_flush_queues(struct ishtp_cl *cl);
92 bool ishtp_cl_tx_empty(struct ishtp_cl *cl);
93 struct ishtp_cl_rb *ishtp_cl_rx_get_rb(struct ishtp_cl *cl);
[all …]
/kernel/linux/linux-5.10/arch/openrisc/kernel/
Ddma.c27 unsigned long cl; in page_set_nocache() local
39 for (cl = __pa(addr); cl < __pa(next); cl += cpuinfo->dcache_block_size) in page_set_nocache()
40 mtspr(SPR_DCBFR, cl); in page_set_nocache()
101 unsigned long cl; in arch_sync_dma_for_device() local
107 for (cl = addr; cl < addr + size; in arch_sync_dma_for_device()
108 cl += cpuinfo->dcache_block_size) in arch_sync_dma_for_device()
109 mtspr(SPR_DCBFR, cl); in arch_sync_dma_for_device()
113 for (cl = addr; cl < addr + size; in arch_sync_dma_for_device()
114 cl += cpuinfo->dcache_block_size) in arch_sync_dma_for_device()
115 mtspr(SPR_DCBIR, cl); in arch_sync_dma_for_device()

12345678910>>...14