/drivers/gpu/drm/amd/amdgpu/ |
D | amdgpu_res_cursor.h | 56 struct amdgpu_res_cursor *cur) in amdgpu_res_first() argument 67 cur->mem_type = res->mem_type; in amdgpu_res_first() 69 switch (cur->mem_type) { in amdgpu_res_first() 87 cur->start = amdgpu_vram_mgr_block_start(block) + start; in amdgpu_res_first() 88 cur->size = min(amdgpu_vram_mgr_block_size(block) - start, size); in amdgpu_res_first() 89 cur->remaining = size; in amdgpu_res_first() 90 cur->node = block; in amdgpu_res_first() 97 cur->start = (node->start << PAGE_SHIFT) + start; in amdgpu_res_first() 98 cur->size = min((node->size << PAGE_SHIFT) - start, size); in amdgpu_res_first() 99 cur->remaining = size; in amdgpu_res_first() [all …]
|
/drivers/char/ |
D | bsr.c | 154 struct bsr_dev *cur, *n; in bsr_cleanup_devs() local 156 list_for_each_entry_safe(cur, n, &bsr_devs, bsr_list) { in bsr_cleanup_devs() 157 if (cur->bsr_device) { in bsr_cleanup_devs() 158 cdev_del(&cur->bsr_cdev); in bsr_cleanup_devs() 159 device_del(cur->bsr_device); in bsr_cleanup_devs() 161 list_del(&cur->bsr_list); in bsr_cleanup_devs() 162 kfree(cur); in bsr_cleanup_devs() 186 struct bsr_dev *cur = kzalloc(sizeof(struct bsr_dev), in bsr_add_node() local 191 if (!cur) { in bsr_add_node() 200 kfree(cur); in bsr_add_node() [all …]
|
/drivers/block/drbd/ |
D | drbd_vli.h | 199 static inline void bitstream_cursor_reset(struct bitstream_cursor *cur, void *s) in bitstream_cursor_reset() argument 201 cur->b = s; in bitstream_cursor_reset() 202 cur->bit = 0; in bitstream_cursor_reset() 207 static inline void bitstream_cursor_advance(struct bitstream_cursor *cur, unsigned int bits) in bitstream_cursor_advance() argument 209 bits += cur->bit; in bitstream_cursor_advance() 210 cur->b = cur->b + (bits >> 3); in bitstream_cursor_advance() 211 cur->bit = bits & 7; in bitstream_cursor_advance() 216 struct bitstream_cursor cur; member 231 bitstream_cursor_reset(&bs->cur, bs->buf); in bitstream_init() 236 bitstream_cursor_reset(&bs->cur, bs->buf); in bitstream_rewind() [all …]
|
/drivers/mtd/nand/ |
D | ecc-sw-hamming.c | 121 u32 cur; in ecc_sw_hamming_calculate() local 152 cur = *bp++; in ecc_sw_hamming_calculate() 153 tmppar = cur; in ecc_sw_hamming_calculate() 154 rp4 ^= cur; in ecc_sw_hamming_calculate() 155 cur = *bp++; in ecc_sw_hamming_calculate() 156 tmppar ^= cur; in ecc_sw_hamming_calculate() 158 cur = *bp++; in ecc_sw_hamming_calculate() 159 tmppar ^= cur; in ecc_sw_hamming_calculate() 160 rp4 ^= cur; in ecc_sw_hamming_calculate() 161 cur = *bp++; in ecc_sw_hamming_calculate() [all …]
|
/drivers/clk/rockchip/ |
D | clk-pll.c | 167 struct rockchip_pll_rate_table cur; in rockchip_rk3036_pll_recalc_rate() local 170 rockchip_rk3036_pll_get_params(pll, &cur); in rockchip_rk3036_pll_recalc_rate() 172 rate64 *= cur.fbdiv; in rockchip_rk3036_pll_recalc_rate() 173 do_div(rate64, cur.refdiv); in rockchip_rk3036_pll_recalc_rate() 175 if (cur.dsmpd == 0) { in rockchip_rk3036_pll_recalc_rate() 177 u64 frac_rate64 = prate * cur.frac; in rockchip_rk3036_pll_recalc_rate() 179 do_div(frac_rate64, cur.refdiv); in rockchip_rk3036_pll_recalc_rate() 183 do_div(rate64, cur.postdiv1); in rockchip_rk3036_pll_recalc_rate() 184 do_div(rate64, cur.postdiv2); in rockchip_rk3036_pll_recalc_rate() 194 struct rockchip_pll_rate_table cur; in rockchip_rk3036_pll_set_params() local [all …]
|
/drivers/char/agp/ |
D | isoch.c | 25 struct agp_3_5_dev *cur, *n = list_entry(new, struct agp_3_5_dev, list); in agp_3_5_dev_list_insert() local 29 cur = list_entry(pos, struct agp_3_5_dev, list); in agp_3_5_dev_list_insert() 30 if (cur->maxbw > n->maxbw) in agp_3_5_dev_list_insert() 38 struct agp_3_5_dev *cur; in agp_3_5_dev_list_sort() local 46 cur = list_entry(pos, struct agp_3_5_dev, list); in agp_3_5_dev_list_sort() 47 dev = cur->dev; in agp_3_5_dev_list_sort() 49 pci_read_config_dword(dev, cur->capndx+AGPNISTAT, &nistat); in agp_3_5_dev_list_sort() 50 cur->maxbw = (nistat >> 16) & 0xff; in agp_3_5_dev_list_sort() 82 struct agp_3_5_dev *cur; in agp_3_5_isochronous_node_enable() local 137 cur = list_entry(pos, struct agp_3_5_dev, list); in agp_3_5_isochronous_node_enable() [all …]
|
/drivers/video/backlight/ |
D | cr_bllcd.c | 63 u32 cur = inl(addr); in cr_backlight_set_intensity() local 67 cur |= CRVML_BACKLIGHT_OFF; in cr_backlight_set_intensity() 68 outl(cur, addr); in cr_backlight_set_intensity() 71 cur &= ~CRVML_BACKLIGHT_OFF; in cr_backlight_set_intensity() 72 outl(cur, addr); in cr_backlight_set_intensity() 81 u32 cur = inl(addr); in cr_backlight_get_intensity() local 84 if (cur & CRVML_BACKLIGHT_OFF) in cr_backlight_get_intensity() 100 u32 cur = inl(addr); in cr_panel_on() local 102 if (!(cur & CRVML_PANEL_ON)) { in cr_panel_on() 104 if (cur & 0x00000001) { in cr_panel_on() [all …]
|
/drivers/misc/ocxl/ |
D | pasid.c | 15 struct id_range *cur; in dump_list() local 18 list_for_each_entry(cur, head, list) { in dump_list() 19 pr_debug("Range %d->%d\n", cur->start, cur->end); in dump_list() 28 struct id_range *cur, *new; in range_alloc() local 37 list_for_each_entry(cur, head, list) { in range_alloc() 38 if ((cur->start - last_end) > size) in range_alloc() 40 last_end = cur->end; in range_alloc() 41 pos = &cur->list; in range_alloc() 65 struct id_range *cur, *tmp; in range_free() local 67 list_for_each_entry_safe(cur, tmp, head, list) { in range_free() [all …]
|
/drivers/misc/vmw_vmci/ |
D | vmci_event.c | 49 struct vmci_subscription *cur, *p2; in vmci_event_exit() local 50 list_for_each_entry_safe(cur, p2, &subscriber_array[e], node) { in vmci_event_exit() 58 list_del(&cur->node); in vmci_event_exit() 59 kfree(cur); in vmci_event_exit() 72 struct vmci_subscription *cur; in event_find() local 73 list_for_each_entry(cur, &subscriber_array[e], node) { in event_find() 74 if (cur->id == sub_id) in event_find() 75 return cur; in event_find() 87 struct vmci_subscription *cur; in event_deliver() local 92 list_for_each_entry_rcu(cur, subscriber_list, node) { in event_deliver() [all …]
|
/drivers/net/ethernet/mediatek/ |
D | mtk_wed_debugfs.c | 50 const struct reg_dump *cur; in dump_wed_regs() local 53 for (cur = regs; cur < ®s[n_regs]; cur++) { in dump_wed_regs() 54 switch (cur->type) { in dump_wed_regs() 57 cur > regs ? "\n" : "", in dump_wed_regs() 58 cur->name); in dump_wed_regs() 61 val = wed_r32(dev, cur->offset); in dump_wed_regs() 64 val = wdma_r32(dev, cur->offset); in dump_wed_regs() 67 val = wpdma_tx_r32(dev, cur->base, cur->offset); in dump_wed_regs() 70 val = wpdma_txfree_r32(dev, cur->offset); in dump_wed_regs() 73 print_reg_val(s, cur->name, val); in dump_wed_regs()
|
/drivers/net/wireless/mediatek/mt76/ |
D | util.c | 12 u32 cur; in __mt76_poll() local 16 cur = __mt76_rr(dev, offset) & mask; in __mt76_poll() 17 if (cur == val) in __mt76_poll() 30 u32 cur; in ____mt76_poll_msec() local 34 cur = __mt76_rr(dev, offset) & mask; in ____mt76_poll_msec() 35 if (cur == val) in ____mt76_poll_msec() 47 int i, idx = 0, cur; in mt76_wcid_alloc() local 55 cur = i * 32 + idx; in mt76_wcid_alloc() 56 if (cur >= size) in mt76_wcid_alloc() 60 return cur; in mt76_wcid_alloc()
|
D | eeprom.c | 145 struct device_node *cur, *fallback = NULL; in mt76_find_power_limits_node() local 155 for_each_child_of_node(np, cur) { in mt76_find_power_limits_node() 156 struct property *country = of_find_property(cur, "country", NULL); in mt76_find_power_limits_node() 157 struct property *regd = of_find_property(cur, "regdomain", NULL); in mt76_find_power_limits_node() 160 fallback = cur; in mt76_find_power_limits_node() 167 return cur; in mt76_find_power_limits_node() 191 struct device_node *cur; in mt76_find_channel_node() local 195 for_each_child_of_node(np, cur) { in mt76_find_channel_node() 196 val = mt76_get_of_array(cur, "channels", &len, 2); in mt76_find_channel_node() 203 return cur; in mt76_find_channel_node() [all …]
|
/drivers/net/wireless/mediatek/mt76/mt7915/ |
D | soc.c | 44 u32 cur; in mt76_wmac_spi_read() local 46 ret = read_poll_timeout(mt76_rr, cur, !(cur & MT_TOP_SPI_POLLING_BIT), in mt76_wmac_spi_read() 56 ret = read_poll_timeout(mt76_rr, cur, !(cur & MT_TOP_SPI_POLLING_BIT), in mt76_wmac_spi_read() 71 u32 cur; in mt76_wmac_spi_write() local 73 ret = read_poll_timeout(mt76_rr, cur, !(cur & MT_TOP_SPI_POLLING_BIT), in mt76_wmac_spi_write() 83 return read_poll_timeout(mt76_rr, cur, !(cur & MT_TOP_SPI_POLLING_BIT), in mt76_wmac_spi_write() 92 u32 cur, ret; in mt76_wmac_spi_rmw() local 94 ret = mt76_wmac_spi_read(dev, adie, addr, &cur); in mt76_wmac_spi_rmw() 98 cur &= ~mask; in mt76_wmac_spi_rmw() 99 cur |= val; in mt76_wmac_spi_rmw() [all …]
|
/drivers/scsi/sym53c8xx_2/ |
D | sym_fw.c | 351 u32 *end, *cur; in sym_fw_bind_script() local 354 cur = start; in sym_fw_bind_script() 357 while (cur < end) { in sym_fw_bind_script() 359 opcode = *cur; in sym_fw_bind_script() 369 sym_name(np), (int) (cur-start)); in sym_fw_bind_script() 370 ++cur; in sym_fw_bind_script() 379 *cur++ = 0; in sym_fw_bind_script() 384 printf ("%d: <%x>\n", (int) (cur-start), in sym_fw_bind_script() 408 tmp1 = cur[1]; in sym_fw_bind_script() 409 tmp2 = cur[2]; in sym_fw_bind_script() [all …]
|
/drivers/gpu/drm/savage/ |
D | savage_bci.c | 374 unsigned int cur = dev_priv->current_dma_page; in savage_dma_alloc() local 376 dev_priv->dma_pages[cur].used; in savage_dma_alloc() 383 cur, dev_priv->dma_pages[cur].used, n, rest, nr_pages); in savage_dma_alloc() 385 if (cur + nr_pages < dev_priv->nr_dma_pages) { in savage_dma_alloc() 387 cur * SAVAGE_DMA_PAGE_SIZE + dev_priv->dma_pages[cur].used; in savage_dma_alloc() 390 dev_priv->dma_pages[cur].used += rest; in savage_dma_alloc() 392 cur++; in savage_dma_alloc() 397 for (i = cur; i < dev_priv->nr_dma_pages; ++i) { in savage_dma_alloc() 403 dev_priv->first_dma_page = cur = 0; in savage_dma_alloc() 405 for (i = cur; nr_pages > 0; ++i, --nr_pages) { in savage_dma_alloc() [all …]
|
/drivers/power/supply/ |
D | sc2731_charger.c | 110 static int sc2731_charger_set_current(struct sc2731_charger_info *info, u32 cur) in sc2731_charger_set_current() argument 115 if (cur > SC2731_CURRENT_LIMIT_2000) in sc2731_charger_set_current() 116 cur = SC2731_CURRENT_LIMIT_2000; in sc2731_charger_set_current() 117 else if (cur < SC2731_CURRENT_PRECHG) in sc2731_charger_set_current() 118 cur = SC2731_CURRENT_PRECHG; in sc2731_charger_set_current() 121 val = (cur - SC2731_CURRENT_PRECHG) / SC2731_CURRENT_STEP; in sc2731_charger_set_current() 150 u32 *cur) in sc2731_charger_get_current() argument 160 *cur = val * SC2731_CURRENT_STEP + SC2731_CURRENT_PRECHG; in sc2731_charger_get_current() 166 u32 *cur) in sc2731_charger_get_current_limit() argument 179 *cur = SC2731_CURRENT_LIMIT_100; in sc2731_charger_get_current_limit() [all …]
|
/drivers/i2c/busses/ |
D | i2c-qcom-geni.c | 87 struct i2c_msg *cur; member 203 if (gi2c->cur) in geni_i2c_err() 205 gi2c->cur->len, gi2c->cur->addr, gi2c->cur->flags); in geni_i2c_err() 233 struct i2c_msg *cur; in geni_i2c_irq() local 241 cur = gi2c->cur; in geni_i2c_irq() 243 if (!cur || in geni_i2c_irq() 267 } else if (cur->flags & I2C_M_RD && in geni_i2c_irq() 274 while (gi2c->cur_rd < cur->len && p < sizeof(val)) { in geni_i2c_irq() 275 cur->buf[gi2c->cur_rd++] = val & 0xff; in geni_i2c_irq() 279 if (gi2c->cur_rd == cur->len) in geni_i2c_irq() [all …]
|
/drivers/gpu/drm/amd/display/amdgpu_dm/ |
D | amdgpu_dm_crtc.c | 224 struct dm_crtc_state *cur = to_dm_crtc_state(state); in dm_crtc_destroy_state() local 227 if (cur->stream) in dm_crtc_destroy_state() 228 dc_stream_release(cur->stream); in dm_crtc_destroy_state() 239 struct dm_crtc_state *state, *cur; in dm_crtc_duplicate_state() local 241 cur = to_dm_crtc_state(crtc->state); in dm_crtc_duplicate_state() 252 if (cur->stream) { in dm_crtc_duplicate_state() 253 state->stream = cur->stream; in dm_crtc_duplicate_state() 257 state->active_planes = cur->active_planes; in dm_crtc_duplicate_state() 258 state->vrr_infopacket = cur->vrr_infopacket; in dm_crtc_duplicate_state() 259 state->abm_level = cur->abm_level; in dm_crtc_duplicate_state() [all …]
|
/drivers/infiniband/hw/cxgb4/ |
D | resource.c | 126 rdev->stats.qid.cur += rdev->qpmask + 1; in c4iw_get_cqid() 157 if (rdev->stats.qid.cur > rdev->stats.qid.max) in c4iw_get_cqid() 158 rdev->stats.qid.max = rdev->stats.qid.cur; in c4iw_get_cqid() 200 rdev->stats.qid.cur += rdev->qpmask + 1; in c4iw_get_qpid() 231 if (rdev->stats.qid.cur > rdev->stats.qid.max) in c4iw_get_qpid() 232 rdev->stats.qid.max = rdev->stats.qid.cur; in c4iw_get_qpid() 271 rdev->stats.pbl.cur += roundup(size, 1 << MIN_PBL_SHIFT); in c4iw_pblpool_alloc() 272 if (rdev->stats.pbl.cur > rdev->stats.pbl.max) in c4iw_pblpool_alloc() 273 rdev->stats.pbl.max = rdev->stats.pbl.cur; in c4iw_pblpool_alloc() 294 rdev->stats.pbl.cur -= roundup(size, 1 << MIN_PBL_SHIFT); in c4iw_pblpool_free() [all …]
|
/drivers/infiniband/core/ |
D | umem_dmabuf.c | 19 unsigned long start, end, cur = 0; in ib_umem_dmabuf_map_pages() local 39 if (start < cur + sg_dma_len(sg) && cur < end) in ib_umem_dmabuf_map_pages() 41 if (cur <= start && start < cur + sg_dma_len(sg)) { in ib_umem_dmabuf_map_pages() 42 unsigned long offset = start - cur; in ib_umem_dmabuf_map_pages() 48 cur += offset; in ib_umem_dmabuf_map_pages() 50 if (cur < end && end <= cur + sg_dma_len(sg)) { in ib_umem_dmabuf_map_pages() 51 unsigned long trim = cur + sg_dma_len(sg) - end; in ib_umem_dmabuf_map_pages() 58 cur += sg_dma_len(sg); in ib_umem_dmabuf_map_pages()
|
/drivers/cpufreq/ |
D | amd_freq_sensitivity.c | 63 freq_next = policy->cur; in amd_powersave_bias_target() 72 freq_next = policy->cur; in amd_powersave_bias_target() 83 if (data->freq_prev == policy->cur) in amd_powersave_bias_target() 84 freq_next = policy->cur; in amd_powersave_bias_target() 86 if (freq_next > policy->cur) in amd_powersave_bias_target() 87 freq_next = policy->cur; in amd_powersave_bias_target() 88 else if (freq_next < policy->cur) in amd_powersave_bias_target() 94 policy->cur - 1, in amd_powersave_bias_target()
|
/drivers/gpu/drm/msm/disp/mdp5/ |
D | mdp5_pipe.c | 31 struct mdp5_hw_pipe *cur = mdp5_kms->hwpipes[i]; in mdp5_pipe_assign() local 40 if (new_state->hwpipe_to_plane[cur->idx] || in mdp5_pipe_assign() 41 old_state->hwpipe_to_plane[cur->idx]) in mdp5_pipe_assign() 45 if (caps & ~cur->caps) in mdp5_pipe_assign() 52 if (cur->caps & MDP_PIPE_CAP_CURSOR && in mdp5_pipe_assign() 59 if (!(*hwpipe) || (hweight_long(cur->caps & ~caps) < in mdp5_pipe_assign() 70 if (r_cur->caps != cur->caps) in mdp5_pipe_assign() 74 if (cur->pipe > r_cur->pipe) in mdp5_pipe_assign() 84 *hwpipe = cur; in mdp5_pipe_assign()
|
/drivers/net/wireless/mediatek/mt7601u/ |
D | core.c | 31 u32 cur; in mt76_poll() local 38 cur = mt7601u_rr(dev, offset) & mask; in mt76_poll() 39 if (cur == val) in mt76_poll() 53 u32 cur; in mt76_poll_msec() local 60 cur = mt7601u_rr(dev, offset) & mask; in mt76_poll_msec() 61 if (cur == val) in mt76_poll_msec()
|
/drivers/media/test-drivers/vivid/ |
D | vivid-radio-common.c | 70 rds->picode = dev->radio_tx_rds_pi->cur.val; in vivid_radio_rds_init() 71 rds->pty = dev->radio_tx_rds_pty->cur.val; in vivid_radio_rds_init() 72 rds->mono_stereo = dev->radio_tx_rds_mono_stereo->cur.val; in vivid_radio_rds_init() 73 rds->art_head = dev->radio_tx_rds_art_head->cur.val; in vivid_radio_rds_init() 74 rds->compressed = dev->radio_tx_rds_compressed->cur.val; in vivid_radio_rds_init() 75 rds->dyn_pty = dev->radio_tx_rds_dyn_pty->cur.val; in vivid_radio_rds_init() 76 rds->ta = dev->radio_tx_rds_ta->cur.val; in vivid_radio_rds_init() 77 rds->tp = dev->radio_tx_rds_tp->cur.val; in vivid_radio_rds_init() 78 rds->ms = dev->radio_tx_rds_ms->cur.val; in vivid_radio_rds_init()
|
/drivers/gpu/drm/vmwgfx/ |
D | vmwgfx_cmdbuf.c | 122 struct vmw_cmdbuf_header *cur; member 663 struct vmw_cmdbuf_header *cur = man->cur; in __vmw_cmdbuf_cur_flush() local 667 if (!cur) in __vmw_cmdbuf_cur_flush() 672 __vmw_cmdbuf_header_free(cur); in __vmw_cmdbuf_cur_flush() 676 man->cur->cb_header->length = man->cur_pos; in __vmw_cmdbuf_cur_flush() 677 vmw_cmdbuf_ctx_add(man, man->cur, SVGA_CB_CONTEXT_0); in __vmw_cmdbuf_cur_flush() 680 man->cur = NULL; in __vmw_cmdbuf_cur_flush() 1004 struct vmw_cmdbuf_header *cur; in vmw_cmdbuf_reserve_cur() local 1010 cur = man->cur; in vmw_cmdbuf_reserve_cur() 1011 if (cur && (size + man->cur_pos > cur->size || in vmw_cmdbuf_reserve_cur() [all …]
|