/drivers/nvdimm/ |
D | badrange.c | 165 static void set_badblock(struct badblocks *bb, sector_t s, int num) in set_badblock() argument 167 dev_dbg(bb->dev, "Found a bad range (0x%llx, 0x%llx)\n", in set_badblock() 170 if (badblocks_set(bb, s, num, 1)) in set_badblock() 171 dev_info_once(bb->dev, "%s: failed for sector %llx\n", in set_badblock() 185 static void __add_badblock_range(struct badblocks *bb, u64 ns_offset, u64 len) in __add_badblock_range() argument 205 set_badblock(bb, s, done); in __add_badblock_range() 210 set_badblock(bb, start_sector, num_sectors); in __add_badblock_range() 214 struct badblocks *bb, const struct range *range) in badblocks_populate() argument 239 __add_badblock_range(bb, start - range->start, len); in badblocks_populate() 253 __add_badblock_range(bb, 0, len); in badblocks_populate() [all …]
|
D | pmem.c | 40 return pmem->bb.dev; in to_dev() 91 badblocks_clear(&pmem->bb, sector, cleared); in pmem_clear_poison() 149 if (unlikely(is_bad_pmem(&pmem->bb, sector, len))) in pmem_do_read() 166 if (unlikely(is_bad_pmem(&pmem->bb, sector, len))) in pmem_do_write() 263 if (unlikely(is_bad_pmem(&pmem->bb, PFN_PHYS(pgoff) / 512, in __pmem_direct_access() 276 if (unlikely(pmem->bb.count)) in __pmem_direct_access() 451 if (devm_init_badblocks(dev, &pmem->bb)) in pmem_attach_disk() 453 nvdimm_badblocks_populate(nd_region, &pmem->bb, &bb_range); in pmem_attach_disk() 454 disk->bb = &pmem->bb; in pmem_attach_disk() 562 struct badblocks *bb; in pmem_revalidate_poison() local [all …]
|
D | claim.c | 270 if (unlikely(is_bad_pmem(&nsio->bb, sector, sz_align))) in nsio_rw_bytes() 277 if (unlikely(is_bad_pmem(&nsio->bb, sector, sz_align))) { in nsio_rw_bytes() 289 badblocks_clear(&nsio->bb, sector, cleared); in nsio_rw_bytes() 321 if (devm_init_badblocks(dev, &nsio->bb)) in devm_nsio_enable() 323 nvdimm_badblocks_populate(to_nd_region(ndns->dev.parent), &nsio->bb, in devm_nsio_enable() 336 devm_exit_badblocks(dev, &nsio->bb); in devm_nsio_disable()
|
D | region.c | 43 if (devm_init_badblocks(dev, &nd_region->bb)) in nd_region_probe() 50 nvdimm_badblocks_populate(nd_region, &nd_region->bb, &range); in nd_region_probe() 132 &nd_region->bb, &range); in nd_region_notify()
|
D | nd.h | 306 struct badblocks bb; member 532 struct badblocks *bb, const struct range *range); 550 static inline bool is_bad_pmem(struct badblocks *bb, sector_t sector, in is_bad_pmem() argument 553 if (bb->count) { in is_bad_pmem() 557 return !!badblocks_check(bb, sector, len / 512, &first_bad, in is_bad_pmem()
|
D | pmem.h | 23 struct badblocks bb; member
|
/drivers/spi/ |
D | spi-gpio.c | 369 struct spi_bitbang *bb; in spi_gpio_probe() local 406 bb = &spi_gpio->bitbang; in spi_gpio_probe() 407 bb->master = master; in spi_gpio_probe() 414 bb->chipselect = spi_gpio_chipselect; in spi_gpio_probe() 415 bb->set_line_direction = spi_gpio_set_direction; in spi_gpio_probe() 418 bb->txrx_word[SPI_MODE_0] = spi_gpio_spec_txrx_word_mode0; in spi_gpio_probe() 419 bb->txrx_word[SPI_MODE_1] = spi_gpio_spec_txrx_word_mode1; in spi_gpio_probe() 420 bb->txrx_word[SPI_MODE_2] = spi_gpio_spec_txrx_word_mode2; in spi_gpio_probe() 421 bb->txrx_word[SPI_MODE_3] = spi_gpio_spec_txrx_word_mode3; in spi_gpio_probe() 423 bb->txrx_word[SPI_MODE_0] = spi_gpio_txrx_word_mode0; in spi_gpio_probe() [all …]
|
/drivers/gpu/drm/i915/gt/ |
D | selftest_ring_submission.c | 207 struct i915_vma *bb; in __live_ctx_switch_wa() local 211 bb = create_wally(engine); in __live_ctx_switch_wa() 212 if (IS_ERR(bb)) in __live_ctx_switch_wa() 213 return PTR_ERR(bb); in __live_ctx_switch_wa() 215 result = i915_gem_object_pin_map_unlocked(bb->obj, I915_MAP_WC); in __live_ctx_switch_wa() 217 intel_context_put(bb->private); in __live_ctx_switch_wa() 218 i915_vma_unpin_and_release(&bb, 0); in __live_ctx_switch_wa() 223 engine->wa_ctx.vma = bb; in __live_ctx_switch_wa()
|
/drivers/gpu/drm/i915/gvt/ |
D | scheduler.c | 524 struct intel_vgpu_shadow_bb *bb; in prepare_shadow_batch_buffer() local 528 list_for_each_entry(bb, &workload->shadow_bb, list) { in prepare_shadow_batch_buffer() 536 if (bb->bb_offset) in prepare_shadow_batch_buffer() 537 bb->bb_start_cmd_va = workload->shadow_ring_buffer_va in prepare_shadow_batch_buffer() 538 + bb->bb_offset; in prepare_shadow_batch_buffer() 549 if (!bb->ppgtt) { in prepare_shadow_batch_buffer() 552 i915_gem_object_lock(bb->obj, &ww); in prepare_shadow_batch_buffer() 554 bb->vma = i915_gem_object_ggtt_pin_ww(bb->obj, &ww, in prepare_shadow_batch_buffer() 556 if (IS_ERR(bb->vma)) { in prepare_shadow_batch_buffer() 557 ret = PTR_ERR(bb->vma); in prepare_shadow_batch_buffer() [all …]
|
D | cmd_parser.c | 1892 struct intel_vgpu_shadow_bb *bb; in perform_bb_shadow() local 1910 bb = kzalloc(sizeof(*bb), GFP_KERNEL); in perform_bb_shadow() 1911 if (!bb) in perform_bb_shadow() 1914 bb->ppgtt = (s->buf_addr_type == GTT_BUFFER) ? false : true; in perform_bb_shadow() 1927 if (bb->ppgtt) in perform_bb_shadow() 1930 bb->obj = i915_gem_object_create_shmem(s->engine->i915, in perform_bb_shadow() 1933 if (IS_ERR(bb->obj)) { in perform_bb_shadow() 1934 ret = PTR_ERR(bb->obj); in perform_bb_shadow() 1938 bb->va = i915_gem_object_pin_map(bb->obj, I915_MAP_WB); in perform_bb_shadow() 1939 if (IS_ERR(bb->va)) { in perform_bb_shadow() [all …]
|
/drivers/pci/ |
D | p2pdma.c | 454 struct pci_dev *a = provider, *b = client, *bb; in calc_map_type_and_dist() local 478 bb = b; in calc_map_type_and_dist() 480 while (bb) { in calc_map_type_and_dist() 481 if (a == bb) in calc_map_type_and_dist() 484 bb = pci_upstream_bridge(bb); in calc_map_type_and_dist() 496 bb = b; in calc_map_type_and_dist() 498 while (bb) { in calc_map_type_and_dist() 499 if (a == bb) in calc_map_type_and_dist() 502 if (pci_bridge_has_acs_redir(bb)) { in calc_map_type_and_dist() 503 seq_buf_print_bus_devfn(&acs_list, bb); in calc_map_type_and_dist() [all …]
|
/drivers/gpu/drm/amd/display/dc/dcn20/ |
D | dcn20_resource.h | 100 struct _vcs_dpi_soc_bounding_box_st *bb, 102 void dcn20_update_bounding_box(struct dc *dc, struct _vcs_dpi_soc_bounding_box_st *bb, 175 struct _vcs_dpi_soc_bounding_box_st *bb); 177 struct _vcs_dpi_soc_bounding_box_st *bb,
|
D | dcn20_resource.c | 3444 struct _vcs_dpi_soc_bounding_box_st *bb, in dcn20_cap_soc_clocks() argument 3450 for (i = 0; i < bb->num_states; i++) { in dcn20_cap_soc_clocks() 3451 if ((bb->clock_limits[i].dcfclk_mhz > (max_clocks.dcfClockInKhz / 1000)) in dcn20_cap_soc_clocks() 3453 bb->clock_limits[i].dcfclk_mhz = (max_clocks.dcfClockInKhz / 1000); in dcn20_cap_soc_clocks() 3455 if ((bb->clock_limits[i].dram_speed_mts > (max_clocks.uClockInKhz / 1000) * 16) in dcn20_cap_soc_clocks() 3457 bb->clock_limits[i].dram_speed_mts = (max_clocks.uClockInKhz / 1000) * 16; in dcn20_cap_soc_clocks() 3459 if ((bb->clock_limits[i].fabricclk_mhz > (max_clocks.fabricClockInKhz / 1000)) in dcn20_cap_soc_clocks() 3461 bb->clock_limits[i].fabricclk_mhz = (max_clocks.fabricClockInKhz / 1000); in dcn20_cap_soc_clocks() 3463 if ((bb->clock_limits[i].dispclk_mhz > (max_clocks.displayClockInKhz / 1000)) in dcn20_cap_soc_clocks() 3465 bb->clock_limits[i].dispclk_mhz = (max_clocks.displayClockInKhz / 1000); in dcn20_cap_soc_clocks() [all …]
|
/drivers/md/ |
D | dm-dust.c | 19 sector_t bb; member 43 if (bblk->bb > blk) in dust_rb_search() 45 else if (bblk->bb < blk) in dust_rb_search() 58 sector_t value = new->bb; in dust_rb_insert() 64 if (bblk->bb > value) in dust_rb_insert() 66 else if (bblk->bb < value) in dust_rb_insert() 119 bblock->bb = block; in dust_add_block() 301 DMEMIT("%llu\n", bblk->bb); in dust_list_badblocks()
|
/drivers/net/wireless/ti/wl1251/ |
D | acx.c | 644 struct acx_beacon_broadcast *bb; in wl1251_acx_bcn_dtim_options() local 649 bb = kzalloc(sizeof(*bb), GFP_KERNEL); in wl1251_acx_bcn_dtim_options() 650 if (!bb) in wl1251_acx_bcn_dtim_options() 653 bb->beacon_rx_timeout = BCN_RX_TIMEOUT_DEF_VALUE; in wl1251_acx_bcn_dtim_options() 654 bb->broadcast_timeout = BROADCAST_RX_TIMEOUT_DEF_VALUE; in wl1251_acx_bcn_dtim_options() 655 bb->rx_broadcast_in_ps = RX_BROADCAST_IN_PS_DEF_VALUE; in wl1251_acx_bcn_dtim_options() 656 bb->ps_poll_threshold = CONSECUTIVE_PS_POLL_FAILURE_DEF; in wl1251_acx_bcn_dtim_options() 658 ret = wl1251_cmd_configure(wl, ACX_BCN_DTIM_OPTIONS, bb, sizeof(*bb)); in wl1251_acx_bcn_dtim_options() 665 kfree(bb); in wl1251_acx_bcn_dtim_options()
|
/drivers/gpu/drm/vmwgfx/ |
D | vmwgfx_stdu.c | 1236 void *cmd, struct drm_rect *bb) in vmw_stdu_bo_populate_update() argument 1248 vmw_stdu_populate_update(&suffix[1], stdu->base.unit, bb->x1, bb->x2, in vmw_stdu_bo_populate_update() 1249 bb->y1, bb->y2); in vmw_stdu_bo_populate_update() 1281 struct drm_rect *bb) in vmw_stdu_bo_populate_update_cpu() argument 1298 width = bb->x2 - bb->x1; in vmw_stdu_bo_populate_update_cpu() 1299 height = bb->y2 - bb->y1; in vmw_stdu_bo_populate_update_cpu() 1305 dst_offset = bb->y1 * dst_pitch + bb->x1 * stdu->cpp; in vmw_stdu_bo_populate_update_cpu() 1510 struct drm_rect *bb) in vmw_stdu_surface_populate_update() argument 1512 vmw_stdu_populate_update(cmd, update->du->unit, bb->x1, bb->x2, bb->y1, in vmw_stdu_surface_populate_update() 1513 bb->y2); in vmw_stdu_surface_populate_update()
|
D | vmwgfx_scrn.c | 525 void *cmd, struct drm_rect *bb) in vmw_stud_bo_post_clip() argument 641 void *cmd, struct drm_rect *bb) in vmw_sou_surface_post_clip() argument 661 src_bb = *bb; in vmw_sou_surface_post_clip() 674 blit->body.destRect.left = bb->x1; in vmw_sou_surface_post_clip() 675 blit->body.destRect.top = bb->y1; in vmw_sou_surface_post_clip() 676 blit->body.destRect.right = bb->x2; in vmw_sou_surface_post_clip() 677 blit->body.destRect.bottom = bb->y2; in vmw_sou_surface_post_clip() 681 rect->left -= bb->x1; in vmw_sou_surface_post_clip() 682 rect->top -= bb->y1; in vmw_sou_surface_post_clip() 683 rect->right -= bb->x1; in vmw_sou_surface_post_clip() [all …]
|
D | vmwgfx_kms.c | 2751 struct drm_rect bb; in vmw_du_helper_plane_update() local 2815 bb.x1 = INT_MAX; in vmw_du_helper_plane_update() 2816 bb.y1 = INT_MAX; in vmw_du_helper_plane_update() 2817 bb.x2 = INT_MIN; in vmw_du_helper_plane_update() 2818 bb.y2 = INT_MIN; in vmw_du_helper_plane_update() 2832 bb.x1 = min_t(int, bb.x1, clip.x1); in vmw_du_helper_plane_update() 2833 bb.y1 = min_t(int, bb.y1, clip.y1); in vmw_du_helper_plane_update() 2834 bb.x2 = max_t(int, bb.x2, clip.x2); in vmw_du_helper_plane_update() 2835 bb.y2 = max_t(int, bb.y2, clip.y2); in vmw_du_helper_plane_update() 2838 curr_size = update->post_clip(update, cmd_next, &bb); in vmw_du_helper_plane_update()
|
/drivers/input/mouse/ |
D | cypress_ps2.h | 10 #define ENCODE_CMD(aa, bb, cc, dd) \ argument 11 (COMPOSIT((aa), 6) | COMPOSIT((bb), 4) | COMPOSIT((cc), 2) | COMPOSIT((dd), 0))
|
/drivers/net/wireless/ti/wlcore/ |
D | acx.c | 565 struct acx_beacon_broadcast *bb; in wl1271_acx_bcn_dtim_options() local 570 bb = kzalloc(sizeof(*bb), GFP_KERNEL); in wl1271_acx_bcn_dtim_options() 571 if (!bb) { in wl1271_acx_bcn_dtim_options() 576 bb->role_id = wlvif->role_id; in wl1271_acx_bcn_dtim_options() 577 bb->beacon_rx_timeout = cpu_to_le16(wl->conf.conn.beacon_rx_timeout); in wl1271_acx_bcn_dtim_options() 578 bb->broadcast_timeout = cpu_to_le16(wl->conf.conn.broadcast_timeout); in wl1271_acx_bcn_dtim_options() 579 bb->rx_broadcast_in_ps = wl->conf.conn.rx_broadcast_in_ps; in wl1271_acx_bcn_dtim_options() 580 bb->ps_poll_threshold = wl->conf.conn.ps_poll_threshold; in wl1271_acx_bcn_dtim_options() 582 ret = wl1271_cmd_configure(wl, ACX_BCN_DTIM_OPTIONS, bb, sizeof(*bb)); in wl1271_acx_bcn_dtim_options() 589 kfree(bb); in wl1271_acx_bcn_dtim_options()
|
/drivers/media/dvb-frontends/ |
D | dib0090.h | 85 extern void dib0090_get_current_gain(struct dvb_frontend *fe, u16 * rf, u16 * bb, u16 * rf_gain_lim… 146 static inline void dib0090_get_current_gain(struct dvb_frontend *fe, u16 * rf, u16 * bb, u16 * rf_g… in dib0090_get_current_gain() argument
|
D | dib0090.c | 918 u16 rf, bb, ref; in dib0090_gain_apply() local 956 bb = gain - rf; in dib0090_gain_apply() 957 if (bb > state->bb_ramp[0]) in dib0090_gain_apply() 958 bb = state->bb_ramp[0]; in dib0090_gain_apply() 961 bb = 0; in dib0090_gain_apply() 965 state->gain[1] = bb; in dib0090_gain_apply() 999 ref = bb; in dib0090_gain_apply() 1003 gain_reg[3] |= ((bb % 10) * 100) / 125; in dib0090_gain_apply() 1006 …(bb) = %3d gain_reg[0]=%04x gain_reg[1]=%04x gain_reg[2]=%04x gain_reg[0]=%04x\n", rf, bb, rf + bb, in dib0090_gain_apply() 1313 void dib0090_get_current_gain(struct dvb_frontend *fe, u16 * rf, u16 * bb, u16 * rf_gain_limit, u16… in dib0090_get_current_gain() argument [all …]
|
/drivers/video/fbdev/omap2/omapfb/dss/ |
D | manager-sysfs.c | 371 info.cpr_coefs.bb); in manager_cpr_coef_show() 388 &coefs.br, &coefs.bg, &coefs.bb) != 9) in manager_cpr_coef_store() 393 coefs.br, coefs.bg, coefs.bb }; in manager_cpr_coef_store()
|
/drivers/net/ethernet/qlogic/qede/ |
D | qede_main.c | 380 struct qede_stats_bb *p_bb = &edev->stats.bb; in qede_fill_by_demand_stats() 383 stats.bb.rx_1519_to_1522_byte_packets; in qede_fill_by_demand_stats() 385 stats.bb.rx_1519_to_2047_byte_packets; in qede_fill_by_demand_stats() 387 stats.bb.rx_2048_to_4095_byte_packets; in qede_fill_by_demand_stats() 389 stats.bb.rx_4096_to_9216_byte_packets; in qede_fill_by_demand_stats() 391 stats.bb.rx_9217_to_16383_byte_packets; in qede_fill_by_demand_stats() 393 stats.bb.tx_1519_to_2047_byte_packets; in qede_fill_by_demand_stats() 395 stats.bb.tx_2048_to_4095_byte_packets; in qede_fill_by_demand_stats() 397 stats.bb.tx_4096_to_9216_byte_packets; in qede_fill_by_demand_stats() 399 stats.bb.tx_9217_to_16383_byte_packets; in qede_fill_by_demand_stats() [all …]
|
/drivers/net/wireless/broadcom/b43/ |
D | lo.c | 905 struct b43_bbatt bb; in b43_lo_g_adjust_to() local 909 memset(&bb, 0, sizeof(bb)); in b43_lo_g_adjust_to() 911 bb.att = bbatt; in b43_lo_g_adjust_to() 913 cal = b43_get_calib_lo_settings(dev, &bb, &rf); in b43_lo_g_adjust_to()
|