/drivers/gpu/drm/amd/display/dc/dcn10/ |
D | dcn10_hw_sequencer.h | 32 struct dc; 34 void dcn10_hw_sequencer_construct(struct dc *dc); 38 struct dc *dc, 42 void dcn10_setup_vupdate_interrupt(struct dc *dc, struct pipe_ctx *pipe_ctx); 46 struct dc *dc); 48 struct dc *dc, 51 struct dc *dc, 54 struct dc *dc, 57 void dcn10_cursor_lock(struct dc *dc, struct pipe_ctx *pipe, bool lock); 59 struct dc *dc, [all …]
|
D | dcn10_hw_sequencer.c | 86 const uint32_t ref_clk_mhz = dc_ctx->dc->res_pool->ref_clocks.dchub_ref_clock_inKhz / 1000; in print_microsec() 95 void dcn10_lock_all_pipes(struct dc *dc, in dcn10_lock_all_pipes() argument 103 for (i = 0; i < dc->res_pool->pipe_count; i++) { in dcn10_lock_all_pipes() 118 dc->hwss.pipe_control_lock(dc, pipe_ctx, true); in dcn10_lock_all_pipes() 120 dc->hwss.pipe_control_lock(dc, pipe_ctx, false); in dcn10_lock_all_pipes() 124 static void log_mpc_crc(struct dc *dc, in log_mpc_crc() argument 127 struct dc_context *dc_ctx = dc->ctx; in log_mpc_crc() 128 struct dce_hwseq *hws = dc->hwseq; in log_mpc_crc() 138 static void dcn10_log_hubbub_state(struct dc *dc, in dcn10_log_hubbub_state() argument 141 struct dc_context *dc_ctx = dc->ctx; in dcn10_log_hubbub_state() [all …]
|
/drivers/dma/ |
D | txx9dmac.c | 24 static struct txx9dmac_cregs __iomem *__dma_regs(const struct txx9dmac_chan *dc) in __dma_regs() argument 26 return dc->ch_regs; in __dma_regs() 30 const struct txx9dmac_chan *dc) in __dma_regs32() argument 32 return dc->ch_regs; in __dma_regs32() 35 #define channel64_readq(dc, name) \ argument 36 __raw_readq(&(__dma_regs(dc)->name)) 37 #define channel64_writeq(dc, name, val) \ argument 38 __raw_writeq((val), &(__dma_regs(dc)->name)) 39 #define channel64_readl(dc, name) \ argument 40 __raw_readl(&(__dma_regs(dc)->name)) [all …]
|
/drivers/gpu/drm/amd/display/dc/inc/ |
D | hw_sequencer.h | 49 void (*hardware_release)(struct dc *dc); 56 void (*init_hw)(struct dc *dc); 57 void (*power_down_on_boot)(struct dc *dc); 58 void (*enable_accelerated_mode)(struct dc *dc, 60 enum dc_status (*apply_ctx_to_hw)(struct dc *dc, 62 void (*disable_plane)(struct dc *dc, struct pipe_ctx *pipe_ctx); 63 void (*disable_pixel_data)(struct dc *dc, struct pipe_ctx *pipe_ctx, bool blank); 64 void (*apply_ctx_for_surface)(struct dc *dc, 67 void (*program_front_end_for_ctx)(struct dc *dc, 69 void (*wait_for_pending_cleared)(struct dc *dc, [all …]
|
D | hw_sequencer_private.h | 75 void (*disable_stream_gating)(struct dc *dc, struct pipe_ctx *pipe_ctx); 76 void (*enable_stream_gating)(struct dc *dc, struct pipe_ctx *pipe_ctx); 77 void (*init_pipes)(struct dc *dc, struct dc_state *context); 78 void (*reset_hw_ctx_wrap)(struct dc *dc, struct dc_state *context); 79 void (*update_plane_addr)(const struct dc *dc, 81 void (*plane_atomic_disconnect)(struct dc *dc, 83 void (*update_mpcc)(struct dc *dc, struct pipe_ctx *pipe_ctx); 84 bool (*set_input_transfer_func)(struct dc *dc, 87 bool (*set_output_transfer_func)(struct dc *dc, 90 void (*power_down)(struct dc *dc); [all …]
|
D | link_enc_cfg.h | 39 const struct dc *dc, 58 struct dc *dc, 78 struct dc *dc, 83 struct dc *dc, 88 struct dc *dc, 93 struct dc *dc, 97 struct link_encoder *link_enc_cfg_get_next_avail_link_enc(struct dc *dc); 101 struct dc *dc, 109 struct dc *dc, 113 bool link_enc_cfg_is_link_enc_avail(struct dc *dc, enum engine_id eng_id, struct dc_link *link); [all …]
|
/drivers/tty/ |
D | nozomi.c | 331 struct nozomi *dc; member 479 static void nozomi_setup_memory(struct nozomi *dc) in nozomi_setup_memory() argument 481 void __iomem *offset = dc->base_addr + dc->config_table.dl_start; in nozomi_setup_memory() 488 dc->port[PORT_MDM].dl_addr[CH_A] = offset; in nozomi_setup_memory() 489 dc->port[PORT_MDM].dl_addr[CH_B] = in nozomi_setup_memory() 490 (offset += dc->config_table.dl_mdm_len1); in nozomi_setup_memory() 491 dc->port[PORT_MDM].dl_size[CH_A] = in nozomi_setup_memory() 492 dc->config_table.dl_mdm_len1 - buff_offset; in nozomi_setup_memory() 493 dc->port[PORT_MDM].dl_size[CH_B] = in nozomi_setup_memory() 494 dc->config_table.dl_mdm_len2 - buff_offset; in nozomi_setup_memory() [all …]
|
/drivers/gpu/drm/amd/display/dc/core/ |
D | dc.c | 82 dc->ctx 85 dc->ctx->logger 146 static void destroy_links(struct dc *dc) in destroy_links() argument 150 for (i = 0; i < dc->link_count; i++) { in destroy_links() 151 if (NULL != dc->links[i]) in destroy_links() 152 link_destroy(&dc->links[i]); in destroy_links() 183 struct dc *dc, in create_links() argument 188 struct dc_bios *bios = dc->ctx->dc_bios; in create_links() 190 dc->link_count = 0; in create_links() 214 …ct table - printing link object info for connector number: %d, link_index: %d", i, dc->link_count); in create_links() [all …]
|
D | dc_stream.c | 35 #define DC_LOGGER dc->ctx->logger 48 if (stream->ctx->dc->caps.dual_link_dvi && in update_stream_signal() 204 if (new_stream->ctx->dc->res_pool->funcs->link_encs_assign) in dc_copy_stream() 247 struct dc *dc = stream->ctx->dc; in dc_stream_get_status() local 248 return dc_stream_get_status_from_state(dc->current_state, stream); in dc_stream_get_status() 252 struct dc *dc, in program_cursor_attributes() argument 263 res_ctx = &dc->current_state->res_ctx; in program_cursor_attributes() 273 dc->hwss.cursor_lock(dc, pipe_to_program, true); in program_cursor_attributes() 275 dc->hwss.cursor_lock(dc, pipe_to_program->next_odm_pipe, true); in program_cursor_attributes() 278 dc->hwss.set_cursor_attribute(pipe_ctx); in program_cursor_attributes() [all …]
|
/drivers/gpu/drm/amd/display/dc/dcn20/ |
D | dcn20_hwseq.h | 36 struct dc *dc, 39 struct dc *dc, 41 void dcn20_update_plane_addr(const struct dc *dc, struct pipe_ctx *pipe_ctx); 42 void dcn20_update_mpcc(struct dc *dc, struct pipe_ctx *pipe_ctx); 43 bool dcn20_set_input_transfer_func(struct dc *dc, struct pipe_ctx *pipe_ctx, 45 bool dcn20_set_output_transfer_func(struct dc *dc, struct pipe_ctx *pipe_ctx, 47 void dcn20_program_output_csc(struct dc *dc, 55 void dcn20_disable_plane(struct dc *dc, struct pipe_ctx *pipe_ctx); 57 struct dc *dc, 61 struct dc *dc, [all …]
|
D | dcn20_resource.h | 35 struct dc; 50 struct dc *dc); 73 bool dcn20_get_dcc_compression_cap(const struct dc *dc, 115 struct dc *dc, 119 bool dcn20_validate_bandwidth(struct dc *dc, struct dc_state *context, bool fast_validate); 121 struct dc *dc, 124 struct dc *dc, 132 bool dcn20_validate_dsc(struct dc *dc, struct dc_state *new_ctx); 139 const struct dc *dc, 143 void dcn20_acquire_dsc(const struct dc *dc, [all …]
|
/drivers/md/bcache/ |
D | writeback.c | 30 static uint64_t __calc_target_rate(struct cached_dev *dc) in __calc_target_rate() argument 32 struct cache_set *c = dc->disk.c; in __calc_target_rate() 48 div64_u64(bdev_nr_sectors(dc->bdev) << WRITEBACK_SHARE_SHIFT, in __calc_target_rate() 52 div_u64(cache_sectors * dc->writeback_percent, 100); in __calc_target_rate() 61 static void __update_writeback_rate(struct cached_dev *dc) in __update_writeback_rate() argument 83 int64_t target = __calc_target_rate(dc); in __update_writeback_rate() 84 int64_t dirty = bcache_dev_sectors_dirty(&dc->disk); in __update_writeback_rate() 87 div_s64(error, dc->writeback_rate_p_term_inverse); in __update_writeback_rate() 101 struct cache_set *c = dc->disk.c; in __update_writeback_rate() 105 if (dc->writeback_consider_fragment && in __update_writeback_rate() [all …]
|
/drivers/md/ |
D | dm-delay.c | 53 struct delay_c *dc = from_timer(dc, t, delay_timer); in handle_delayed_timer() local 55 queue_work(dc->kdelayd_wq, &dc->flush_expired_bios); in handle_delayed_timer() 58 static void queue_timeout(struct delay_c *dc, unsigned long expires) in queue_timeout() argument 60 mutex_lock(&dc->timer_lock); in queue_timeout() 62 if (!timer_pending(&dc->delay_timer) || expires < dc->delay_timer.expires) in queue_timeout() 63 mod_timer(&dc->delay_timer, expires); in queue_timeout() 65 mutex_unlock(&dc->timer_lock); in queue_timeout() 80 static struct bio *flush_delayed_bios(struct delay_c *dc, int flush_all) in flush_delayed_bios() argument 88 list_for_each_entry_safe(delayed, next, &dc->delayed_bios, list) { in flush_delayed_bios() 107 queue_timeout(dc, next_expires); in flush_delayed_bios() [all …]
|
/drivers/gpu/drm/amd/display/dc/dcn31/ |
D | dcn31_hwseq.c | 62 dc->ctx->logger 69 static void enable_memory_low_power(struct dc *dc) in enable_memory_low_power() argument 71 struct dce_hwseq *hws = dc->hwseq; in enable_memory_low_power() 74 if (dc->debug.enable_mem_low_power.bits.dmcu) { in enable_memory_low_power() 76 if (dc->debug.disable_dmcu || dc->config.disable_dmcu) { in enable_memory_low_power() 82 if (dc->debug.enable_mem_low_power.bits.optc) { in enable_memory_low_power() 87 if (dc->debug.enable_mem_low_power.bits.vga) { in enable_memory_low_power() 92 if (dc->debug.enable_mem_low_power.bits.mpc) in enable_memory_low_power() 93 dc->res_pool->mpc->funcs->set_mpc_mem_lp_mode(dc->res_pool->mpc); in enable_memory_low_power() 96 …if (dc->debug.enable_mem_low_power.bits.vpg && dc->res_pool->stream_enc[0]->vpg->funcs->vpg_powerd… in enable_memory_low_power() [all …]
|
/drivers/gpu/drm/amd/display/dc/dcn30/ |
D | dcn30_hwseq.c | 66 dc->ctx->logger 99 struct mpc *mpc = pipe_ctx->stream_res.opp->ctx->dc->res_pool->mpc; in dcn30_set_mpc_shaper_3dlut() 147 bool dcn30_set_input_transfer_func(struct dc *dc, in dcn30_set_input_transfer_func() argument 151 struct dce_hwseq *hws = dc->hwseq; in dcn30_set_input_transfer_func() 190 bool dcn30_set_output_transfer_func(struct dc *dc, in dcn30_set_output_transfer_func() argument 195 struct mpc *mpc = pipe_ctx->stream_res.opp->ctx->dc->res_pool->mpc; in dcn30_set_output_transfer_func() 223 struct dc *dc, in dcn30_set_writeback() argument 233 ASSERT(wb_info->mpcc_inst < dc->res_pool->mpcc_count); in dcn30_set_writeback() 234 mcif_wb = dc->res_pool->mcif_wb[wb_info->dwb_pipe_inst]; in dcn30_set_writeback() 238 dc->res_pool->mpc->funcs->set_dwb_mux(dc->res_pool->mpc, in dcn30_set_writeback() [all …]
|
D | dcn30_hwseq.h | 31 struct dc; 33 void dcn30_init_hw(struct dc *dc); 35 struct dc *dc, 39 struct dc *dc, 43 struct dc *dc, 47 struct dc *dc, 51 struct dc *dc, 58 bool dcn30_set_input_transfer_func(struct dc *dc, 61 bool dcn30_set_output_transfer_func(struct dc *dc, 68 bool dcn30_does_plane_fit_in_mall(struct dc *dc, struct dc_plane_state *plane, [all …]
|
D | dcn30_resource.h | 34 struct dc; 46 struct dc *dc); 49 struct dc *dc, 59 bool dcn30_validate_bandwidth(struct dc *dc, struct dc_state *context, 62 struct dc *dc, 70 struct dc *dc, struct dc_state *context, 74 void dcn30_update_soc_for_wm_a(struct dc *dc, struct dc_state *context); 76 struct dc *dc, struct resource_context *res_ctx, display_e2e_pipe_params_st *pipes); 79 struct dc *dc, struct dc_state *context, 97 struct dc *dc, [all …]
|
/drivers/gpu/drm/amd/display/dc/dml/calcs/ |
D | dcn_calcs.c | 40 dc->ctx->logger 318 if (pipe->plane_res.dpp->ctx->dc->debug.optimized_watermark) { in pipe_ctx_to_e2e_pipe_params() 333 input->src.dcc = pipe->plane_res.dpp->ctx->dc->res_pool->hubbub->funcs-> in pipe_ctx_to_e2e_pipe_params() 453 const struct dc *dc, in dcn_bw_calc_rq_dlg_ttu() argument 458 struct display_mode_lib *dml = (struct display_mode_lib *)(&dc->dml); in dcn_bw_calc_rq_dlg_ttu() 497 input->clks_cfg.refclk_mhz = dc->res_pool->ref_clocks.dchub_ref_clock_inKhz / 1000.0; in dcn_bw_calc_rq_dlg_ttu() 638 static bool dcn_bw_apply_registry_override(struct dc *dc) in dcn_bw_apply_registry_override() argument 642 if ((int)(dc->dcn_soc->sr_exit_time * 1000) != dc->debug.sr_exit_time_ns in dcn_bw_apply_registry_override() 643 && dc->debug.sr_exit_time_ns) { in dcn_bw_apply_registry_override() 645 dc->dcn_soc->sr_exit_time = dc->debug.sr_exit_time_ns / 1000.0; in dcn_bw_apply_registry_override() [all …]
|
/drivers/gpu/drm/amd/display/dc/ |
D | dc_stream.h | 346 bool dc_update_planes_and_stream(struct dc *dc, 361 void dc_commit_updates_for_stream(struct dc *dc, 370 void dc_stream_log(const struct dc *dc, const struct dc_stream_state *stream); 372 uint8_t dc_get_current_stream_count(struct dc *dc); 373 struct dc_stream_state *dc_get_stream_at_index(struct dc *dc, uint8_t i); 398 struct dc *dc, 403 struct dc *dc, 409 const struct dc *dc, 415 const struct dc *dc, 421 const struct dc *dc, [all …]
|
D | dc.h | 363 struct dc; 369 bool (*get_dcc_compression_cap)(const struct dc *dc, 572 unsigned long long start_tick = dc->debug.bw_val_profile.enable ? \ 573 dm_get_timestamp(dc->ctx) : 0 576 if (dc->debug.bw_val_profile.enable) \ 577 dc->debug.bw_val_profile.total_count++ 580 if (dc->debug.bw_val_profile.enable) { \ 582 voltage_level_tick = dm_get_timestamp(dc->ctx); \ 583 dc->debug.bw_val_profile.skip_ ## status ## _count++; \ 587 if (dc->debug.bw_val_profile.enable) \ [all …]
|
/drivers/gpu/drm/tegra/ |
D | dc.c | 49 static u32 tegra_dc_readl_active(struct tegra_dc *dc, unsigned long offset) in tegra_dc_readl_active() argument 53 tegra_dc_writel(dc, READ_MUX, DC_CMD_STATE_ACCESS); in tegra_dc_readl_active() 54 value = tegra_dc_readl(dc, offset); in tegra_dc_readl_active() 55 tegra_dc_writel(dc, 0, DC_CMD_STATE_ACCESS); in tegra_dc_readl_active() 78 dev_WARN(plane->dc->dev, "invalid offset: %x\n", offset); in tegra_plane_offset() 86 return tegra_dc_readl(plane->dc, tegra_plane_offset(plane, offset)); in tegra_plane_readl() 92 tegra_dc_writel(plane->dc, value, tegra_plane_offset(plane, offset)); in tegra_plane_writel() 95 bool tegra_dc_has_output(struct tegra_dc *dc, struct device *dev) in tegra_dc_has_output() argument 97 struct device_node *np = dc->dev->of_node; in tegra_dc_has_output() 120 void tegra_dc_commit(struct tegra_dc *dc) in tegra_dc_commit() argument [all …]
|
/drivers/scsi/esas2r/ |
D | esas2r_disc.c | 291 struct esas2r_disc_context *dc = &a->disc_ctx; in esas2r_disc_queue_event() local 298 dc->disc_evt |= disc_evt; in esas2r_disc_queue_event() 314 struct esas2r_disc_context *dc = &a->disc_ctx; in esas2r_disc_start_port() local 326 if (dc->disc_evt) { in esas2r_disc_start_port() 352 esas2r_trace("disc_evt: %d", dc->disc_evt); in esas2r_disc_start_port() 354 dc->flags = 0; in esas2r_disc_start_port() 357 dc->flags |= DCF_POLLED; in esas2r_disc_start_port() 359 rq->interrupt_cx = dc; in esas2r_disc_start_port() 363 if (dc->disc_evt & DCDE_DEV_SCAN) { in esas2r_disc_start_port() 364 dc->disc_evt &= ~DCDE_DEV_SCAN; in esas2r_disc_start_port() [all …]
|
/drivers/gpu/drm/amd/display/dc/dcn32/ |
D | dcn32_hwseq.h | 31 struct dc; 44 bool dcn32_apply_idle_power_optimizations(struct dc *dc, bool enable); 46 void dcn32_cab_for_ss_control(struct dc *dc, bool enable); 48 void dcn32_commit_subvp_config(struct dc *dc, struct dc_state *context); 53 bool dcn32_set_input_transfer_func(struct dc *dc, 57 bool dcn32_set_output_transfer_func(struct dc *dc, 61 void dcn32_init_hw(struct dc *dc); 63 void dcn32_program_mall_pipe_config(struct dc *dc, struct dc_state *context); 65 void dcn32_update_mall_sel(struct dc *dc, struct dc_state *context); 67 void dcn32_subvp_update_force_pstate(struct dc *dc, struct dc_state *context); [all …]
|
D | dcn32_hwseq.c | 63 dc->ctx->logger 79 if (hws->ctx->dc->debug.disable_dsc_power_gate) in dcn32_dsc_pg_control() 82 if (!hws->ctx->dc->debug.enable_double_buffered_dsc_pg_support) in dcn32_dsc_pg_control() 159 if (hws->ctx->dc->debug.disable_hubp_power_gate) in dcn32_hubp_pg_control() 188 static bool dcn32_check_no_memory_request_for_cab(struct dc *dc) in dcn32_check_no_memory_request_for_cab() argument 193 for (i = 0; i < dc->current_state->stream_count; i++) { in dcn32_check_no_memory_request_for_cab() 194 if (dc->current_state->stream_status[i].plane_count) in dcn32_check_no_memory_request_for_cab() 199 if (i == dc->current_state->stream_count) in dcn32_check_no_memory_request_for_cab() 210 static uint32_t dcn32_calculate_cab_allocation(struct dc *dc, struct dc_state *ctx) in dcn32_calculate_cab_allocation() argument 230 for (i = 0; i < dc->res_pool->pipe_count; i++) { in dcn32_calculate_cab_allocation() [all …]
|
/drivers/gpu/drm/amd/display/dc/dml/ |
D | Makefile | 57 CFLAGS_$(AMDDALPATH)/dc/dml/display_mode_lib.o := $(dml_ccflags) 60 CFLAGS_$(AMDDALPATH)/dc/dml/display_mode_vba.o := $(dml_ccflags) 61 CFLAGS_$(AMDDALPATH)/dc/dml/dcn10/dcn10_fpu.o := $(dml_ccflags) 62 CFLAGS_$(AMDDALPATH)/dc/dml/dcn20/dcn20_fpu.o := $(dml_ccflags) 63 CFLAGS_$(AMDDALPATH)/dc/dml/dcn20/display_mode_vba_20.o := $(dml_ccflags) $(frame_warn_flag) 64 CFLAGS_$(AMDDALPATH)/dc/dml/dcn20/display_rq_dlg_calc_20.o := $(dml_ccflags) 65 CFLAGS_$(AMDDALPATH)/dc/dml/dcn20/display_mode_vba_20v2.o := $(dml_ccflags) $(frame_warn_flag) 66 CFLAGS_$(AMDDALPATH)/dc/dml/dcn20/display_rq_dlg_calc_20v2.o := $(dml_ccflags) 67 CFLAGS_$(AMDDALPATH)/dc/dml/dcn21/display_mode_vba_21.o := $(dml_ccflags) $(frame_warn_flag) 68 CFLAGS_$(AMDDALPATH)/dc/dml/dcn21/display_rq_dlg_calc_21.o := $(dml_ccflags) [all …]
|