Home
last modified time | relevance | path

Searched refs:dc (Results 1 – 25 of 351) sorted by relevance

12345678910>>...15

/drivers/gpu/drm/amd/display/dc/dcn10/
Ddcn10_hw_sequencer.h32 struct dc;
34 void dcn10_hw_sequencer_construct(struct dc *dc);
38 struct dc *dc,
42 void dcn10_setup_vupdate_interrupt(struct dc *dc, struct pipe_ctx *pipe_ctx);
46 struct dc *dc);
48 struct dc *dc,
51 struct dc *dc,
54 struct dc *dc,
57 void dcn10_cursor_lock(struct dc *dc, struct pipe_ctx *pipe, bool lock);
59 struct dc *dc,
[all …]
Ddcn10_hw_sequencer.c84 const uint32_t ref_clk_mhz = dc_ctx->dc->res_pool->ref_clocks.dchub_ref_clock_inKhz / 1000; in print_microsec()
93 void dcn10_lock_all_pipes(struct dc *dc, in dcn10_lock_all_pipes() argument
101 for (i = 0; i < dc->res_pool->pipe_count; i++) { in dcn10_lock_all_pipes()
115 dc->hwss.pipe_control_lock(dc, pipe_ctx, true); in dcn10_lock_all_pipes()
117 dc->hwss.pipe_control_lock(dc, pipe_ctx, false); in dcn10_lock_all_pipes()
121 static void log_mpc_crc(struct dc *dc, in log_mpc_crc() argument
124 struct dc_context *dc_ctx = dc->ctx; in log_mpc_crc()
125 struct dce_hwseq *hws = dc->hwseq; in log_mpc_crc()
135 void dcn10_log_hubbub_state(struct dc *dc, struct dc_log_buffer_ctx *log_ctx) in dcn10_log_hubbub_state() argument
137 struct dc_context *dc_ctx = dc->ctx; in dcn10_log_hubbub_state()
[all …]
/drivers/dma/
Dtxx9dmac.c24 static struct txx9dmac_cregs __iomem *__dma_regs(const struct txx9dmac_chan *dc) in __dma_regs() argument
26 return dc->ch_regs; in __dma_regs()
30 const struct txx9dmac_chan *dc) in __dma_regs32() argument
32 return dc->ch_regs; in __dma_regs32()
35 #define channel64_readq(dc, name) \ argument
36 __raw_readq(&(__dma_regs(dc)->name))
37 #define channel64_writeq(dc, name, val) \ argument
38 __raw_writeq((val), &(__dma_regs(dc)->name))
39 #define channel64_readl(dc, name) \ argument
40 __raw_readl(&(__dma_regs(dc)->name))
[all …]
/drivers/gpu/drm/amd/display/dc/inc/
Dhw_sequencer.h48 void (*hardware_release)(struct dc *dc);
55 void (*init_hw)(struct dc *dc);
56 void (*power_down_on_boot)(struct dc *dc);
57 void (*enable_accelerated_mode)(struct dc *dc,
59 enum dc_status (*apply_ctx_to_hw)(struct dc *dc,
61 void (*disable_plane)(struct dc *dc, struct pipe_ctx *pipe_ctx);
62 void (*apply_ctx_for_surface)(struct dc *dc,
65 void (*program_front_end_for_ctx)(struct dc *dc,
67 void (*wait_for_pending_cleared)(struct dc *dc,
69 void (*post_unlock_program_front_end)(struct dc *dc,
[all …]
Dhw_sequencer_private.h72 void (*disable_stream_gating)(struct dc *dc, struct pipe_ctx *pipe_ctx);
73 void (*enable_stream_gating)(struct dc *dc, struct pipe_ctx *pipe_ctx);
74 void (*init_pipes)(struct dc *dc, struct dc_state *context);
75 void (*reset_hw_ctx_wrap)(struct dc *dc, struct dc_state *context);
76 void (*update_plane_addr)(const struct dc *dc,
78 void (*plane_atomic_disconnect)(struct dc *dc,
80 void (*update_mpcc)(struct dc *dc, struct pipe_ctx *pipe_ctx);
81 bool (*set_input_transfer_func)(struct dc *dc,
84 bool (*set_output_transfer_func)(struct dc *dc,
87 void (*power_down)(struct dc *dc);
[all …]
/drivers/tty/
Dnozomi.c331 struct nozomi *dc; member
479 static void nozomi_setup_memory(struct nozomi *dc) in nozomi_setup_memory() argument
481 void __iomem *offset = dc->base_addr + dc->config_table.dl_start; in nozomi_setup_memory()
488 dc->port[PORT_MDM].dl_addr[CH_A] = offset; in nozomi_setup_memory()
489 dc->port[PORT_MDM].dl_addr[CH_B] = in nozomi_setup_memory()
490 (offset += dc->config_table.dl_mdm_len1); in nozomi_setup_memory()
491 dc->port[PORT_MDM].dl_size[CH_A] = in nozomi_setup_memory()
492 dc->config_table.dl_mdm_len1 - buff_offset; in nozomi_setup_memory()
493 dc->port[PORT_MDM].dl_size[CH_B] = in nozomi_setup_memory()
494 dc->config_table.dl_mdm_len2 - buff_offset; in nozomi_setup_memory()
[all …]
/drivers/gpu/drm/amd/display/dc/dcn20/
Ddcn20_hwseq.h36 struct dc *dc,
39 struct dc *dc,
41 void dcn20_update_plane_addr(const struct dc *dc, struct pipe_ctx *pipe_ctx);
42 void dcn20_update_mpcc(struct dc *dc, struct pipe_ctx *pipe_ctx);
43 bool dcn20_set_input_transfer_func(struct dc *dc, struct pipe_ctx *pipe_ctx,
45 bool dcn20_set_output_transfer_func(struct dc *dc, struct pipe_ctx *pipe_ctx,
47 void dcn20_program_output_csc(struct dc *dc,
55 void dcn20_disable_plane(struct dc *dc, struct pipe_ctx *pipe_ctx);
57 struct dc *dc,
61 struct dc *dc,
[all …]
Ddcn20_resource.h34 struct dc;
43 struct dc *dc);
53 struct dc *dc,
69 bool dcn20_get_dcc_compression_cap(const struct dc *dc,
102 void dcn20_update_bounding_box(struct dc *dc, struct _vcs_dpi_soc_bounding_box_st *bb,
117 struct dc *dc,
121 bool dcn20_validate_bandwidth(struct dc *dc, struct dc_state *context, bool fast_validate);
123 struct dc *dc,
126 struct dc *dc,
134 bool dcn20_validate_dsc(struct dc *dc, struct dc_state *new_ctx);
[all …]
Ddcn20_hwseq.c67 static int find_free_gsl_group(const struct dc *dc) in find_free_gsl_group() argument
69 if (dc->res_pool->gsl_groups.gsl_0 == 0) in find_free_gsl_group()
71 if (dc->res_pool->gsl_groups.gsl_1 == 0) in find_free_gsl_group()
73 if (dc->res_pool->gsl_groups.gsl_2 == 0) in find_free_gsl_group()
95 const struct dc *dc, in dcn20_setup_gsl_group_as_lock() argument
111 group_idx = find_free_gsl_group(dc); in dcn20_setup_gsl_group_as_lock()
119 dc->res_pool->gsl_groups.gsl_0 = 1; in dcn20_setup_gsl_group_as_lock()
123 dc->res_pool->gsl_groups.gsl_1 = 1; in dcn20_setup_gsl_group_as_lock()
127 dc->res_pool->gsl_groups.gsl_2 = 1; in dcn20_setup_gsl_group_as_lock()
145 dc->res_pool->gsl_groups.gsl_0 = 0; in dcn20_setup_gsl_group_as_lock()
[all …]
/drivers/gpu/drm/amd/display/dc/core/
Ddc.c80 dc->ctx
83 dc->ctx->logger
146 static void destroy_links(struct dc *dc) in destroy_links() argument
150 for (i = 0; i < dc->link_count; i++) { in destroy_links()
151 if (NULL != dc->links[i]) in destroy_links()
152 link_destroy(&dc->links[i]); in destroy_links()
183 struct dc *dc, in create_links() argument
188 struct dc_bios *bios = dc->ctx->dc_bios; in create_links()
190 dc->link_count = 0; in create_links()
214 …ct table - printing link object info for connector number: %d, link_index: %d", i, dc->link_count); in create_links()
[all …]
Ddc_stream.c37 #define DC_LOGGER dc->ctx->logger
50 if (stream->ctx->dc->caps.dual_link_dvi && in update_stream_signal()
245 struct dc *dc = stream->ctx->dc; in dc_stream_get_status() local
246 return dc_stream_get_status_from_state(dc->current_state, stream); in dc_stream_get_status()
250 struct dc *dc, in program_cursor_attributes() argument
261 res_ctx = &dc->current_state->res_ctx; in program_cursor_attributes()
271 dc->hwss.cursor_lock(dc, pipe_to_program, true); in program_cursor_attributes()
274 dc->hwss.set_cursor_attribute(pipe_ctx); in program_cursor_attributes()
275 if (dc->hwss.set_cursor_sdr_white_level) in program_cursor_attributes()
276 dc->hwss.set_cursor_sdr_white_level(pipe_ctx); in program_cursor_attributes()
[all …]
/drivers/md/bcache/
Dwriteback.c30 static uint64_t __calc_target_rate(struct cached_dev *dc) in __calc_target_rate() argument
32 struct cache_set *c = dc->disk.c; in __calc_target_rate()
48 div64_u64(bdev_sectors(dc->bdev) << WRITEBACK_SHARE_SHIFT, in __calc_target_rate()
52 div_u64(cache_sectors * dc->writeback_percent, 100); in __calc_target_rate()
61 static void __update_writeback_rate(struct cached_dev *dc) in __update_writeback_rate() argument
83 int64_t target = __calc_target_rate(dc); in __update_writeback_rate()
84 int64_t dirty = bcache_dev_sectors_dirty(&dc->disk); in __update_writeback_rate()
87 div_s64(error, dc->writeback_rate_p_term_inverse); in __update_writeback_rate()
101 struct cache_set *c = dc->disk.c; in __update_writeback_rate()
105 if (dc->writeback_consider_fragment && in __update_writeback_rate()
[all …]
/drivers/md/
Ddm-delay.c53 struct delay_c *dc = from_timer(dc, t, delay_timer); in handle_delayed_timer() local
55 queue_work(dc->kdelayd_wq, &dc->flush_expired_bios); in handle_delayed_timer()
58 static void queue_timeout(struct delay_c *dc, unsigned long expires) in queue_timeout() argument
60 mutex_lock(&dc->timer_lock); in queue_timeout()
62 if (!timer_pending(&dc->delay_timer) || expires < dc->delay_timer.expires) in queue_timeout()
63 mod_timer(&dc->delay_timer, expires); in queue_timeout()
65 mutex_unlock(&dc->timer_lock); in queue_timeout()
80 static struct bio *flush_delayed_bios(struct delay_c *dc, int flush_all) in flush_delayed_bios() argument
88 list_for_each_entry_safe(delayed, next, &dc->delayed_bios, list) { in flush_delayed_bios()
107 queue_timeout(dc, next_expires); in flush_delayed_bios()
[all …]
/drivers/gpu/drm/amd/display/dc/dcn31/
Ddcn31_hwseq.c59 dc->ctx->logger
66 void dcn31_init_hw(struct dc *dc) in dcn31_init_hw() argument
68 struct abm **abms = dc->res_pool->multiple_abms; in dcn31_init_hw()
69 struct dce_hwseq *hws = dc->hwseq; in dcn31_init_hw()
70 struct dc_bios *dcb = dc->ctx->dc_bios; in dcn31_init_hw()
71 struct resource_pool *res_pool = dc->res_pool; in dcn31_init_hw()
76 if (dc->clk_mgr && dc->clk_mgr->funcs->init_clocks) in dcn31_init_hw()
77 dc->clk_mgr->funcs->init_clocks(dc->clk_mgr); in dcn31_init_hw()
79 if (IS_FPGA_MAXIMUS_DC(dc->ctx->dce_environment)) { in dcn31_init_hw()
85 if (!dc->debug.disable_clock_gate) { in dcn31_init_hw()
[all …]
/drivers/gpu/drm/amd/display/dc/
Ddc_stream.h304 bool dc_update_planes_and_stream(struct dc *dc,
319 void dc_commit_updates_for_stream(struct dc *dc,
328 void dc_stream_log(const struct dc *dc, const struct dc_stream_state *stream);
330 uint8_t dc_get_current_stream_count(struct dc *dc);
331 struct dc_stream_state *dc_get_stream_at_index(struct dc *dc, uint8_t i);
357 struct dc *dc,
362 struct dc *dc,
368 const struct dc *dc,
374 const struct dc *dc,
380 const struct dc *dc,
[all …]
Ddc.h285 struct dc;
291 bool (*get_dcc_compression_cap)(const struct dc *dc,
423 unsigned long long start_tick = dc->debug.bw_val_profile.enable ? \
424 dm_get_timestamp(dc->ctx) : 0
427 if (dc->debug.bw_val_profile.enable) \
428 dc->debug.bw_val_profile.total_count++
431 if (dc->debug.bw_val_profile.enable) { \
433 voltage_level_tick = dm_get_timestamp(dc->ctx); \
434 dc->debug.bw_val_profile.skip_ ## status ## _count++; \
438 if (dc->debug.bw_val_profile.enable) \
[all …]
/drivers/gpu/drm/amd/display/dc/calcs/
Ddcn_calcs.c40 dc->ctx->logger
318 if (pipe->plane_res.dpp->ctx->dc->debug.optimized_watermark) { in pipe_ctx_to_e2e_pipe_params()
333 input->src.dcc = pipe->plane_res.dpp->ctx->dc->res_pool->hubbub->funcs-> in pipe_ctx_to_e2e_pipe_params()
453 const struct dc *dc, in dcn_bw_calc_rq_dlg_ttu() argument
458 struct display_mode_lib *dml = (struct display_mode_lib *)(&dc->dml); in dcn_bw_calc_rq_dlg_ttu()
494 input.clks_cfg.refclk_mhz = dc->res_pool->ref_clocks.dchub_ref_clock_inKhz / 1000.0; in dcn_bw_calc_rq_dlg_ttu()
503 dlg_sys_param.t_srx_delay_us = dc->dcn_ip->dcfclk_cstate_latency / v->dcf_clk_deep_sleep; in dcn_bw_calc_rq_dlg_ttu()
636 static bool dcn_bw_apply_registry_override(struct dc *dc) in dcn_bw_apply_registry_override() argument
641 if ((int)(dc->dcn_soc->sr_exit_time * 1000) != dc->debug.sr_exit_time_ns in dcn_bw_apply_registry_override()
642 && dc->debug.sr_exit_time_ns) { in dcn_bw_apply_registry_override()
[all …]
/drivers/gpu/drm/amd/display/dc/dcn30/
Ddcn30_hwseq.c64 dc->ctx->logger
97 struct mpc *mpc = pipe_ctx->stream_res.opp->ctx->dc->res_pool->mpc; in dcn30_set_mpc_shaper_3dlut()
144 bool dcn30_set_input_transfer_func(struct dc *dc, in dcn30_set_input_transfer_func() argument
148 struct dce_hwseq *hws = dc->hwseq; in dcn30_set_input_transfer_func()
187 bool dcn30_set_output_transfer_func(struct dc *dc, in dcn30_set_output_transfer_func() argument
192 struct mpc *mpc = pipe_ctx->stream_res.opp->ctx->dc->res_pool->mpc; in dcn30_set_output_transfer_func()
220 struct dc *dc, in dcn30_set_writeback() argument
230 ASSERT(wb_info->mpcc_inst < dc->res_pool->mpcc_count); in dcn30_set_writeback()
231 mcif_wb = dc->res_pool->mcif_wb[wb_info->dwb_pipe_inst]; in dcn30_set_writeback()
235 dc->res_pool->mpc->funcs->set_dwb_mux(dc->res_pool->mpc, in dcn30_set_writeback()
[all …]
Ddcn30_hwseq.h31 struct dc;
33 void dcn30_init_hw(struct dc *dc);
35 struct dc *dc,
39 struct dc *dc,
43 struct dc *dc,
47 struct dc *dc,
51 struct dc *dc,
58 bool dcn30_set_input_transfer_func(struct dc *dc,
61 bool dcn30_set_output_transfer_func(struct dc *dc,
68 bool dcn30_does_plane_fit_in_mall(struct dc *dc, struct dc_plane_state *plane,
[all …]
Ddcn30_resource.h34 struct dc;
43 struct dc *dc);
46 struct dc *dc,
56 bool dcn30_validate_bandwidth(struct dc *dc, struct dc_state *context,
59 struct dc *dc,
66 struct dc *dc, struct dc_state *context,
70 void dcn30_update_soc_for_wm_a(struct dc *dc, struct dc_state *context);
72 struct dc *dc, struct resource_context *res_ctx, display_e2e_pipe_params_st *pipes);
75 struct dc *dc, struct dc_state *context,
93 struct dc *dc,
[all …]
/drivers/scsi/esas2r/
Desas2r_disc.c291 struct esas2r_disc_context *dc = &a->disc_ctx; in esas2r_disc_queue_event() local
298 dc->disc_evt |= disc_evt; in esas2r_disc_queue_event()
314 struct esas2r_disc_context *dc = &a->disc_ctx; in esas2r_disc_start_port() local
326 if (dc->disc_evt) { in esas2r_disc_start_port()
352 esas2r_trace("disc_evt: %d", dc->disc_evt); in esas2r_disc_start_port()
354 dc->flags = 0; in esas2r_disc_start_port()
357 dc->flags |= DCF_POLLED; in esas2r_disc_start_port()
359 rq->interrupt_cx = dc; in esas2r_disc_start_port()
363 if (dc->disc_evt & DCDE_DEV_SCAN) { in esas2r_disc_start_port()
364 dc->disc_evt &= ~DCDE_DEV_SCAN; in esas2r_disc_start_port()
[all …]
/drivers/gpu/drm/tegra/
Ddc.c44 static u32 tegra_dc_readl_active(struct tegra_dc *dc, unsigned long offset) in tegra_dc_readl_active() argument
48 tegra_dc_writel(dc, READ_MUX, DC_CMD_STATE_ACCESS); in tegra_dc_readl_active()
49 value = tegra_dc_readl(dc, offset); in tegra_dc_readl_active()
50 tegra_dc_writel(dc, 0, DC_CMD_STATE_ACCESS); in tegra_dc_readl_active()
73 dev_WARN(plane->dc->dev, "invalid offset: %x\n", offset); in tegra_plane_offset()
81 return tegra_dc_readl(plane->dc, tegra_plane_offset(plane, offset)); in tegra_plane_readl()
87 tegra_dc_writel(plane->dc, value, tegra_plane_offset(plane, offset)); in tegra_plane_writel()
90 bool tegra_dc_has_output(struct tegra_dc *dc, struct device *dev) in tegra_dc_has_output() argument
92 struct device_node *np = dc->dev->of_node; in tegra_dc_has_output()
115 void tegra_dc_commit(struct tegra_dc *dc) in tegra_dc_commit() argument
[all …]
Drgb.c18 struct tegra_dc *dc; member
78 static void tegra_dc_write_regs(struct tegra_dc *dc, in tegra_dc_write_regs() argument
85 tegra_dc_writel(dc, table[i].value, table[i].offset); in tegra_dc_write_regs()
93 tegra_dc_write_regs(rgb->dc, rgb_disable, ARRAY_SIZE(rgb_disable)); in tegra_rgb_encoder_disable()
94 tegra_dc_commit(rgb->dc); in tegra_rgb_encoder_disable()
103 tegra_dc_write_regs(rgb->dc, rgb_enable, ARRAY_SIZE(rgb_enable)); in tegra_rgb_encoder_enable()
106 tegra_dc_writel(rgb->dc, value, DC_DISP_DATA_ENABLE_OPTIONS); in tegra_rgb_encoder_enable()
109 value = tegra_dc_readl(rgb->dc, DC_COM_PIN_OUTPUT_POLARITY(1)); in tegra_rgb_encoder_enable()
112 tegra_dc_writel(rgb->dc, value, DC_COM_PIN_OUTPUT_POLARITY(1)); in tegra_rgb_encoder_enable()
117 tegra_dc_writel(rgb->dc, value, DC_DISP_DISP_INTERFACE_CONTROL); in tegra_rgb_encoder_enable()
[all …]
/drivers/gpu/drm/amd/display/dc/dml/
DMakefile57 CFLAGS_$(AMDDALPATH)/dc/dml/display_mode_lib.o := $(dml_ccflags)
60 CFLAGS_$(AMDDALPATH)/dc/dml/display_mode_vba.o := $(dml_ccflags)
61 CFLAGS_$(AMDDALPATH)/dc/dml/dcn2x/dcn2x.o := $(dml_ccflags)
62 CFLAGS_$(AMDDALPATH)/dc/dml/dcn20/display_mode_vba_20.o := $(dml_ccflags)
63 CFLAGS_$(AMDDALPATH)/dc/dml/dcn20/display_mode_vba_20.o := $(dml_ccflags)
64 CFLAGS_$(AMDDALPATH)/dc/dml/dcn20/display_rq_dlg_calc_20.o := $(dml_ccflags)
65 CFLAGS_$(AMDDALPATH)/dc/dml/dcn20/display_mode_vba_20v2.o := $(dml_ccflags)
66 CFLAGS_$(AMDDALPATH)/dc/dml/dcn20/display_rq_dlg_calc_20v2.o := $(dml_ccflags)
67 CFLAGS_$(AMDDALPATH)/dc/dml/dcn21/display_mode_vba_21.o := $(dml_ccflags)
68 CFLAGS_$(AMDDALPATH)/dc/dml/dcn21/display_rq_dlg_calc_21.o := $(dml_ccflags)
[all …]
/drivers/clk/mvebu/
Ddove-divider.c51 static unsigned int dove_get_divider(struct dove_clk *dc) in dove_get_divider() argument
56 val = readl_relaxed(dc->base + DIV_CTRL0); in dove_get_divider()
57 val >>= dc->div_bit_start; in dove_get_divider()
59 divider = val & ~(~0 << dc->div_bit_size); in dove_get_divider()
61 if (dc->divider_table) in dove_get_divider()
62 divider = dc->divider_table[divider]; in dove_get_divider()
67 static int dove_calc_divider(const struct dove_clk *dc, unsigned long rate, in dove_calc_divider() argument
74 if (dc->divider_table) { in dove_calc_divider()
77 for (i = 0; dc->divider_table[i]; i++) in dove_calc_divider()
78 if (divider == dc->divider_table[i]) { in dove_calc_divider()
[all …]

12345678910>>...15