/drivers/scsi/device_handler/ |
D | scsi_dh_alua.c | 85 struct alua_port_group __rcu *pg; member 103 static bool alua_rtpg_queue(struct alua_port_group *pg, 110 struct alua_port_group *pg; in release_port_group() local 112 pg = container_of(kref, struct alua_port_group, kref); in release_port_group() 113 if (pg->rtpg_sdev) in release_port_group() 114 flush_delayed_work(&pg->rtpg_work); in release_port_group() 116 list_del(&pg->node); in release_port_group() 118 kfree_rcu(pg, rcu); in release_port_group() 181 struct alua_port_group *pg; in alua_find_get_pg() local 186 list_for_each_entry(pg, &port_group_list, node) { in alua_find_get_pg() [all …]
|
/drivers/md/ |
D | dm-mpath.c | 41 struct priority_group *pg; /* Owning PG */ member 168 struct priority_group *pg; in alloc_priority_group() local 170 pg = kzalloc(sizeof(*pg), GFP_KERNEL); in alloc_priority_group() 172 if (pg) in alloc_priority_group() 173 INIT_LIST_HEAD(&pg->pgpaths); in alloc_priority_group() 175 return pg; in alloc_priority_group() 189 static void free_priority_group(struct priority_group *pg, in free_priority_group() argument 192 struct path_selector *ps = &pg->ps; in free_priority_group() 199 free_pgpaths(&pg->pgpaths, ti); in free_priority_group() 200 kfree(pg); in free_priority_group() [all …]
|
/drivers/xen/ |
D | unpopulated-alloc.c | 86 struct page *pg = virt_to_page(vaddr + PAGE_SIZE * i); in fill_list() local 89 pg->zone_device_data = page_list; in fill_list() 90 page_list = pg; in fill_list() 124 struct page *pg = page_list; in xen_alloc_unpopulated_pages() local 126 BUG_ON(!pg); in xen_alloc_unpopulated_pages() 127 page_list = pg->zone_device_data; in xen_alloc_unpopulated_pages() 129 pages[i] = pg; in xen_alloc_unpopulated_pages() 133 ret = xen_alloc_p2m_entry(page_to_pfn(pg)); in xen_alloc_unpopulated_pages() 192 struct page *pg = in init() local 195 pg->zone_device_data = page_list; in init() [all …]
|
D | xen-scsiback.c | 381 struct page **pg, grant_handle_t *grant, int cnt) in scsiback_gnttab_data_map_batch() argument 388 err = gnttab_map_refs(map, NULL, pg, cnt); in scsiback_gnttab_data_map_batch() 396 get_page(pg[i]); in scsiback_gnttab_data_map_batch() 404 struct scsiif_request_segment *seg, struct page **pg, in scsiback_gnttab_data_map_list() argument 412 if (gnttab_page_cache_get(&info->free_pages, pg + mapcount)) { in scsiback_gnttab_data_map_list() 413 gnttab_page_cache_put(&info->free_pages, pg, mapcount); in scsiback_gnttab_data_map_list() 417 gnttab_set_map_op(&map[mapcount], vaddr_page(pg[mapcount]), in scsiback_gnttab_data_map_list() 422 err = scsiback_gnttab_data_map_batch(map, pg, grant, mapcount); in scsiback_gnttab_data_map_list() 423 pg += mapcount; in scsiback_gnttab_data_map_list() 430 err = scsiback_gnttab_data_map_batch(map, pg, grant, mapcount); in scsiback_gnttab_data_map_list() [all …]
|
/drivers/soc/tegra/ |
D | pmc.c | 644 static void tegra_powergate_disable_clocks(struct tegra_powergate *pg) in tegra_powergate_disable_clocks() argument 648 for (i = 0; i < pg->num_clks; i++) in tegra_powergate_disable_clocks() 649 clk_disable_unprepare(pg->clks[i]); in tegra_powergate_disable_clocks() 652 static int tegra_powergate_enable_clocks(struct tegra_powergate *pg) in tegra_powergate_enable_clocks() argument 657 for (i = 0; i < pg->num_clks; i++) { in tegra_powergate_enable_clocks() 658 err = clk_prepare_enable(pg->clks[i]); in tegra_powergate_enable_clocks() 667 clk_disable_unprepare(pg->clks[i]); in tegra_powergate_enable_clocks() 677 static int tegra_powergate_power_up(struct tegra_powergate *pg, in tegra_powergate_power_up() argument 682 err = reset_control_assert(pg->reset); in tegra_powergate_power_up() 688 err = tegra_powergate_set(pg->pmc, pg->id, true); in tegra_powergate_power_up() [all …]
|
/drivers/block/paride/ |
D | pg.c | 208 struct pg { struct 223 static struct pg devices[PG_UNITS]; argument 225 static int pg_identify(struct pg *dev, int log); 250 struct pg *dev = &devices[unit]; in pg_init_units() 263 static inline int status_reg(struct pg *dev) in status_reg() 268 static inline int read_reg(struct pg *dev, int reg) in read_reg() 273 static inline void write_reg(struct pg *dev, int reg, int val) in write_reg() 278 static inline u8 DRIVE(struct pg *dev) in DRIVE() 288 static int pg_wait(struct pg *dev, int go, int stop, unsigned long tmo, char *msg) in pg_wait() 320 static int pg_command(struct pg *dev, char *cmd, int dlen, unsigned long tmo) in pg_command() [all …]
|
/drivers/gpu/drm/gma500/ |
D | gtt.c | 413 struct psb_gtt *pg; in psb_gtt_init() local 424 pg = &dev_priv->gtt; in psb_gtt_init() 438 pg->gtt_phys_start = dev_priv->pge_ctl & PAGE_MASK; in psb_gtt_init() 446 pg->mmu_gatt_start = 0xE0000000; in psb_gtt_init() 448 pg->gtt_start = pci_resource_start(dev->pdev, PSB_GTT_RESOURCE); in psb_gtt_init() 452 if (pg->gtt_start == 0 || gtt_pages == 0) { in psb_gtt_init() 455 pg->gtt_start = dev_priv->pge_ctl; in psb_gtt_init() 458 pg->gatt_start = pci_resource_start(dev->pdev, PSB_GATT_RESOURCE); in psb_gtt_init() 459 pg->gatt_pages = pci_resource_len(dev->pdev, PSB_GATT_RESOURCE) in psb_gtt_init() 463 if (pg->gatt_pages == 0 || pg->gatt_start == 0) { in psb_gtt_init() [all …]
|
D | psb_drv.c | 110 struct psb_gtt *pg = &dev_priv->gtt; in psb_do_init() local 114 if (pg->mmu_gatt_start & 0x0FFFFFFF) { in psb_do_init() 119 stolen_gtt = (pg->stolen_size >> PAGE_SHIFT) * 4; in psb_do_init() 121 stolen_gtt = (stolen_gtt < pg->gtt_pages) ? stolen_gtt : pg->gtt_pages; in psb_do_init() 123 dev_priv->gatt_free_offset = pg->mmu_gatt_start + in psb_do_init() 141 PSB_WSGX32(pg->gatt_start, PSB_CR_BIF_TWOD_REQ_BASE); in psb_do_init() 168 struct psb_gtt *pg = &dev_priv->gtt; in psb_driver_unload() local 170 down_read(&pg->sem); in psb_driver_unload() 174 pg->mmu_gatt_start, in psb_driver_unload() 176 up_read(&pg->sem); in psb_driver_unload() [all …]
|
/drivers/gpio/ |
D | gpio-palmas.c | 29 struct palmas_gpio *pg = gpiochip_get_data(gc); in palmas_gpio_get() local 30 struct palmas *palmas = pg->palmas; in palmas_gpio_get() 61 struct palmas_gpio *pg = gpiochip_get_data(gc); in palmas_gpio_set() local 62 struct palmas *palmas = pg->palmas; in palmas_gpio_set() 83 struct palmas_gpio *pg = gpiochip_get_data(gc); in palmas_gpio_output() local 84 struct palmas *palmas = pg->palmas; in palmas_gpio_output() 105 struct palmas_gpio *pg = gpiochip_get_data(gc); in palmas_gpio_input() local 106 struct palmas *palmas = pg->palmas; in palmas_gpio_input() 123 struct palmas_gpio *pg = gpiochip_get_data(gc); in palmas_gpio_to_irq() local 124 struct palmas *palmas = pg->palmas; in palmas_gpio_to_irq()
|
/drivers/net/wireless/marvell/mwifiex/ |
D | sta_ioctl.c | 678 struct mwifiex_power_group *pg; in mwifiex_set_tx_power() local 710 pg = (struct mwifiex_power_group *) in mwifiex_set_tx_power() 714 pg->first_rate_code = 0x00; in mwifiex_set_tx_power() 715 pg->last_rate_code = 0x03; in mwifiex_set_tx_power() 716 pg->modulation_class = MOD_CLASS_HR_DSSS; in mwifiex_set_tx_power() 717 pg->power_step = 0; in mwifiex_set_tx_power() 718 pg->power_min = (s8) dbm_min; in mwifiex_set_tx_power() 719 pg->power_max = (s8) dbm; in mwifiex_set_tx_power() 720 pg++; in mwifiex_set_tx_power() 722 pg->first_rate_code = 0x00; in mwifiex_set_tx_power() [all …]
|
D | sta_cmdresp.c | 339 struct mwifiex_power_group *pg; in mwifiex_get_power_level() local 345 pg = (struct mwifiex_power_group *) in mwifiex_get_power_level() 353 max_power = pg->power_max; in mwifiex_get_power_level() 354 min_power = pg->power_min; in mwifiex_get_power_level() 358 pg++; in mwifiex_get_power_level() 359 if (max_power < pg->power_max) in mwifiex_get_power_level() 360 max_power = pg->power_max; in mwifiex_get_power_level() 362 if (min_power > pg->power_min) in mwifiex_get_power_level() 363 min_power = pg->power_min; in mwifiex_get_power_level() 386 struct mwifiex_power_group *pg; in mwifiex_ret_tx_power_cfg() local [all …]
|
/drivers/net/ethernet/huawei/hinic/ |
D | hinic_hw_eqs.c | 634 int err, pg; in alloc_eq_pages() local 648 for (pg = 0; pg < eq->num_pages; pg++) { in alloc_eq_pages() 649 eq->virt_addr[pg] = dma_alloc_coherent(&pdev->dev, in alloc_eq_pages() 651 &eq->dma_addr[pg], in alloc_eq_pages() 653 if (!eq->virt_addr[pg]) { in alloc_eq_pages() 658 addr = EQ_HI_PHYS_ADDR_REG(eq, pg); in alloc_eq_pages() 659 val = upper_32_bits(eq->dma_addr[pg]); in alloc_eq_pages() 663 addr = EQ_LO_PHYS_ADDR_REG(eq, pg); in alloc_eq_pages() 664 val = lower_32_bits(eq->dma_addr[pg]); in alloc_eq_pages() 679 while (--pg >= 0) in alloc_eq_pages() [all …]
|
/drivers/mmc/host/ |
D | usdhi6rol0.c | 167 struct usdhi6_page pg; /* current page from an SG */ member 324 host->head_pg.page = host->pg.page; in usdhi6_blk_bounce() 325 host->head_pg.mapped = host->pg.mapped; in usdhi6_blk_bounce() 326 host->pg.page = nth_page(host->pg.page, 1); in usdhi6_blk_bounce() 327 host->pg.mapped = kmap(host->pg.page); in usdhi6_blk_bounce() 337 memcpy(host->bounce_buf + blk_head, host->pg.mapped, in usdhi6_blk_bounce() 362 WARN(host->pg.page, "%p not properly unmapped!\n", host->pg.page); in usdhi6_sg_map() 368 host->pg.page = sg_page(sg); in usdhi6_sg_map() 369 host->pg.mapped = kmap(host->pg.page); in usdhi6_sg_map() 385 host->blk_page = host->pg.mapped; in usdhi6_sg_map() [all …]
|
D | tifm_sd.c | 112 static void tifm_sd_read_fifo(struct tifm_sd *host, struct page *pg, in tifm_sd_read_fifo() argument 119 buf = kmap_atomic(pg) + off; in tifm_sd_read_fifo() 138 static void tifm_sd_write_fifo(struct tifm_sd *host, struct page *pg, in tifm_sd_write_fifo() argument 145 buf = kmap_atomic(pg) + off; in tifm_sd_write_fifo() 171 struct page *pg; in tifm_sd_transfer_data() local 193 pg = nth_page(sg_page(&sg[host->sg_pos]), off >> PAGE_SHIFT); in tifm_sd_transfer_data() 200 tifm_sd_read_fifo(host, pg, p_off, p_cnt); in tifm_sd_transfer_data() 202 tifm_sd_write_fifo(host, pg, p_off, p_cnt); in tifm_sd_transfer_data() 228 struct page *pg; in tifm_sd_bounce_block() local 242 pg = nth_page(sg_page(&sg[host->sg_pos]), off >> PAGE_SHIFT); in tifm_sd_bounce_block() [all …]
|
/drivers/net/ethernet/broadcom/bnx2x/ |
D | bnx2x_dcb.c | 1127 data[i].pg = DCBX_ILLEGAL_PG; in bnx2x_dcbx_get_num_pg_traf_type() 1137 if (data[traf_type].pg == add_pg) { in bnx2x_dcbx_get_num_pg_traf_type() 1149 data[help_data->num_of_pg].pg = add_pg; in bnx2x_dcbx_get_num_pg_traf_type() 1249 if (pg_help_data->data[0].pg < DCBX_MAX_NUM_PG_BW_ENTRIES) { in bnx2x_dcbx_2cos_limit_cee_single_pg_to_cos_params() 1376 u8 pg[DCBX_COS_MAX_NUM_E2] = { 0 }; in bnx2x_dcbx_2cos_limit_cee_two_pg_to_cos_params() local 1405 pg[0] = pg_help_data->data[0].pg; in bnx2x_dcbx_2cos_limit_cee_two_pg_to_cos_params() 1408 pg[1] = pg_help_data->data[1].pg; in bnx2x_dcbx_2cos_limit_cee_two_pg_to_cos_params() 1412 pg[0] = pg_help_data->data[1].pg; in bnx2x_dcbx_2cos_limit_cee_two_pg_to_cos_params() 1415 pg[1] = pg_help_data->data[0].pg; in bnx2x_dcbx_2cos_limit_cee_two_pg_to_cos_params() 1425 pg[0] = pg_help_data->data[0].pg; in bnx2x_dcbx_2cos_limit_cee_two_pg_to_cos_params() [all …]
|
/drivers/hv/ |
D | hv_balloon.c | 670 static void hv_page_online_one(struct hv_hotadd_state *has, struct page *pg) in hv_page_online_one() argument 672 if (!has_pfn_is_backed(has, page_to_pfn(pg))) { in hv_page_online_one() 673 if (!PageOffline(pg)) in hv_page_online_one() 674 __SetPageOffline(pg); in hv_page_online_one() 677 if (PageOffline(pg)) in hv_page_online_one() 678 __ClearPageOffline(pg); in hv_page_online_one() 681 generic_online_page(pg, 0); in hv_page_online_one() 763 static void hv_online_page(struct page *pg, unsigned int order) in hv_online_page() argument 767 unsigned long pfn = page_to_pfn(pg); in hv_online_page() 1193 struct page *pg; in free_balloon_pages() local [all …]
|
/drivers/target/ |
D | target_core_rd.c | 66 struct page *pg; in rd_release_sgl_table() local 75 pg = sg_page(&sg[j]); in rd_release_sgl_table() 76 if (pg) { in rd_release_sgl_table() 77 __free_page(pg); in rd_release_sgl_table() 118 struct page *pg; in rd_allocate_sgl_table() local 153 pg = alloc_pages(GFP_KERNEL, 0); in rd_allocate_sgl_table() 154 if (!pg) { in rd_allocate_sgl_table() 159 sg_assign_page(&sg[j], pg); in rd_allocate_sgl_table() 162 p = kmap(pg); in rd_allocate_sgl_table() 164 kunmap(pg); in rd_allocate_sgl_table()
|
/drivers/power/supply/ |
D | bq24257_charger.c | 92 struct gpio_desc *pg; member 438 if (bq->pg) in bq24257_get_chip_state() 439 state->power_good = !gpiod_get_value_cansleep(bq->pg); in bq24257_get_chip_state() 866 bq->pg = devm_gpiod_get_optional(bq->dev, BQ24257_PG_GPIO, GPIOD_IN); in bq24257_pg_gpio_probe() 868 if (PTR_ERR(bq->pg) == -EPROBE_DEFER) { in bq24257_pg_gpio_probe() 871 } else if (IS_ERR(bq->pg)) { in bq24257_pg_gpio_probe() 873 bq->pg = NULL; in bq24257_pg_gpio_probe() 877 if (bq->pg) in bq24257_pg_gpio_probe() 878 dev_dbg(bq->dev, "probed PG pin = %d\n", desc_to_gpio(bq->pg)); in bq24257_pg_gpio_probe() 1036 if (PTR_ERR(bq->pg) == -EPROBE_DEFER) in bq24257_probe() [all …]
|
/drivers/dma-buf/heaps/ |
D | cma_heap.c | 287 pgoff_t pg; in cma_heap_allocate() local 333 for (pg = 0; pg < pagecount; pg++) in cma_heap_allocate() 334 buffer->pages[pg] = &cma_pages[pg]; in cma_heap_allocate()
|
/drivers/mtd/tests/ |
D | readtest.c | 83 int pg, oob; in dump_eraseblock() local 100 for (pg = 0, i = 0; pg < pgcnt; pg++) in dump_eraseblock()
|
/drivers/char/agp/ |
D | i460-agp.c | 403 int i, start_offset, end_offset, idx, pg, num_entries; in i460_insert_memory_large_io_page() local 443 pg = lp - i460.lp_desc; in i460_insert_memory_large_io_page() 444 WR_GATT(pg, i460_mask_memory(agp_bridge, in i460_insert_memory_large_io_page() 446 WR_FLUSH_GATT(pg); in i460_insert_memory_large_io_page() 464 int i, pg, start_offset, end_offset, idx, num_entries; in i460_remove_memory_large_io_page() local 489 pg = lp - i460.lp_desc; in i460_remove_memory_large_io_page() 490 WR_GATT(pg, 0); in i460_remove_memory_large_io_page() 491 WR_FLUSH_GATT(pg); in i460_remove_memory_large_io_page()
|
/drivers/pinctrl/ |
D | pinctrl-oxnas.c | 591 const struct oxnas_pin_group *pg = &pctl->groups[group]; in oxnas_ox810se_pinmux_enable() local 594 struct oxnas_desc_function *functions = pg->functions; in oxnas_ox810se_pinmux_enable() 595 u32 mask = BIT(pg->pin); in oxnas_ox810se_pinmux_enable() 601 fname, pg->bank, pg->pin, in oxnas_ox810se_pinmux_enable() 605 (pg->bank ? in oxnas_ox810se_pinmux_enable() 612 (pg->bank ? in oxnas_ox810se_pinmux_enable() 619 (pg->bank ? in oxnas_ox810se_pinmux_enable() 641 const struct oxnas_pin_group *pg = &pctl->groups[group]; in oxnas_ox820_pinmux_enable() local 644 struct oxnas_desc_function *functions = pg->functions; in oxnas_ox820_pinmux_enable() 645 unsigned int offset = (pg->bank ? PINMUX_820_BANK_OFFSET : 0); in oxnas_ox820_pinmux_enable() [all …]
|
/drivers/dma-buf/ |
D | udmabuf.c | 105 pgoff_t pg; in release_udmabuf() local 110 for (pg = 0; pg < ubuf->pagecount; pg++) in release_udmabuf() 111 put_page(ubuf->pages[pg]); in release_udmabuf()
|
/drivers/net/ethernet/mellanox/mlx4/ |
D | fw_qos.c | 56 __be16 pg; member 112 u8 *pg, u16 *ratelimit) in mlx4_SET_PORT_SCHEDULER() argument 146 tc->pg = htons(pg[i]); in mlx4_SET_PORT_SCHEDULER()
|
/drivers/usb/gadget/function/ |
D | uvc_configfs.c | 280 char *pg = page; in uvcg_default_processing_bm_controls_show() local 290 result += sprintf(pg, "%u\n", pd->bmControls[i]); in uvcg_default_processing_bm_controls_show() 291 pg = page + result; in uvcg_default_processing_bm_controls_show() 391 char *pg = page; in uvcg_default_camera_bm_controls_show() local 402 result += sprintf(pg, "%u\n", cd->bmControls[i]); in uvcg_default_camera_bm_controls_show() 403 pg = page + result; in uvcg_default_camera_bm_controls_show() 813 char *pg = page; in uvcg_format_bma_controls_show() local 821 result = sprintf(pg, "0x"); in uvcg_format_bma_controls_show() 822 pg += result; in uvcg_format_bma_controls_show() 824 result += sprintf(pg, "%x\n", f->bmaControls[i]); in uvcg_format_bma_controls_show() [all …]
|