Home
last modified time | relevance | path

Searched refs:page (Results 1 – 25 of 1177) sorted by relevance

12345678910>>...48

/drivers/dma-buf/heaps/
Dpage_pool.c53 struct page *dmabuf_page_pool_alloc_pages(struct dmabuf_page_pool *pool) in dmabuf_page_pool_alloc_pages()
61 struct page *page) in dmabuf_page_pool_free_pages() argument
63 __free_pages(page, pool->order); in dmabuf_page_pool_free_pages()
66 static void dmabuf_page_pool_add(struct dmabuf_page_pool *pool, struct page *page) in dmabuf_page_pool_add() argument
70 if (PageHighMem(page)) in dmabuf_page_pool_add()
76 list_add_tail(&page->lru, &pool->items[index]); in dmabuf_page_pool_add()
79 mod_node_page_state(page_pgdat(page), NR_KERNEL_MISC_RECLAIMABLE, in dmabuf_page_pool_add()
83 static struct page *dmabuf_page_pool_remove(struct dmabuf_page_pool *pool, int index) in dmabuf_page_pool_remove()
85 struct page *page; in dmabuf_page_pool_remove() local
88 page = list_first_entry_or_null(&pool->items[index], struct page, lru); in dmabuf_page_pool_remove()
[all …]
/drivers/target/
Dtarget_core_stat.c48 static ssize_t target_stat_inst_show(struct config_item *item, char *page) in target_stat_inst_show() argument
52 return snprintf(page, PAGE_SIZE, "%u\n", hba->hba_index); in target_stat_inst_show()
55 static ssize_t target_stat_indx_show(struct config_item *item, char *page) in target_stat_indx_show() argument
57 return snprintf(page, PAGE_SIZE, "%u\n", to_stat_dev(item)->dev_index); in target_stat_indx_show()
60 static ssize_t target_stat_role_show(struct config_item *item, char *page) in target_stat_role_show() argument
62 return snprintf(page, PAGE_SIZE, "Target\n"); in target_stat_role_show()
65 static ssize_t target_stat_ports_show(struct config_item *item, char *page) in target_stat_ports_show() argument
67 return snprintf(page, PAGE_SIZE, "%u\n", to_stat_dev(item)->export_count); in target_stat_ports_show()
98 static ssize_t target_stat_tgt_inst_show(struct config_item *item, char *page) in target_stat_tgt_inst_show() argument
102 return snprintf(page, PAGE_SIZE, "%u\n", hba->hba_index); in target_stat_tgt_inst_show()
[all …]
Dtarget_core_configfs.c89 char *page) in target_core_item_version_show() argument
91 return sprintf(page, "Target Engine Core ConfigFS Infrastructure %s" in target_core_item_version_show()
102 char *page) in target_core_item_dbroot_show() argument
104 return sprintf(page, "%s\n", db_root); in target_core_item_dbroot_show()
108 const char *page, size_t count) in target_core_item_dbroot_store() argument
126 read_bytes = snprintf(db_root_stage, DB_ROOT_LEN, "%s", page); in target_core_item_dbroot_store()
543 static ssize_t _name##_show(struct config_item *item, char *page) \
545 return snprintf(page, PAGE_SIZE, "%u\n", to_attrib(item)->_name); \
582 static ssize_t _name##_store(struct config_item *item, const char *page,\
589 ret = kstrtou32(page, 0, &val); \
[all …]
/drivers/target/iscsi/
Discsi_target_stat.c51 char *page) in iscsi_stat_instance_inst_show() argument
53 return snprintf(page, PAGE_SIZE, "%u\n", in iscsi_stat_instance_inst_show()
58 char *page) in iscsi_stat_instance_min_ver_show() argument
60 return snprintf(page, PAGE_SIZE, "%u\n", ISCSI_DRAFT20_VERSION); in iscsi_stat_instance_min_ver_show()
64 char *page) in iscsi_stat_instance_max_ver_show() argument
66 return snprintf(page, PAGE_SIZE, "%u\n", ISCSI_DRAFT20_VERSION); in iscsi_stat_instance_max_ver_show()
70 char *page) in iscsi_stat_instance_portals_show() argument
72 return snprintf(page, PAGE_SIZE, "%u\n", in iscsi_stat_instance_portals_show()
77 char *page) in iscsi_stat_instance_nodes_show() argument
79 return snprintf(page, PAGE_SIZE, "%u\n", ISCSI_INST_NUM_NODES); in iscsi_stat_instance_nodes_show()
[all …]
Discsi_target_configfs.c39 static ssize_t lio_target_np_driver_show(struct config_item *item, char *page, in lio_target_np_driver_show() argument
48 rb = sysfs_emit(page, "1\n"); in lio_target_np_driver_show()
50 rb = sysfs_emit(page, "0\n"); in lio_target_np_driver_show()
56 const char *page, size_t count, enum iscsit_transport_type type, in lio_target_np_driver_store() argument
66 rc = kstrtou32(page, 0, &op); in lio_target_np_driver_store()
116 static ssize_t lio_target_np_iser_show(struct config_item *item, char *page) in lio_target_np_iser_show() argument
118 return lio_target_np_driver_show(item, page, ISCSI_INFINIBAND); in lio_target_np_iser_show()
122 const char *page, size_t count) in lio_target_np_iser_store() argument
124 return lio_target_np_driver_store(item, page, count, in lio_target_np_iser_store()
129 static ssize_t lio_target_np_cxgbit_show(struct config_item *item, char *page) in lio_target_np_cxgbit_show() argument
[all …]
/drivers/nvme/target/
Dconfigfs.c63 static ssize_t nvmet_addr_adrfam_show(struct config_item *item, char *page) in nvmet_addr_adrfam_show() argument
70 return snprintf(page, PAGE_SIZE, "%s\n", in nvmet_addr_adrfam_show()
74 return snprintf(page, PAGE_SIZE, "\n"); in nvmet_addr_adrfam_show()
78 const char *page, size_t count) in nvmet_addr_adrfam_store() argument
87 if (sysfs_streq(page, nvmet_addr_family[i].name)) in nvmet_addr_adrfam_store()
91 pr_err("Invalid value '%s' for adrfam\n", page); in nvmet_addr_adrfam_store()
102 char *page) in nvmet_addr_portid_show() argument
106 return snprintf(page, PAGE_SIZE, "%d\n", le16_to_cpu(portid)); in nvmet_addr_portid_show()
110 const char *page, size_t count) in nvmet_addr_portid_store() argument
115 if (kstrtou16(page, 0, &portid)) { in nvmet_addr_portid_store()
[all …]
/drivers/xen/
Dballoon.c148 static void balloon_append(struct page *page) in balloon_append() argument
150 __SetPageOffline(page); in balloon_append()
153 if (PageHighMem(page)) { in balloon_append()
154 list_add_tail(&page->lru, &ballooned_pages); in balloon_append()
157 list_add(&page->lru, &ballooned_pages); in balloon_append()
164 static struct page *balloon_retrieve(bool require_lowmem) in balloon_retrieve()
166 struct page *page; in balloon_retrieve() local
171 page = list_entry(ballooned_pages.next, struct page, lru); in balloon_retrieve()
172 if (require_lowmem && PageHighMem(page)) in balloon_retrieve()
174 list_del(&page->lru); in balloon_retrieve()
[all …]
/drivers/infiniband/hw/mthca/
Dmthca_memfree.c61 } page[]; member
109 struct page *page; in mthca_alloc_icm_pages() local
115 page = alloc_pages(gfp_mask | __GFP_ZERO, order); in mthca_alloc_icm_pages()
116 if (!page) in mthca_alloc_icm_pages()
119 sg_set_page(mem, page, PAGE_SIZE << order, 0); in mthca_alloc_icm_pages()
283 struct page *page = NULL; in mthca_table_find() local
309 page = sg_page(&chunk->mem[i]); in mthca_table_find()
318 return page ? lowmem_page_address(page) + offset : NULL; in mthca_table_find()
439 static u64 mthca_uarc_virt(struct mthca_dev *dev, struct mthca_uar *uar, int page) in mthca_uarc_virt() argument
443 page * MTHCA_ICM_PAGE_SIZE; in mthca_uarc_virt()
[all …]
/drivers/block/
Dbrd.c52 static struct page *brd_lookup_page(struct brd_device *brd, sector_t sector) in brd_lookup_page()
55 struct page *page; in brd_lookup_page() local
58 page = xa_load(&brd->brd_pages, idx); in brd_lookup_page()
60 BUG_ON(page && page->index != idx); in brd_lookup_page()
62 return page; in brd_lookup_page()
71 struct page *page, *cur; in brd_insert_page() local
74 page = brd_lookup_page(brd, sector); in brd_insert_page()
75 if (page) in brd_insert_page()
78 page = alloc_page(gfp | __GFP_ZERO | __GFP_HIGHMEM); in brd_insert_page()
79 if (!page) in brd_insert_page()
[all …]
/drivers/video/fbdev/core/
Dfb_defio.c26 static struct page *fb_deferred_io_page(struct fb_info *info, unsigned long offs) in fb_deferred_io_page()
29 struct page *page; in fb_deferred_io_page() local
32 page = vmalloc_to_page(screen_base + offs); in fb_deferred_io_page()
34 page = pfn_to_page((info->fix.smem_start + offs) >> PAGE_SHIFT); in fb_deferred_io_page()
36 return page; in fb_deferred_io_page()
41 struct page *page) in fb_deferred_io_pageref_get() argument
63 pageref->page = page; in fb_deferred_io_pageref_get()
97 struct page *page; in fb_deferred_io_fault() local
104 page = fb_deferred_io_page(info, offset); in fb_deferred_io_fault()
105 if (!page) in fb_deferred_io_fault()
[all …]
/drivers/net/dsa/b53/
Db53_mmap.c33 static int b53_mmap_read8(struct b53_device *dev, u8 page, u8 reg, u8 *val) in b53_mmap_read8() argument
38 *val = readb(regs + (page << 8) + reg); in b53_mmap_read8()
43 static int b53_mmap_read16(struct b53_device *dev, u8 page, u8 reg, u16 *val) in b53_mmap_read16() argument
52 *val = ioread16be(regs + (page << 8) + reg); in b53_mmap_read16()
54 *val = readw(regs + (page << 8) + reg); in b53_mmap_read16()
59 static int b53_mmap_read32(struct b53_device *dev, u8 page, u8 reg, u32 *val) in b53_mmap_read32() argument
68 *val = ioread32be(regs + (page << 8) + reg); in b53_mmap_read32()
70 *val = readl(regs + (page << 8) + reg); in b53_mmap_read32()
75 static int b53_mmap_read48(struct b53_device *dev, u8 page, u8 reg, u64 *val) in b53_mmap_read48() argument
88 lo = ioread16be(regs + (page << 8) + reg); in b53_mmap_read48()
[all …]
Db53_spi.c76 static inline int b53_spi_set_page(struct spi_device *spi, u8 page) in b53_spi_set_page() argument
82 txbuf[2] = page; in b53_spi_set_page()
87 static inline int b53_prepare_reg_access(struct spi_device *spi, u8 page) in b53_prepare_reg_access() argument
94 return b53_spi_set_page(spi, page); in b53_prepare_reg_access()
124 static int b53_spi_read(struct b53_device *dev, u8 page, u8 reg, u8 *data, in b53_spi_read() argument
130 ret = b53_prepare_reg_access(spi, page); in b53_spi_read()
141 static int b53_spi_read8(struct b53_device *dev, u8 page, u8 reg, u8 *val) in b53_spi_read8() argument
143 return b53_spi_read(dev, page, reg, val, 1); in b53_spi_read8()
146 static int b53_spi_read16(struct b53_device *dev, u8 page, u8 reg, u16 *val) in b53_spi_read16() argument
151 ret = b53_spi_read(dev, page, reg, (u8 *)&value, 2); in b53_spi_read16()
[all …]
/drivers/infiniband/hw/mlx5/
Ddoorbell.c51 struct mlx5_ib_user_db_page *page; in mlx5_ib_db_map_user() local
56 list_for_each_entry(page, &context->db_page_list, list) in mlx5_ib_db_map_user()
57 if ((current->mm == page->mm) && in mlx5_ib_db_map_user()
58 (page->user_virt == (virt & PAGE_MASK))) in mlx5_ib_db_map_user()
61 page = kmalloc(sizeof(*page), GFP_KERNEL); in mlx5_ib_db_map_user()
62 if (!page) { in mlx5_ib_db_map_user()
67 page->user_virt = (virt & PAGE_MASK); in mlx5_ib_db_map_user()
68 page->refcnt = 0; in mlx5_ib_db_map_user()
69 page->umem = ib_umem_get(context->ibucontext.device, virt & PAGE_MASK, in mlx5_ib_db_map_user()
71 if (IS_ERR(page->umem)) { in mlx5_ib_db_map_user()
[all …]
/drivers/gpu/drm/nouveau/nvkm/subdev/mmu/
Dvmm.c41 const struct nvkm_vmm_page *page) in nvkm_vmm_pt_new() argument
49 const struct nvkm_vmm_desc *pair = page[-1].desc; in nvkm_vmm_pt_new()
58 pgt->page = page ? page->shift : 0; in nvkm_vmm_pt_new()
73 const struct nvkm_vmm_page *page; member
200 const struct nvkm_vmm_desc *pair = it->page[-1].desc; in nvkm_vmm_unref_sptes()
299 const struct nvkm_vmm_desc *pair = it->page[-1].desc; in nvkm_vmm_ref_sptes()
489 pgt = nvkm_vmm_pt_new(desc, NVKM_VMM_PDE_SPARSED(pgt), it->page); in nvkm_vmm_ref_swpt()
501 nvkm_vmm_iter(struct nvkm_vmm *vmm, const struct nvkm_vmm_page *page, in nvkm_vmm_iter() argument
507 const struct nvkm_vmm_desc *desc = page->desc; in nvkm_vmm_iter()
509 u64 bits = addr >> page->shift; in nvkm_vmm_iter()
[all …]
/drivers/most/
Dconfigfs.c114 const char *page, size_t count) in mdev_link_create_link_store() argument
120 ret = kstrtobool(page, &tmp); in mdev_link_create_link_store()
136 const char *page, size_t count) in mdev_link_destroy_link_store() argument
142 ret = kstrtobool(page, &tmp); in mdev_link_destroy_link_store()
160 static ssize_t mdev_link_direction_show(struct config_item *item, char *page) in mdev_link_direction_show() argument
162 return snprintf(page, PAGE_SIZE, "%s\n", to_mdev_link(item)->direction); in mdev_link_direction_show()
166 const char *page, size_t count) in mdev_link_direction_store() argument
170 if (!sysfs_streq(page, "dir_rx") && !sysfs_streq(page, "rx") && in mdev_link_direction_store()
171 !sysfs_streq(page, "dir_tx") && !sysfs_streq(page, "tx")) in mdev_link_direction_store()
173 strcpy(mdev_link->direction, page); in mdev_link_direction_store()
[all …]
/drivers/hwmon/pmbus/
Dltc2978.c176 static int ltc_read_word_data(struct i2c_client *client, int page, int phase, in ltc_read_word_data() argument
185 return pmbus_read_word_data(client, page, 0xff, reg); in ltc_read_word_data()
188 static int ltc_read_byte_data(struct i2c_client *client, int page, int reg) in ltc_read_byte_data() argument
196 return pmbus_read_byte_data(client, page, reg); in ltc_read_byte_data()
199 static int ltc_write_byte_data(struct i2c_client *client, int page, int reg, u8 value) in ltc_write_byte_data() argument
207 return pmbus_write_byte_data(client, page, reg, value); in ltc_write_byte_data()
210 static int ltc_write_byte(struct i2c_client *client, int page, u8 byte) in ltc_write_byte() argument
218 return pmbus_write_byte(client, page, byte); in ltc_write_byte()
235 int page, int reg, u16 *pmax) in ltc_get_max() argument
239 ret = ltc_read_word_data(client, page, 0xff, reg); in ltc_get_max()
[all …]
Dpmbus_core.c37 u8 page; /* page number */ member
74 #define pb_reg_to_index(page, reg, mask) (((page) << PB_PAGE_SHIFT) | \ argument
104 int (*read_status)(struct i2c_client *client, int page);
115 u8 page; member
161 int pmbus_set_page(struct i2c_client *client, int page, int phase) in pmbus_set_page() argument
166 if (page < 0) in pmbus_set_page()
169 if (!(data->info->func[page] & PMBUS_PAGE_VIRTUAL) && in pmbus_set_page()
170 data->info->pages > 1 && page != data->currpage) { in pmbus_set_page()
171 rv = i2c_smbus_write_byte_data(client, PMBUS_PAGE, page); in pmbus_set_page()
179 if (rv != page) in pmbus_set_page()
[all …]
Dmp2975.c130 mp2975_read_word_helper(struct i2c_client *client, int page, int phase, u8 reg, in mp2975_read_word_helper() argument
133 int ret = pmbus_read_word_data(client, page, phase, reg); in mp2975_read_word_helper()
197 int page, int phase, u8 reg) in mp2975_read_phase() argument
201 ret = pmbus_read_word_data(client, page, phase, reg); in mp2975_read_phase()
228 ret = pmbus_read_word_data(client, page, phase, PMBUS_READ_IOUT); in mp2975_read_phase()
232 return max_t(int, DIV_ROUND_CLOSEST(ret, data->info.phases[page]), in mp2975_read_phase()
233 DIV_ROUND_CLOSEST(ph_curr, data->curr_sense_gain[page])); in mp2975_read_phase()
238 int page, int phase) in mp2975_read_phases() argument
242 if (page) { in mp2975_read_phases()
245 ret = mp2975_read_phase(client, data, page, phase, in mp2975_read_phases()
[all …]
Dmax31785.c27 static int max31785_read_byte_data(struct i2c_client *client, int page, in max31785_read_byte_data() argument
30 if (page < MAX31785_NR_PAGES) in max31785_read_byte_data()
37 return pmbus_read_byte_data(client, page - MAX31785_NR_PAGES, in max31785_read_byte_data()
44 static int max31785_write_byte(struct i2c_client *client, int page, u8 value) in max31785_write_byte() argument
46 if (page < MAX31785_NR_PAGES) in max31785_write_byte()
52 static int max31785_read_long_data(struct i2c_client *client, int page, in max31785_read_long_data() argument
76 rc = pmbus_set_page(client, page, 0xff); in max31785_read_long_data()
90 static int max31785_get_pwm(struct i2c_client *client, int page) in max31785_get_pwm() argument
94 rv = pmbus_get_fan_rate_device(client, page, 0, percent); in max31785_get_pwm()
105 static int max31785_get_pwm_mode(struct i2c_client *client, int page) in max31785_get_pwm_mode() argument
[all …]
/drivers/infiniband/hw/mlx4/
Ddoorbell.c48 struct mlx4_ib_user_db_page *page; in mlx4_ib_db_map_user() local
55 list_for_each_entry(page, &context->db_page_list, list) in mlx4_ib_db_map_user()
56 if (page->user_virt == (virt & PAGE_MASK)) in mlx4_ib_db_map_user()
59 page = kmalloc(sizeof *page, GFP_KERNEL); in mlx4_ib_db_map_user()
60 if (!page) { in mlx4_ib_db_map_user()
65 page->user_virt = (virt & PAGE_MASK); in mlx4_ib_db_map_user()
66 page->refcnt = 0; in mlx4_ib_db_map_user()
67 page->umem = ib_umem_get(context->ibucontext.device, virt & PAGE_MASK, in mlx4_ib_db_map_user()
69 if (IS_ERR(page->umem)) { in mlx4_ib_db_map_user()
70 err = PTR_ERR(page->umem); in mlx4_ib_db_map_user()
[all …]
/drivers/gpu/drm/i915/gt/
Dshmem_utils.c59 struct page **pages; in shmem_pin_map()
102 struct page *page; in __shmem_rw() local
105 page = shmem_read_mapping_page_gfp(file->f_mapping, pfn, in __shmem_rw()
107 if (IS_ERR(page)) in __shmem_rw()
108 return PTR_ERR(page); in __shmem_rw()
110 vaddr = kmap(page); in __shmem_rw()
113 set_page_dirty(page); in __shmem_rw()
117 mark_page_accessed(page); in __shmem_rw()
118 kunmap(page); in __shmem_rw()
119 put_page(page); in __shmem_rw()
[all …]
/drivers/misc/
Dvmw_balloon.c333 struct page *page; member
597 vmballoon_mark_page_offline(struct page *page, in vmballoon_mark_page_offline() argument
603 __SetPageOffline(page + i); in vmballoon_mark_page_offline()
612 vmballoon_mark_page_online(struct page *page, in vmballoon_mark_page_online() argument
618 __ClearPageOffline(page + i); in vmballoon_mark_page_online()
663 struct page *page; in vmballoon_alloc_page_list() local
673 page = list_first_entry(&ctl->prealloc_pages, in vmballoon_alloc_page_list()
674 struct page, lru); in vmballoon_alloc_page_list()
675 list_del(&page->lru); in vmballoon_alloc_page_list()
678 page = alloc_pages(__GFP_HIGHMEM|__GFP_NOWARN| in vmballoon_alloc_page_list()
[all …]
/drivers/infiniband/hw/hns/
Dhns_roce_db.c14 struct hns_roce_user_db_page *page; in hns_roce_db_map_user() local
20 list_for_each_entry(page, &context->page_list, list) in hns_roce_db_map_user()
21 if (page->user_virt == page_addr) in hns_roce_db_map_user()
24 page = kmalloc(sizeof(*page), GFP_KERNEL); in hns_roce_db_map_user()
25 if (!page) { in hns_roce_db_map_user()
30 refcount_set(&page->refcount, 1); in hns_roce_db_map_user()
31 page->user_virt = page_addr; in hns_roce_db_map_user()
32 page->umem = ib_umem_get(context->ibucontext.device, page_addr, in hns_roce_db_map_user()
34 if (IS_ERR(page->umem)) { in hns_roce_db_map_user()
35 ret = PTR_ERR(page->umem); in hns_roce_db_map_user()
[all …]
/drivers/mtd/devices/
Dblock2mtd.c50 static struct page *page_read(struct address_space *mapping, pgoff_t index) in page_read()
59 struct page *page; in _block2mtd_erase() local
66 page = page_read(mapping, index); in _block2mtd_erase()
67 if (IS_ERR(page)) in _block2mtd_erase()
68 return PTR_ERR(page); in _block2mtd_erase()
70 max = page_address(page) + PAGE_SIZE; in _block2mtd_erase()
71 for (p=page_address(page); p<max; p++) in _block2mtd_erase()
73 lock_page(page); in _block2mtd_erase()
74 memset(page_address(page), 0xff, PAGE_SIZE); in _block2mtd_erase()
75 set_page_dirty(page); in _block2mtd_erase()
[all …]
/drivers/char/agp/
Defficeon-agp.c69 static inline unsigned long efficeon_mask_memory(struct page *page) in efficeon_mask_memory() argument
71 unsigned long addr = page_to_phys(page); in efficeon_mask_memory()
163 unsigned long page = efficeon_private.l1_table[index]; in efficeon_free_gatt_table() local
164 if (page) { in efficeon_free_gatt_table()
166 free_page(page); in efficeon_free_gatt_table()
210 unsigned long page; in efficeon_create_gatt_table() local
213 page = efficeon_private.l1_table[index]; in efficeon_create_gatt_table()
214 BUG_ON(page); in efficeon_create_gatt_table()
216 page = get_zeroed_page(GFP_KERNEL); in efficeon_create_gatt_table()
217 if (!page) { in efficeon_create_gatt_table()
[all …]

12345678910>>...48