/block/ |
D | blk-sysfs.c | 30 queue_var_show(unsigned long var, char *page) in queue_var_show() argument 32 return sprintf(page, "%lu\n", var); in queue_var_show() 36 queue_var_store(unsigned long *var, const char *page, size_t count) in queue_var_store() argument 41 err = kstrtoul(page, 10, &v); in queue_var_store() 50 static ssize_t queue_var_store64(s64 *var, const char *page) in queue_var_store64() argument 55 err = kstrtos64(page, 10, &v); in queue_var_store64() 63 static ssize_t queue_requests_show(struct request_queue *q, char *page) in queue_requests_show() argument 65 return queue_var_show(q->nr_requests, page); in queue_requests_show() 69 queue_requests_store(struct request_queue *q, const char *page, size_t count) in queue_requests_store() argument 77 ret = queue_var_store(&nr, page, count); in queue_requests_store() [all …]
|
D | blk-integrity.c | 222 char *page) in integrity_attr_show() argument 229 return entry->show(bi, page); in integrity_attr_show() 233 struct attribute *attr, const char *page, in integrity_attr_store() argument 243 ret = entry->store(bi, page, count); in integrity_attr_store() 248 static ssize_t integrity_format_show(struct blk_integrity *bi, char *page) in integrity_format_show() argument 251 return sprintf(page, "%s\n", bi->profile->name); in integrity_format_show() 253 return sprintf(page, "none\n"); in integrity_format_show() 256 static ssize_t integrity_tag_size_show(struct blk_integrity *bi, char *page) in integrity_tag_size_show() argument 258 return sprintf(page, "%u\n", bi->tag_size); in integrity_tag_size_show() 261 static ssize_t integrity_interval_show(struct blk_integrity *bi, char *page) in integrity_interval_show() argument [all …]
|
D | bio.c | 859 struct page *page, unsigned int len, unsigned int off, in page_is_mergeable() argument 864 phys_addr_t page_addr = page_to_phys(page); in page_is_mergeable() 868 if (xen_domain() && !xen_biovec_phys_mergeable(bv, page)) in page_is_mergeable() 876 return (bv->bv_page + bv_end / PAGE_SIZE) == (page + off / PAGE_SIZE); in page_is_mergeable() 895 static bool __bio_try_merge_page(struct bio *bio, struct page *page, in __bio_try_merge_page() argument 904 if (page_is_mergeable(bv, page, len, off, same_page)) { in __bio_try_merge_page() 923 struct page *page, unsigned len, in bio_try_merge_hw_seg() argument 929 phys_addr_t addr2 = page_to_phys(page) + offset + len - 1; in bio_try_merge_hw_seg() 935 return __bio_try_merge_page(bio, page, len, offset, same_page); in bio_try_merge_hw_seg() 952 struct page *page, unsigned int len, unsigned int offset, in bio_add_hw_page() argument [all …]
|
D | blk-crypto-sysfs.c | 21 struct blk_crypto_attr *attr, char *page); 35 struct blk_crypto_attr *attr, char *page) in max_dun_bits_show() argument 37 return sysfs_emit(page, "%u\n", 8 * profile->max_dun_bytes_supported); in max_dun_bits_show() 41 struct blk_crypto_attr *attr, char *page) in num_keyslots_show() argument 43 return sysfs_emit(page, "%u\n", profile->num_slots); in num_keyslots_show() 82 struct blk_crypto_attr *attr, char *page) in blk_crypto_mode_show() argument 86 return sysfs_emit(page, "0x%x\n", profile->modes_supported[mode_num]); in blk_crypto_mode_show() 102 struct attribute *attr, char *page) in blk_crypto_attr_show() argument 107 return a->show(profile, a, page); in blk_crypto_attr_show()
|
D | blk-map.c | 133 struct page *page; in bio_copy_user_iov() local 178 page = map_data->pages[i / nr_pages]; in bio_copy_user_iov() 179 page += (i % nr_pages); in bio_copy_user_iov() 183 page = alloc_page(GFP_NOIO | gfp_mask); in bio_copy_user_iov() 184 if (!page) { in bio_copy_user_iov() 190 if (bio_add_pc_page(rq->q, bio, page, bytes, offset) < bytes) { in bio_copy_user_iov() 192 __free_page(page); in bio_copy_user_iov() 289 struct page **pages, *stack_pages[UIO_FASTIOV]; in bio_map_user_iov() 313 struct page *page = pages[j]; in bio_map_user_iov() local 320 if (!bio_add_hw_page(rq->q, bio, page, n, offs, in bio_map_user_iov() [all …]
|
D | badblocks.c | 58 u64 *p = bb->page; in badblocks_check() 129 u64 *p = bb->page; in badblocks_update_acked() 186 p = bb->page; in badblocks_set() 351 p = bb->page; in badblocks_clear() 433 if (bb->page == NULL || bb->changed) in ack_all_badblocks() 439 u64 *p = bb->page; in ack_all_badblocks() 465 ssize_t badblocks_show(struct badblocks *bb, char *page, int unack) in badblocks_show() argument 469 u64 *p = bb->page; in badblocks_show() 491 len += snprintf(page+len, PAGE_SIZE-len, "%llu %u\n", in badblocks_show() 515 ssize_t badblocks_store(struct badblocks *bb, const char *page, size_t len, in badblocks_store() argument [all …]
|
D | blk-mq-sysfs.c | 53 struct attribute *attr, char *page) in blk_mq_hw_sysfs_show() argument 68 res = entry->show(hctx, page); in blk_mq_hw_sysfs_show() 74 struct attribute *attr, const char *page, in blk_mq_hw_sysfs_store() argument 90 res = entry->store(hctx, page, length); in blk_mq_hw_sysfs_store() 96 char *page) in blk_mq_hw_sysfs_nr_tags_show() argument 98 return sprintf(page, "%u\n", hctx->tags->nr_tags); in blk_mq_hw_sysfs_nr_tags_show() 102 char *page) in blk_mq_hw_sysfs_nr_reserved_tags_show() argument 104 return sprintf(page, "%u\n", hctx->tags->nr_reserved_tags); in blk_mq_hw_sysfs_nr_reserved_tags_show() 107 static ssize_t blk_mq_hw_sysfs_cpus_show(struct blk_mq_hw_ctx *hctx, char *page) in blk_mq_hw_sysfs_cpus_show() argument 115 ret = snprintf(pos + page, size - pos, "%u", i); in blk_mq_hw_sysfs_cpus_show() [all …]
|
D | bdev.c | 329 struct page *page) in bdev_read_page() argument 340 result = ops->rw_page(bdev, sector + get_start_sect(bdev), page, in bdev_read_page() 366 struct page *page, struct writeback_control *wbc) in bdev_write_page() argument 377 set_page_writeback(page); in bdev_write_page() 378 result = ops->rw_page(bdev, sector + get_start_sect(bdev), page, in bdev_write_page() 381 end_page_writeback(page); in bdev_write_page() 383 clean_page_buffers(page); in bdev_write_page() 384 unlock_page(page); in bdev_write_page()
|
D | fops.c | 389 static int blkdev_writepage(struct page *page, struct writeback_control *wbc) in blkdev_writepage() argument 391 return block_write_full_page(page, blkdev_get_block, wbc); in blkdev_writepage() 405 loff_t pos, unsigned len, struct page **pagep, void **fsdata) in blkdev_write_begin() 411 loff_t pos, unsigned len, unsigned copied, struct page *page, in blkdev_write_end() argument 415 ret = block_write_end(file, mapping, pos, len, copied, page, fsdata); in blkdev_write_end() 417 unlock_page(page); in blkdev_write_end() 418 put_page(page); in blkdev_write_end()
|
D | blk-crypto-fallback.c | 320 struct page *plaintext_page = enc_bvec->bv_page; in blk_crypto_fallback_encrypt_bio() 321 struct page *ciphertext_page = in blk_crypto_fallback_encrypt_bio() 420 struct page *page = bv.bv_page; in blk_crypto_fallback_decrypt_bio() local 422 sg_set_page(&sg, page, data_unit_size, bv.bv_offset); in blk_crypto_fallback_decrypt_bio()
|
D | blk.h | 399 extern ssize_t blk_throtl_sample_time_show(struct request_queue *q, char *page); 401 const char *page, size_t count); 456 struct page *page, unsigned int len, unsigned int offset,
|
D | bio-integrity.c | 123 int bio_integrity_add_page(struct bio *bio, struct page *page, in bio_integrity_add_page() argument 138 bvec_set_page(&bip->bip_vec[bip->bip_vcnt], page, len, offset); in bio_integrity_add_page()
|
D | blk-mq.c | 3219 struct page *page; in blk_mq_clear_rq_mapping() local 3229 list_for_each_entry(page, &tags->page_list, lru) { in blk_mq_clear_rq_mapping() 3230 unsigned long start = (unsigned long)page_address(page); in blk_mq_clear_rq_mapping() 3231 unsigned long end = start + order_to_size(page->private); in blk_mq_clear_rq_mapping() 3259 struct page *page; in blk_mq_free_rqs() local 3285 page = list_first_entry(&tags->page_list, struct page, lru); in blk_mq_free_rqs() 3286 list_del_init(&page->lru); in blk_mq_free_rqs() 3291 kmemleak_free(page_address(page)); in blk_mq_free_rqs() 3292 __free_pages(page, page->private); in blk_mq_free_rqs() 3407 struct page *page; in blk_mq_alloc_rqs() local [all …]
|
D | bfq-iosched.c | 7231 static ssize_t bfq_var_show(unsigned int var, char *page) in bfq_var_show() argument 7233 return sprintf(page, "%u\n", var); in bfq_var_show() 7236 static int bfq_var_store(unsigned long *var, const char *page) in bfq_var_store() argument 7239 int ret = kstrtoul(page, 10, &new_val); in bfq_var_store() 7248 static ssize_t __FUNC(struct elevator_queue *e, char *page) \ 7256 return bfq_var_show(__data, (page)); \ 7270 static ssize_t __FUNC(struct elevator_queue *e, char *page) \ 7275 return bfq_var_show(__data, (page)); \ 7282 __FUNC(struct elevator_queue *e, const char *page, size_t count) \ 7288 ret = bfq_var_store(&__data, (page)); \ [all …]
|
D | elevator.c | 451 elv_attr_show(struct kobject *kobj, struct attribute *attr, char *page) in elv_attr_show() argument 462 error = e->type ? entry->show(e, page) : -ENOENT; in elv_attr_show() 469 const char *page, size_t length) in elv_attr_store() argument 480 error = e->type ? entry->store(e, page, length) : -ENOENT; in elv_attr_store()
|
D | blk-merge.c | 188 struct page *start_page, unsigned long offset) in get_max_segment_size() 462 struct page *page = bvec->bv_page; in blk_bvec_map_sg() local 472 page += (offset >> PAGE_SHIFT); in blk_bvec_map_sg() 476 sg_set_page(*sg, page, len, offset); in blk_bvec_map_sg()
|
D | kyber-iosched.c | 866 char *page) \ 870 return sprintf(page, "%llu\n", kqd->latency_targets[domain]); \ 874 const char *page, size_t count) \ 880 ret = kstrtoull(page, 10, &nsec); \
|
D | genhd.c | 730 char *page) in disk_badblocks_show() argument 735 return sprintf(page, "\n"); in disk_badblocks_show() 737 return badblocks_show(disk->bb, page, 0); in disk_badblocks_show() 742 const char *page, size_t len) in disk_badblocks_store() argument 749 return badblocks_store(disk->bb, page, len, 0); in disk_badblocks_store()
|
D | mq-deadline.c | 932 static ssize_t __FUNC(struct elevator_queue *e, char *page) \ 936 return sysfs_emit(page, "%d\n", __VAR); \ 950 static ssize_t __FUNC(struct elevator_queue *e, const char *page, size_t count) \ 955 __ret = kstrtoint(page, 0, &__data); \
|
D | bounce.c | 242 struct page *bounce_page; in __blk_queue_bounce()
|
D | blk-iocost.c | 866 u64 *page, u64 *seqio, u64 *randio) in calc_lcoefs() argument 870 *page = *seqio = *randio = 0; in calc_lcoefs() 876 *page = DIV64_U64_ROUND_UP(VTIME_PER_SEC, bps_pages); in calc_lcoefs() 878 *page = 1; in calc_lcoefs() 883 if (v > *page) in calc_lcoefs() 884 *seqio = v - *page; in calc_lcoefs() 889 if (v > *page) in calc_lcoefs() 890 *randio = v - *page; in calc_lcoefs()
|
D | blk-throttle.c | 2452 ssize_t blk_throtl_sample_time_show(struct request_queue *q, char *page) in blk_throtl_sample_time_show() argument 2456 return sprintf(page, "%u\n", jiffies_to_msecs(q->td->throtl_slice)); in blk_throtl_sample_time_show() 2460 const char *page, size_t count) in blk_throtl_sample_time_store() argument 2467 if (kstrtoul(page, 10, &v)) in blk_throtl_sample_time_store()
|