/block/ |
D | blk-sysfs.c | 27 queue_var_store(unsigned long *var, const char *page, size_t count) in queue_var_store() argument 38 return count; in queue_var_store() 47 queue_requests_store(struct request_queue *q, const char *page, size_t count) in queue_requests_store() argument 56 ret = queue_var_store(&nr, page, count); in queue_requests_store() 70 if (rl->count[BLK_RW_SYNC] >= queue_congestion_on_threshold(q)) in queue_requests_store() 72 else if (rl->count[BLK_RW_SYNC] < queue_congestion_off_threshold(q)) in queue_requests_store() 75 if (rl->count[BLK_RW_ASYNC] >= queue_congestion_on_threshold(q)) in queue_requests_store() 77 else if (rl->count[BLK_RW_ASYNC] < queue_congestion_off_threshold(q)) in queue_requests_store() 81 if (rl->count[BLK_RW_SYNC] >= q->nr_requests) { in queue_requests_store() 88 if (rl->count[BLK_RW_ASYNC] >= q->nr_requests) { in queue_requests_store() [all …]
|
D | blk-integrity.c | 238 size_t count) in integrity_attr_store() argument 247 ret = entry->store(bi, page, count); in integrity_attr_store() 269 const char *page, size_t count) in integrity_read_store() argument 279 return count; in integrity_read_store() 288 const char *page, size_t count) in integrity_write_store() argument 298 return count; in integrity_write_store()
|
D | blk-timeout.c | 49 const char *buf, size_t count) in part_timeout_store() argument 54 if (count) { in part_timeout_store() 67 return count; in part_timeout_store()
|
D | bsg.c | 533 __bsg_read(char __user *buf, size_t count, struct bsg_device *bd, in __bsg_read() argument 539 if (count % sizeof(struct sg_io_v4)) in __bsg_read() 543 nr_commands = count / sizeof(struct sg_io_v4); in __bsg_read() 595 bsg_read(struct file *file, char __user *buf, size_t count, loff_t *ppos) in bsg_read() argument 601 dprintk("%s: read %Zd bytes\n", bd->name, count); in bsg_read() 606 ret = __bsg_read(buf, count, bd, NULL, &bytes_read); in bsg_read() 616 size_t count, ssize_t *bytes_written, in __bsg_write() argument 623 if (count % sizeof(struct sg_io_v4)) in __bsg_write() 626 nr_commands = count / sizeof(struct sg_io_v4); in __bsg_write() 670 bsg_write(struct file *file, const char __user *buf, size_t count, loff_t *ppos) in bsg_write() argument [all …]
|
D | deadline-iosched.c | 373 deadline_var_store(int *var, const char *page, size_t count) in deadline_var_store() argument 378 return count; in deadline_var_store() 398 static ssize_t __FUNC(struct elevator_queue *e, const char *page, size_t count) \ 402 int ret = deadline_var_store(&__data, (page), count); \
|
D | blk-core.c | 562 rl->count[BLK_RW_SYNC] = rl->count[BLK_RW_ASYNC] = 0; in blk_init_rl() 814 rl->count[sync] < queue_congestion_off_threshold(q)) in __freed_request() 817 if (rl->count[sync] + 1 <= q->nr_requests) { in __freed_request() 835 rl->count[sync]--; in freed_request() 912 if (rl->count[is_sync]+1 >= queue_congestion_on_threshold(q)) { in __get_request() 913 if (rl->count[is_sync]+1 >= q->nr_requests) { in __get_request() 948 if (rl->count[is_sync] >= (3 * q->nr_requests / 2)) in __get_request() 952 rl->count[is_sync]++; in __get_request() 1052 if (unlikely(rl->count[is_sync] == 0)) in __get_request() 1858 unsigned int count; in submit_bio() local [all …]
|
D | cfq-iosched.c | 87 unsigned count; member 845 return cfqg->service_tree_idle.count; in cfq_group_busy_queues_wl() 847 return cfqg->service_trees[wl_class][ASYNC_WORKLOAD].count + in cfq_group_busy_queues_wl() 848 cfqg->service_trees[wl_class][SYNC_NOIDLE_WORKLOAD].count + in cfq_group_busy_queues_wl() 849 cfqg->service_trees[wl_class][SYNC_WORKLOAD].count; in cfq_group_busy_queues_wl() 855 return cfqg->service_trees[RT_WORKLOAD][ASYNC_WORKLOAD].count + in cfqg_busy_async_queues() 856 cfqg->service_trees[BE_WORKLOAD][ASYNC_WORKLOAD].count; in cfqg_busy_async_queues() 1167 if (!root->count) in cfq_rb_first() 1201 --root->count; in cfq_rb_erase() 1450 - cfqg->service_tree_idle.count; in cfq_group_served() [all …]
|
D | blk-throttle.c | 34 unsigned int count; member 39 .count = 0, .min_disptime = 0} 315 if (!root->count) in throtl_rb_first() 338 --root->count; in throtl_rb_erase() 386 st->count++; in __throtl_enqueue_tg() 417 BUG_ON(!st->count); in throtl_schedule_next_dispatch()
|
D | partition-generic.c | 158 const char *buf, size_t count) in part_fail_store() argument 163 if (count > 0 && sscanf(buf, "%d", &i) > 0) in part_fail_store() 166 return count; in part_fail_store()
|
D | elevator.c | 1002 size_t count) in elv_iosched_store() argument 1007 return count; in elv_iosched_store() 1011 return count; in elv_iosched_store()
|
D | genhd.c | 1717 const char *buf, size_t count) in disk_events_poll_msecs_store() argument 1722 if (!count || !sscanf(buf, "%ld", &intv)) in disk_events_poll_msecs_store() 1732 return count; in disk_events_poll_msecs_store()
|
/block/partitions/ |
D | efi.c | 193 u64 lba, u8 *buffer, size_t count) in read_lba() argument 202 while (count) { in read_lba() 208 if (copied > count) in read_lba() 209 copied = count; in read_lba() 214 count -= copied; in read_lba() 231 size_t count; in alloc_read_gpt_entries() local 237 count = le32_to_cpu(gpt->num_partition_entries) * in alloc_read_gpt_entries() 239 if (!count) in alloc_read_gpt_entries() 241 pte = kmalloc(count, GFP_KERNEL); in alloc_read_gpt_entries() 247 count) < count) { in alloc_read_gpt_entries()
|