/block/partitions/ |
D | msdos.c | 35 #define SYS_IND(p) get_unaligned(&p->sys_ind) argument 37 static inline sector_t nr_sects(struct partition *p) in nr_sects() argument 39 return (sector_t)get_unaligned_le32(&p->nr_sects); in nr_sects() 42 static inline sector_t start_sect(struct partition *p) in start_sect() argument 44 return (sector_t)get_unaligned_le32(&p->start_sect); in start_sect() 47 static inline int is_extended_partition(struct partition *p) in is_extended_partition() argument 49 return (SYS_IND(p) == DOS_EXTENDED_PARTITION || in is_extended_partition() 50 SYS_IND(p) == WIN98_EXTENDED_PARTITION || in is_extended_partition() 51 SYS_IND(p) == LINUX_EXTENDED_PARTITION); in is_extended_partition() 58 msdos_magic_present(unsigned char *p) in msdos_magic_present() argument [all …]
|
D | aix.c | 132 struct pvd *p; in alloc_pvd() local 134 p = kmalloc(count, GFP_KERNEL); in alloc_pvd() 135 if (!p) in alloc_pvd() 138 if (read_lba(state, lba, (u8 *) p, count) < count) { in alloc_pvd() 139 kfree(p); in alloc_pvd() 142 return p; in alloc_pvd() 157 struct lvname *p; in alloc_lvn() local 159 p = kmalloc(count, GFP_KERNEL); in alloc_lvn() 160 if (!p) in alloc_lvn() 163 if (read_lba(state, lba, (u8 *) p, count) < count) { in alloc_lvn() [all …]
|
D | check.h | 31 sector_t n, Sector *p) in read_part_sector() argument 37 return read_dev_sector(state->bdev, n, p); in read_part_sector() 41 put_partition(struct parsed_partitions *p, int n, sector_t from, sector_t size) in put_partition() argument 43 if (n < p->limit) { in put_partition() 46 p->parts[n].from = from; in put_partition() 47 p->parts[n].size = size; in put_partition() 48 snprintf(tmp, sizeof(tmp), " %s%d", p->name, n); in put_partition() 49 strlcat(p->pp_buf, tmp, PAGE_SIZE); in put_partition()
|
D | karma.c | 31 struct d_partition *p; in karma_partition() local 43 p = label->d_partitions; in karma_partition() 44 for (i = 0 ; i < 2; i++, p++) { in karma_partition() 48 if (p->p_fstype == 0x4d && le32_to_cpu(p->p_size)) { in karma_partition() 49 put_partition(state, slot, le32_to_cpu(p->p_offset), in karma_partition() 50 le32_to_cpu(p->p_size)); in karma_partition()
|
D | acorn.c | 359 const struct ics_part *p; in adfspart_check_ICS() local 377 for (slot = 1, p = (const struct ics_part *)data; p->size; p++) { in adfspart_check_ICS() 378 u32 start = le32_to_cpu(p->start); in adfspart_check_ICS() 379 s32 size = le32_to_cpu(p->size); /* yes, it's signed. */ in adfspart_check_ICS() 456 const struct ptec_part *p; in adfspart_check_POWERTEC() local 471 for (i = 0, p = (const struct ptec_part *)data; i < 12; i++, p++) { in adfspart_check_POWERTEC() 472 u32 start = le32_to_cpu(p->start); in adfspart_check_POWERTEC() 473 u32 size = le32_to_cpu(p->size); in adfspart_check_POWERTEC() 518 struct eesox_part *p; in adfspart_check_EESOX() local 534 for (i = 0, p = (struct eesox_part *)buffer; i < 8; i++, p++) { in adfspart_check_EESOX() [all …]
|
D | sgi.c | 39 struct sgi_partition *p; in sgi_partition() local 45 p = &label->partitions[0]; in sgi_partition() 69 for(i = 0; i < 16; i++, p++) { in sgi_partition() 70 blocks = be32_to_cpu(p->num_blocks); in sgi_partition() 71 start = be32_to_cpu(p->first_block); in sgi_partition() 74 if (be32_to_cpu(p->type) == LINUX_RAID_PARTITION) in sgi_partition()
|
D | sun.c | 58 struct sun_partition *p; in sun_partition() local 68 p = label->partitions; in sun_partition() 101 for (i = 0; i < nparts; i++, p++) { in sun_partition() 105 st_sector = be32_to_cpu(p->start_cylinder) * spc; in sun_partition() 106 num_sectors = be32_to_cpu(p->num_sectors); in sun_partition()
|
D | ldm.h | 102 #define SYS_IND(p) (get_unaligned(&(p)->sys_ind)) argument
|
D | ldm.c | 562 struct partition *p; in ldm_validate_partition_table() local 577 p = (struct partition*)(data + 0x01BE); in ldm_validate_partition_table() 578 for (i = 0; i < 4; i++, p++) in ldm_validate_partition_table() 579 if (SYS_IND (p) == LDM_PARTITION) { in ldm_validate_partition_table()
|
/block/ |
D | partition-generic.c | 70 struct hd_struct *p = dev_to_part(dev); in part_partition_show() local 72 return sprintf(buf, "%d\n", p->partno); in part_partition_show() 78 struct hd_struct *p = dev_to_part(dev); in part_start_show() local 80 return sprintf(buf, "%llu\n",(unsigned long long)p->start_sect); in part_start_show() 86 struct hd_struct *p = dev_to_part(dev); in part_size_show() local 87 return sprintf(buf, "%llu\n",(unsigned long long)part_nr_sects_read(p)); in part_size_show() 93 struct hd_struct *p = dev_to_part(dev); in part_ro_show() local 94 return sprintf(buf, "%d\n", p->policy ? 1 : 0); in part_ro_show() 100 struct hd_struct *p = dev_to_part(dev); in part_alignment_offset_show() local 101 return sprintf(buf, "%llu\n", (unsigned long long)p->alignment_offset); in part_alignment_offset_show() [all …]
|
D | ioprio.c | 65 struct task_struct *p, *g; in SYSCALL_DEFINE3() local 96 p = current; in SYSCALL_DEFINE3() 98 p = find_task_by_vpid(who); in SYSCALL_DEFINE3() 99 if (p) in SYSCALL_DEFINE3() 100 ret = set_task_ioprio(p, ioprio); in SYSCALL_DEFINE3() 107 do_each_pid_thread(pgrp, PIDTYPE_PGID, p) { in SYSCALL_DEFINE3() 108 ret = set_task_ioprio(p, ioprio); in SYSCALL_DEFINE3() 111 } while_each_pid_thread(pgrp, PIDTYPE_PGID, p); in SYSCALL_DEFINE3() 125 do_each_thread(g, p) { in SYSCALL_DEFINE3() 126 if (!uid_eq(task_uid(p), uid) || in SYSCALL_DEFINE3() [all …]
|
D | ioctl.c | 19 struct blkpg_partition p; in blkpg_ioctl() local 28 if (copy_from_user(&p, a.data, sizeof(struct blkpg_partition))) in blkpg_ioctl() 33 partno = p.pno; in blkpg_ioctl() 38 start = p.start >> 9; in blkpg_ioctl() 39 length = p.length >> 9; in blkpg_ioctl() 97 start = p.start >> 9; in blkpg_ioctl() 99 length = p.length >> 9; in blkpg_ioctl() 143 i_size_write(bdevp->bd_inode, p.length); in blkpg_ioctl() 357 struct pr_preempt p; in blkdev_pr_preempt() local 363 if (copy_from_user(&p, arg, sizeof(p))) in blkdev_pr_preempt() [all …]
|
D | scsi_ioctl.c | 54 static int sg_get_version(int __user *p) in sg_get_version() argument 57 return put_user(sg_version_num, p); in sg_get_version() 60 static int scsi_get_idlun(struct request_queue *q, int __user *p) in scsi_get_idlun() argument 62 return put_user(0, p); in scsi_get_idlun() 65 static int scsi_get_bus(struct request_queue *q, int __user *p) in scsi_get_bus() argument 67 return put_user(0, p); in scsi_get_bus() 75 static int sg_set_timeout(struct request_queue *q, int __user *p) in sg_set_timeout() argument 77 int timeout, err = get_user(timeout, p); in sg_set_timeout() 94 static int sg_get_reserved_size(struct request_queue *q, int __user *p) in sg_get_reserved_size() argument 98 return put_user(val, p); in sg_get_reserved_size() [all …]
|
D | genhd.c | 291 struct blk_major_name **n, *p; in register_blkdev() local 313 p = kmalloc(sizeof(struct blk_major_name), GFP_KERNEL); in register_blkdev() 314 if (p == NULL) { in register_blkdev() 319 p->major = major; in register_blkdev() 320 strlcpy(p->name, name, sizeof(p->name)); in register_blkdev() 321 p->next = NULL; in register_blkdev() 329 *n = p; in register_blkdev() 336 kfree(p); in register_blkdev() 348 struct blk_major_name *p = NULL; in unregister_blkdev() local 358 p = *n; in unregister_blkdev() [all …]
|
D | blk-integrity.c | 294 char *p = (char *) page; in integrity_verify_store() local 295 unsigned long val = simple_strtoul(p, &p, 10); in integrity_verify_store() 313 char *p = (char *) page; in integrity_generate_store() local 314 unsigned long val = simple_strtoul(p, &p, 10); in integrity_generate_store()
|
D | bio.c | 248 void *p; in bio_free() local 259 p = bio; in bio_free() 260 p -= bs->front_pad; in bio_free() 262 mempool_free(p, bs->bio_pool); in bio_free() 435 void *p; in bio_alloc_bioset() local 441 p = kmalloc(sizeof(struct bio) + in bio_alloc_bioset() 476 p = mempool_alloc(bs->bio_pool, gfp_mask); in bio_alloc_bioset() 477 if (!p && gfp_mask != saved_gfp) { in bio_alloc_bioset() 480 p = mempool_alloc(bs->bio_pool, gfp_mask); in bio_alloc_bioset() 487 if (unlikely(!p)) in bio_alloc_bioset() [all …]
|
D | blk-timeout.c | 57 char *p = (char *) buf; in part_timeout_store() local 59 val = simple_strtoul(p, &p, 10); in part_timeout_store()
|
D | elevator.c | 299 struct rb_node **p = &root->rb_node; in elv_rb_add() local 303 while (*p) { in elv_rb_add() 304 parent = *p; in elv_rb_add() 308 p = &(*p)->rb_left; in elv_rb_add() 310 p = &(*p)->rb_right; in elv_rb_add() 313 rb_link_node(&rq->rb_node, parent, p); in elv_rb_add()
|
D | deadline-iosched.c | 385 char *p = (char *) page; in deadline_var_store() local 387 *var = simple_strtol(p, &p, 10); in deadline_var_store()
|
D | cfq-iosched.c | 2202 struct rb_node **p, *parent; in cfq_service_tree_add() local 2249 p = &st->rb.rb_node; in cfq_service_tree_add() 2250 while (*p) { in cfq_service_tree_add() 2251 parent = *p; in cfq_service_tree_add() 2258 p = &parent->rb_left; in cfq_service_tree_add() 2260 p = &parent->rb_right; in cfq_service_tree_add() 2269 rb_link_node(&cfqq->rb_node, parent, p); in cfq_service_tree_add() 2282 struct rb_node **p, *parent; in cfq_prio_tree_lookup() local 2286 p = &root->rb_node; in cfq_prio_tree_lookup() 2287 while (*p) { in cfq_prio_tree_lookup() [all …]
|
D | blk-throttle.c | 1335 char *p; in tg_set_max() local 1346 p = tok; in tg_set_max() 1347 strsep(&p, "="); in tg_set_max() 1348 if (!p || (sscanf(p, "%llu", &val) != 1 && strcmp(p, "max"))) in tg_set_max()
|
D | blk-mq.c | 1509 void *p; in blk_mq_init_rq_map() local 1532 p = page_address(page); in blk_mq_init_rq_map() 1537 kmemleak_alloc(p, order_to_size(this_order), 1, GFP_NOIO); in blk_mq_init_rq_map() 1542 tags->rqs[i] = p; in blk_mq_init_rq_map() 1552 p += rq_size; in blk_mq_init_rq_map()
|
D | blk-core.c | 1829 struct hd_struct *p = bdev->bd_part; in blk_partition_remap() local 1831 bio->bi_iter.bi_sector += p->start_sect; in blk_partition_remap() 1836 bio->bi_iter.bi_sector - p->start_sect); in blk_partition_remap()
|