Home
last modified time | relevance | path

Searched refs:p (Results 1 – 25 of 28) sorted by relevance

12

/block/
Dbadblocks.c58 u64 *p = bb->page; in badblocks_check() local
88 sector_t a = BB_OFFSET(p[mid]); in badblocks_check()
105 BB_OFFSET(p[lo]) + BB_LEN(p[lo]) > s) { in badblocks_check()
106 if (BB_OFFSET(p[lo]) < target) { in badblocks_check()
110 if (rv != -1 && BB_ACK(p[lo])) in badblocks_check()
114 *first_bad = BB_OFFSET(p[lo]); in badblocks_check()
115 *bad_sectors = BB_LEN(p[lo]); in badblocks_check()
130 u64 *p = bb->page; in badblocks_update_acked() local
138 if (!BB_ACK(p[i])) { in badblocks_update_acked()
166 u64 *p; in badblocks_set() local
[all …]
Dioprio.c103 struct task_struct *p, *g; in SYSCALL_DEFINE3() local
118 p = current; in SYSCALL_DEFINE3()
120 p = find_task_by_vpid(who); in SYSCALL_DEFINE3()
121 if (p) in SYSCALL_DEFINE3()
122 ret = set_task_ioprio(p, ioprio); in SYSCALL_DEFINE3()
129 do_each_pid_thread(pgrp, PIDTYPE_PGID, p) { in SYSCALL_DEFINE3()
130 ret = set_task_ioprio(p, ioprio); in SYSCALL_DEFINE3()
133 } while_each_pid_thread(pgrp, PIDTYPE_PGID, p); in SYSCALL_DEFINE3()
147 for_each_process_thread(g, p) { in SYSCALL_DEFINE3()
148 if (!uid_eq(task_uid(p), uid) || in SYSCALL_DEFINE3()
[all …]
Dioctl.c19 struct blkpg_partition p; in blkpg_do_ioctl() local
24 if (copy_from_user(&p, upart, sizeof(struct blkpg_partition))) in blkpg_do_ioctl()
29 if (p.pno <= 0) in blkpg_do_ioctl()
33 return bdev_del_partition(bdev, p.pno); in blkpg_do_ioctl()
35 start = p.start >> SECTOR_SHIFT; in blkpg_do_ioctl()
36 length = p.length >> SECTOR_SHIFT; in blkpg_do_ioctl()
43 plength < 0 || p.pno > 65535) in blkpg_do_ioctl()
50 if (p.start & (bdev_logical_block_size(bdev) - 1)) in blkpg_do_ioctl()
52 return bdev_add_partition(bdev, p.pno, start, length); in blkpg_do_ioctl()
54 return bdev_resize_partition(bdev, p.pno, start, length); in blkpg_do_ioctl()
[all …]
Dt10-pi.c150 void *p, *pmap; in t10_pi_type1_prepare() local
154 p = pmap + iv.bv_offset; in t10_pi_type1_prepare()
156 struct t10_pi_tuple *pi = p; in t10_pi_type1_prepare()
162 p += tuple_sz; in t10_pi_type1_prepare()
198 void *p, *pmap; in t10_pi_type1_complete() local
202 p = pmap + iv.bv_offset; in t10_pi_type1_complete()
204 struct t10_pi_tuple *pi = p; in t10_pi_type1_complete()
211 p += tuple_sz; in t10_pi_type1_complete()
Dgenhd.c442 struct blk_major_name **n, *p; in register_blkdev() local
472 p = kmalloc(sizeof(struct blk_major_name), GFP_KERNEL); in register_blkdev()
473 if (p == NULL) { in register_blkdev()
478 p->major = major; in register_blkdev()
479 strlcpy(p->name, name, sizeof(p->name)); in register_blkdev()
480 p->next = NULL; in register_blkdev()
488 *n = p; in register_blkdev()
495 kfree(p); in register_blkdev()
507 struct blk_major_name *p = NULL; in unregister_blkdev() local
517 p = *n; in unregister_blkdev()
[all …]
Dscsi_ioctl.c40 static int sg_get_version(int __user *p) in sg_get_version() argument
43 return put_user(sg_version_num, p); in sg_get_version()
46 static int scsi_get_idlun(struct request_queue *q, int __user *p) in scsi_get_idlun() argument
48 return put_user(0, p); in scsi_get_idlun()
51 static int scsi_get_bus(struct request_queue *q, int __user *p) in scsi_get_bus() argument
53 return put_user(0, p); in scsi_get_bus()
61 static int sg_set_timeout(struct request_queue *q, int __user *p) in sg_set_timeout() argument
63 int timeout, err = get_user(timeout, p); in sg_set_timeout()
80 static int sg_get_reserved_size(struct request_queue *q, int __user *p) in sg_get_reserved_size() argument
84 return put_user(val, p); in sg_get_reserved_size()
[all …]
Dblk-integrity.c270 char *p = (char *) page; in integrity_verify_store() local
271 unsigned long val = simple_strtoul(p, &p, 10); in integrity_verify_store()
289 char *p = (char *) page; in integrity_generate_store() local
290 unsigned long val = simple_strtoul(p, &p, 10); in integrity_generate_store()
Dblk-timeout.c56 char *p = (char *) buf; in part_timeout_store() local
58 val = simple_strtoul(p, &p, 10); in part_timeout_store()
Delevator.c256 struct rb_node **p = &root->rb_node; in elv_rb_add() local
260 while (*p) { in elv_rb_add()
261 parent = *p; in elv_rb_add()
265 p = &(*p)->rb_left; in elv_rb_add()
267 p = &(*p)->rb_right; in elv_rb_add()
270 rb_link_node(&rq->rb_node, parent, p); in elv_rb_add()
Dbio.c254 void *p; in bio_free() local
265 p = bio; in bio_free()
266 p -= bs->front_pad; in bio_free()
268 mempool_free(p, &bs->bio_pool); in bio_free()
447 void *p; in bio_alloc_bioset() local
453 p = kmalloc(struct_size(bio, bi_inline_vecs, nr_iovecs), gfp_mask); in bio_alloc_bioset()
488 p = mempool_alloc(&bs->bio_pool, gfp_mask); in bio_alloc_bioset()
489 if (!p && gfp_mask != saved_gfp) { in bio_alloc_bioset()
492 p = mempool_alloc(&bs->bio_pool, gfp_mask); in bio_alloc_bioset()
499 if (unlikely(!p)) in bio_alloc_bioset()
[all …]
Dblk-map.c440 char *p = bio->bi_private; in bio_copy_kern_endio_read() local
445 memcpy(p, page_address(bvec->bv_page), bvec->bv_len); in bio_copy_kern_endio_read()
446 p += bvec->bv_len; in bio_copy_kern_endio_read()
470 void *p = data; in bio_copy_kern() local
496 memcpy(page_address(page), p, bytes); in bio_copy_kern()
502 p += bytes; in bio_copy_kern()
Dsed-opal.c2649 void *p; in sed_ioctl() local
2659 p = memdup_user(arg, _IOC_SIZE(cmd)); in sed_ioctl()
2660 if (IS_ERR(p)) in sed_ioctl()
2661 return PTR_ERR(p); in sed_ioctl()
2665 ret = opal_save(dev, p); in sed_ioctl()
2668 ret = opal_lock_unlock(dev, p); in sed_ioctl()
2671 ret = opal_take_ownership(dev, p); in sed_ioctl()
2674 ret = opal_activate_lsp(dev, p); in sed_ioctl()
2677 ret = opal_set_new_pw(dev, p); in sed_ioctl()
2680 ret = opal_activate_user(dev, p); in sed_ioctl()
[all …]
Dblk-iocost.c814 const struct ioc_params *p = &autop[idx]; in ioc_autop_idx() local
838 if (p->too_fast_vrate_pct && p->too_fast_vrate_pct <= vrate_pct) { in ioc_autop_idx()
847 if (p->too_slow_vrate_pct && p->too_slow_vrate_pct >= vrate_pct) { in ioc_autop_idx()
914 const struct ioc_params *p; in ioc_refresh_params() local
920 p = &autop[idx]; in ioc_refresh_params()
933 memcpy(ioc->params.qos, p->qos, sizeof(p->qos)); in ioc_refresh_params()
935 memcpy(ioc->params.i_lcoefs, p->i_lcoefs, sizeof(p->i_lcoefs)); in ioc_refresh_params()
3178 char *p; in ioc_qos_write() local
3199 while ((p = strsep(&input, " \t\n"))) { in ioc_qos_write()
3205 if (!*p) in ioc_qos_write()
[all …]
Dblk-core.c755 struct hd_struct *p; in blk_partition_remap() local
759 p = __disk_get_part(bio->bi_disk, bio->bi_partno); in blk_partition_remap()
760 if (unlikely(!p)) in blk_partition_remap()
762 if (unlikely(should_fail_request(p, bio->bi_iter.bi_size))) in blk_partition_remap()
764 if (unlikely(bio_check_ro(bio, p))) in blk_partition_remap()
768 if (bio_check_eod(bio, part_nr_sects_read(p))) in blk_partition_remap()
770 bio->bi_iter.bi_sector += p->start_sect; in blk_partition_remap()
771 trace_block_bio_remap(bio->bi_disk->queue, bio, part_devt(p), in blk_partition_remap()
772 bio->bi_iter.bi_sector - p->start_sect); in blk_partition_remap()
Dmq-deadline-main.c800 enum dd_prio p; in dd_has_write_work() local
802 for (p = 0; p <= DD_PRIO_MAX; p++) in dd_has_write_work()
803 if (!list_empty_careful(&dd->per_prio[p].fifo_list[DD_WRITE])) in dd_has_write_work()
Dblk-iolatency.c837 char *p, *tok; in iolatency_set_limit() local
847 p = ctx.body; in iolatency_set_limit()
850 while ((tok = strsep(&p, " "))) { in iolatency_set_limit()
/block/partitions/
Dmsdos.c41 static inline sector_t nr_sects(struct msdos_partition *p) in nr_sects() argument
43 return (sector_t)get_unaligned_le32(&p->nr_sects); in nr_sects()
46 static inline sector_t start_sect(struct msdos_partition *p) in start_sect() argument
48 return (sector_t)get_unaligned_le32(&p->start_sect); in start_sect()
51 static inline int is_extended_partition(struct msdos_partition *p) in is_extended_partition() argument
53 return (p->sys_ind == DOS_EXTENDED_PARTITION || in is_extended_partition()
54 p->sys_ind == WIN98_EXTENDED_PARTITION || in is_extended_partition()
55 p->sys_ind == LINUX_EXTENDED_PARTITION); in is_extended_partition()
62 msdos_magic_present(unsigned char *p) in msdos_magic_present() argument
64 return (p[0] == MSDOS_LABEL_MAGIC1 && p[1] == MSDOS_LABEL_MAGIC2); in msdos_magic_present()
[all …]
Dcore.c178 struct hd_struct *p = dev_to_part(dev); in part_partition_show() local
180 return sprintf(buf, "%d\n", p->partno); in part_partition_show()
186 struct hd_struct *p = dev_to_part(dev); in part_start_show() local
188 return sprintf(buf, "%llu\n",(unsigned long long)p->start_sect); in part_start_show()
194 struct hd_struct *p = dev_to_part(dev); in part_ro_show() local
195 return sprintf(buf, "%d\n", p->policy ? 1 : 0); in part_ro_show()
201 struct hd_struct *p = dev_to_part(dev); in part_alignment_offset_show() local
204 queue_limit_alignment_offset(&part_to_disk(p)->queue->limits, in part_alignment_offset_show()
205 p->start_sect)); in part_alignment_offset_show()
211 struct hd_struct *p = dev_to_part(dev); in part_discard_alignment_show() local
[all …]
Daix.c132 struct pvd *p; in alloc_pvd() local
134 p = kmalloc(count, GFP_KERNEL); in alloc_pvd()
135 if (!p) in alloc_pvd()
138 if (read_lba(state, lba, (u8 *) p, count) < count) { in alloc_pvd()
139 kfree(p); in alloc_pvd()
142 return p; in alloc_pvd()
157 struct lvname *p; in alloc_lvn() local
159 p = kmalloc(count, GFP_KERNEL); in alloc_lvn()
160 if (!p) in alloc_lvn()
163 if (read_lba(state, lba, (u8 *) p, count) < count) { in alloc_lvn()
[all …]
Dcheck.h31 void *read_part_sector(struct parsed_partitions *state, sector_t n, Sector *p);
32 static inline void put_dev_sector(Sector p) in put_dev_sector() argument
34 put_page(p.v); in put_dev_sector()
38 put_partition(struct parsed_partitions *p, int n, sector_t from, sector_t size) in put_partition() argument
40 if (n < p->limit) { in put_partition()
43 p->parts[n].from = from; in put_partition()
44 p->parts[n].size = size; in put_partition()
45 snprintf(tmp, sizeof(tmp), " %s%d", p->name, n); in put_partition()
46 strlcat(p->pp_buf, tmp, PAGE_SIZE); in put_partition()
Dkarma.c33 struct d_partition *p; in karma_partition() local
45 p = label->d_partitions; in karma_partition()
46 for (i = 0 ; i < 2; i++, p++) { in karma_partition()
50 if (p->p_fstype == 0x4d && le32_to_cpu(p->p_size)) { in karma_partition()
51 put_partition(state, slot, le32_to_cpu(p->p_offset), in karma_partition()
52 le32_to_cpu(p->p_size)); in karma_partition()
Dacorn.c353 const struct ics_part *p; in adfspart_check_ICS() local
371 for (slot = 1, p = (const struct ics_part *)data; p->size; p++) { in adfspart_check_ICS()
372 u32 start = le32_to_cpu(p->start); in adfspart_check_ICS()
373 s32 size = le32_to_cpu(p->size); /* yes, it's signed. */ in adfspart_check_ICS()
450 const struct ptec_part *p; in adfspart_check_POWERTEC() local
465 for (i = 0, p = (const struct ptec_part *)data; i < 12; i++, p++) { in adfspart_check_POWERTEC()
466 u32 start = le32_to_cpu(p->start); in adfspart_check_POWERTEC()
467 u32 size = le32_to_cpu(p->size); in adfspart_check_POWERTEC()
512 struct eesox_part *p; in adfspart_check_EESOX() local
528 for (i = 0, p = (struct eesox_part *)buffer; i < 8; i++, p++) { in adfspart_check_EESOX()
[all …]
Dsgi.c45 struct sgi_partition *p; in sgi_partition() local
51 p = &label->partitions[0]; in sgi_partition()
75 for(i = 0; i < 16; i++, p++) { in sgi_partition()
76 blocks = be32_to_cpu(p->num_blocks); in sgi_partition()
77 start = be32_to_cpu(p->first_block); in sgi_partition()
80 if (be32_to_cpu(p->type) == LINUX_RAID_PARTITION) in sgi_partition()
Dsun.c66 struct sun_partition *p; in sun_partition() local
76 p = label->partitions; in sun_partition()
109 for (i = 0; i < nparts; i++, p++) { in sun_partition()
113 st_sector = be32_to_cpu(p->start_cylinder) * spc; in sun_partition()
114 num_sectors = be32_to_cpu(p->num_sectors); in sun_partition()
Dldm.c496 struct msdos_partition *p; in ldm_validate_partition_table() local
511 p = (struct msdos_partition *)(data + 0x01BE); in ldm_validate_partition_table()
512 for (i = 0; i < 4; i++, p++) in ldm_validate_partition_table()
513 if (p->sys_ind == LDM_PARTITION) { in ldm_validate_partition_table()

12