/fs/xfs/ |
D | xfs_message.h | 33 #define xfs_printk_ratelimited(func, dev, fmt, ...) \ argument 39 func(dev, fmt, ##__VA_ARGS__); \ 42 #define xfs_emerg_ratelimited(dev, fmt, ...) \ argument 43 xfs_printk_ratelimited(xfs_emerg, dev, fmt, ##__VA_ARGS__) 44 #define xfs_alert_ratelimited(dev, fmt, ...) \ argument 45 xfs_printk_ratelimited(xfs_alert, dev, fmt, ##__VA_ARGS__) 46 #define xfs_crit_ratelimited(dev, fmt, ...) \ argument 47 xfs_printk_ratelimited(xfs_crit, dev, fmt, ##__VA_ARGS__) 48 #define xfs_err_ratelimited(dev, fmt, ...) \ argument 49 xfs_printk_ratelimited(xfs_err, dev, fmt, ##__VA_ARGS__) [all …]
|
D | xfs_trace.h | 46 __field(dev_t, dev) 59 __entry->dev = VFS_I(ctx->dp)->i_sb->s_dev; 72 MAJOR(__entry->dev), MINOR(__entry->dev), 107 __field(dev_t, dev) 113 __entry->dev = mp->m_super->s_dev; 119 MAJOR(__entry->dev), MINOR(__entry->dev), 143 __field(dev_t, dev) 158 __entry->dev = VFS_I(ctx->dp)->i_sb->s_dev; 174 MAJOR(__entry->dev), MINOR(__entry->dev), 195 __field(dev_t, dev) [all …]
|
/fs/gfs2/ |
D | trace_gfs2.h | 96 __field( dev_t, dev ) 107 __entry->dev = gl->gl_sbd->sd_vfs->s_dev; 118 MAJOR(__entry->dev), MINOR(__entry->dev), __entry->gltype, 135 __field( dev_t, dev ) 143 __entry->dev = gl->gl_sbd->sd_vfs->s_dev; 151 MAJOR(__entry->dev), MINOR(__entry->dev), 167 __field( dev_t, dev ) 177 __entry->dev = gl->gl_sbd->sd_vfs->s_dev; 187 MAJOR(__entry->dev), MINOR(__entry->dev), __entry->gltype, 204 __field( dev_t, dev ) [all …]
|
/fs/fuse/ |
D | cuse.c | 60 struct device *dev; /* device representing @cdev */ member 117 dev_t devt = inode->i_cdev->dev; in cuse_open() 124 if (pos->dev->devt == devt) { in cuse_open() 299 static void cuse_gendev_release(struct device *dev) in cuse_gendev_release() argument 301 kfree(dev); in cuse_gendev_release() 317 struct device *dev; in cuse_process_init_reply() local 352 dev = kzalloc(sizeof(*dev), GFP_KERNEL); in cuse_process_init_reply() 353 if (!dev) in cuse_process_init_reply() 356 device_initialize(dev); in cuse_process_init_reply() 357 dev_set_uevent_suppress(dev, 1); in cuse_process_init_reply() [all …]
|
/fs/btrfs/ |
D | reada.c | 250 struct btrfs_device *dev, u64 logical, in reada_find_zone() argument 262 ret = radix_tree_gang_lookup(&dev->reada_zones, (void **)&zone, in reada_find_zone() 295 zone->device = dev; /* our device always sits at index 0 */ in reada_find_zone() 298 zone->devs[i] = bbio->stripes[i].dev; in reada_find_zone() 303 ret = radix_tree_insert(&dev->reada_zones, in reada_find_zone() 309 ret = radix_tree_gang_lookup(&dev->reada_zones, (void **)&zone, in reada_find_zone() 328 struct btrfs_device *dev; in reada_find_extent() local 376 dev = bbio->stripes[nzones].dev; in reada_find_extent() 377 zone = reada_find_zone(fs_info, dev, logical, bbio); in reada_find_extent() 418 dev = bbio->stripes[i].dev; in reada_find_extent() [all …]
|
D | volumes.h | 176 struct btrfs_device *dev; member 197 struct btrfs_device *dev; member 321 void btrfs_dev_stat_inc_and_print(struct btrfs_device *dev, int index); 339 static inline void btrfs_dev_stat_inc(struct btrfs_device *dev, in btrfs_dev_stat_inc() argument 342 atomic_inc(dev->dev_stat_values + index); in btrfs_dev_stat_inc() 343 dev->dev_stats_dirty = 1; in btrfs_dev_stat_inc() 346 static inline int btrfs_dev_stat_read(struct btrfs_device *dev, in btrfs_dev_stat_read() argument 349 return atomic_read(dev->dev_stat_values + index); in btrfs_dev_stat_read() 352 static inline int btrfs_dev_stat_read_and_reset(struct btrfs_device *dev, in btrfs_dev_stat_read_and_reset() argument 357 ret = atomic_xchg(dev->dev_stat_values + index, 0); in btrfs_dev_stat_read_and_reset() [all …]
|
D | scrub.c | 69 struct btrfs_device *dev; member 87 struct btrfs_device *dev; member 154 struct btrfs_device *dev; member 178 struct btrfs_device *dev; member 223 u64 physical, struct btrfs_device *dev, u64 flags, 237 struct btrfs_device *dev, 351 struct scrub_ctx *scrub_setup_ctx(struct btrfs_device *dev, int is_dev_replace) in scrub_setup_ctx() argument 355 struct btrfs_fs_info *fs_info = dev->dev_root->fs_info; in scrub_setup_ctx() 366 if (dev->bdev) in scrub_setup_ctx() 368 bio_get_nr_vecs(dev->bdev)); in scrub_setup_ctx() [all …]
|
D | volumes.c | 48 static void __btrfs_reset_dev_stats(struct btrfs_device *dev); 49 static void btrfs_dev_stat_print_on_error(struct btrfs_device *dev); 107 struct btrfs_device *dev; in __find_device() local 109 list_for_each_entry(dev, head, dev_list) { in __find_device() 110 if (dev->devid == devid && in __find_device() 111 (!uuid || !memcmp(dev->uuid, uuid, BTRFS_UUID_SIZE))) { in __find_device() 112 return dev; in __find_device() 2410 ret = btrfs_free_dev_extent(trans, map->stripes[i].dev, in btrfs_relocate_chunk() 2414 if (map->stripes[i].dev) { in btrfs_relocate_chunk() 2415 ret = btrfs_update_device(trans, map->stripes[i].dev); in btrfs_relocate_chunk() [all …]
|
/fs/afs/ |
D | netdevices.c | 20 struct net_device *dev; in afs_get_MAC_address() local 26 dev = __dev_getfirstbyhwtype(&init_net, ARPHRD_ETHER); in afs_get_MAC_address() 27 if (dev) { in afs_get_MAC_address() 28 memcpy(mac, dev->dev_addr, maclen); in afs_get_MAC_address() 43 struct net_device *dev; in afs_get_ipv4_interfaces() local 50 for_each_netdev(&init_net, dev) { in afs_get_ipv4_interfaces() 51 if (dev->type == ARPHRD_LOOPBACK && !wantloopback) in afs_get_ipv4_interfaces() 53 idev = __in_dev_get_rtnl(dev); in afs_get_ipv4_interfaces() 59 bufs[n].mtu = dev->mtu; in afs_get_ipv4_interfaces()
|
/fs/exofs/ |
D | ore.c | 448 unsigned dev = per_dev->dev - ios->oc->first_dev; in ore_check_io() local 449 struct ore_dev *od = ios->oc->ods[dev]; in ore_check_io() 451 on_dev_error(ios, od, dev, osi.osd_err_pri, in ore_check_io() 556 si->dev = (group_width + C - RxP) % group_width + first_dev; in ore_calc_stripe_info() 562 si->dev = C; in ore_calc_stripe_info() 565 si->dev *= layout->mirrors_p1; in ore_calc_stripe_info() 581 osd_request_queue(_ios_od(ios, per_dev->dev)); in _ore_add_stripe_unit() 642 unsigned dev = si->dev; in _prepare_for_striping() local 643 unsigned first_dev = dev - (dev % devs_in_group); in _prepare_for_striping() 656 dev_order = _dev_order(devs_in_group, mirrors_p1, si->par_dev, dev); in _prepare_for_striping() [all …]
|
D | ore_raid.h | 38 unsigned par_dev, unsigned dev) in _dev_order() 40 unsigned first_dev = dev - dev % devs_in_group; in _dev_order() 42 dev -= first_dev; in _dev_order() 46 return dev / mirrors_p1; in _dev_order() 48 return ((devs_in_group + dev - par_dev - mirrors_p1) % devs_in_group) / in _dev_order()
|
D | ore_raid.c | 271 per_dev->dev, cur_len, not_last, per_dev->cur_sg, in _ore_add_sg_seg() 337 unsigned first_dev = si->dev - (si->dev % in _add_to_r4w() 339 unsigned comp = si->dev - first_dev; in _add_to_r4w() 362 per_dev->dev = si->dev; in _add_to_r4w() 368 q = osd_request_queue(ore_comp_dev(read_ios->oc, per_dev->dev)); in _add_to_r4w() 393 _LLU(si.obj_offset), pg_len, page->index, si.dev); in _add_to_r4w_first_page() 409 ios->layout->mirrors_p1, si.par_dev, si.dev); in _add_to_r4w_last_page() 416 p, c, _LLU(*offset), pg_len, si.dev, si.par_dev); in _add_to_r4w_last_page() 539 ios->layout->mirrors_p1, read_si.par_dev, read_si.dev); in _read_4_write_last_stripe()
|
/fs/ |
D | char_dev.c | 232 int alloc_chrdev_region(dev_t *dev, unsigned baseminor, unsigned count, in alloc_chrdev_region() argument 239 *dev = MKDEV(cd->major, cd->baseminor); in alloc_chrdev_region() 450 static struct kobject *exact_match(dev_t dev, int *part, void *data) in exact_match() argument 456 static int exact_lock(dev_t dev, void *data) in exact_lock() argument 472 int cdev_add(struct cdev *p, dev_t dev, unsigned count) in cdev_add() argument 476 p->dev = dev; in cdev_add() 479 error = kobj_map(cdev_map, dev, count, NULL, in cdev_add() 489 static void cdev_unmap(dev_t dev, unsigned count) in cdev_unmap() argument 491 kobj_unmap(cdev_map, dev, count); in cdev_unmap() 503 cdev_unmap(p->dev, p->count); in cdev_del() [all …]
|
/fs/nfs/blocklayout/ |
D | blocklayoutdev.c | 97 struct pnfs_device *dev) in nfs4_blk_decode_device() argument 105 .totallen = dev->mincount, in nfs4_blk_decode_device() 115 dprintk("%s: deviceid: %s, mincount: %d\n", __func__, dev->dev_id.data, in nfs4_blk_decode_device() 116 dev->mincount); in nfs4_blk_decode_device() 120 msg->data = kzalloc(sizeof(bl_msg) + dev->mincount, GFP_NOFS); in nfs4_blk_decode_device() 128 len = dev->mincount; in nfs4_blk_decode_device() 131 memcpy(&dataptr[offset], page_address(dev->pages[i]), in nfs4_blk_decode_device() 136 msg->len = sizeof(bl_msg) + dev->mincount; in nfs4_blk_decode_device() 175 memcpy(&rv->bm_mdevid, &dev->dev_id, sizeof(struct nfs4_deviceid)); in nfs4_blk_decode_device() 193 struct pnfs_block_dev *dev; in translate_devid() local [all …]
|
D | blocklayout.c | 1033 struct pnfs_block_dev *dev, *tmp; in free_blk_mountid() local 1036 list_for_each_entry_safe(dev, tmp, &mid->bm_devlist, bm_node) { in free_blk_mountid() 1037 list_del(&dev->bm_node); in free_blk_mountid() 1038 bl_free_block_dev(dev); in free_blk_mountid() 1051 struct pnfs_device *dev; in nfs4_blk_get_deviceinfo() local 1067 dev = kmalloc(sizeof(*dev), GFP_NOFS); in nfs4_blk_get_deviceinfo() 1068 if (!dev) { in nfs4_blk_get_deviceinfo() 1075 kfree(dev); in nfs4_blk_get_deviceinfo() 1086 memcpy(&dev->dev_id, d_id, sizeof(*d_id)); in nfs4_blk_get_deviceinfo() 1087 dev->layout_type = LAYOUT_BLOCK_VOLUME; in nfs4_blk_get_deviceinfo() [all …]
|
D | blocklayoutdm.c | 41 static void dev_remove(struct net *net, dev_t dev) in dev_remove() argument 64 bl_umount_request.major = MAJOR(dev); in dev_remove() 65 bl_umount_request.minor = MINOR(dev); in dev_remove()
|
/fs/proc/ |
D | consoles.c | 33 dev_t dev = 0; in show_console_dev() local 40 dev = MKDEV(driver->major, driver->minor_start); in show_console_dev() 41 dev += index; in show_console_dev() 57 if (dev) in show_console_dev() 58 seq_printf(m, " %4d:%d", MAJOR(dev), MINOR(dev)); in show_console_dev()
|
D | nommu.c | 41 dev_t dev = 0; in nommu_region_show() local 49 dev = inode->i_sb->s_dev; in nommu_region_show() 62 MAJOR(dev), MINOR(dev), ino, &len); in nommu_region_show()
|
/fs/ufs/ |
D | util.c | 184 dev_t dev; in ufs_get_inode_dev() local 195 dev = old_decode_dev(fs32 & 0x7fff); in ufs_get_inode_dev() 197 dev = MKDEV(sysv_major(fs32), sysv_minor(fs32)); in ufs_get_inode_dev() 201 dev = old_decode_dev(fs32); in ufs_get_inode_dev() 204 return dev; in ufs_get_inode_dev() 208 ufs_set_inode_dev(struct super_block *sb, struct ufs_inode_info *ufsi, dev_t dev) in ufs_set_inode_dev() argument 215 fs32 = sysv_encode_dev(dev); in ufs_set_inode_dev() 217 fs32 = old_encode_dev(dev); in ufs_set_inode_dev() 222 fs32 = old_encode_dev(dev); in ufs_set_inode_dev()
|
/fs/pstore/ |
D | ram.c | 351 static int ramoops_init_przs(struct device *dev, struct ramoops_context *cxt, in ramoops_init_przs() argument 361 dev_err(dev, "no room for dumps\n"); in ramoops_init_przs() 372 dev_err(dev, "failed to initialize a prz array for dumps\n"); in ramoops_init_przs() 383 dev_err(dev, "failed to request mem region (0x%zx@0x%llx): %d\n", in ramoops_init_przs() 396 static int ramoops_init_prz(struct device *dev, struct ramoops_context *cxt, in ramoops_init_prz() argument 404 dev_err(dev, "no room for mem region (0x%zx@0x%llx) in (0x%lx@0x%llx)\n", in ramoops_init_prz() 414 dev_err(dev, "failed to request mem region (0x%zx@0x%llx): %d\n", in ramoops_init_prz() 434 struct device *dev = &pdev->dev; in ramoops_probe() local 435 struct ramoops_platform_data *pdata = pdev->dev.platform_data; in ramoops_probe() 478 err = ramoops_init_przs(dev, cxt, &paddr, dump_mem_sz); in ramoops_probe() [all …]
|
/fs/nfsd/ |
D | nfsfh.h | 30 static inline void mk_fsid(int vers, u32 *fsidv, dev_t dev, ino_t ino, in mk_fsid() argument 36 fsidv[0] = htonl((MAJOR(dev)<<16) | in mk_fsid() 37 MINOR(dev)); in mk_fsid() 44 fsidv[0] = htonl(MAJOR(dev)); in mk_fsid() 45 fsidv[1] = htonl(MINOR(dev)); in mk_fsid() 50 fsidv[0] = new_encode_dev(dev); in mk_fsid()
|
/fs/ubifs/ |
D | misc.h | 156 static inline int ubifs_encode_dev(union ubifs_dev_desc *dev, dev_t rdev) in ubifs_encode_dev() argument 159 dev->new = cpu_to_le32(new_encode_dev(rdev)); in ubifs_encode_dev() 160 return sizeof(dev->new); in ubifs_encode_dev() 162 dev->huge = cpu_to_le64(huge_encode_dev(rdev)); in ubifs_encode_dev() 163 return sizeof(dev->huge); in ubifs_encode_dev()
|
/fs/cifs/ |
D | cifs_dfs_ref.c | 68 char *dev; in cifs_build_devname() local 92 dev = kmalloc(2 + unclen + 1 + pplen + 1, GFP_KERNEL); in cifs_build_devname() 93 if (!dev) in cifs_build_devname() 96 pos = dev; in cifs_build_devname() 118 convert_delimiter(dev, '/'); in cifs_build_devname() 119 return dev; in cifs_build_devname()
|
/fs/quota/ |
D | netlink.c | 33 void quota_send_warning(struct kqid qid, dev_t dev, in quota_send_warning() argument 69 ret = nla_put_u32(skb, QUOTA_NL_A_DEV_MAJOR, MAJOR(dev)); in quota_send_warning() 72 ret = nla_put_u32(skb, QUOTA_NL_A_DEV_MINOR, MINOR(dev)); in quota_send_warning()
|
/fs/nfs/ |
D | callback_xdr.c | 324 struct cb_devicenotifyitem *dev = &args->devs[i]; in decode_devicenotify_args() local 337 dev->cbd_notify_type = ntohl(*p++); in decode_devicenotify_args() 338 if (dev->cbd_notify_type != NOTIFY_DEVICEID4_CHANGE && in decode_devicenotify_args() 339 dev->cbd_notify_type != NOTIFY_DEVICEID4_DELETE) { in decode_devicenotify_args() 345 if (((dev->cbd_notify_type == NOTIFY_DEVICEID4_CHANGE) && in decode_devicenotify_args() 347 ((dev->cbd_notify_type == NOTIFY_DEVICEID4_DELETE) && in decode_devicenotify_args() 352 dev->cbd_layout_type = ntohl(*p++); in decode_devicenotify_args() 353 memcpy(dev->cbd_dev_id.data, p, NFS4_DEVICEID4_SIZE); in decode_devicenotify_args() 356 if (dev->cbd_layout_type == NOTIFY_DEVICEID4_CHANGE) { in decode_devicenotify_args() 362 dev->cbd_immediate = ntohl(*p++); in decode_devicenotify_args() [all …]
|