Home
last modified time | relevance | path

Searched refs:dev (Results 1 – 25 of 104) sorted by relevance

12345

/fs/xfs/
Dxfs_message.h33 #define xfs_printk_ratelimited(func, dev, fmt, ...) \ argument
39 func(dev, fmt, ##__VA_ARGS__); \
42 #define xfs_emerg_ratelimited(dev, fmt, ...) \ argument
43 xfs_printk_ratelimited(xfs_emerg, dev, fmt, ##__VA_ARGS__)
44 #define xfs_alert_ratelimited(dev, fmt, ...) \ argument
45 xfs_printk_ratelimited(xfs_alert, dev, fmt, ##__VA_ARGS__)
46 #define xfs_crit_ratelimited(dev, fmt, ...) \ argument
47 xfs_printk_ratelimited(xfs_crit, dev, fmt, ##__VA_ARGS__)
48 #define xfs_err_ratelimited(dev, fmt, ...) \ argument
49 xfs_printk_ratelimited(xfs_err, dev, fmt, ##__VA_ARGS__)
[all …]
Dxfs_trace.h46 __field(dev_t, dev)
59 __entry->dev = VFS_I(ctx->dp)->i_sb->s_dev;
72 MAJOR(__entry->dev), MINOR(__entry->dev),
107 __field(dev_t, dev)
113 __entry->dev = mp->m_super->s_dev;
119 MAJOR(__entry->dev), MINOR(__entry->dev),
143 __field(dev_t, dev)
158 __entry->dev = VFS_I(ctx->dp)->i_sb->s_dev;
174 MAJOR(__entry->dev), MINOR(__entry->dev),
195 __field(dev_t, dev)
[all …]
/fs/gfs2/
Dtrace_gfs2.h96 __field( dev_t, dev )
107 __entry->dev = gl->gl_sbd->sd_vfs->s_dev;
118 MAJOR(__entry->dev), MINOR(__entry->dev), __entry->gltype,
135 __field( dev_t, dev )
143 __entry->dev = gl->gl_sbd->sd_vfs->s_dev;
151 MAJOR(__entry->dev), MINOR(__entry->dev),
167 __field( dev_t, dev )
177 __entry->dev = gl->gl_sbd->sd_vfs->s_dev;
187 MAJOR(__entry->dev), MINOR(__entry->dev), __entry->gltype,
204 __field( dev_t, dev )
[all …]
/fs/fuse/
Dcuse.c60 struct device *dev; /* device representing @cdev */ member
117 dev_t devt = inode->i_cdev->dev; in cuse_open()
124 if (pos->dev->devt == devt) { in cuse_open()
299 static void cuse_gendev_release(struct device *dev) in cuse_gendev_release() argument
301 kfree(dev); in cuse_gendev_release()
317 struct device *dev; in cuse_process_init_reply() local
352 dev = kzalloc(sizeof(*dev), GFP_KERNEL); in cuse_process_init_reply()
353 if (!dev) in cuse_process_init_reply()
356 device_initialize(dev); in cuse_process_init_reply()
357 dev_set_uevent_suppress(dev, 1); in cuse_process_init_reply()
[all …]
/fs/btrfs/
Dreada.c250 struct btrfs_device *dev, u64 logical, in reada_find_zone() argument
262 ret = radix_tree_gang_lookup(&dev->reada_zones, (void **)&zone, in reada_find_zone()
295 zone->device = dev; /* our device always sits at index 0 */ in reada_find_zone()
298 zone->devs[i] = bbio->stripes[i].dev; in reada_find_zone()
303 ret = radix_tree_insert(&dev->reada_zones, in reada_find_zone()
309 ret = radix_tree_gang_lookup(&dev->reada_zones, (void **)&zone, in reada_find_zone()
328 struct btrfs_device *dev; in reada_find_extent() local
376 dev = bbio->stripes[nzones].dev; in reada_find_extent()
377 zone = reada_find_zone(fs_info, dev, logical, bbio); in reada_find_extent()
418 dev = bbio->stripes[i].dev; in reada_find_extent()
[all …]
Dvolumes.h176 struct btrfs_device *dev; member
197 struct btrfs_device *dev; member
321 void btrfs_dev_stat_inc_and_print(struct btrfs_device *dev, int index);
339 static inline void btrfs_dev_stat_inc(struct btrfs_device *dev, in btrfs_dev_stat_inc() argument
342 atomic_inc(dev->dev_stat_values + index); in btrfs_dev_stat_inc()
343 dev->dev_stats_dirty = 1; in btrfs_dev_stat_inc()
346 static inline int btrfs_dev_stat_read(struct btrfs_device *dev, in btrfs_dev_stat_read() argument
349 return atomic_read(dev->dev_stat_values + index); in btrfs_dev_stat_read()
352 static inline int btrfs_dev_stat_read_and_reset(struct btrfs_device *dev, in btrfs_dev_stat_read_and_reset() argument
357 ret = atomic_xchg(dev->dev_stat_values + index, 0); in btrfs_dev_stat_read_and_reset()
[all …]
Dscrub.c69 struct btrfs_device *dev; member
87 struct btrfs_device *dev; member
154 struct btrfs_device *dev; member
178 struct btrfs_device *dev; member
223 u64 physical, struct btrfs_device *dev, u64 flags,
237 struct btrfs_device *dev,
351 struct scrub_ctx *scrub_setup_ctx(struct btrfs_device *dev, int is_dev_replace) in scrub_setup_ctx() argument
355 struct btrfs_fs_info *fs_info = dev->dev_root->fs_info; in scrub_setup_ctx()
366 if (dev->bdev) in scrub_setup_ctx()
368 bio_get_nr_vecs(dev->bdev)); in scrub_setup_ctx()
[all …]
Dvolumes.c48 static void __btrfs_reset_dev_stats(struct btrfs_device *dev);
49 static void btrfs_dev_stat_print_on_error(struct btrfs_device *dev);
107 struct btrfs_device *dev; in __find_device() local
109 list_for_each_entry(dev, head, dev_list) { in __find_device()
110 if (dev->devid == devid && in __find_device()
111 (!uuid || !memcmp(dev->uuid, uuid, BTRFS_UUID_SIZE))) { in __find_device()
112 return dev; in __find_device()
2410 ret = btrfs_free_dev_extent(trans, map->stripes[i].dev, in btrfs_relocate_chunk()
2414 if (map->stripes[i].dev) { in btrfs_relocate_chunk()
2415 ret = btrfs_update_device(trans, map->stripes[i].dev); in btrfs_relocate_chunk()
[all …]
/fs/afs/
Dnetdevices.c20 struct net_device *dev; in afs_get_MAC_address() local
26 dev = __dev_getfirstbyhwtype(&init_net, ARPHRD_ETHER); in afs_get_MAC_address()
27 if (dev) { in afs_get_MAC_address()
28 memcpy(mac, dev->dev_addr, maclen); in afs_get_MAC_address()
43 struct net_device *dev; in afs_get_ipv4_interfaces() local
50 for_each_netdev(&init_net, dev) { in afs_get_ipv4_interfaces()
51 if (dev->type == ARPHRD_LOOPBACK && !wantloopback) in afs_get_ipv4_interfaces()
53 idev = __in_dev_get_rtnl(dev); in afs_get_ipv4_interfaces()
59 bufs[n].mtu = dev->mtu; in afs_get_ipv4_interfaces()
/fs/exofs/
Dore.c448 unsigned dev = per_dev->dev - ios->oc->first_dev; in ore_check_io() local
449 struct ore_dev *od = ios->oc->ods[dev]; in ore_check_io()
451 on_dev_error(ios, od, dev, osi.osd_err_pri, in ore_check_io()
556 si->dev = (group_width + C - RxP) % group_width + first_dev; in ore_calc_stripe_info()
562 si->dev = C; in ore_calc_stripe_info()
565 si->dev *= layout->mirrors_p1; in ore_calc_stripe_info()
581 osd_request_queue(_ios_od(ios, per_dev->dev)); in _ore_add_stripe_unit()
642 unsigned dev = si->dev; in _prepare_for_striping() local
643 unsigned first_dev = dev - (dev % devs_in_group); in _prepare_for_striping()
656 dev_order = _dev_order(devs_in_group, mirrors_p1, si->par_dev, dev); in _prepare_for_striping()
[all …]
Dore_raid.h38 unsigned par_dev, unsigned dev) in _dev_order()
40 unsigned first_dev = dev - dev % devs_in_group; in _dev_order()
42 dev -= first_dev; in _dev_order()
46 return dev / mirrors_p1; in _dev_order()
48 return ((devs_in_group + dev - par_dev - mirrors_p1) % devs_in_group) / in _dev_order()
Dore_raid.c271 per_dev->dev, cur_len, not_last, per_dev->cur_sg, in _ore_add_sg_seg()
337 unsigned first_dev = si->dev - (si->dev % in _add_to_r4w()
339 unsigned comp = si->dev - first_dev; in _add_to_r4w()
362 per_dev->dev = si->dev; in _add_to_r4w()
368 q = osd_request_queue(ore_comp_dev(read_ios->oc, per_dev->dev)); in _add_to_r4w()
393 _LLU(si.obj_offset), pg_len, page->index, si.dev); in _add_to_r4w_first_page()
409 ios->layout->mirrors_p1, si.par_dev, si.dev); in _add_to_r4w_last_page()
416 p, c, _LLU(*offset), pg_len, si.dev, si.par_dev); in _add_to_r4w_last_page()
539 ios->layout->mirrors_p1, read_si.par_dev, read_si.dev); in _read_4_write_last_stripe()
/fs/
Dchar_dev.c232 int alloc_chrdev_region(dev_t *dev, unsigned baseminor, unsigned count, in alloc_chrdev_region() argument
239 *dev = MKDEV(cd->major, cd->baseminor); in alloc_chrdev_region()
450 static struct kobject *exact_match(dev_t dev, int *part, void *data) in exact_match() argument
456 static int exact_lock(dev_t dev, void *data) in exact_lock() argument
472 int cdev_add(struct cdev *p, dev_t dev, unsigned count) in cdev_add() argument
476 p->dev = dev; in cdev_add()
479 error = kobj_map(cdev_map, dev, count, NULL, in cdev_add()
489 static void cdev_unmap(dev_t dev, unsigned count) in cdev_unmap() argument
491 kobj_unmap(cdev_map, dev, count); in cdev_unmap()
503 cdev_unmap(p->dev, p->count); in cdev_del()
[all …]
/fs/nfs/blocklayout/
Dblocklayoutdev.c97 struct pnfs_device *dev) in nfs4_blk_decode_device() argument
105 .totallen = dev->mincount, in nfs4_blk_decode_device()
115 dprintk("%s: deviceid: %s, mincount: %d\n", __func__, dev->dev_id.data, in nfs4_blk_decode_device()
116 dev->mincount); in nfs4_blk_decode_device()
120 msg->data = kzalloc(sizeof(bl_msg) + dev->mincount, GFP_NOFS); in nfs4_blk_decode_device()
128 len = dev->mincount; in nfs4_blk_decode_device()
131 memcpy(&dataptr[offset], page_address(dev->pages[i]), in nfs4_blk_decode_device()
136 msg->len = sizeof(bl_msg) + dev->mincount; in nfs4_blk_decode_device()
175 memcpy(&rv->bm_mdevid, &dev->dev_id, sizeof(struct nfs4_deviceid)); in nfs4_blk_decode_device()
193 struct pnfs_block_dev *dev; in translate_devid() local
[all …]
Dblocklayout.c1033 struct pnfs_block_dev *dev, *tmp; in free_blk_mountid() local
1036 list_for_each_entry_safe(dev, tmp, &mid->bm_devlist, bm_node) { in free_blk_mountid()
1037 list_del(&dev->bm_node); in free_blk_mountid()
1038 bl_free_block_dev(dev); in free_blk_mountid()
1051 struct pnfs_device *dev; in nfs4_blk_get_deviceinfo() local
1067 dev = kmalloc(sizeof(*dev), GFP_NOFS); in nfs4_blk_get_deviceinfo()
1068 if (!dev) { in nfs4_blk_get_deviceinfo()
1075 kfree(dev); in nfs4_blk_get_deviceinfo()
1086 memcpy(&dev->dev_id, d_id, sizeof(*d_id)); in nfs4_blk_get_deviceinfo()
1087 dev->layout_type = LAYOUT_BLOCK_VOLUME; in nfs4_blk_get_deviceinfo()
[all …]
Dblocklayoutdm.c41 static void dev_remove(struct net *net, dev_t dev) in dev_remove() argument
64 bl_umount_request.major = MAJOR(dev); in dev_remove()
65 bl_umount_request.minor = MINOR(dev); in dev_remove()
/fs/proc/
Dconsoles.c33 dev_t dev = 0; in show_console_dev() local
40 dev = MKDEV(driver->major, driver->minor_start); in show_console_dev()
41 dev += index; in show_console_dev()
57 if (dev) in show_console_dev()
58 seq_printf(m, " %4d:%d", MAJOR(dev), MINOR(dev)); in show_console_dev()
Dnommu.c41 dev_t dev = 0; in nommu_region_show() local
49 dev = inode->i_sb->s_dev; in nommu_region_show()
62 MAJOR(dev), MINOR(dev), ino, &len); in nommu_region_show()
/fs/ufs/
Dutil.c184 dev_t dev; in ufs_get_inode_dev() local
195 dev = old_decode_dev(fs32 & 0x7fff); in ufs_get_inode_dev()
197 dev = MKDEV(sysv_major(fs32), sysv_minor(fs32)); in ufs_get_inode_dev()
201 dev = old_decode_dev(fs32); in ufs_get_inode_dev()
204 return dev; in ufs_get_inode_dev()
208 ufs_set_inode_dev(struct super_block *sb, struct ufs_inode_info *ufsi, dev_t dev) in ufs_set_inode_dev() argument
215 fs32 = sysv_encode_dev(dev); in ufs_set_inode_dev()
217 fs32 = old_encode_dev(dev); in ufs_set_inode_dev()
222 fs32 = old_encode_dev(dev); in ufs_set_inode_dev()
/fs/pstore/
Dram.c351 static int ramoops_init_przs(struct device *dev, struct ramoops_context *cxt, in ramoops_init_przs() argument
361 dev_err(dev, "no room for dumps\n"); in ramoops_init_przs()
372 dev_err(dev, "failed to initialize a prz array for dumps\n"); in ramoops_init_przs()
383 dev_err(dev, "failed to request mem region (0x%zx@0x%llx): %d\n", in ramoops_init_przs()
396 static int ramoops_init_prz(struct device *dev, struct ramoops_context *cxt, in ramoops_init_prz() argument
404 dev_err(dev, "no room for mem region (0x%zx@0x%llx) in (0x%lx@0x%llx)\n", in ramoops_init_prz()
414 dev_err(dev, "failed to request mem region (0x%zx@0x%llx): %d\n", in ramoops_init_prz()
434 struct device *dev = &pdev->dev; in ramoops_probe() local
435 struct ramoops_platform_data *pdata = pdev->dev.platform_data; in ramoops_probe()
478 err = ramoops_init_przs(dev, cxt, &paddr, dump_mem_sz); in ramoops_probe()
[all …]
/fs/nfsd/
Dnfsfh.h30 static inline void mk_fsid(int vers, u32 *fsidv, dev_t dev, ino_t ino, in mk_fsid() argument
36 fsidv[0] = htonl((MAJOR(dev)<<16) | in mk_fsid()
37 MINOR(dev)); in mk_fsid()
44 fsidv[0] = htonl(MAJOR(dev)); in mk_fsid()
45 fsidv[1] = htonl(MINOR(dev)); in mk_fsid()
50 fsidv[0] = new_encode_dev(dev); in mk_fsid()
/fs/ubifs/
Dmisc.h156 static inline int ubifs_encode_dev(union ubifs_dev_desc *dev, dev_t rdev) in ubifs_encode_dev() argument
159 dev->new = cpu_to_le32(new_encode_dev(rdev)); in ubifs_encode_dev()
160 return sizeof(dev->new); in ubifs_encode_dev()
162 dev->huge = cpu_to_le64(huge_encode_dev(rdev)); in ubifs_encode_dev()
163 return sizeof(dev->huge); in ubifs_encode_dev()
/fs/cifs/
Dcifs_dfs_ref.c68 char *dev; in cifs_build_devname() local
92 dev = kmalloc(2 + unclen + 1 + pplen + 1, GFP_KERNEL); in cifs_build_devname()
93 if (!dev) in cifs_build_devname()
96 pos = dev; in cifs_build_devname()
118 convert_delimiter(dev, '/'); in cifs_build_devname()
119 return dev; in cifs_build_devname()
/fs/quota/
Dnetlink.c33 void quota_send_warning(struct kqid qid, dev_t dev, in quota_send_warning() argument
69 ret = nla_put_u32(skb, QUOTA_NL_A_DEV_MAJOR, MAJOR(dev)); in quota_send_warning()
72 ret = nla_put_u32(skb, QUOTA_NL_A_DEV_MINOR, MINOR(dev)); in quota_send_warning()
/fs/nfs/
Dcallback_xdr.c324 struct cb_devicenotifyitem *dev = &args->devs[i]; in decode_devicenotify_args() local
337 dev->cbd_notify_type = ntohl(*p++); in decode_devicenotify_args()
338 if (dev->cbd_notify_type != NOTIFY_DEVICEID4_CHANGE && in decode_devicenotify_args()
339 dev->cbd_notify_type != NOTIFY_DEVICEID4_DELETE) { in decode_devicenotify_args()
345 if (((dev->cbd_notify_type == NOTIFY_DEVICEID4_CHANGE) && in decode_devicenotify_args()
347 ((dev->cbd_notify_type == NOTIFY_DEVICEID4_DELETE) && in decode_devicenotify_args()
352 dev->cbd_layout_type = ntohl(*p++); in decode_devicenotify_args()
353 memcpy(dev->cbd_dev_id.data, p, NFS4_DEVICEID4_SIZE); in decode_devicenotify_args()
356 if (dev->cbd_layout_type == NOTIFY_DEVICEID4_CHANGE) { in decode_devicenotify_args()
362 dev->cbd_immediate = ntohl(*p++); in decode_devicenotify_args()
[all …]

12345