| /drivers/android/binder/ |
| D | trace.rs | 32 unsafe fn rust_binder_transaction_fd_send(t_debug_id: c_int, fd: c_int, offset: usize); 33 unsafe fn rust_binder_transaction_fd_recv(t_debug_id: c_int, fd: c_int, offset: usize); 38 unsafe fn rust_binder_update_page_range(pid: c_int, allocate: bool, start: usize, end: usize); 39 unsafe fn rust_binder_alloc_lru_start(pid: c_int, page_index: usize); 40 unsafe fn rust_binder_alloc_lru_end(pid: c_int, page_index: usize); 41 unsafe fn rust_binder_free_lru_start(pid: c_int, page_index: usize); 42 unsafe fn rust_binder_free_lru_end(pid: c_int, page_index: usize); 43 unsafe fn rust_binder_alloc_page_start(pid: c_int, page_index: usize); 44 unsafe fn rust_binder_alloc_page_end(pid: c_int, page_index: usize); 45 unsafe fn rust_binder_unmap_user_start(pid: c_int, page_index: usize); [all …]
|
| D | allocation.rs | 29 pub(crate) offsets: Option<Range<usize>>, 53 pub(crate) offset: usize, 54 size: usize, 55 pub(crate) ptr: usize, 61 pub(crate) debug_id: usize, 67 debug_id: usize, in new() argument 68 offset: usize, in new() argument 69 size: usize, in new() argument 70 ptr: usize, in new() argument 85 fn size_check(&self, offset: usize, size: usize) -> Result { in size_check() argument [all …]
|
| D | stats.rs | 11 const BC_COUNT: usize = _IOC_NR(BC_REPLY_SG) as usize + 1; 12 const BR_COUNT: usize = _IOC_NR(BR_TRANSACTION_PENDING_FROZEN) as usize + 1; 32 let idx = _IOC_NR(bc) as usize; in inc_bc() 39 let idx = _IOC_NR(br) as usize; in inc_br() 70 pub(super) fn command_string(i: usize) -> &'static str { in command_string() 79 pub(super) fn return_string(i: usize) -> &'static str { in return_string()
|
| D | page_range.rs | 155 size: usize, 157 vma_addr: usize, 285 pub(crate) fn register_with_vma(&self, vma: &virt::VmaNew) -> Result<usize> { in register_with_vma() argument 286 let num_bytes = usize::min(vma.end() - vma.start(), bindings::SZ_4M as usize); in register_with_vma() 334 pub(crate) fn use_range(&self, start: usize, end: usize) -> Result<()> { in use_range() argument 387 fn use_page_slow(&self, i: usize) -> Result<()> { in use_page_slow() 462 pub(crate) fn stop_using_range(&self, start: usize, end: usize) { in stop_using_range() argument 492 unsafe fn iterate<T>(&self, mut offset: usize, mut size: usize, mut cb: T) -> Result in iterate() argument 494 T: FnMut(&Page, usize, usize) -> Result, in iterate() 518 let available = usize::min(size, PAGE_SIZE - offset); in iterate() [all …]
|
| D | thread.rs | 52 ancestors: KVec<usize>, 59 obj_index: usize, 61 offset: usize, 63 sender_uaddr: usize, 65 length: usize, 67 fixup_min_offset: usize, 78 skip: usize, 83 target_offset: usize, 89 parent_sg_index: usize, 94 num_ancestors: usize, [all …]
|
| D | node.rs | 130 count: usize, 192 pub(crate) debug_id: usize, 346 pub(crate) fn global_id(&self) -> usize { in global_id() argument 347 self as *const Node as usize in global_id() 405 count: usize, in update_refcount_locked() argument 502 pub(crate) fn update_refcount(self: &DArc<Self>, inc: bool, count: usize, strong: bool) { in update_refcount() argument 790 strong_node_count: usize, 791 weak_node_count: usize, 793 strong_count: usize, 794 weak_count: usize, [all …]
|
| D | process.rs | 59 address: usize, 64 fn new(address: usize, size: usize) -> Self { in new() argument 197 count: usize, in update_node_refcount() argument 314 debug_id: usize, 378 by_node: RBTree<usize, u32>, 960 debug_id: usize, in buffer_alloc() argument 961 size: usize, in buffer_alloc() argument 1020 pub(crate) fn buffer_get(self: &Arc<Self>, ptr: usize) -> Option<Allocation> { in buffer_get() 1032 pub(crate) fn buffer_raw_free(&self, ptr: usize) { in buffer_raw_free() argument 1061 pub(crate) fn buffer_make_freeable(&self, offset: usize, mut data: Option<AllocationInfo>) { in buffer_make_freeable() argument [all …]
|
| D | rust_binder_main.rs | 95 fn next_debug_id() -> usize { in next_debug_id() 127 fn len(&self) -> usize { in len() argument 286 fn ptr_align(value: usize) -> Option<usize> { in ptr_align() argument 287 let size = core::mem::size_of::<usize>() - 1; in ptr_align() 527 let pid = (unsafe { (*ptr).private }) as usize as Pid; in rust_binder_proc_show()
|
| D | transaction.rs | 41 pub(crate) debug_id: usize, 55 data_size: usize, 56 offsets_size: usize, 57 data_address: usize, 59 txn_security_ctx_off: Option<usize>,
|
| /drivers/android/binder/range_alloc/ |
| D | mod.rs | 19 fn new(is_oneway: bool, debug_id: usize, pid: Pid) -> Self { in new() argument 44 debug_id: usize, 68 fn debug_id(&self) -> usize { in debug_id() argument 79 const TREE_THRESHOLD: usize = 8; 84 pub(crate) start_page_idx: usize, 85 pub(crate) end_page_idx: usize, 89 fn interior_pages(offset: usize, size: usize) -> FreedRange { in interior_pages() argument 100 offset: usize, 101 size: usize, 106 fn endpoint(&self) -> usize { in endpoint() argument [all …]
|
| D | array.rs | 26 size: usize, 27 free_oneway_space: usize, 34 insert_at_idx: usize, 36 insert_at_offset: usize, 40 pub(crate) fn new(size: usize, alloc: EmptyArrayAlloc<T>) -> Self { in new() argument 48 pub(crate) fn free_oneway_space(&self) -> usize { in free_oneway_space() argument 52 pub(crate) fn count_buffers(&self) -> usize { in count_buffers() argument 56 pub(crate) fn total_size(&self) -> usize { in total_size() argument 90 fn find_empty_range(&self, size: usize) -> Option<FindEmptyRes> { in find_empty_range() 117 debug_id: usize, in reserve_new() argument [all …]
|
| D | tree.rs | 25 tree: RBTree<usize, Descriptor<T>>, 29 size: usize, 30 free_oneway_space: usize, 35 size: usize, in from_array() argument 99 pub(crate) fn total_size(&self) -> usize { in total_size() argument 103 pub(crate) fn free_oneway_space(&self) -> usize { in free_oneway_space() argument 107 pub(crate) fn count_buffers(&self) -> usize { in count_buffers() argument 142 fn find_best_match(&mut self, size: usize) -> Option<&mut Descriptor<T>> { in find_best_match() 151 debug_id: usize, in reserve_new() argument 152 size: usize, in reserve_new() argument [all …]
|
| /drivers/gpu/drm/ |
| D | drm_panic_qr.rs | 33 struct Version(usize); 119 const MAX_EC_SIZE: usize = 30; 120 const MAX_BLK_SIZE: usize = 123; 221 fn max_data(&self) -> usize { in max_data() argument 225 fn ec_size(&self) -> usize { in ec_size() argument 229 fn g1_blocks(&self) -> usize { in g1_blocks() argument 230 VPARAM[self.0 - 1].1 as usize in g1_blocks() 233 fn g2_blocks(&self) -> usize { in g2_blocks() argument 234 VPARAM[self.0 - 1].2 as usize in g2_blocks() 237 fn g1_blk_size(&self) -> usize { in g1_blk_size() argument [all …]
|
| /drivers/staging/android/ |
| D | ashmem_range.rs | 54 fn shrink_range(&mut self, range: &Range, pgstart: usize, pgend: usize) { in shrink_range() argument 121 pub(crate) pgstart: usize, 122 pub(crate) pgend: usize, 135 pub(crate) fn pgstart(&self, guard: &AshmemGuard) -> usize { in pgstart() argument 139 pub(crate) fn pgend(&self, guard: &AshmemGuard) -> usize { in pgend() argument 143 pub(crate) fn size(&self, guard: &AshmemGuard) -> usize { in size() argument 148 pub(crate) fn is_before_page(&self, page: usize, guard: &AshmemGuard) -> bool { in is_before_page() argument 153 pub(crate) fn contains_page(&self, page: usize, guard: &AshmemGuard) -> bool { in contains_page() argument 160 pgstart: usize, in is_superset_of_range() argument 161 pgend: usize, in is_superset_of_range() argument [all …]
|
| D | ashmem_rust.rs | 37 const ASHMEM_NAME_LEN: usize = bindings::ASHMEM_NAME_LEN as usize; 38 const ASHMEM_FULL_NAME_LEN: usize = bindings::ASHMEM_FULL_NAME_LEN as usize; 39 const ASHMEM_NAME_PREFIX_LEN: usize = bindings::ASHMEM_NAME_PREFIX_LEN as usize; 42 const ASHMEM_MAX_SIZE: usize = usize::MAX >> 1; 44 const PROT_READ: usize = bindings::PROT_READ as usize; 45 const PROT_EXEC: usize = bindings::PROT_EXEC as usize; 46 const PROT_WRITE: usize = bindings::PROT_WRITE as usize; 47 const PROT_MASK: usize = PROT_EXEC | PROT_READ | PROT_WRITE; 139 size: usize, 140 prot_mask: usize, [all …]
|
| D | ashmem_shrinker.rs | 87 pub fn set_batch(&mut self, batch: usize) { in set_batch() argument 217 pub fn new(count: usize) -> Self { in new() 218 if count > Self::MAX.inner as usize { in new() 246 pub fn from_count(count: usize) -> Self { in from_count() 247 if count > Self::MAX.inner as usize { in from_count() 298 pub fn nr_to_scan(&self) -> usize { in nr_to_scan() argument 300 unsafe { (*self.ptr.as_ptr()).nr_to_scan as usize } in nr_to_scan() 304 pub fn set_nr_scanned(&mut self, val: usize) { in set_nr_scanned() argument
|
| D | shmem.rs | 56 pub(crate) fn new(name: &CStr, size: usize, flags: vm_flags_t) -> Result<Self> { in new() argument 111 pub(crate) fn punch_hole(&self, start: usize, len: usize) { in punch_hole() argument 130 pub(crate) fn inode_ino(&self) -> usize { in inode_ino() argument 132 unsafe { (*(*self.inner.as_ptr()).f_inode).i_ino as usize } in inode_ino()
|
| D | ashmem_toggle.rs | 46 fn read_iter(mut kiocb: Kiocb<'_, Self::Ptr>, iov: &mut IovIter) -> Result<usize> { in read_iter() argument 61 fn write_iter(_kiocb: Kiocb<'_, Self::Ptr>, iov: &mut IovIter) -> Result<usize> { in write_iter() argument
|
| /drivers/acpi/apei/ |
| D | erst-dbg.c | 83 size_t usize, loff_t *off) in erst_dbg_read() argument 140 if (len > usize) in erst_dbg_read() 153 size_t usize, loff_t *off) in erst_dbg_write() argument 161 if (usize > ERST_DBG_RECORD_LEN_MAX) { in erst_dbg_write() 168 if (usize > erst_dbg_buf_len) { in erst_dbg_write() 171 p = kmalloc(usize, GFP_KERNEL); in erst_dbg_write() 176 erst_dbg_buf_len = usize; in erst_dbg_write() 178 rc = copy_from_user(erst_dbg_buf, ubuf, usize); in erst_dbg_write() 185 if (rcd->record_length != usize) in erst_dbg_write() 192 return rc < 0 ? rc : usize; in erst_dbg_write()
|
| /drivers/pci/pcie/ |
| D | aer_inject.c | 487 size_t usize, loff_t *off) in aer_inject_write() argument 494 if (usize < offsetof(struct aer_error_inj, domain) || in aer_inject_write() 495 usize > sizeof(einj)) in aer_inject_write() 499 if (copy_from_user(&einj, ubuf, usize)) in aer_inject_write() 503 return ret ? ret : usize; in aer_inject_write()
|
| /drivers/gpu/drm/nouveau/nvkm/subdev/fb/ |
| D | ramgf100.c | 501 u64 total = 0, lcomm = ~0, lower, ubase, usize; in gf100_ram_ctor() local 520 usize = total - lower; in gf100_ram_ctor() 523 nvkm_debug(subdev, "Upper: %4lld MiB @ %010llx\n", usize >> 20, ubase); in gf100_ram_ctor() 550 (usize - rsvd_tail) >> NVKM_RAM_MM_SHIFT, 1); in gf100_ram_ctor()
|
| /drivers/xen/ |
| D | mcelog.c | 107 size_t usize, loff_t *off) in xen_mce_chrdev_read() argument 119 if (*off != 0 || usize < XEN_MCE_LOG_LEN*sizeof(struct xen_mce)) in xen_mce_chrdev_read()
|
| /drivers/gpu/drm/amd/amdkfd/ |
| D | kfd_chardev.c | 3252 unsigned int usize, asize; in kfd_ioctl() local 3265 usize = asize = _IOC_SIZE(cmd); in kfd_ioctl() 3326 if (asize > usize) in kfd_ioctl() 3327 memset(kdata + usize, 0, asize - usize); in kfd_ioctl() 3331 if (copy_from_user(kdata, (void __user *)arg, usize) != 0) { in kfd_ioctl() 3336 memset(kdata, 0, usize); in kfd_ioctl() 3342 if (copy_to_user((void __user *)arg, kdata, usize) != 0) in kfd_ioctl()
|
| /drivers/accel/habanalabs/common/ |
| D | habanalabs_ioctl.c | 1236 unsigned int usize, asize; in _hl_ioctl() local 1251 usize = asize = _IOC_SIZE(cmd); in _hl_ioctl() 1270 if (copy_from_user(kdata, (void __user *)arg, usize)) { in _hl_ioctl() 1278 if ((cmd & IOC_OUT) && copy_to_user((void __user *)arg, kdata, usize)) in _hl_ioctl()
|
| /drivers/crypto/intel/qat/qat_common/ |
| D | qat_hal.c | 1235 int usize = 0; in qat_hal_concat_micro_code() local 1241 usize = ARRAY_SIZE(inst_4b); in qat_hal_concat_micro_code() 1243 for (i = 0; i < usize; i++) in qat_hal_concat_micro_code()
|