/mm/kasan/ |
D | generic.c | 163 size_t size, bool write, in check_region_inline() argument 170 return !kasan_report(addr, size, write, ret_ip); in check_region_inline() 174 return !kasan_report(addr, size, write, ret_ip); in check_region_inline() 180 return !kasan_report(addr, size, write, ret_ip); in check_region_inline() 183 bool kasan_check_range(unsigned long addr, size_t size, bool write, in kasan_check_range() argument 186 return check_region_inline(addr, size, write, ret_ip); in kasan_check_range()
|
D | sw_tags.c | 70 bool kasan_check_range(unsigned long addr, size_t size, bool write, in kasan_check_range() argument 81 return !kasan_report(addr, size, write, ret_ip); in kasan_check_range() 108 return !kasan_report(addr, size, write, ret_ip); in kasan_check_range() 114 return !kasan_report(addr, size, write, ret_ip); in kasan_check_range()
|
D | kasan.h | 221 bool kasan_check_range(unsigned long addr, size_t size, bool write,
|
/mm/ |
D | util.c | 854 int overcommit_ratio_handler(struct ctl_table *table, int write, void *buffer, in overcommit_ratio_handler() argument 859 ret = proc_dointvec(table, write, buffer, lenp, ppos); in overcommit_ratio_handler() 860 if (ret == 0 && write) in overcommit_ratio_handler() 870 int overcommit_policy_handler(struct ctl_table *table, int write, void *buffer, in overcommit_policy_handler() argument 888 if (write) { in overcommit_policy_handler() 891 ret = proc_dointvec_minmax(&t, write, buffer, lenp, ppos); in overcommit_policy_handler() 900 ret = proc_dointvec_minmax(table, write, buffer, lenp, ppos); in overcommit_policy_handler() 906 int overcommit_kbytes_handler(struct ctl_table *table, int write, void *buffer, in overcommit_kbytes_handler() argument 911 ret = proc_doulongvec_minmax(table, write, buffer, lenp, ppos); in overcommit_kbytes_handler() 912 if (ret == 0 && write) in overcommit_kbytes_handler()
|
D | page-writeback.c | 515 int dirty_background_ratio_handler(struct ctl_table *table, int write, in dirty_background_ratio_handler() argument 520 ret = proc_dointvec_minmax(table, write, buffer, lenp, ppos); in dirty_background_ratio_handler() 521 if (ret == 0 && write) in dirty_background_ratio_handler() 526 int dirty_background_bytes_handler(struct ctl_table *table, int write, in dirty_background_bytes_handler() argument 531 ret = proc_doulongvec_minmax(table, write, buffer, lenp, ppos); in dirty_background_bytes_handler() 532 if (ret == 0 && write) in dirty_background_bytes_handler() 537 int dirty_ratio_handler(struct ctl_table *table, int write, void *buffer, in dirty_ratio_handler() argument 543 ret = proc_dointvec_minmax(table, write, buffer, lenp, ppos); in dirty_ratio_handler() 544 if (ret == 0 && write && vm_dirty_ratio != old_ratio) { in dirty_ratio_handler() 551 int dirty_bytes_handler(struct ctl_table *table, int write, in dirty_bytes_handler() argument [all …]
|
D | hugetlb_cgroup.c | 632 cft->write = hugetlb_cgroup_write_dfl; in __hugetlb_cgroup_file_dfl_init() 640 cft->write = hugetlb_cgroup_write_dfl; in __hugetlb_cgroup_file_dfl_init() 696 cft->write = hugetlb_cgroup_write_legacy; in __hugetlb_cgroup_file_legacy_init() 703 cft->write = hugetlb_cgroup_write_legacy; in __hugetlb_cgroup_file_legacy_init() 721 cft->write = hugetlb_cgroup_reset; in __hugetlb_cgroup_file_legacy_init() 728 cft->write = hugetlb_cgroup_reset; in __hugetlb_cgroup_file_legacy_init() 735 cft->write = hugetlb_cgroup_reset; in __hugetlb_cgroup_file_legacy_init() 742 cft->write = hugetlb_cgroup_reset; in __hugetlb_cgroup_file_legacy_init()
|
D | huge_memory.c | 773 pmd_t *pmd, pfn_t pfn, pgprot_t prot, bool write, in insert_pfn_pmd() argument 782 if (write) { in insert_pfn_pmd() 799 if (write) { in insert_pfn_pmd() 833 pgprot_t pgprot, bool write) in vmf_insert_pfn_pmd_prot() argument 861 insert_pfn_pmd(vma, addr, vmf->pmd, pfn, pgprot, write, pgtable); in vmf_insert_pfn_pmd_prot() 875 pud_t *pud, pfn_t pfn, pgprot_t prot, bool write) in insert_pfn_pud() argument 883 if (write) { in insert_pfn_pud() 899 if (write) { in insert_pfn_pud() 924 pgprot_t pgprot, bool write) in vmf_insert_pfn_pud_prot() argument 945 insert_pfn_pud(vma, addr, vmf->pud, pfn, pgprot, write); in vmf_insert_pfn_pud_prot() [all …]
|
D | memcontrol.c | 5051 .write = mem_cgroup_reset, 5057 .write = mem_cgroup_write, 5063 .write = mem_cgroup_write, 5069 .write = mem_cgroup_reset, 5078 .write = mem_cgroup_force_empty_write, 5087 .write = memcg_write_event_control, 5118 .write = mem_cgroup_write, 5129 .write = mem_cgroup_reset, 5135 .write = mem_cgroup_reset, 5148 .write = mem_cgroup_write, [all …]
|
D | madvise.c | 1103 int write; in do_madvise() local 1133 write = madvise_need_mmap_write(behavior); in do_madvise() 1134 if (write) { in do_madvise() 1187 if (write) in do_madvise()
|
D | vmstat.c | 78 int sysctl_vm_numa_stat_handler(struct ctl_table *table, int write, in sysctl_vm_numa_stat_handler() argument 84 if (write) in sysctl_vm_numa_stat_handler() 86 ret = proc_dointvec_minmax(table, write, buffer, length, ppos); in sysctl_vm_numa_stat_handler() 87 if (ret || !write) in sysctl_vm_numa_stat_handler() 1818 int vmstat_refresh(struct ctl_table *table, int write, in vmstat_refresh() argument 1860 if (write) in vmstat_refresh()
|
D | hugetlb.c | 3439 static int proc_hugetlb_doulongvec_minmax(struct ctl_table *table, int write, in proc_hugetlb_doulongvec_minmax() argument 3452 return proc_doulongvec_minmax(&dup_table, write, buffer, length, ppos); in proc_hugetlb_doulongvec_minmax() 3456 struct ctl_table *table, int write, in hugetlb_sysctl_handler_common() argument 3466 ret = proc_hugetlb_doulongvec_minmax(table, write, buffer, length, ppos, in hugetlb_sysctl_handler_common() 3471 if (write) in hugetlb_sysctl_handler_common() 3478 int hugetlb_sysctl_handler(struct ctl_table *table, int write, in hugetlb_sysctl_handler() argument 3482 return hugetlb_sysctl_handler_common(false, table, write, in hugetlb_sysctl_handler() 3487 int hugetlb_mempolicy_sysctl_handler(struct ctl_table *table, int write, in hugetlb_mempolicy_sysctl_handler() argument 3490 return hugetlb_sysctl_handler_common(true, table, write, in hugetlb_mempolicy_sysctl_handler() 3495 int hugetlb_overcommit_handler(struct ctl_table *table, int write, in hugetlb_overcommit_handler() argument [all …]
|
D | page_alloc.c | 5951 int numa_zonelist_order_handler(struct ctl_table *table, int write, in numa_zonelist_order_handler() argument 5954 if (write) in numa_zonelist_order_handler() 5956 return proc_dostring(table, write, buffer, length, ppos); in numa_zonelist_order_handler() 8274 int min_free_kbytes_sysctl_handler(struct ctl_table *table, int write, in postcore_initcall() 8279 rc = proc_dointvec_minmax(table, write, buffer, length, ppos); in postcore_initcall() 8283 if (write) { in postcore_initcall() 8290 int watermark_scale_factor_sysctl_handler(struct ctl_table *table, int write, in watermark_scale_factor_sysctl_handler() argument 8295 rc = proc_dointvec_minmax(table, write, buffer, length, ppos); in watermark_scale_factor_sysctl_handler() 8299 if (write) in watermark_scale_factor_sysctl_handler() 8320 int sysctl_min_unmapped_ratio_sysctl_handler(struct ctl_table *table, int write, in sysctl_min_unmapped_ratio_sysctl_handler() argument [all …]
|
D | swap.c | 169 int get_kernel_pages(const struct kvec *kiov, int nr_segs, int write, in get_kernel_pages() argument 197 int get_kernel_page(unsigned long start, int write, struct page **pages) in get_kernel_page() argument 204 return get_kernel_pages(&kiov, 1, write, pages); in get_kernel_page()
|
D | gup.c | 954 int write = (gup_flags & FOLL_WRITE); in check_vma_flags() local 966 if (write) { in check_vma_flags() 996 if (!arch_vma_access_permitted(vma, write, false, foreign)) in check_vma_flags() 1193 bool write = !!(fault_flags & FAULT_FLAG_WRITE); in vma_permits_fault() local 1195 vm_flags_t vm_flags = write ? VM_WRITE : VM_READ; in vma_permits_fault() 1207 if (!arch_vma_access_permitted(vma, write, false, foreign)) in vma_permits_fault()
|
D | compaction.c | 2662 int compaction_proactiveness_sysctl_handler(struct ctl_table *table, int write, in compaction_proactiveness_sysctl_handler() argument 2667 rc = proc_dointvec_minmax(table, write, buffer, length, ppos); in compaction_proactiveness_sysctl_handler() 2671 if (write && sysctl_compaction_proactiveness) { in compaction_proactiveness_sysctl_handler() 2690 int sysctl_compaction_handler(struct ctl_table *table, int write, in sysctl_compaction_handler() argument 2693 if (write) in sysctl_compaction_handler()
|
D | memory.c | 4106 bool write = vmf->flags & FAULT_FLAG_WRITE; in do_set_pmd() local 4138 if (write) in do_set_pmd() 4170 bool write = vmf->flags & FAULT_FLAG_WRITE; in do_set_pte() local 4182 if (write) in do_set_pte() 4185 if (write && !(vmf->vma_flags & VM_SHARED)) { in do_set_pte() 5575 void *buf, int len, int write) in generic_access_phys() argument 5582 if (follow_phys(vma, addr, write, &prot, &phys_addr)) in generic_access_phys() 5589 if (write) in generic_access_phys() 5609 int write = gup_flags & FOLL_WRITE; in __access_remote_vm() local 5635 len, write); in __access_remote_vm() [all …]
|
D | nommu.c | 1691 int write = gup_flags & FOLL_WRITE; in __access_remote_vm() local 1704 if (write && vma->vm_flags & VM_MAYWRITE) in __access_remote_vm() 1707 else if (!write && vma->vm_flags & VM_MAYREAD) in __access_remote_vm()
|
D | kmemleak.c | 1847 .write = kmemleak_write,
|
/mm/damon/ |
D | dbgfs.c | 621 .write = dbgfs_attrs_write, 627 .write = dbgfs_schemes_write, 633 .write = dbgfs_target_ids_write, 639 .write = dbgfs_init_regions_write, 938 .write = dbgfs_mk_context_write, 942 .write = dbgfs_rm_context_write, 947 .write = dbgfs_monitor_on_write,
|