Home
last modified time | relevance | path

Searched refs:write (Results 1 – 19 of 19) sorted by relevance

/mm/kasan/
Dgeneric.c163 size_t size, bool write, in check_region_inline() argument
170 return !kasan_report(addr, size, write, ret_ip); in check_region_inline()
174 return !kasan_report(addr, size, write, ret_ip); in check_region_inline()
180 return !kasan_report(addr, size, write, ret_ip); in check_region_inline()
183 bool kasan_check_range(unsigned long addr, size_t size, bool write, in kasan_check_range() argument
186 return check_region_inline(addr, size, write, ret_ip); in kasan_check_range()
Dsw_tags.c70 bool kasan_check_range(unsigned long addr, size_t size, bool write, in kasan_check_range() argument
81 return !kasan_report(addr, size, write, ret_ip); in kasan_check_range()
108 return !kasan_report(addr, size, write, ret_ip); in kasan_check_range()
114 return !kasan_report(addr, size, write, ret_ip); in kasan_check_range()
Dkasan.h221 bool kasan_check_range(unsigned long addr, size_t size, bool write,
/mm/
Dutil.c854 int overcommit_ratio_handler(struct ctl_table *table, int write, void *buffer, in overcommit_ratio_handler() argument
859 ret = proc_dointvec(table, write, buffer, lenp, ppos); in overcommit_ratio_handler()
860 if (ret == 0 && write) in overcommit_ratio_handler()
870 int overcommit_policy_handler(struct ctl_table *table, int write, void *buffer, in overcommit_policy_handler() argument
888 if (write) { in overcommit_policy_handler()
891 ret = proc_dointvec_minmax(&t, write, buffer, lenp, ppos); in overcommit_policy_handler()
900 ret = proc_dointvec_minmax(table, write, buffer, lenp, ppos); in overcommit_policy_handler()
906 int overcommit_kbytes_handler(struct ctl_table *table, int write, void *buffer, in overcommit_kbytes_handler() argument
911 ret = proc_doulongvec_minmax(table, write, buffer, lenp, ppos); in overcommit_kbytes_handler()
912 if (ret == 0 && write) in overcommit_kbytes_handler()
Dpage-writeback.c515 int dirty_background_ratio_handler(struct ctl_table *table, int write, in dirty_background_ratio_handler() argument
520 ret = proc_dointvec_minmax(table, write, buffer, lenp, ppos); in dirty_background_ratio_handler()
521 if (ret == 0 && write) in dirty_background_ratio_handler()
526 int dirty_background_bytes_handler(struct ctl_table *table, int write, in dirty_background_bytes_handler() argument
531 ret = proc_doulongvec_minmax(table, write, buffer, lenp, ppos); in dirty_background_bytes_handler()
532 if (ret == 0 && write) in dirty_background_bytes_handler()
537 int dirty_ratio_handler(struct ctl_table *table, int write, void *buffer, in dirty_ratio_handler() argument
543 ret = proc_dointvec_minmax(table, write, buffer, lenp, ppos); in dirty_ratio_handler()
544 if (ret == 0 && write && vm_dirty_ratio != old_ratio) { in dirty_ratio_handler()
551 int dirty_bytes_handler(struct ctl_table *table, int write, in dirty_bytes_handler() argument
[all …]
Dhugetlb_cgroup.c632 cft->write = hugetlb_cgroup_write_dfl; in __hugetlb_cgroup_file_dfl_init()
640 cft->write = hugetlb_cgroup_write_dfl; in __hugetlb_cgroup_file_dfl_init()
696 cft->write = hugetlb_cgroup_write_legacy; in __hugetlb_cgroup_file_legacy_init()
703 cft->write = hugetlb_cgroup_write_legacy; in __hugetlb_cgroup_file_legacy_init()
721 cft->write = hugetlb_cgroup_reset; in __hugetlb_cgroup_file_legacy_init()
728 cft->write = hugetlb_cgroup_reset; in __hugetlb_cgroup_file_legacy_init()
735 cft->write = hugetlb_cgroup_reset; in __hugetlb_cgroup_file_legacy_init()
742 cft->write = hugetlb_cgroup_reset; in __hugetlb_cgroup_file_legacy_init()
Dhuge_memory.c773 pmd_t *pmd, pfn_t pfn, pgprot_t prot, bool write, in insert_pfn_pmd() argument
782 if (write) { in insert_pfn_pmd()
799 if (write) { in insert_pfn_pmd()
833 pgprot_t pgprot, bool write) in vmf_insert_pfn_pmd_prot() argument
861 insert_pfn_pmd(vma, addr, vmf->pmd, pfn, pgprot, write, pgtable); in vmf_insert_pfn_pmd_prot()
875 pud_t *pud, pfn_t pfn, pgprot_t prot, bool write) in insert_pfn_pud() argument
883 if (write) { in insert_pfn_pud()
899 if (write) { in insert_pfn_pud()
924 pgprot_t pgprot, bool write) in vmf_insert_pfn_pud_prot() argument
945 insert_pfn_pud(vma, addr, vmf->pud, pfn, pgprot, write); in vmf_insert_pfn_pud_prot()
[all …]
Dmemcontrol.c5051 .write = mem_cgroup_reset,
5057 .write = mem_cgroup_write,
5063 .write = mem_cgroup_write,
5069 .write = mem_cgroup_reset,
5078 .write = mem_cgroup_force_empty_write,
5087 .write = memcg_write_event_control,
5118 .write = mem_cgroup_write,
5129 .write = mem_cgroup_reset,
5135 .write = mem_cgroup_reset,
5148 .write = mem_cgroup_write,
[all …]
Dmadvise.c1103 int write; in do_madvise() local
1133 write = madvise_need_mmap_write(behavior); in do_madvise()
1134 if (write) { in do_madvise()
1187 if (write) in do_madvise()
Dvmstat.c78 int sysctl_vm_numa_stat_handler(struct ctl_table *table, int write, in sysctl_vm_numa_stat_handler() argument
84 if (write) in sysctl_vm_numa_stat_handler()
86 ret = proc_dointvec_minmax(table, write, buffer, length, ppos); in sysctl_vm_numa_stat_handler()
87 if (ret || !write) in sysctl_vm_numa_stat_handler()
1818 int vmstat_refresh(struct ctl_table *table, int write, in vmstat_refresh() argument
1860 if (write) in vmstat_refresh()
Dhugetlb.c3439 static int proc_hugetlb_doulongvec_minmax(struct ctl_table *table, int write, in proc_hugetlb_doulongvec_minmax() argument
3452 return proc_doulongvec_minmax(&dup_table, write, buffer, length, ppos); in proc_hugetlb_doulongvec_minmax()
3456 struct ctl_table *table, int write, in hugetlb_sysctl_handler_common() argument
3466 ret = proc_hugetlb_doulongvec_minmax(table, write, buffer, length, ppos, in hugetlb_sysctl_handler_common()
3471 if (write) in hugetlb_sysctl_handler_common()
3478 int hugetlb_sysctl_handler(struct ctl_table *table, int write, in hugetlb_sysctl_handler() argument
3482 return hugetlb_sysctl_handler_common(false, table, write, in hugetlb_sysctl_handler()
3487 int hugetlb_mempolicy_sysctl_handler(struct ctl_table *table, int write, in hugetlb_mempolicy_sysctl_handler() argument
3490 return hugetlb_sysctl_handler_common(true, table, write, in hugetlb_mempolicy_sysctl_handler()
3495 int hugetlb_overcommit_handler(struct ctl_table *table, int write, in hugetlb_overcommit_handler() argument
[all …]
Dpage_alloc.c5951 int numa_zonelist_order_handler(struct ctl_table *table, int write, in numa_zonelist_order_handler() argument
5954 if (write) in numa_zonelist_order_handler()
5956 return proc_dostring(table, write, buffer, length, ppos); in numa_zonelist_order_handler()
8274 int min_free_kbytes_sysctl_handler(struct ctl_table *table, int write, in postcore_initcall()
8279 rc = proc_dointvec_minmax(table, write, buffer, length, ppos); in postcore_initcall()
8283 if (write) { in postcore_initcall()
8290 int watermark_scale_factor_sysctl_handler(struct ctl_table *table, int write, in watermark_scale_factor_sysctl_handler() argument
8295 rc = proc_dointvec_minmax(table, write, buffer, length, ppos); in watermark_scale_factor_sysctl_handler()
8299 if (write) in watermark_scale_factor_sysctl_handler()
8320 int sysctl_min_unmapped_ratio_sysctl_handler(struct ctl_table *table, int write, in sysctl_min_unmapped_ratio_sysctl_handler() argument
[all …]
Dswap.c169 int get_kernel_pages(const struct kvec *kiov, int nr_segs, int write, in get_kernel_pages() argument
197 int get_kernel_page(unsigned long start, int write, struct page **pages) in get_kernel_page() argument
204 return get_kernel_pages(&kiov, 1, write, pages); in get_kernel_page()
Dgup.c954 int write = (gup_flags & FOLL_WRITE); in check_vma_flags() local
966 if (write) { in check_vma_flags()
996 if (!arch_vma_access_permitted(vma, write, false, foreign)) in check_vma_flags()
1193 bool write = !!(fault_flags & FAULT_FLAG_WRITE); in vma_permits_fault() local
1195 vm_flags_t vm_flags = write ? VM_WRITE : VM_READ; in vma_permits_fault()
1207 if (!arch_vma_access_permitted(vma, write, false, foreign)) in vma_permits_fault()
Dcompaction.c2662 int compaction_proactiveness_sysctl_handler(struct ctl_table *table, int write, in compaction_proactiveness_sysctl_handler() argument
2667 rc = proc_dointvec_minmax(table, write, buffer, length, ppos); in compaction_proactiveness_sysctl_handler()
2671 if (write && sysctl_compaction_proactiveness) { in compaction_proactiveness_sysctl_handler()
2690 int sysctl_compaction_handler(struct ctl_table *table, int write, in sysctl_compaction_handler() argument
2693 if (write) in sysctl_compaction_handler()
Dmemory.c4106 bool write = vmf->flags & FAULT_FLAG_WRITE; in do_set_pmd() local
4138 if (write) in do_set_pmd()
4170 bool write = vmf->flags & FAULT_FLAG_WRITE; in do_set_pte() local
4182 if (write) in do_set_pte()
4185 if (write && !(vmf->vma_flags & VM_SHARED)) { in do_set_pte()
5575 void *buf, int len, int write) in generic_access_phys() argument
5582 if (follow_phys(vma, addr, write, &prot, &phys_addr)) in generic_access_phys()
5589 if (write) in generic_access_phys()
5609 int write = gup_flags & FOLL_WRITE; in __access_remote_vm() local
5635 len, write); in __access_remote_vm()
[all …]
Dnommu.c1691 int write = gup_flags & FOLL_WRITE; in __access_remote_vm() local
1704 if (write && vma->vm_flags & VM_MAYWRITE) in __access_remote_vm()
1707 else if (!write && vma->vm_flags & VM_MAYREAD) in __access_remote_vm()
Dkmemleak.c1847 .write = kmemleak_write,
/mm/damon/
Ddbgfs.c621 .write = dbgfs_attrs_write,
627 .write = dbgfs_schemes_write,
633 .write = dbgfs_target_ids_write,
639 .write = dbgfs_init_regions_write,
938 .write = dbgfs_mk_context_write,
942 .write = dbgfs_rm_context_write,
947 .write = dbgfs_monitor_on_write,