/mm/ |
D | page_reporting.c | 26 unsigned int state; in __page_reporting_request() local 29 state = atomic_read(&prdev->state); in __page_reporting_request() 30 if (state == PAGE_REPORTING_REQUESTED) in __page_reporting_request() 37 state = atomic_xchg(&prdev->state, PAGE_REPORTING_REQUESTED); in __page_reporting_request() 38 if (state != PAGE_REPORTING_IDLE) in __page_reporting_request() 157 atomic_set(&prdev->state, PAGE_REPORTING_REQUESTED); in page_reporting_cycle() 275 int err = 0, state = PAGE_REPORTING_ACTIVE; in page_reporting_process() local 285 atomic_set(&prdev->state, state); in page_reporting_process() 307 state = atomic_cmpxchg(&prdev->state, state, PAGE_REPORTING_IDLE); in page_reporting_process() 308 if (state == PAGE_REPORTING_REQUESTED) in page_reporting_process() [all …]
|
D | backing-dev.c | 100 !list_empty(&bdi->bdi_list), bdi->wb.state); in bdi_debug_stats_show() 270 if (test_bit(WB_registered, &wb->state)) in wb_wakeup_delayed() 339 if (!test_and_clear_bit(WB_registered, &wb->state)) { in wb_shutdown() 485 set_bit(WB_registered, &wb->state); in cgwb_create() 495 if (test_bit(WB_registered, &bdi->wb.state) && in cgwb_create() 625 WARN_ON(test_bit(WB_registered, &bdi->wb.state)); in cgwb_bdi_unregister() 826 set_bit(WB_registered, &bdi->wb.state); in bdi_register_va() 907 if (test_bit(WB_registered, &bdi->wb.state)) in release_bdi()
|
D | slab_common.c | 882 static void freelist_randomize(struct rnd_state *state, unsigned int *list, in freelist_randomize() argument 893 rand = prandom_u32_state(state); in freelist_randomize() 903 struct rnd_state state; in cache_random_seq_create() local 913 prandom_seed_state(&state, get_random_long()); in cache_random_seq_create() 915 freelist_randomize(&state, cachep->random_seq, count); in cache_random_seq_create()
|
D | slab.c | 2386 static bool freelist_state_initialize(union freelist_init_state *state, in freelist_state_initialize() argument 2398 prandom_seed_state(&state->rnd_state, rand); in freelist_state_initialize() 2401 state->list = cachep->random_seq; in freelist_state_initialize() 2402 state->count = count; in freelist_state_initialize() 2403 state->pos = rand % count; in freelist_state_initialize() 2410 static freelist_idx_t next_random_slot(union freelist_init_state *state) in next_random_slot() argument 2412 if (state->pos >= state->count) in next_random_slot() 2413 state->pos = 0; in next_random_slot() 2414 return state->list[state->pos++]; in next_random_slot() 2431 union freelist_init_state state; in shuffle_freelist() local [all …]
|
D | Kconfig.debug | 21 Also, the state of page tracking structures is checked more often as 22 pages are being allocated and freed, as unexpected state changes
|
D | Kconfig | 176 'online' state by default. 178 memory blocks in 'offline' state. 764 bit writers can set the state of the bit in the flags so that PTE
|
D | filemap.c | 1202 struct page *page, int bit_nr, int state, enum behavior behavior) in wait_on_page_bit_common() argument 1274 set_current_state(state); in wait_on_page_bit_common() 1279 if (signal_pending_state(state, current)) in wait_on_page_bit_common()
|
D | page_alloc.c | 1667 struct mminit_pfnnid_cache *state) in __early_pfn_to_nid() argument 1672 if (state->last_start <= pfn && pfn < state->last_end) in __early_pfn_to_nid() 1673 return state->last_nid; in __early_pfn_to_nid() 1677 state->last_start = start_pfn; in __early_pfn_to_nid() 1678 state->last_end = end_pfn; in __early_pfn_to_nid() 1679 state->last_nid = nid; in __early_pfn_to_nid()
|
D | compaction.c | 1911 return pgdat->kswapd && (pgdat->kswapd->state == TASK_RUNNING); in kswapd_is_running()
|
/mm/kasan/ |
D | sw_tags.c | 62 u32 state = this_cpu_read(prng_state); in kasan_random_tag() local 64 state = 1664525 * state + 1013904223; in kasan_random_tag() 65 this_cpu_write(prng_state, state); in kasan_random_tag() 67 return (u8)(state % (KASAN_TAG_MAX + 1)); in kasan_random_tag()
|
D | kasan.h | 294 #define arch_set_tagging_report_once(state) argument 311 #define hw_set_tagging_report_once(state) arch_set_tagging_report_once(state) argument 322 #define hw_set_tagging_report_once(state) argument 328 void kasan_set_tagging_report_once(bool state); 334 static inline void kasan_set_tagging_report_once(bool state) { } in kasan_set_tagging_report_once() argument
|
D | hw_tags.c | 273 void kasan_set_tagging_report_once(bool state) in kasan_set_tagging_report_once() argument 275 hw_set_tagging_report_once(state); in kasan_set_tagging_report_once()
|
/mm/kfence/ |
D | core.c | 205 WRITE_ONCE(meta->state, next); in metadata_update_state() 294 if (meta->state == KFENCE_OBJECT_FREED) in kfence_guarded_alloc() 358 if (meta->state != KFENCE_OBJECT_ALLOCATED || meta->addr != (unsigned long)addr) { in kfence_guarded_free() 476 meta->state = KFENCE_OBJECT_UNUSED; in kfence_init_pool() 695 READ_ONCE(meta->state) != KFENCE_OBJECT_ALLOCATED) in kfence_shutdown_cache() 699 in_use = meta->cache == s && meta->state == KFENCE_OBJECT_ALLOCATED; in kfence_shutdown_cache() 725 if (READ_ONCE(meta->cache) != s || READ_ONCE(meta->state) != KFENCE_OBJECT_FREED) in kfence_shutdown_cache() 729 if (meta->cache == s && meta->state == KFENCE_OBJECT_FREED) in kfence_shutdown_cache() 839 if (meta && READ_ONCE(meta->state) == KFENCE_OBJECT_ALLOCATED) { in kfence_handle_page_fault() 846 if (meta && READ_ONCE(meta->state) == KFENCE_OBJECT_ALLOCATED) { in kfence_handle_page_fault()
|
D | report.c | 124 if (meta->state == KFENCE_OBJECT_UNUSED) { in kfence_print_object() 136 if (meta->state == KFENCE_OBJECT_FREED) { in kfence_print_object()
|
D | kfence.h | 58 enum kfence_object_state state; member
|