/tools/include/asm-generic/ |
D | barrier.h | 27 #ifndef mb 28 #define mb() barrier() macro 32 #define rmb() mb() 36 #define wmb() mb()
|
/tools/perf/ui/browsers/ |
D | map.c | 28 struct map_browser *mb = container_of(browser, struct map_browser, b); in map_browser__write() local 34 mb->addrlen, sym->start, mb->addrlen, sym->end, in map_browser__write() 37 width = browser->width - ((mb->addrlen * 2) + 4); in map_browser__write() 107 struct map_browser mb = { in map__browse() local 120 for (nd = rb_first(mb.b.entries); nd; nd = rb_next(nd)) { in map__browse() 127 *idx = mb.b.nr_entries; in map__browse() 129 ++mb.b.nr_entries; in map__browse() 132 mb.addrlen = snprintf(tmp, sizeof(tmp), "%" PRIx64, maxaddr); in map__browse() 133 return map_browser__run(&mb); in map__browse()
|
/tools/memory-model/ |
D | linux-kernel.def | 17 smp_store_mb(X,V) { __store{once}(X,V); __fence{mb}; } 20 smp_mb() { __fence{mb}; } 30 xchg(X,V) __xchg{mb}(X,V) 34 cmpxchg(X,V,W) __cmpxchg{mb}(X,V,W) 68 atomic_add_return(V,X) __atomic_op_return{mb}(X,+,V) 72 atomic_fetch_add(V,X) __atomic_fetch_op{mb}(X,+,V) 77 atomic_inc_return(X) __atomic_op_return{mb}(X,+,1) 81 atomic_fetch_inc(X) __atomic_fetch_op{mb}(X,+,1) 86 atomic_sub_return(V,X) __atomic_op_return{mb}(X,-,V) 90 atomic_fetch_sub(V,X) __atomic_fetch_op{mb}(X,-,V) [all …]
|
D | linux-kernel.cfg | 19 edgeattr mb,color,darkgreen
|
D | linux-kernel.bell | 26 'mb (*smp_mb*) ||
|
/tools/arch/mips/include/asm/ |
D | barrier.h | 11 #define mb() asm volatile( \ macro 18 #define wmb() mb() 19 #define rmb() mb()
|
/tools/arch/sh/include/asm/ |
D | barrier.h | 26 #define mb() __asm__ __volatile__ ("synco": : :"memory") macro 27 #define rmb() mb() 28 #define wmb() mb()
|
/tools/arch/ia64/include/asm/ |
D | barrier.h | 45 #define mb() ia64_mf() macro 46 #define rmb() mb() 47 #define wmb() mb()
|
/tools/testing/selftests/powerpc/mm/ |
D | segv_errors.c | 55 mb(); in test_segv_errors() 57 mb(); in test_segv_errors() 65 mb(); in test_segv_errors() 67 mb(); in test_segv_errors()
|
/tools/arch/s390/include/asm/ |
D | barrier.h | 26 #define mb() do { asm volatile(__ASM_BARRIER : : : "memory"); } while (0) macro 28 #define rmb() mb() 29 #define wmb() mb()
|
/tools/arch/arm64/include/asm/ |
D | barrier.h | 13 #define mb() asm volatile("dmb ish" ::: "memory") macro 59 mb(); \ 92 mb(); \
|
/tools/arch/xtensa/include/asm/ |
D | barrier.h | 14 #define mb() ({ __asm__ __volatile__("memw" : : : "memory"); }) macro 16 #define wmb() mb()
|
/tools/arch/x86/include/asm/ |
D | barrier.h | 20 #define mb() asm volatile("lock; addl $0,0(%%esp)" ::: "memory") macro 24 #define mb() asm volatile("mfence" ::: "memory") macro
|
/tools/testing/selftests/powerpc/pmu/ebb/ |
D | ebb.c | 152 mb(); in setup_ebb_handler() 156 mb(); in setup_ebb_handler() 280 mb(); in ebb_event_enable() 289 mb(); in ebb_event_enable() 297 mb(); in ebb_freeze_pmcs() 304 mb(); in ebb_unfreeze_pmcs() 311 mb(); in ebb_global_enable() 318 mb(); in ebb_global_disable()
|
D | cycles_with_freeze_test.c | 83 mb(); in cycles_with_freeze() 89 mb(); in cycles_with_freeze()
|
D | instruction_count_test.c | 34 mb(); in do_count_loop() 40 mb(); in do_count_loop()
|
D | no_handler_test.c | 43 mb(); in no_handler_test()
|
/tools/arch/arm/include/asm/ |
D | barrier.h | 8 #define mb() ((void(*)(void))0xffff0fa0)() macro
|
/tools/arch/alpha/include/asm/ |
D | barrier.h | 5 #define mb() __asm__ __volatile__("mb": : :"memory") macro
|
/tools/arch/sparc/include/asm/ |
D | barrier_64.h | 39 #define mb() membar_safe("#StoreLoad") macro
|
/tools/arch/powerpc/include/asm/ |
D | barrier.h | 26 #define mb() __asm__ __volatile__ ("sync" : : : "memory") macro
|
/tools/virtio/asm/ |
D | barrier.h | 15 # define mb() abort() macro
|
/tools/include/asm/ |
D | barrier.h | 47 # define smp_mb() mb()
|
/tools/testing/selftests/powerpc/include/ |
D | reg.h | 19 #define mb() asm volatile("sync" : : : "memory"); macro
|
/tools/testing/selftests/net/ |
D | tcp_mmap.c | 207 unsigned long mb = total >> 20; in child_thread() local 218 (double)total_usec/mb, in child_thread()
|