/tools/perf/util/ |
D | block-range.c | 25 assert(old < entry->start); in block_range__debug() 26 assert(entry->start <= entry->end); /* single instruction block; jump to a jump */ in block_range__debug() 43 if (addr < entry->start) in block_range__find() 81 struct block_range_iter block_range__create(u64 start, u64 end) in block_range__create() argument 92 if (start < entry->start) in block_range__create() 94 else if (start > entry->end) in block_range__create() 112 if (entry->end < start) { in block_range__create() 119 if (next->start <= end) { /* add head: [start...][n->start...] */ in block_range__create() 125 .start = start, in block_range__create() 126 .end = next->start - 1, in block_range__create() [all …]
|
D | svghelper.h | 9 void open_svg(const char *filename, int cpus, int rows, u64 start, u64 end); 10 void svg_ubox(int Yslot, u64 start, u64 end, double height, const char *type, int fd, int err, int … 11 void svg_lbox(int Yslot, u64 start, u64 end, double height, const char *type, int fd, int err, int … 12 void svg_fbox(int Yslot, u64 start, u64 end, double height, const char *type, int fd, int err, int … 13 void svg_box(int Yslot, u64 start, u64 end, const char *type); 14 void svg_blocked(int Yslot, int cpu, u64 start, u64 end, const char *backtrace); 15 void svg_running(int Yslot, int cpu, u64 start, u64 end, const char *backtrace); 16 void svg_waiting(int Yslot, int cpu, u64 start, u64 end, const char *backtrace); 20 void svg_process(int cpu, u64 start, u64 end, int pid, const char *name, const char *backtrace); 21 void svg_cstate(int cpu, u64 start, u64 end, int type); [all …]
|
D | svghelper.c | 85 void open_svg(const char *filename, int cpus, int rows, u64 start, u64 end) in open_svg() argument 94 first_time = start; in open_svg() 153 void svg_ubox(int Yslot, u64 start, u64 end, double height, const char *type, int fd, int err, int … in svg_ubox() argument 155 double w = time2pixels(end) - time2pixels(start); in svg_ubox() 164 time2pixels(start), in svg_ubox() 172 void svg_lbox(int Yslot, u64 start, u64 end, double height, const char *type, int fd, int err, int … in svg_lbox() argument 174 double w = time2pixels(end) - time2pixels(start); in svg_lbox() 183 time2pixels(start), in svg_lbox() 191 void svg_fbox(int Yslot, u64 start, u64 end, double height, const char *type, int fd, int err, int … in svg_fbox() argument 193 double w = time2pixels(end) - time2pixels(start); in svg_fbox() [all …]
|
D | block-range.h | 27 u64 start; member 47 struct block_range *start; member 53 return iter->start; in block_range_iter() 58 if (iter->start == iter->end) in block_range_iter__next() 61 iter->start = block_range__next(iter->start); in block_range_iter__next() 67 if (!iter->start || !iter->end) in block_range_iter__valid() 73 extern struct block_range_iter block_range__create(u64 start, u64 end);
|
D | time-utils.c | 55 (parse_nsec_time(start_str, &ptime->start) != 0)) { in parse_timestr_sec_nsec() 67 static int split_start_end(char **start, char **end, const char *ostr, char ch) in split_start_end() argument 88 *start = start_str; in split_start_end() 103 ptime->start = 0; in perf_time__parse_str() 111 if (rc == 0 && ptime->end && ptime->end < ptime->start) in perf_time__parse_str() 114 pr_debug("start time %" PRIu64 ", ", ptime->start); in perf_time__parse_str() 167 if (ptime[i].end >= ptime[i + 1].start) { in perf_time__parse_strs() 200 double end_pcnt, u64 start, u64 end) in set_percent_time() argument 202 u64 total = end - start; in set_percent_time() 209 ptime->start = start + round(start_pcnt * total); in set_percent_time() [all …]
|
D | mem2node.c | 13 u64 start; member 28 if (entry->start < e->start) in phys_entry__insert() 39 phys_entry__init(struct phys_entry *entry, u64 start, u64 bsize, u64 node) in phys_entry__init() argument 41 entry->start = start; in phys_entry__init() 42 entry->end = start + bsize; in phys_entry__init() 72 u64 start; in mem2node__init() local 77 start = bit * bsize; in mem2node__init() 86 if ((prev->end == start) && in mem2node__init() 93 phys_entry__init(&entries[j++], start, bsize, n->node); in mem2node__init() 104 entries[i].node, entries[i].start, entries[i].end); in mem2node__init() [all …]
|
/tools/lib/ |
D | find_bit.c | 33 unsigned long start, unsigned long invert) in _find_next_bit() argument 37 if (unlikely(start >= nbits)) in _find_next_bit() 40 tmp = addr1[start / BITS_PER_LONG]; in _find_next_bit() 42 tmp &= addr2[start / BITS_PER_LONG]; in _find_next_bit() 46 tmp &= BITMAP_FIRST_WORD_MASK(start); in _find_next_bit() 47 start = round_down(start, BITS_PER_LONG); in _find_next_bit() 50 start += BITS_PER_LONG; in _find_next_bit() 51 if (start >= nbits) in _find_next_bit() 54 tmp = addr1[start / BITS_PER_LONG]; in _find_next_bit() 56 tmp &= addr2[start / BITS_PER_LONG]; in _find_next_bit() [all …]
|
D | string.c | 172 static void *check_bytes8(const u8 *start, u8 value, unsigned int bytes) in check_bytes8() argument 175 if (*start != value) in check_bytes8() 176 return (void *)start; in check_bytes8() 177 start++; in check_bytes8() 192 void *memchr_inv(const void *start, int c, size_t bytes) in memchr_inv() argument 199 return check_bytes8(start, value, bytes); in memchr_inv() 206 prefix = (unsigned long)start % 8; in memchr_inv() 211 r = check_bytes8(start, value, prefix); in memchr_inv() 214 start += prefix; in memchr_inv() 221 if (*(u64 *)start != value64) in memchr_inv() [all …]
|
/tools/testing/nvdimm/test/ |
D | iomap.c | 72 - nfit_res->res.start; in __nfit_test_ioremap() 83 - nfit_res->res.start; in __wrap_devm_ioremap() 94 return nfit_res->buf + offset - nfit_res->res.start; in __wrap_devm_memremap() 129 resource_size_t offset = pgmap->range.start; in __wrap_devm_memremap_pages() 155 return nfit_res->buf + offset - nfit_res->res.start; in __wrap_devm_memremap_pages() 175 return nfit_res->buf + offset - nfit_res->res.start; in __wrap_memremap() 222 struct resource *parent, resource_size_t start, 229 WARN_ON(!nfit_test_release_region(NULL, &iomem_resource, res->start, in nfit_devres_release() 236 resource_size_t start = *((resource_size_t *) match_data); in match() local 238 return res->start == start; in match() [all …]
|
/tools/testing/radix-tree/ |
D | benchmark.c | 18 struct timespec start, finish; in benchmark_iter() local 26 clock_gettime(CLOCK_MONOTONIC, &start); in benchmark_iter() 38 nsec = (finish.tv_sec - start.tv_sec) * NSEC_PER_SEC + in benchmark_iter() 39 (finish.tv_nsec - start.tv_nsec); in benchmark_iter() 55 struct timespec start, finish; in benchmark_insert() local 59 clock_gettime(CLOCK_MONOTONIC, &start); in benchmark_insert() 66 nsec = (finish.tv_sec - start.tv_sec) * NSEC_PER_SEC + in benchmark_insert() 67 (finish.tv_nsec - start.tv_nsec); in benchmark_insert() 76 struct timespec start, finish; in benchmark_tagging() local 80 clock_gettime(CLOCK_MONOTONIC, &start); in benchmark_tagging() [all …]
|
D | main.c | 50 unsigned long start; in __big_gang_check() local 53 start = 0; in __big_gang_check() 58 __gang_check(start, rand() % 113 + 1, rand() % 71, in __big_gang_check() 60 old_start = start; in __big_gang_check() 61 start += rand() % 1000000; in __big_gang_check() 62 start %= 1ULL << 33; in __big_gang_check() 63 if (start < old_start) in __big_gang_check() 126 void check_copied_tags(struct radix_tree_root *tree, unsigned long start, unsigned long end, unsign… in check_copied_tags() argument 133 if (idx[i] < start || idx[i] > end) { in check_copied_tags() 135 printv(2, "%lu-%lu: %lu, tags %d-%d\n", start, in check_copied_tags() [all …]
|
D | bitmap.c | 5 void bitmap_clear(unsigned long *map, unsigned int start, int len) in bitmap_clear() argument 7 unsigned long *p = map + BIT_WORD(start); in bitmap_clear() 8 const unsigned int size = start + len; in bitmap_clear() 9 int bits_to_clear = BITS_PER_LONG - (start % BITS_PER_LONG); in bitmap_clear() 10 unsigned long mask_to_clear = BITMAP_FIRST_WORD_MASK(start); in bitmap_clear()
|
/tools/perf/pmu-events/ |
D | jsmn.c | 39 tok->start = tok->end = -1; in jsmn_alloc_token() 48 int start, int end) in jsmn_fill_token() argument 51 token->start = start; in jsmn_fill_token() 64 int start; in jsmn_parse_primitive() local 66 start = parser->pos; in jsmn_parse_primitive() 89 parser->pos = start; in jsmn_parse_primitive() 98 parser->pos = start; in jsmn_parse_primitive() 105 parser->pos = start; in jsmn_parse_primitive() 108 jsmn_fill_token(token, JSMN_PRIMITIVE, start, parser->pos); in jsmn_parse_primitive() 121 int start = parser->pos; in jsmn_parse_string() local [all …]
|
/tools/testing/selftests/netfilter/ |
D | nft_concat_range.sh | 994 __start="$(eval format_"${f}" "${start}")" 1029 __expr="${__expr}$(eval format_"${f}" "${start}")" 1032 __expr="${__expr} . $(eval format_"${f}" "${start}")" 1041 __start="$(eval format_"${f}" "${start}")" 1205 for i in $(seq "${start}" $((start + count))); do 1206 end=$((start + range_size)) 1209 if [ $((end / 65534)) -gt $((start / 65534)) ]; then 1210 start=${end} 1213 srcstart=$((start + src_delta)) 1217 for j in $(seq ${start} $((range_size / 2 + 1)) ${end}); do [all …]
|
/tools/perf/ |
D | perf-read-vdso.c | 15 void *start, *end; in main() local 18 if (find_map(&start, &end, VDSO__MAP_NAME)) in main() 21 size = end - start; in main() 24 written = fwrite(start, 1, size, stdout); in main() 27 start += written; in main()
|
/tools/perf/arch/s390/util/ |
D | machine.c | 11 int arch__fix_module_text_start(u64 *start, u64 *size, const char *name) in arch__fix_module_text_start() argument 13 u64 m_start = *start; in arch__fix_module_text_start() 18 if (sysfs__read_ull(path, (unsigned long long *)start) < 0) { in arch__fix_module_text_start() 20 *start = m_start; in arch__fix_module_text_start() 32 *size -= (*start - m_start); in arch__fix_module_text_start()
|
/tools/testing/selftests/bpf/ |
D | trace_helpers.c | 53 int start = 0, end = sym_cnt; in ksym_search() local 60 while (start < end) { in ksym_search() 61 size_t mid = start + (end - start) / 2; in ksym_search() 67 start = mid + 1; in ksym_search() 72 if (start >= 1 && syms[start - 1].addr < key && in ksym_search() 73 key < syms[start].addr) in ksym_search() 75 return &syms[start - 1]; in ksym_search()
|
/tools/testing/selftests/timens/ |
D | gettime_perf.c | 41 struct timespec tp, start; in test() local 45 vdso_clock_gettime(clockid, &start); in test() 46 tp = start; in test() 47 for (tp = start; start.tv_sec + timeout > tp.tv_sec || in test() 48 (start.tv_sec + timeout == tp.tv_sec && in test() 49 start.tv_nsec > tp.tv_nsec); i++) { in test()
|
/tools/testing/selftests/timers/ |
D | posix_timers.c | 62 static int check_diff(struct timeval start, struct timeval end) in check_diff() argument 66 diff = end.tv_usec - start.tv_usec; in check_diff() 67 diff += (end.tv_sec - start.tv_sec) * USECS_PER_SEC; in check_diff() 80 struct timeval start, end; in check_itimer() local 105 err = gettimeofday(&start, NULL); in check_itimer() 130 if (!check_diff(start, end)) in check_itimer() 142 struct timeval start, end; in check_timer_create() local 163 err = gettimeofday(&start, NULL); in check_timer_create() 183 if (!check_diff(start, end)) in check_timer_create()
|
D | raw_skew.c | 60 long long diff_timespec(struct timespec start, struct timespec end) in diff_timespec() argument 64 start_ns = ts_to_nsec(start); in diff_timespec() 71 struct timespec start, mid, end; in get_monotonic_and_raw() local 78 clock_gettime(CLOCK_MONOTONIC, &start); in get_monotonic_and_raw() 82 newdiff = diff_timespec(start, end); in get_monotonic_and_raw() 86 tmp = (ts_to_nsec(start) + ts_to_nsec(end))/2; in get_monotonic_and_raw() 94 struct timespec mon, raw, start, end; in main() local 108 start = mon; in main() 124 interval = diff_timespec(start, end); in main()
|
/tools/perf/scripts/python/ |
D | net_dropmonitor.py | 38 start, end = -1, len(kallsyms) 39 while end != start + 1: 40 pivot = (start + end) // 2 44 start = pivot 48 if start >= 0: 49 symloc, name = kallsyms[start]
|
/tools/perf/tests/ |
D | maps.c | 12 u64 start; member 25 TEST_ASSERT_VAL("wrong map start", map->start == merged[i].start); in check_maps() 75 map->start = bpf_progs[i].start; in test__maps__merge_in() 91 map_kcore1->start = 100; in test__maps__merge_in() 95 map_kcore2->start = 550; in test__maps__merge_in() 99 map_kcore3->start = 880; in test__maps__merge_in()
|
/tools/testing/selftests/vm/ |
D | mlock2.h | 15 static int mlock2_(void *start, size_t len, int flags) in mlock2_() argument 18 return syscall(__NR_mlock2, start, len, flags); in mlock2_() 30 unsigned long start, end; in seek_to_smaps_entry() local 45 &start, &end, perms, &offset, dev, &inode, path) < 6) in seek_to_smaps_entry() 48 if (start <= addr && addr < end) in seek_to_smaps_entry()
|
D | mlock2-tests.c | 15 unsigned long start; member 26 unsigned long start; in get_vm_area() local 55 sscanf(line, "%lx", &start); in get_vm_area() 58 if (start <= addr && end > addr) { in get_vm_area() 59 area->start = start; in get_vm_area() 447 if (page1.start != page2.start || page2.start != page3.start) { in test_vma_management() 466 if (page1.start == page2.start || page2.start == page3.start) { in test_vma_management() 485 if (page1.start != page2.start || page2.start != page3.start) { in test_vma_management()
|
/tools/perf/arch/x86/util/ |
D | machine.c | 25 static int add_extra_kernel_map(struct extra_kernel_map_info *mi, u64 start, in add_extra_kernel_map() argument 40 mi->maps[mi->cnt].start = start; in add_extra_kernel_map() 51 u64 start) in find_extra_kernel_maps() argument 57 mi->entry_trampoline = start; in find_extra_kernel_maps() 62 u64 end = start + page_size; in find_extra_kernel_maps() 64 return add_extra_kernel_map(mi, start, end, 0, name); in find_extra_kernel_maps()
|