/external/icu/android_icu4j/src/main/java/android/icu/text/ |
D | BidiLine.java | 42 * The implementation of the access to same-level-runs and of the reordering 48 * same-level-runs is created. Reordering then is done on this vector. 53 * This is inefficient if there are many very short runs. If the average run 135 lineBidi.runs = new BidiRun[0]; in setLine() 281 /* this is done based on runs rather than on levels since levels have in getLogicalRun() 288 iRun = bidi.runs[0]; in getLogicalRun() 291 iRun = bidi.runs[i]; in getLogicalRun() 307 int start = bidi.runs[runIndex].start; in getVisualRun() 309 byte level = bidi.runs[runIndex].level; in getVisualRun() 313 bidi.runs[runIndex].limit - in getVisualRun() [all …]
|
/external/icu/icu4j/main/classes/core/src/com/ibm/icu/text/ |
D | BidiLine.java | 41 * The implementation of the access to same-level-runs and of the reordering 47 * same-level-runs is created. Reordering then is done on this vector. 52 * This is inefficient if there are many very short runs. If the average run 134 lineBidi.runs = new BidiRun[0]; in setLine() 280 /* this is done based on runs rather than on levels since levels have in getLogicalRun() 287 iRun = bidi.runs[0]; in getLogicalRun() 290 iRun = bidi.runs[i]; in getLogicalRun() 306 int start = bidi.runs[runIndex].start; in getVisualRun() 308 byte level = bidi.runs[runIndex].level; in getVisualRun() 312 bidi.runs[runIndex].limit - in getVisualRun() [all …]
|
/external/autotest/client/site_tests/kernel_LTP/ |
D | ltp-diff.py | 11 # where the results of those runs differ 13 # 0 if all runs had identical results 32 Return dictionary keyed on unique testnames across all runs. 38 runs = {} 41 runs[i] = {} 54 runs[i][testname] = status 59 return (runs, testnames) 63 def compare_results(runs): argument 66 Print any testnames with differing results across runs. 67 Return 1 if any test results across runs differ. [all …]
|
/external/autotest/client/tests/ltp/ |
D | ltp-diff.py | 11 # where the results of those runs differ 13 # 0 if all runs had identical results 31 Return dictionary keyed on unique testnames across all runs. 35 runs = {} 38 runs[i] = {} 51 runs[i][testname] = status 56 return (runs, testnames) 60 def compare_results(runs): argument 63 Print any testnames with differing results across runs. 64 Return 1 if any test results across runs differ. [all …]
|
/external/icu/icu4c/source/common/ |
D | ubidiln.c | 48 * The implementation of the access to same-level-runs and of the reordering 54 * same-level-runs is created. Reordering then is done on this vector. 59 * This is inefficient if there are many very short runs. If the average run 155 pLineBiDi->runs=NULL; in ubidi_setLine() 318 /* this is done based on runs rather than on levels since levels have in ubidi_getLogicalRun() 322 iRun=pBiDi->runs[0]; in ubidi_getLogicalRun() 325 iRun = pBiDi->runs[i]; in ubidi_getLogicalRun() 349 /* runs API functions ------------------------------------------------------- */ 375 start=pBiDi->runs[runIndex].logicalStart; in ubidi_getVisualRun() 381 *pLength=pBiDi->runs[runIndex].visualLimit- in ubidi_getVisualRun() [all …]
|
/external/skia/src/core/ |
D | SkRegionPriv.h | 30 static int compute_intervalcount(const SkRegion::RunType runs[]) { in compute_intervalcount() argument 31 const SkRegion::RunType* curr = runs; in compute_intervalcount() 37 return SkToInt((curr - runs) >> 1); in compute_intervalcount() 131 * Given a scanline (including its Bottom value at runs[0]), return the next 132 * scanline. Asserts that there is one (i.e. runs[0] < Sentinel) 134 static SkRegion::RunType* SkipEntireScanline(const SkRegion::RunType runs[]) { in SkipEntireScanline() 136 SkASSERT(runs[0] < SkRegion::kRunTypeSentinel); in SkipEntireScanline() 138 const int intervals = runs[1]; in SkipEntireScanline() 139 SkASSERT(runs[2 + intervals * 2] == SkRegion::kRunTypeSentinel); in SkipEntireScanline() 142 int n = compute_intervalcount(&runs[2]); in SkipEntireScanline() [all …]
|
D | SkAntiRun.h | 60 int16_t* runs = fRuns + offsetX; in add() local 66 SkAlphaRuns::Break(runs, alpha, x, 1); in add() 76 runs += x + 1; in add() 83 SkAlphaRuns::Break(runs, alpha, x, middleCount); in add() 85 runs += x; in add() 89 int n = runs[0]; in add() 92 runs += n; in add() 100 SkAlphaRuns::Break(runs, alpha, x, 1); in add() 114 * Break the runs in the buffer at offsets x and x+count, properly 115 * updating the runs to the right and left. [all …]
|
D | SkRegion.cpp | 32 static SkRegion::RunType* skip_intervals(const SkRegion::RunType runs[]) { 33 int intervals = runs[-1]; 36 SkASSERT(runs[0] < runs[1]); 37 SkASSERT(runs[1] < SkRegion::kRunTypeSentinel); 40 SkASSERT(SkRegion::kRunTypeSentinel == runs[0]); 43 runs += intervals * 2 + 1; 44 return const_cast<SkRegion::RunType*>(runs); 47 bool SkRegion::RunsAreARect(const SkRegion::RunType runs[], int count, in RunsAreARect() argument 49 assert_sentinel(runs[0], false); // top in RunsAreARect() 53 assert_sentinel(runs[1], false); // bottom in RunsAreARect() [all …]
|
D | SkAlphaRuns.cpp | 30 const int16_t* runs = fRuns; in assertValid() local 33 while (*runs) { in assertValid() 35 alpha += *runs; in assertValid() 36 runs += *runs; in assertValid() 41 const int16_t* runs = fRuns; in dump() local 44 SkDebugf("Runs"); in dump() 45 while (*runs) { in dump() 46 int n = *runs; in dump() 53 runs += n; in dump() 62 const int16_t* runs = fRuns; in validate() local [all …]
|
D | SkBlitter.h | 37 /// Blit a horizontal run of antialiased pixels; runs[] is a *sparse* 39 /// The runs[] and antialias[] work together to represent long runs of pixels with the same 40 /// alphas. The runs[] contains the number of pixels with the same alpha, and antialias[] 41 /// contain the coverage value for that number of pixels. The runs[] (and antialias[]) are 42 /// encoded in a clever way. The runs array is zero terminated, and has enough entries for 44 /// in the runs array contains the number of pixels (np) that have the same alpha value. The 45 /// next np value is found np entries away. For example, if runs[0] = 7, then the next valid 46 /// entry will by at runs[7]. The runs array and antialias[] are coupled by index. So, if the 47 /// np entry is at runs[45] = 12 then the alpha value can be found at antialias[45] = 0x88. 49 virtual void blitAntiH(int x, int y, const SkAlpha antialias[], const int16_t runs[]) = 0; [all …]
|
D | SkBlitter.cpp | 39 const int16_t runs[]) { 48 int16_t runs[2]; in blitV() local 49 runs[0] = 1; in blitV() 50 runs[1] = 0; in blitV() 53 this->blitAntiH(x, y++, &alpha, runs); in blitV() 191 int16_t* runs = runStorage.get(); in blitMask() local 194 sk_memset16((uint16_t*)runs, 1, width); in blitMask() 195 runs[width] = 0; in blitMask() 200 this->blitAntiH(clip.fLeft, y, aa, runs); in blitMask() 248 const int16_t runs[]) {} in blitAntiH() argument [all …]
|
/external/trappy/trappy/ |
D | plot_utils.py | 124 def number_freq_plots(runs, map_label): argument 132 for run in runs: 143 def plot_temperature(runs, width=None, height=None, ylim="range", tz_id=None): argument 146 runs is an array of FTrace() instances. Extract the control_temp 155 for run in runs: 189 def plot_load(runs, map_label, width=None, height=None): argument 191 num_runs = len(runs) 199 for ax, run in zip(axis, runs): 203 def plot_allfreqs(runs, map_label, width=None, height=None): argument 205 num_runs = len(runs) [all …]
|
/external/llvm/unittests/Analysis/ |
D | CGSCCPassManagerTest.cpp | 35 TestModuleAnalysis(int &Runs) : Runs(Runs) {} in TestModuleAnalysis() argument 38 ++Runs; in run() 45 int &Runs; member in __anon40a790b50111::TestModuleAnalysis 60 TestSCCAnalysis(int &Runs) : Runs(Runs) {} in TestSCCAnalysis() argument 63 ++Runs; in run() 70 int &Runs; member in __anon40a790b50111::TestSCCAnalysis 85 TestFunctionAnalysis(int &Runs) : Runs(Runs) {} in TestFunctionAnalysis() argument 88 ++Runs; in run() 100 int &Runs; member in __anon40a790b50111::TestFunctionAnalysis 114 TestImmutableFunctionAnalysis(int &Runs) : Runs(Runs) {} in TestImmutableFunctionAnalysis() argument [all …]
|
/external/ltp/utils/benchmark/kernbench-0.42/ |
D | README | 12 It runs a kernel at various numbers of concurrent jobs: 1/2 number of cpus, 15 for the average of each group of runs and logs them to kernbench.log 37 kernbench [-n runs] [-o jobs] [-s] [-H] [-O] [-M] [-h] [-v] 40 s : perform single threaded runs (default don't) 41 H : don't perform half load runs (default do) 42 O : don't perform optimal load runs (default do) 43 M : don't perform maximal load runs (default do) 55 v0.40 Made all runs use the oldconfig if it exists. Changed to only do one 61 preparation and drops number of runs to 3. Modified half loads to 65 v0.20 Change to average of runs, add options to choose which runs to perform [all …]
|
/external/autotest/client/tests/compilebench/ |
D | compilebench.py | 53 # intial create total runs 10 avg 149.82 MB/s (user 0.63s sys 0.85s) 54 # create total runs 5 avg 27.50 MB/s (user 0.62s sys 0.83s) 55 # patch total runs 4 avg 15.01 MB/s (user 0.33s sys 0.63s) 56 # compile total runs 7 avg 41.47 MB/s (user 0.14s sys 0.75s) 57 # clean total runs 4 avg 697.77 MB/s (user 0.02s sys 0.08s) 58 # read tree total runs 2 avg 23.68 MB/s (user 0.85s sys 1.59s) 59 # read compiled tree total runs 1 avg 25.27 MB/s (user 0.98s sys 2.84s) 60 # delete tree total runs 2 avg 1.48 seconds (user 0.35s sys 0.45s) 61 # no runs for delete compiled tree 62 # stat tree total runs 4 avg 1.46 seconds (user 0.35s sys 0.26s) [all …]
|
/external/llvm/lib/Fuzzer/test/ |
D | fuzzer-dfsan.test | 11 RUN: not LLVMFuzzer-SimpleCmpTest-DFSan -use_traces=1 -seed=1 -runs=10000000 -timeout=5 2>&1 | File… 12 RUN: LLVMFuzzer-SimpleCmpTest-DFSan -use_traces=1 -seed=1 -runs=100 -timeout=5 -verbosity=3 2>&1 | … 14 RUN: not LLVMFuzzer-MemcmpTest-DFSan -use_traces=1 -seed=1 -runs=100000 -timeout=5 2>&1 | FileCheck… 15 RUN: LLVMFuzzer-MemcmpTest-DFSan -use_traces=1 -seed=1 -runs=2 -timeout=5 -verbosity=3 2>&1 | FileC… 17 RUN: not LLVMFuzzer-StrncmpTest-DFSan -use_traces=1 -seed=1 -runs=10000 -timeout=5 2>&1 | FileCheck… 18 RUN: LLVMFuzzer-StrncmpTest-DFSan -use_traces=1 -seed=1 -runs=2 -timeout=5 -verbosity=3 2>&1 | File… 20 RUN: not LLVMFuzzer-StrcmpTest-DFSan -use_traces=1 -seed=1 -runs=10000 -timeout=5 2>&1 | FileCheck … 21 RUN: LLVMFuzzer-StrcmpTest-DFSan -use_traces=1 -seed=1 -runs=2 -timeout=5 -verbosity=3 2>&1… 23 RUN: not LLVMFuzzer-SwitchTest-DFSan -use_traces=1 -seed=1 -runs=100000 -timeout=5 2>&… 24 RUN: LLVMFuzzer-SwitchTest-DFSan -use_traces=1 -seed=1 -runs=2 -timeout=5 -verbosity=3 2>&1…
|
D | fuzzer-traces-hooks.test | 6 Done1000000: Done 1000000 runs in 8 RUN: not LLVMFuzzer-MemcmpTest -seed=4294967295 -runs=100000 2>&1 | FileCheck %s 9 RUN: LLVMFuzzer-MemcmpTest -use_memcmp=0 -seed=4294967295 -runs=1000000 2>&1 | FileCheck %s --… 11 RUN: not LLVMFuzzer-StrncmpTest -seed=2 -runs=100000 2>&1 | FileCheck %s 12 RUN: LLVMFuzzer-StrncmpTest -use_memcmp=0 -seed=3 -runs=1000000 2>&1 | FileCheck %s --check-pr… 14 RUN: not LLVMFuzzer-StrcmpTest -seed=4 -runs=200000 2>&1 | FileCheck %s 15 RUN: LLVMFuzzer-StrcmpTest -use_memcmp=0 -seed=5 -runs=1000000 2>&1 | FileCheck %s --check-pre… 18 RUN: LLVMFuzzer-RepeatedMemcmp -seed=10 -runs=100000 2>&1 | FileCheck %s --check-prefix=RECOMMENDED…
|
D | fuzzer-leak.test | 2 RUN: not LLVMFuzzer-LeakTest -runs=100000 -detect_leaks=1 2>&1 | FileCheck %s --check-prefix=LEAK_D… 8 RUN: not LLVMFuzzer-LeakTest -runs=0 -detect_leaks=1 %S 2>&1 | FileCheck %s --check-prefix=LEAK_IN_… 13 RUN: not LLVMFuzzer-LeakTest -runs=100000 -detect_leaks=0 2>&1 | FileCheck %s --check-prefix=LEAK_A… 14 RUN: not LLVMFuzzer-LeakTest -runs=100000 2>&1 | FileCheck %s --check-prefix=LEAK_D… 15 LEAK_AFTER: Done 100000 runs in 18 RUN: not LLVMFuzzer-LeakTest -runs=100000 -max_len=1 2>&1 | FileCheck %s --check-prefix=MAX_LEN_1 25 RUN: LLVMFuzzer-AccumulateAllocationsTest -detect_leaks=1 -runs=100000 2>&1 | FileCheck %s --check-…
|
D | fuzzer-traces.test | 2 Done1000000: Done 1000000 runs in 3 RUN: not LLVMFuzzer-SimpleCmpTest -use_traces=1 -seed=1 -runs=10000001 2>&1 | FileCheck %s 5 RUN: not LLVMFuzzer-SwitchTest -use_traces=1 -seed=6 -runs=1000002 2>&1 | FileCheck %s 6 RUN: LLVMFuzzer-SwitchTest -seed=7 -runs=1000000 2>&1 | FileCheck %s --check-pre… 8 RUN: not LLVMFuzzer-SimpleHashTest -use_traces=1 -seed=8 -runs=1000000 -max_len=16 2>&1 | FileChec… 9 RUN: LLVMFuzzer-SimpleHashTest -seed=9 -runs=1000000 -max_len=16 2>&1 | FileCheck…
|
D | fuzzer-threaded.test | 1 CHECK: Done 1000 runs in 3 RUN: LLVMFuzzer-ThreadedTest -use_traces=1 -runs=1000 2>&1 | FileCheck %s 4 RUN: LLVMFuzzer-ThreadedTest -use_traces=1 -runs=1000 2>&1 | FileCheck %s 5 RUN: LLVMFuzzer-ThreadedTest -use_traces=1 -runs=1000 2>&1 | FileCheck %s 6 RUN: LLVMFuzzer-ThreadedTest -use_traces=1 -runs=1000 2>&1 | FileCheck %s
|
/external/apache-harmony/jdwp/src/test/java/org/apache/harmony/jpda/tests/jdwp/Events/ |
D | Breakpoint002Test.java | 35 * <BR>It runs Breakpoint002Debuggee and sets breakpoint in the 45 * <BR>It runs Breakpoint002Debuggee and sets breakpoint in the 56 * <BR>It runs Breakpoint002Debuggee and sets breakpoint in the 67 * <BR>It runs Breakpoint002Debuggee and sets breakpoint in the 78 * <BR>It runs Breakpoint002Debuggee and sets breakpoint in the 89 * <BR>It runs Breakpoint002Debuggee and sets breakpoint in the 100 * <BR>It runs Breakpoint002Debuggee and sets breakpoint in the 110 * <BR>It runs Breakpoint002Debuggee and sets breakpoint in the 120 * <BR>It runs Breakpoint002Debuggee and sets breakpoint in the 130 * <BR>It runs Breakpoint002Debuggee and sets breakpoint in the [all …]
|
/external/autotest/tko/ |
D | nightly.py | 50 runs = {} # platform --> list of test runs 64 runs.setdefault(platform, []).append(testrunx) 65 return runs 77 # partition all twoway runs based on name of antagonist progs 81 runs = {} 85 runs.setdefault(antag, []).append(testrunx) 87 twoway_runs[kernel] = runs 91 def collect_raw_scores(runs, metric): argument 92 # get unscaled scores of test runs for 1 test on certain jobs 95 for platform in runs: [all …]
|
/external/llvm/unittests/IR/ |
D | PassManagerTest.cpp | 29 TestFunctionAnalysis(int &Runs) : Runs(Runs) {} in TestFunctionAnalysis() argument 33 ++Runs; in run() 46 int &Runs; member in __anon832003490111::TestFunctionAnalysis 58 TestModuleAnalysis(int &Runs) : Runs(Runs) {} in TestModuleAnalysis() argument 61 ++Runs; in run() 72 int &Runs; member in __anon832003490111::TestModuleAnalysis 130 // Typical path just runs the analysis as needed. in run() 230 // Count the runs over a Function. in TEST_F() 249 // Count the runs over a module. in TEST_F() 253 // Count the runs over a Function in a separate manager. in TEST_F()
|
/external/autotest/client/deps/glbench/src/ |
D | testbase.h | 20 // Runs test->TestFunc() passing it sequential powers of two recording time it 21 // took until reaching a minimum amount of testing time. The last two runs are 25 // Runs Bench on an instance of TestBase and prints out results. 46 // Runs the test case n times. 68 // Runs the test and reports results in mpixels per second, assuming each 71 // Runs the test and reports results in mpixels per second, assuming each 75 // Runs the test three times: with blending on; with depth test enabled and
|
/external/swiftshader/third_party/subzero/tests_lit/llvm2ice_tests/ |
D | icmp-with-zero.ll | 21 ; lines actually runs the output through the assembler. 42 ; lines actually runs the output through the assembler. 63 ; lines actually runs the output through the assembler. 84 ; lines actually runs the output through the assembler. 105 ; lines actually runs the output through the assembler. 126 ; lines actually runs the output through the assembler. 148 ; lines actually runs the output through the assembler. 170 ; lines actually runs the output through the assembler. 192 ; lines actually runs the output through the assembler. 214 ; lines actually runs the output through the assembler. [all …]
|