| /third_party/flutter/engine/flutter/third_party/txt/tests/ |
| D | FontCollectionItemizeTest.cpp | 87 std::vector<FontCollection::Run> runs; in TEST_F() local 94 itemize(collection, "'a' 'b' 'c' 'd' 'e'", kRegularStyle, &runs); in TEST_F() 95 ASSERT_EQ(1U, runs.size()); in TEST_F() 96 EXPECT_EQ(0, runs[0].start); in TEST_F() 97 EXPECT_EQ(5, runs[0].end); in TEST_F() 98 EXPECT_EQ(kLatinFont, getFontPath(runs[0])); in TEST_F() 99 EXPECT_FALSE(runs[0].fakedFont.fakery.isFakeBold()); in TEST_F() 100 EXPECT_FALSE(runs[0].fakedFont.fakery.isFakeItalic()); in TEST_F() 102 itemize(collection, "'a' 'b' 'c' 'd' 'e'", kItalicStyle, &runs); in TEST_F() 103 ASSERT_EQ(1U, runs.size()); in TEST_F() [all …]
|
| /third_party/icu/ohos_icu4j/src/main/java/ohos/global/icu/text/ |
| D | BidiLine.java | 42 * The implementation of the access to same-level-runs and of the reordering 48 * same-level-runs is created. Reordering then is done on this vector. 53 * This is inefficient if there are many very short runs. If the average run 135 lineBidi.runs = new BidiRun[0]; in setLine() 281 /* this is done based on runs rather than on levels since levels have in getLogicalRun() 288 iRun = bidi.runs[0]; in getLogicalRun() 291 iRun = bidi.runs[i]; in getLogicalRun() 307 int start = bidi.runs[runIndex].start; in getVisualRun() 309 byte level = bidi.runs[runIndex].level; in getVisualRun() 313 bidi.runs[runIndex].limit - in getVisualRun() [all …]
|
| /third_party/icu/icu4j/main/classes/core/src/com/ibm/icu/text/ |
| D | BidiLine.java | 41 * The implementation of the access to same-level-runs and of the reordering 47 * same-level-runs is created. Reordering then is done on this vector. 52 * This is inefficient if there are many very short runs. If the average run 134 lineBidi.runs = new BidiRun[0]; in setLine() 280 /* this is done based on runs rather than on levels since levels have in getLogicalRun() 287 iRun = bidi.runs[0]; in getLogicalRun() 290 iRun = bidi.runs[i]; in getLogicalRun() 306 int start = bidi.runs[runIndex].start; in getVisualRun() 308 byte level = bidi.runs[runIndex].level; in getVisualRun() 312 bidi.runs[runIndex].limit - in getVisualRun() [all …]
|
| /third_party/flutter/skia/src/core/ |
| D | SkAntiRun.h | 60 int16_t* runs = fRuns + offsetX; in add() local 66 SkAlphaRuns::Break(runs, alpha, x, 1); in add() 76 runs += x + 1; in add() 83 SkAlphaRuns::Break(runs, alpha, x, middleCount); in add() 85 runs += x; in add() 89 int n = runs[0]; in add() 92 runs += n; in add() 100 SkAlphaRuns::Break(runs, alpha, x, 1); in add() 114 * Break the runs in the buffer at offsets x and x+count, properly 115 * updating the runs to the right and left. [all …]
|
| D | SkRegionPriv.h | 46 static int compute_intervalcount(const SkRegionPriv::RunType runs[]) { in compute_intervalcount() argument 47 const SkRegionPriv::RunType* curr = runs; in compute_intervalcount() 53 return SkToInt((curr - runs) >> 1); in compute_intervalcount() 143 * Given a scanline (including its Bottom value at runs[0]), return the next 144 * scanline. Asserts that there is one (i.e. runs[0] < Sentinel) 146 static SkRegion::RunType* SkipEntireScanline(const SkRegion::RunType runs[]) { in SkipEntireScanline() 148 SkASSERT(runs[0] < SkRegion_kRunTypeSentinel); in SkipEntireScanline() 150 const int intervals = runs[1]; in SkipEntireScanline() 151 SkASSERT(runs[2 + intervals * 2] == SkRegion_kRunTypeSentinel); in SkipEntireScanline() 154 int n = compute_intervalcount(&runs[2]); in SkipEntireScanline() [all …]
|
| D | SkAlphaRuns.cpp | 30 const int16_t* runs = fRuns; in assertValid() local 33 while (*runs) { in assertValid() 35 alpha += *runs; in assertValid() 36 runs += *runs; in assertValid() 41 const int16_t* runs = fRuns; in dump() local 44 SkDebugf("Runs"); in dump() 45 while (*runs) { in dump() 46 int n = *runs; in dump() 53 runs += n; in dump() 62 const int16_t* runs = fRuns; in validate() local [all …]
|
| D | SkRegion.cpp | 77 static SkRegionPriv::RunType* skip_intervals(const SkRegionPriv::RunType runs[]) { in skip_intervals() argument 78 int intervals = runs[-1]; in skip_intervals() 81 SkASSERT(runs[0] < runs[1]); in skip_intervals() 82 SkASSERT(runs[1] < SkRegion_kRunTypeSentinel); in skip_intervals() 85 SkASSERT(SkRegion_kRunTypeSentinel == runs[0]); in skip_intervals() 88 runs += intervals * 2 + 1; in skip_intervals() 89 return const_cast<SkRegionPriv::RunType*>(runs); in skip_intervals() 92 bool SkRegion::RunsAreARect(const SkRegion::RunType runs[], int count, in RunsAreARect() argument 94 assert_sentinel(runs[0], false); // top in RunsAreARect() 98 assert_sentinel(runs[1], false); // bottom in RunsAreARect() [all …]
|
| D | SkBlitter.h | 38 /// Blit a horizontal run of antialiased pixels; runs[] is a *sparse* 40 /// The runs[] and antialias[] work together to represent long runs of pixels with the same 41 /// alphas. The runs[] contains the number of pixels with the same alpha, and antialias[] 42 /// contain the coverage value for that number of pixels. The runs[] (and antialias[]) are 43 /// encoded in a clever way. The runs array is zero terminated, and has enough entries for 45 /// in the runs array contains the number of pixels (np) that have the same alpha value. The 46 /// next np value is found np entries away. For example, if runs[0] = 7, then the next valid 47 /// entry will by at runs[7]. The runs array and antialias[] are coupled by index. So, if the 48 /// np entry is at runs[45] = 12 then the alpha value can be found at antialias[45] = 0x88. 50 virtual void blitAntiH(int x, int y, const SkAlpha antialias[], const int16_t runs[]) = 0; [all …]
|
| /third_party/skia/src/core/ |
| D | SkAntiRun.h | 60 int16_t* runs = fRuns + offsetX; in add() local 66 SkAlphaRuns::Break(runs, alpha, x, 1); in add() 76 runs += x + 1; in add() 83 SkAlphaRuns::Break(runs, alpha, x, middleCount); in add() 85 runs += x; in add() 89 int n = runs[0]; in add() 92 runs += n; in add() 100 SkAlphaRuns::Break(runs, alpha, x, 1); in add() 114 * Break the runs in the buffer at offsets x and x+count, properly 115 * updating the runs to the right and left. [all …]
|
| D | SkRegionPriv.h | 46 static int compute_intervalcount(const SkRegionPriv::RunType runs[]) { in compute_intervalcount() argument 47 const SkRegionPriv::RunType* curr = runs; in compute_intervalcount() 53 return SkToInt((curr - runs) >> 1); in compute_intervalcount() 143 * Given a scanline (including its Bottom value at runs[0]), return the next 144 * scanline. Asserts that there is one (i.e. runs[0] < Sentinel) 146 static SkRegion::RunType* SkipEntireScanline(const SkRegion::RunType runs[]) { in SkipEntireScanline() 148 SkASSERT(runs[0] < SkRegion_kRunTypeSentinel); in SkipEntireScanline() 150 const int intervals = runs[1]; in SkipEntireScanline() 151 SkASSERT(runs[2 + intervals * 2] == SkRegion_kRunTypeSentinel); in SkipEntireScanline() 154 int n = compute_intervalcount(&runs[2]); in SkipEntireScanline() [all …]
|
| D | SkAlphaRuns.cpp | 30 const int16_t* runs = fRuns; in assertValid() local 33 while (*runs) { in assertValid() 35 alpha += *runs; in assertValid() 36 runs += *runs; in assertValid() 41 const int16_t* runs = fRuns; in dump() local 44 SkDebugf("Runs"); in dump() 45 while (*runs) { in dump() 46 int n = *runs; in dump() 53 runs += n; in dump() 62 const int16_t* runs = fRuns; in validate() local [all …]
|
| D | SkRegion.cpp | 76 static SkRegionPriv::RunType* skip_intervals(const SkRegionPriv::RunType runs[]) { in skip_intervals() argument 77 int intervals = runs[-1]; in skip_intervals() 80 SkASSERT(runs[0] < runs[1]); in skip_intervals() 81 SkASSERT(runs[1] < SkRegion_kRunTypeSentinel); in skip_intervals() 84 SkASSERT(SkRegion_kRunTypeSentinel == runs[0]); in skip_intervals() 87 runs += intervals * 2 + 1; in skip_intervals() 88 return const_cast<SkRegionPriv::RunType*>(runs); in skip_intervals() 91 bool SkRegion::RunsAreARect(const SkRegion::RunType runs[], int count, in RunsAreARect() argument 93 assert_sentinel(runs[0], false); // top in RunsAreARect() 97 assert_sentinel(runs[1], false); // bottom in RunsAreARect() [all …]
|
| D | SkBlitter.h | 39 /// Blit a horizontal run of antialiased pixels; runs[] is a *sparse* 41 /// The runs[] and antialias[] work together to represent long runs of pixels with the same 42 /// alphas. The runs[] contains the number of pixels with the same alpha, and antialias[] 43 /// contain the coverage value for that number of pixels. The runs[] (and antialias[]) are 44 /// encoded in a clever way. The runs array is zero terminated, and has enough entries for 46 /// in the runs array contains the number of pixels (np) that have the same alpha value. The 47 /// next np value is found np entries away. For example, if runs[0] = 7, then the next valid 48 /// entry will by at runs[7]. The runs array and antialias[] are coupled by index. So, if the 49 /// np entry is at runs[45] = 12 then the alpha value can be found at antialias[45] = 0x88. 51 virtual void blitAntiH(int x, int y, const SkAlpha antialias[], const int16_t runs[]) = 0; [all …]
|
| /third_party/skia/third_party/externals/icu/source/common/ |
| D | ubidiln.cpp | 48 * The implementation of the access to same-level-runs and of the reordering 54 * same-level-runs is created. Reordering then is done on this vector. 59 * This is inefficient if there are many very short runs. If the average run 155 pLineBiDi->runs=NULL; in ubidi_setLine() 318 /* this is done based on runs rather than on levels since levels have in ubidi_getLogicalRun() 322 iRun=pBiDi->runs[0]; in ubidi_getLogicalRun() 325 iRun = pBiDi->runs[i]; in ubidi_getLogicalRun() 349 /* runs API functions ------------------------------------------------------- */ 375 start=pBiDi->runs[runIndex].logicalStart; in ubidi_getVisualRun() 381 *pLength=pBiDi->runs[runIndex].visualLimit- in ubidi_getVisualRun() [all …]
|
| /third_party/icu/icu4c/source/common/ |
| D | ubidiln.cpp | 48 * The implementation of the access to same-level-runs and of the reordering 54 * same-level-runs is created. Reordering then is done on this vector. 59 * This is inefficient if there are many very short runs. If the average run 155 pLineBiDi->runs=NULL; in ubidi_setLine() 318 /* this is done based on runs rather than on levels since levels have in ubidi_getLogicalRun() 322 iRun=pBiDi->runs[0]; in ubidi_getLogicalRun() 325 iRun = pBiDi->runs[i]; in ubidi_getLogicalRun() 349 /* runs API functions ------------------------------------------------------- */ 375 start=pBiDi->runs[runIndex].logicalStart; in ubidi_getVisualRun() 381 *pLength=pBiDi->runs[runIndex].visualLimit- in ubidi_getVisualRun() [all …]
|
| /third_party/flutter/skia/third_party/externals/icu/source/common/ |
| D | ubidiln.cpp | 48 * The implementation of the access to same-level-runs and of the reordering 54 * same-level-runs is created. Reordering then is done on this vector. 59 * This is inefficient if there are many very short runs. If the average run 155 pLineBiDi->runs=NULL; in ubidi_setLine() 318 /* this is done based on runs rather than on levels since levels have in ubidi_getLogicalRun() 322 iRun=pBiDi->runs[0]; in ubidi_getLogicalRun() 325 iRun = pBiDi->runs[i]; in ubidi_getLogicalRun() 349 /* runs API functions ------------------------------------------------------- */ 375 start=pBiDi->runs[runIndex].logicalStart; in ubidi_getVisualRun() 381 *pLength=pBiDi->runs[runIndex].visualLimit- in ubidi_getVisualRun() [all …]
|
| /third_party/node/deps/icu-small/source/common/ |
| D | ubidiln.cpp | 48 * The implementation of the access to same-level-runs and of the reordering 54 * same-level-runs is created. Reordering then is done on this vector. 59 * This is inefficient if there are many very short runs. If the average run 155 pLineBiDi->runs=NULL; in ubidi_setLine() 318 /* this is done based on runs rather than on levels since levels have in ubidi_getLogicalRun() 322 iRun=pBiDi->runs[0]; in ubidi_getLogicalRun() 325 iRun = pBiDi->runs[i]; in ubidi_getLogicalRun() 349 /* runs API functions ------------------------------------------------------- */ 375 start=pBiDi->runs[runIndex].logicalStart; in ubidi_getVisualRun() 381 *pLength=pBiDi->runs[runIndex].visualLimit- in ubidi_getVisualRun() [all …]
|
| /third_party/json/test/thirdparty/Fuzzer/test/ |
| D | fuzzer-traces-hooks.test | 6 Done1000000: Done 1000000 runs in 8 RUN: not LLVMFuzzer-MemcmpTest -seed=4294967295 -runs=100000 2>&1 | FileCheck %s 9 RUN: LLVMFuzzer-MemcmpTest -use_memcmp=0 -seed=4294967295 -runs=1000000 2>&1 | FileCheck %s --… 11 RUN: not LLVMFuzzer-StrncmpTest -seed=2 -runs=100000 2>&1 | FileCheck %s 12 RUN: LLVMFuzzer-StrncmpTest -use_memcmp=0 -seed=3 -runs=1000000 2>&1 | FileCheck %s --check-pr… 14 RUN: not LLVMFuzzer-StrcmpTest -seed=4 -runs=200000 2>&1 | FileCheck %s 15 RUN: LLVMFuzzer-StrcmpTest -use_memcmp=0 -seed=5 -runs=1000000 2>&1 | FileCheck %s --check-pre… 17 RUN: not LLVMFuzzer-StrstrTest -seed=6 -runs=200000 2>&1 | FileCheck %s 18 RUN: LLVMFuzzer-StrstrTest -use_memmem=0 -seed=7 -runs=1000000 2>&1 | FileCheck %s --check-pre… 20 RUN: LLVMFuzzer-RepeatedMemcmp -seed=10 -runs=100000 2>&1 | FileCheck %s --check-prefix=RECOMMENDED…
|
| D | fuzzer-leak.test | 2 RUN: not LLVMFuzzer-LeakTest -runs=100000 -detect_leaks=1 2>&1 | FileCheck %s --check-prefix=LEAK_D… 10 RUN: not LLVMFuzzer-LeakTest -runs=0 -detect_leaks=1 %S 2>&1 | FileCheck %s --check-prefix=LEAK_IN_… 14 RUN: not LLVMFuzzer-LeakTest -runs=100000000 %S/hi.txt 2>&1 | FileCheck %s --check-prefix=MULTI_RUN… 18 RUN: not LLVMFuzzer-LeakTest -runs=100000 -detect_leaks=0 2>&1 | FileCheck %s --check-prefix=LEAK_A… 19 RUN: not LLVMFuzzer-LeakTest -runs=100000 2>&1 | FileCheck %s --check-prefix=LEAK_D… 20 RUN: not LLVMFuzzer-ThreadedLeakTest -runs=100000 -detect_leaks=0 2>&1 | FileCheck %s --check-prefi… 21 RUN: not LLVMFuzzer-ThreadedLeakTest -runs=100000 2>&1 | FileCheck %s --check-prefi… 22 LEAK_AFTER: Done 100000 runs in 25 RUN: not LLVMFuzzer-LeakTest -runs=100000 -max_len=1 2>&1 | FileCheck %s --check-prefix=MAX_LEN_1 32 RUN: LLVMFuzzer-AccumulateAllocationsTest -detect_leaks=1 -runs=100000 2>&1 | FileCheck %s --check-…
|
| /third_party/ltp/utils/benchmark/kernbench-0.42/ |
| D | README | 12 It runs a kernel at various numbers of concurrent jobs: 1/2 number of cpus, 15 for the average of each group of runs and logs them to kernbench.log 37 kernbench [-n runs] [-o jobs] [-s] [-H] [-O] [-M] [-h] [-v] 40 s : perform single threaded runs (default don't) 41 H : don't perform half load runs (default do) 42 O : don't perform optimal load runs (default do) 43 M : don't perform maximal load runs (default do) 55 v0.40 Made all runs use the oldconfig if it exists. Changed to only do one 61 preparation and drops number of runs to 3. Modified half loads to 65 v0.20 Change to average of runs, add options to choose which runs to perform [all …]
|
| /third_party/ntfs-3g/ntfsprogs/ |
| D | cluster.c | 53 runlist *runs; in cluster_find() local 81 runs = ntfs_mapping_pairs_decompress(vol, a_ctx->attr, NULL); in cluster_find() 82 if (!runs) { in cluster_find() 83 ntfs_log_error("Couldn't read the data runs.\n"); in cluster_find() 91 for (j = 0; runs[j].length > 0; j++) { in cluster_find() 92 LCN a_begin = runs[j].lcn; in cluster_find() 93 LCN a_end = a_begin + runs[j].length - 1; in cluster_find() 99 (long long)runs[j].vcn, in cluster_find() 100 (long long)runs[j].lcn, in cluster_find() 101 (long long)(runs[j].lcn + in cluster_find() [all …]
|
| /third_party/node/deps/npm/node_modules/har-validator/node_modules/fast-deep-equal/ |
| D | README.md | 57 fast-deep-equal x 261,950 ops/sec ±0.52% (89 runs sampled) 58 fast-deep-equal/es6 x 212,991 ops/sec ±0.34% (92 runs sampled) 59 fast-equals x 230,957 ops/sec ±0.83% (85 runs sampled) 60 nano-equal x 187,995 ops/sec ±0.53% (88 runs sampled) 61 shallow-equal-fuzzy x 138,302 ops/sec ±0.49% (90 runs sampled) 62 underscore.isEqual x 74,423 ops/sec ±0.38% (89 runs sampled) 63 lodash.isEqual x 36,637 ops/sec ±0.72% (90 runs sampled) 64 deep-equal x 2,310 ops/sec ±0.37% (90 runs sampled) 65 deep-eql x 35,312 ops/sec ±0.67% (91 runs sampled) 66 ramda.equals x 12,054 ops/sec ±0.40% (91 runs sampled) [all …]
|
| /third_party/flatbuffers/.github/workflows/ |
| D | build.yml | 14 runs-on: ubuntu-latest 34 runs-on: windows-latest 53 runs-on: windows-2016 67 runs-on: macos-latest 85 runs-on: ubuntu-latest 103 runs-on: ubuntu-latest 116 runs-on: ubuntu-latest 125 runs-on: macos-latest 148 runs-on: ubuntu-latest 157 # runs-on: ubuntu-latest [all …]
|
| /third_party/ffmpeg/libavcodec/ |
| D | faxcompr.c | 129 unsigned int *pix_left, int **runs, in decode_uncompressed() argument 164 *(*runs)++ = saved_run; in decode_uncompressed() 165 if (*runs >= runend) { in decode_uncompressed() 181 *(*runs)++ = saved_run; in decode_uncompressed() 182 if (*runs >= runend) { in decode_uncompressed() 196 *(*runs)++ = 0; in decode_uncompressed() 197 if (*runs >= runend) { in decode_uncompressed() 207 unsigned int pix_left, int *runs, in decode_group3_1d_line() argument 219 *runs++ = run; in decode_group3_1d_line() 220 if (runs >= runend) { in decode_group3_1d_line() [all …]
|
| /third_party/flatbuffers/tests/FlatBuffers.Benchmarks.swift/Sources/FlatBuffers.Benchmarks.swift/ |
| D | main.swift | 27 func run(name: String, runs: Int, action: () -> Void) -> Benchmark { 30 for _ in 0..<runs { 34 let value = Double(ends - start) / Double(runs) 119 func benchmark(numberOfRuns runs: Int) { in benchmark() 122 benchmarks.append(run(name: "500_000", runs: runs, action: benchmarkFiveHundredAdds)) in benchmark() 123 benchmarks.append(run(name: "10 str", runs: runs, action: create10Strings)) in benchmark() 124 let hundredStr = run(name: "100 str", runs: runs) { in benchmark() 127 benchmarks.append(run(name: "3M strc", runs: 1, action: benchmarkThreeMillionStructs)) in benchmark()
|