/external/e2fsprogs/tests/f_mmp/ |
D | script | 5 TMPFILE=$(mktemp ./tmp-$test_name.XXXXXX) 10 echo "$test_name: $test_description: skipped for tmpfs (no O_DIRECT)" 14 echo "make the test image ..." > $test_name.log 15 $MKE2FS -q -F -o Linux -b 4096 -O mmp -E mmp_update_interval=1 $TMPFILE 100 >> $test_name.log 2>&1 18 echo "mke2fs -O mmp failed" > $test_name.failed 19 echo "$test_name: $test_description: failed" 30 # this will cause debugfs to create the $test_name.mark file once it has 32 MARKFILE=$test_name.new 35 echo "set mmp sequence to EXT2_MMP_SEQ_FSCK..." >> $test_name.log 37 $DEBUGFS -w $TMPFILE >> $test_name.log 2>&1 & ) > /dev/null 2>&1 & [all …]
|
/external/tensorflow/tensorflow/compiler/xla/tests/ |
D | test_macros.h | 80 const string& test_name); 90 #define XLA_GTEST_TEST_(test_case_name, test_name, parent_class, parent_id) \ argument 91 class GTEST_TEST_CLASS_NAME_(test_case_name, test_name) \ 94 GTEST_TEST_CLASS_NAME_(test_case_name, test_name)() {} \ 100 test_name)); \ 104 test_name)::test_info_ = \ 107 ::xla::PrependDisabledIfIndicated(#test_case_name, #test_name) \ 113 test_case_name, test_name)>); \ 114 void GTEST_TEST_CLASS_NAME_(test_case_name, test_name)::TestBody() 122 #define XLA_TEST_F(test_fixture, test_name) \ argument [all …]
|
/external/e2fsprogs/tests/j_corrupt_ext_jnl_sb_csum/ |
D | script | 2 OUT=$test_name.log 4 EXP=$test_name.tmp 13 …d "$JOURNAL_DUMP_DIR" -a -w "$JOURNAL_DUMP_DIR" && cp $test_name.img "$JOURNAL_DUMP_DIR/$test_name… 14 …"$JOURNAL_DUMP_DIR" -a -w "$JOURNAL_DUMP_DIR" && cp $test_name.img.jnl "$JOURNAL_DUMP_DIR/$test_na… 16 $FSCK $FSCK_OPT -N test_filesys -j $test_name.img.jnl $test_name.img > $OUT.new 2>&1 22 $FSCK $FSCK_OPT -N test_filesys -j $test_name.img.jnl $test_name.img > $OUT.new 2>&1 28 rm -f $TMPFILE $test_name.img $test_name.img.jnl 34 echo "$test_name: $test_description: ok" 35 touch $test_name.ok 37 echo "$test_name: $test_description: failed" [all …]
|
/external/e2fsprogs/tests/j_corrupt_ext_jnl_sb_block/ |
D | script | 2 OUT=$test_name.log 4 EXP=$test_name.tmp 13 …d "$JOURNAL_DUMP_DIR" -a -w "$JOURNAL_DUMP_DIR" && cp $test_name.img "$JOURNAL_DUMP_DIR/$test_name… 14 …"$JOURNAL_DUMP_DIR" -a -w "$JOURNAL_DUMP_DIR" && cp $test_name.img.jnl "$JOURNAL_DUMP_DIR/$test_na… 16 $FSCK $FSCK_OPT -N test_filesys -j $test_name.img.jnl $test_name.img > $OUT.new 2>&1 22 rm -f $TMPFILE $test_name.img $test_name.img.jnl 28 echo "$test_name: $test_description: ok" 29 touch $test_name.ok 31 echo "$test_name: $test_description: failed" 32 diff $DIFF_OPTS $EXP $OUT > $test_name.failed [all …]
|
/external/e2fsprogs/tests/t_quota_1on/ |
D | script | 4 echo "$test_name: $test_description: skipped" 8 $MKE2FS -q -F -o Linux -b 4096 $TMPFILE 10000 > $test_name.log 2>&1 11 echo "mke2fs failed" > $test_name.failed 12 echo "$test_name: $test_description: failed" 16 dd if=/dev/zero of=$TMPFILE.2 bs=1048576 count=1 >> $test_name.log 2>&1 17 cat <<- EOF | $DEBUGFS -w -f /dev/stdin $TMPFILE >> $test_name.log 2>&1 24 $TUNE2FS -O quota $TMPFILE >> $test_name.log 2>&1 27 echo "tune2fs -O quota failed with $status" > $test_name.failed 28 echo "$test_name: $test_description: failed" 32 $FSCK $FSCK_OPT $TMPFILE >> $test_name.log 2>&1 [all …]
|
/external/valgrind/ |
D | runtest.sh | 29 test_name=$1 31 test_local=$product_out/data/nativetest$2/$test_name/$test_name 32 test_target=/data/nativetest$2/$test_name/$test_name 37 logfile_native=$test_name.stdout.log 38 logfile_valgrind=$test_name.stdout.vlog 51 … "WARNING: linker: Unsupported flags" | grep -v "^> $" | grep -v "^> 0x421==" > $test_name.diff.log 53 if [ -s $test_name.diff.log ]; then 54 echo "Test $test_name FAILED, please check the diff below" 55 cat $test_name.diff.log | sed "s/^< /expected: /" | sed "s/^> /actual : /" 59 echo "Test $test_name PASSED"
|
/external/autotest/client/site_tests/video_VEAPerf/ |
D | video_VEAPerf.py | 80 def _logperf(self, test_name, key, value, units, higher_is_better=False): argument 81 description = '%s.%s' % (test_name, key) 87 def _analyze_fps(self, test_name, log_file): argument 98 self._logperf(test_name, KEY_FPS, fps[0], UNIT_FPS, True) 101 def _analyze_encode_latency(self, test_name, log_file): argument 116 self._logperf(test_name, KEY_ENCODE_LATENCY_50, latency_50[0], 118 self._logperf(test_name, KEY_ENCODE_LATENCY_75, latency_75[0], 120 self._logperf(test_name, KEY_ENCODE_LATENCY_95, latency_95[0], 124 def _analyze_cpu_usage(self, test_name, time_log_file): argument 131 self._logperf(test_name, KEY_CPU_USER_USAGE, u / r, UNIT_RATIO) [all …]
|
/external/e2fsprogs/tests/t_quota_2off/ |
D | script | 4 echo "$test_name: $test_description: skipped" 8 $MKE2FS -q -F -o Linux -b 4096 -O quota $TMPFILE 100 > $test_name.log 2>&1 11 echo "mke2fs -O quota failed" > $test_name.failed 12 echo "$test_name: $test_description: failed" 16 $TUNE2FS -O ^quota $TMPFILE >> $test_name.log 2>&1 19 echo "tune2fs -O ^quota failed" > $test_name.failed 20 echo "$test_name: $test_description: failed" 24 $FSCK $FSCK_OPT $TMPFILE >> $test_name.log 2>&1 27 echo "$test_name: $test_description: ok" 28 touch $test_name.ok [all …]
|
/external/e2fsprogs/tests/t_mmp_2off/ |
D | script | 5 TMPFILE=$(mktemp ./tmp-$test_name.XXXXXX) 10 echo "$test_name: $test_description: skipped for tmpfs (no O_DIRECT)" 14 $MKE2FS -q -F -o Linux -b 4096 -O mmp $TMPFILE 100 > $test_name.log 2>&1 17 echo "mke2fs -O mmp failed" > $test_name.failed 18 echo "$test_name: $test_description: failed" 22 $TUNE2FS -O ^mmp $TMPFILE > $test_name.log 2>&1 25 echo "tune2fs -O ^mmp failed" > $test_name.failed 26 echo "$test_name: $test_description: failed" 30 $FSCK $FSCK_OPT $TMPFILE > $test_name.log 2>&1 33 echo "$test_name: $test_description: ok" [all …]
|
/external/e2fsprogs/tests/t_mmp_1on/ |
D | script | 5 TMPFILE=$(mktemp ./tmp-$test_name.XXXXXX) 10 echo "$test_name: $test_description: skipped for tmpfs (no O_DIRECT)" 14 $MKE2FS -q -F -o Linux -b 4096 $TMPFILE 100 > $test_name.log 2>&1 17 echo "mke2fs failed" > $test_name.failed 18 echo "$test_name: $test_description: failed" 22 $TUNE2FS -O mmp -E mmp_update_interval=1 $TMPFILE >> $test_name.log 2>&1 25 echo "tune2fs -O mmp failed with $status" > $test_name.failed 26 echo "$test_name: $test_description: failed" 30 $FSCK $FSCK_OPT $TMPFILE >> $test_name.log 2>&1 33 echo "$test_name: $test_description: ok" [all …]
|
/external/harfbuzz_ng/test/api/ |
D | CMakeLists.txt | 2 foreach(test_name ${ARGV}) 3 if (EXISTS ${CMAKE_CURRENT_SOURCE_DIR}/${test_name}.c) 4 add_executable (${test_name} ${test_name}.c) 5 elseif (EXISTS ${CMAKE_CURRENT_SOURCE_DIR}/${test_name}.cc) 6 add_executable (${test_name} ${test_name}.cc) 8 message(FATAL_ERROR "No source file found for test ${test_name}") 10 target_link_libraries (${test_name} harfbuzz) 11 add_test (${test_name} ${test_name})
|
/external/autotest/client/site_tests/security_SeccompSyscallFilters/src/ |
D | test_harness.h | 214 #define _TEST(test_name) __TEST_IMPL(test_name, -1) argument 216 #define _TEST_SIGNAL(test_name, signal) __TEST_IMPL(test_name, signal) argument 218 #define __TEST_IMPL(test_name, _signal) \ argument 219 static void test_name(struct __test_metadata *_metadata); \ 220 static struct __test_metadata _##test_name##_object = \ 221 { name: "global." #test_name, fn: &test_name, termsig: _signal }; \ 222 static void __attribute__((constructor)) _register_##test_name(void) { \ 223 __register_test(&_##test_name##_object); \ 225 static void test_name( \ 255 #define _TEST_F(fixture_name, test_name) \ argument [all …]
|
/external/linux-kselftest/tools/testing/selftests/ |
D | kselftest_harness.h | 126 #define TEST(test_name) __TEST_IMPL(test_name, -1) argument 145 #define TEST_SIGNAL(test_name, signal) __TEST_IMPL(test_name, signal) argument 147 #define __TEST_IMPL(test_name, _signal) \ argument 148 static void test_name(struct __test_metadata *_metadata); \ 149 static struct __test_metadata _##test_name##_object = \ 150 { name: "global." #test_name, \ 151 fn: &test_name, termsig: _signal }; \ 152 static void __attribute__((constructor)) _register_##test_name(void) \ 154 __register_test(&_##test_name##_object); \ 156 static void test_name( \ [all …]
|
/external/seccomp-tests/linux/ |
D | test_harness.h | 219 #define _TEST(test_name) __TEST_IMPL(test_name, -1) argument 221 #define _TEST_SIGNAL(test_name, signal) __TEST_IMPL(test_name, signal) argument 223 #define __TEST_IMPL(test_name, _signal) \ argument 224 static void test_name(struct __test_metadata *_metadata); \ 225 static struct __test_metadata _##test_name##_object = \ 226 { name: "global." #test_name, \ 227 fn: &test_name, termsig: _signal }; \ 228 static void __attribute__((constructor)) _register_##test_name(void) \ 230 __register_test(&_##test_name##_object); \ 232 static void test_name( \ [all …]
|
/external/autotest/client/bin/ |
D | setup_job.py | 60 test_name = os.path.split(testdir)[-1] 61 outputdir = os.path.join(job.resultdir, test_name) 72 import_stmt = 'import %s' % test_name 74 (test_name, test_name)) 79 if re.search(test_name, str(e)): 84 (test_name, e, test_name)) 87 logging.exception("%s: %s", test_name, e) 113 for test_name in sorted(os.listdir(testdir)): 114 client_test = init_test(options, os.path.join(testdir, test_name)) 118 broken_tests.append(test_name) [all …]
|
/external/protobuf/conformance/ |
D | conformance_test.h | 69 virtual void RunTest(const std::string& test_name, 112 void ReportSuccess(const std::string& test_name); 113 void ReportFailure(const string& test_name, 117 void ReportSkip(const string& test_name, 120 void RunTest(const std::string& test_name, 123 void RunValidInputTest(const string& test_name, const string& input, 127 void RunValidJsonTest(const string& test_name, const string& input_json, 129 void RunValidJsonTestWithProtobufInput(const string& test_name, 134 void RunValidJsonTestWithValidator(const string& test_name, 137 void ExpectParseFailureForJson(const string& test_name, [all …]
|
/external/arm-neon-tests/ |
D | stm-arm-neon-ref.h | 448 static void dump_results (char *test_name) in dump_results() argument 452 fprintf(ref_file, "\n%s output:\n", test_name); in dump_results() 453 fprintf(gcc_tests_file, "\n%s output:\n", test_name); in dump_results() 455 DUMP(test_name, int, 8, 8, PRId8); in dump_results() 456 DUMP(test_name, int, 16, 4, PRId16); in dump_results() 457 DUMP(test_name, int, 32, 2, PRId32); in dump_results() 458 DUMP(test_name, int, 64, 1, PRId64); in dump_results() 459 DUMP(test_name, uint, 8, 8, PRIu8); in dump_results() 460 DUMP(test_name, uint, 16, 4, PRIu16); in dump_results() 461 DUMP(test_name, uint, 32, 2, PRIu32); in dump_results() [all …]
|
/external/autotest/site_utils/presubmit_hooks/ |
D | check_control_files.py | 111 def CheckSuites(ctrl_data, test_name, useflags): argument 139 if flag == 'tests_%s' % test_name: 145 '<your_ebuild>. 3. emerge-<board> <your_ebuild>' % test_name) 148 def CheckValidAttr(ctrl_data, whitelist, test_name): argument 167 % (attribute_diff, test_name)) 187 def CheckRetry(ctrl_data, test_name): argument 202 'file for %s.' % (' or '.join(SUITES_NEED_RETRY), test_name)) 205 def CheckDependencies(ctrl_data, test_name): argument 214 if test_name.startswith(TESTS_NEED_ARC): 217 'DEPENDENCIES = \'arc\' for %s is needed' % test_name) [all …]
|
/external/e2fsprogs/tests/i_qcow/ |
D | script | 9 OUT=$test_name.log 10 CRC=$SRCDIR/$test_name/$test_name.crc 11 CRC_TMP=$test_name.crc.tmp 13 rm -f $test_name/_image.* $CRC_TMP $OUT >/dev/null 2>&1 17 ORIG_IMG=$test_name/$i 48 echo "$test_name: $test_description: ok" 49 touch $test_name.ok 51 ln -f $test_name.log $test_name.failed 52 echo "$test_name: $test_description: failed" 58 echo "$test_name: $test_description: skipped"
|
/external/e2fsprogs/tests/f_boundscheck/ |
D | script | 11 OUT1=$test_name.1.log 12 rm -rf $test_name.failed $test_name.ok 19 OUT2=$test_name.2.log 26 echo "$test_name: $test_description: ok" 27 touch $test_name.ok 29 echo "$test_name: $test_description: failed" 30 diff -u $EXP1 $OUT1 >> $test_name.failed 31 diff -u $EXP2 $OUT2 >> $test_name.failed
|
/external/e2fsprogs/tests/j_recover_csum2_64bit/ |
D | script | 10 OUT1=$test_name.1.log 11 rm -rf $test_name.failed $test_name.ok 18 OUT2=$test_name.2.log 25 echo "$test_name: $test_description: ok" 26 touch $test_name.ok 28 echo "$test_name: $test_description: failed" 29 diff -u $EXP1 $OUT1 >> $test_name.failed 30 diff -u $EXP2 $OUT2 >> $test_name.failed
|
/external/e2fsprogs/tests/j_recover_csum2_32bit/ |
D | script | 10 OUT1=$test_name.1.log 11 rm -rf $test_name.failed $test_name.ok 18 OUT2=$test_name.2.log 25 echo "$test_name: $test_description: ok" 26 touch $test_name.ok 28 echo "$test_name: $test_description: failed" 29 diff -u $EXP1 $OUT1 >> $test_name.failed 30 diff -u $EXP2 $OUT2 >> $test_name.failed
|
/external/e2fsprogs/tests/f_itable_collision/ |
D | script | 16 OUT1=$test_name.1.log 17 rm -rf $test_name.failed $test_name.ok 24 OUT2=$test_name.2.log 31 echo "$test_name: $test_description: ok" 32 touch $test_name.ok 34 echo "$test_name: $test_description: failed" 35 diff -u $EXP1 $OUT1 >> $test_name.failed 36 diff -u $EXP2 $OUT2 >> $test_name.failed
|
/external/e2fsprogs/tests/f_super_bad_csum/ |
D | script | 11 OUT1=$test_name.1.log 12 rm -rf $test_name.failed $test_name.ok 19 OUT2=$test_name.2.log 26 echo "$test_name: $test_description: ok" 27 touch $test_name.ok 29 echo "$test_name: $test_description: failed" 30 diff -u $EXP1 $OUT1 >> $test_name.failed 31 diff -u $EXP2 $OUT2 >> $test_name.failed
|
/external/tensorflow/tensorflow/tools/test/ |
D | run_and_gather_logs_lib.py | 51 def process_test_logs(name, test_name, test_args, benchmark_type, argument 71 results.target = test_name 98 def run_and_gather_logs(name, test_name, test_args, argument 121 if not (test_name and test_name.startswith("//") and ".." not in test_name and 122 not test_name.endswith(":") and not test_name.endswith(":all") and 123 not test_name.endswith("...") and len(test_name.split(":")) == 2): 126 test_executable = test_name.rstrip().strip("/").replace(":", "/") 167 test_name=test_name,
|