/external/arm-neon-tests/ |
stm-arm-neon-ref.h | 448 static void dump_results (char *test_name) 452 fprintf(ref_file, "\n%s output:\n", test_name); 453 fprintf(gcc_tests_file, "\n%s output:\n", test_name); 455 DUMP(test_name, int, 8, 8, PRId8); 456 DUMP(test_name, int, 16, 4, PRId16); 457 DUMP(test_name, int, 32, 2, PRId32); 458 DUMP(test_name, int, 64, 1, PRId64); 459 DUMP(test_name, uint, 8, 8, PRIu8); 460 DUMP(test_name, uint, 16, 4, PRIu16); 461 DUMP(test_name, uint, 32, 2, PRIu32) [all...] |
/external/minijail/ |
test_harness.h | 202 #define _TEST(test_name) \ 203 static void test_name(struct __test_metadata *_metadata); \ 204 static struct __test_metadata _##test_name##_object = \ 205 { .name= "global." #test_name, .fn= &(test_name) }; \ 206 static void __attribute__((constructor)) _register_##test_name(void) { \ 207 __register_test(&_##test_name##_object); \ 209 static void test_name( \ 239 #define _TEST_F(fixture_name, test_name) \ 240 static void fixture_name##_##test_name( \ [all...] |
/art/compiler/utils/ |
jni_macro_assembler_test.h | 42 void DriverFn(TestFn f, const std::string& test_name) { 43 DriverWrapper(f(this, assembler_.get()), test_name); local 47 void DriverStr(const std::string& assembly_string, const std::string& test_name) { 48 DriverWrapper(assembly_string, test_name); 131 void DriverWrapper(const std::string& assembly_text, const std::string& test_name) { 138 test_helper_->Driver(*data, assembly_text, test_name);
|
/external/autotest/client/site_tests/video_WebRtcMediaRecorder/ |
video_WebRtcMediaRecorder.py | 30 def launch_recorder_test(self, test_name): 33 @param test_name: Name of test to run. 43 self.tab.EvaluateJavaScript(test_name + "();") 45 logging.error('%s did not complete', test_name) 46 raise error.TestFail('Failed %s' %(test_name)) 51 raise error.TestFail('Failed %s' %(test_name)) 53 raise error.TestFail('Failed %s, got %s' %(test_name,
|
/test/vts/script/ |
create-test-project.py | 41 test_name: string, test case name in UpperCamel 44 test_name: string, test case name in UpperCamel 50 def __init__(self, test_name, test_dir_under_testcases, test_type): 54 test_name: string, test case name in UpperCamel 64 if not self.IsUpperCamel(test_name): 67 self.test_name = test_name 157 test_name=self.test_name, 180 test_name=self.test_name [all...] |
/frameworks/base/libs/hwui/tests/common/ |
TestUtils.h | 56 #define INNER_PIPELINE_TEST(test_case_name, test_name, pipeline, functionCall) \ 57 TEST(test_case_name, test_name##_##pipeline) { \ 67 #define OPENGL_PIPELINE_TEST(test_case_name, test_name) \ 68 class test_case_name##_##test_name##_HwuiTest { \ 72 INNER_PIPELINE_TEST(test_case_name, test_name, OpenGL, \ 73 test_case_name##_##test_name##_HwuiTest::doTheThing()) \ 74 void test_case_name##_##test_name##_HwuiTest::doTheThing() 76 #define INNER_PIPELINE_RENDERTHREAD_TEST(test_case_name, test_name, pipeline) \ 77 INNER_PIPELINE_TEST(test_case_name, test_name, pipeline, \ 78 TestUtils::runOnRenderThread(test_case_name##_##test_name##_RenderThreadTest::doTheThing) [all...] |
/external/autotest/server/site_tests/hardware_StorageQualBase/ |
hardware_StorageQualBase.py | 53 for test_name, argv in self.CLIENT_FUNCTIONAL_TESTS: 54 client_at.run_test(test_name, disable_sysinfo=True, tag=client_tag,
|
/external/e2fsprogs/tests/f_extent_htree/ |
script | 16 OUT=$test_name.log 22 SRC=$TMPDIR/$test_name.tmp 55 OUT1=$test_name.pre.1.log 56 OUT2=$test_name.pre.2.log
|
/external/lisa/libs/utils/android/workloads/ |
vellamo.py | 50 def run(self, out_dir, test_name, collect=''): 57 :param test_name: Name of the test to run 58 :type test_name: str 79 if test_name.upper() not in VELLAMO_TESTS: 80 raise ValueError('Vellamo workload [%s] not supported', test_name) 83 self._log.debug('Start Vellamo Benchmark [%s]', test_name) 86 if test_name.upper() == 'BROWSER': 89 elif test_name.upper() == 'METAL': 92 elif test_name.upper() == 'MULTI': 143 self._log.info("Vellamo - {} started!".format(test_name.upper()) [all...] |
/external/toolchain-utils/deprecated/ |
summarize_results.py | 41 test_name = (':'.join(line[1:])).replace('\t', ' ').strip() 42 count = name_count.get(test_name, 0) + 1 43 name_count[test_name] = count 44 test_name = '%s (%s)' % (test_name, str(count)) 49 result += '%s\t%s\t%s\n' % (test_name, test_result, filename) 98 test_name = (line[0].strip()) 103 result += '%s\t%s\t%s\n' % (test_name, test_result, filename)
|
/external/v8/tools/testrunner/local/ |
junit_output.py | 37 def HasRunTest(self, test_name, test_duration, test_failure): 39 testCaseElement.attrib["name"] = " ".join(test_name)
|
/system/core/storaged/include/ |
storaged_info.h | 22 #define FRIEND_TEST(test_case_name, test_name) \ 23 friend class test_case_name##_##test_name##_Test
|
/tools/test/connectivity/acts/tests/google/bt/car_bt/ |
BtCarToggleTest.py | 32 def on_fail(self, test_name, begin_time): 33 bt_test_utils.take_btsnoop_logs(self.android_devices, self, test_name)
|
/external/autotest/tko/ |
test.cgi | 45 for test_name in uniq_test: 46 header_row.append(display.box(test_name, header=True)) 52 for test_name in uniq_test: 55 and test.testname == test_name]
|
/external/libmojo/mojo/edk/test/ |
test_support_impl.cc | 50 void TestSupportImpl::LogPerfResult(const char* test_name, 54 DCHECK(test_name); 56 std::string name = base::StringPrintf("%s/%s", test_name, sub_test_name); 59 base::LogPerfResult(test_name, value, units);
|
/external/toolchain-utils/crosperf/ |
experiment_factory.py | 100 for test_name in benchmark_list: 101 telemetry_benchmark = Benchmark(test_name, test_name, test_args, 164 test_name = benchmark_settings.GetField('test_name') 165 if not test_name: 166 test_name = benchmark_name 174 if test_name == 'all_perfv2': 178 elif test_name == 'all_pagecyclers': 183 elif test_name == 'all_toolchain_perf' [all...] |
/test/vts/testcases/template/gtest_binary_test/ |
gtest_test_case.py | 31 test_name: string, test case name which does not include test suite 39 def GetRunCommand(self, output_file_path=None, test_name=None): 44 test_name: name of the gtest test case. 50 if not test_name: 51 test_name = self.full_name 55 test=test_name,
|
/external/ImageMagick/config/ |
test-driver | 53 test_name= # Used for reporting. 63 --test-name) test_name=$2; shift;; 77 test x"$test_name" = x && missing_opts="$missing_opts --test-name" 129 echo "$res $test_name (exit status: $estatus)" >>$log_file 132 echo "${col}${res}${std}: $test_name"
|
/external/autotest/puppylab/ |
results_mocker.py | 44 def __init__(self, test_name, results_dir, machine_name): 47 @param test_name: The name of the test, eg: dummy_Pass. 53 self.test_results = os.path.join(results_dir, test_name) 56 self.test_name = test_name 106 test_name/status - core test status 112 'test_name': self.test_name,
|
/external/autotest/server/site_tests/display_SuspendStress/ |
display_SuspendStress.py | 40 test_name = "%s_%dx%d" % testcase_spec 46 ': %s', test_name) 48 edid_path = os.path.join(self.bindir, 'test_data', 'edids', test_name) 61 logging.info('Use EDID: %s', test_name) 71 ' (chameleon) from DUT: %s', test_name)
|
/external/pcre/dist2/ |
test-driver | 53 test_name= # Used for reporting. 63 --test-name) test_name=$2; shift;; 77 test x"$test_name" = x && missing_opts="$missing_opts --test-name" 129 echo "$res $test_name (exit status: $estatus)" >>$log_file 132 echo "${col}${res}${std}: $test_name"
|
/external/autotest/tko/perf_upload/ |
perf_uploader.py | 68 def _gather_presentation_info(config_data, test_name): 74 @param test_name: The name of an autotest. 81 if not test_name in config_data: 84 (test_name, _PRESENTATION_CONFIG_FILE)) 86 presentation_dict = config_data[test_name] 92 (test_name, _PRESENTATION_CONFIG_FILE)) 94 test_name = presentation_dict['dashboard_test_name'] 95 return {'master_name': master_name, 'test_name': test_name} 137 'benchmark_name': presentation_info['test_name'], [all...] |
/external/autotest/client/bin/ |
fio_util.py | 71 test_name = '' variable in class:fio_graph_generator 178 def _write_option(cls, f, test_name, test_type, percentile): 183 @param test_name: name of current workload. i.e. randwrite 190 (cls.graph_title[test_type], test_name)) 193 (cls.graph_title[test_type], test_name)) 203 def _write_graph(cls, test_name, test_type, pass_list, percentile=False): 207 @param test_name: name of current workload. i.e. randwrite 213 test_name, test_type, str(pass_list)) 217 out_file_name = '%s_%s_percentile.html' % (test_name, test_type) 219 out_file_name = '%s_%s.html' % (test_name, test_type [all...] |
/art/test/testrunner/ |
testrunner.py | 360 test_name = 'test-art-' 361 test_name += target + '-run-test-' 362 test_name += run + '-' 363 test_name += prebuild + '-' 364 test_name += compiler + '-' 365 test_name += relocate + '-' 366 test_name += trace + '-' 367 test_name += gc + '-' 368 test_name += jni + '-' 369 test_name += image + '- [all...] |
/external/autotest/server/ |
site_server_job_utils.py | 33 def __init__(self, test_name, test_args, test_attribs=None): 37 test_name: string, name of test to execute. 44 self.test_name = test_name 46 self.tagged_test_name = test_name 48 self.tagged_test_name = test_name + '.' + test_args.get('tag') 59 msg = '%s(%s)' % (self.test_name, params) 107 logging.info('Running Server_Job=%s', self.test_name) 108 server_job.run_test(self.test_name, **self.test_args) 111 '%s.', self.test_name) [all...] |