diff options
Diffstat (limited to 'tests/python/modules/render_report.py')
-rwxr-xr-x | tests/python/modules/render_report.py | 73 |
1 files changed, 33 insertions, 40 deletions
diff --git a/tests/python/modules/render_report.py b/tests/python/modules/render_report.py index 7dfc74904a7..5a2baa354c7 100755 --- a/tests/python/modules/render_report.py +++ b/tests/python/modules/render_report.py @@ -374,48 +374,41 @@ class Report: return not failed - def _run_tests(self, filepaths, render_cb): - # Run all tests together for performance, since Blender - # startup time is a significant factor. - tmp_filepaths = [] - for filepath in filepaths: - testname = test_get_name(filepath) - print_message(testname, 'SUCCESS', 'RUN') - tmp_filepaths.append(os.path.join(self.output_dir, "tmp_" + testname)) - - run_errors = render_cb(filepaths, tmp_filepaths) - errors = [] - - for error, filepath, tmp_filepath in zip(run_errors, filepaths, tmp_filepaths): - if not error: - if os.path.getsize(tmp_filepath) == 0: - error = "VERIFY" - elif not self._diff_output(filepath, tmp_filepath): - error = "VERIFY" + def _run_test(self, filepath, render_cb): + testname = test_get_name(filepath) + print_message(testname, 'SUCCESS', 'RUN') + time_start = time.time() + tmp_filepath = os.path.join(self.output_dir, "tmp_" + testname) - if os.path.exists(tmp_filepath): - os.remove(tmp_filepath) + error = render_cb(filepath, tmp_filepath) + status = "FAIL" + if not error: + if not self._diff_output(filepath, tmp_filepath): + error = "VERIFY" - errors.append(error) + if os.path.exists(tmp_filepath): + os.remove(tmp_filepath) - testname = test_get_name(filepath) - if not error: - print_message(testname, 'SUCCESS', 'OK') + time_end = time.time() + elapsed_ms = int((time_end - time_start) * 1000) + if not error: + print_message("{} ({} ms)" . format(testname, elapsed_ms), + 'SUCCESS', 'OK') + else: + if error == "NO_ENGINE": + print_message("Can't perform tests because the render engine failed to load!") + return error + elif error == "NO_START": + print_message('Can not perform tests because blender fails to start.', + 'Make sure INSTALL target was run.') + return error + elif error == 'VERIFY': + print_message("Rendered result is different from reference image") else: - if error == "SKIPPED": - print_message("Skipped after previous render caused error") - elif error == "NO_ENGINE": - print_message("Can't perform tests because the render engine failed to load!") - elif error == "NO_START": - print_message('Can not perform tests because blender fails to start.', - 'Make sure INSTALL target was run.') - elif error == 'VERIFY': - print_message("Rendered result is different from reference image") - else: - print_message("Unknown error %r" % error) - print_message(testname, 'FAILURE', 'FAILED') - - return errors + print_message("Unknown error %r" % error) + print_message("{} ({} ms)" . format(testname, elapsed_ms), + 'FAILURE', 'FAILED') + return error def _run_all_tests(self, dirname, dirpath, render_cb): passed_tests = [] @@ -426,8 +419,8 @@ class Report: format(len(all_files)), 'SUCCESS', "==========") time_start = time.time() - errors = self._run_tests(all_files, render_cb) - for filepath, error in zip(all_files, errors): + for filepath in all_files: + error = self._run_test(filepath, render_cb) testname = test_get_name(filepath) if error: if error == "NO_ENGINE": |