diff options
Diffstat (limited to 'tests')
-rw-r--r-- | tests/CMakeLists.txt | 7 | ||||
-rw-r--r-- | tests/performance/api/graph.py | 2 | ||||
-rw-r--r-- | tests/performance/tests/eevee.py | 129 | ||||
-rw-r--r-- | tests/python/CMakeLists.txt | 11 | ||||
-rw-r--r-- | tests/python/bl_blendfile_library_overrides.py | 116 | ||||
-rw-r--r-- | tests/python/bl_pyapi_bpy_driver_secure_eval.py | 220 | ||||
-rw-r--r-- | tests/python/bl_rigging_symmetrize.py | 26 | ||||
-rw-r--r-- | tests/python/bl_run_operators.py | 1 | ||||
-rw-r--r-- | tests/python/eevee_render_tests.py | 33 | ||||
-rw-r--r-- | tests/python/gpu_info.py | 26 | ||||
-rwxr-xr-x | tests/python/modules/render_report.py | 19 | ||||
-rw-r--r-- | tests/python/operators.py | 36 | ||||
-rw-r--r-- | tests/python/workbench_render_tests.py | 7 |
13 files changed, 588 insertions, 45 deletions
diff --git a/tests/CMakeLists.txt b/tests/CMakeLists.txt index 484ffd17046..6d1c838ad6d 100644 --- a/tests/CMakeLists.txt +++ b/tests/CMakeLists.txt @@ -5,12 +5,7 @@ # # Getting the install path of the executable is somewhat involved, as there are # no direct CMake generator expressions to get the install paths of executables. -get_property(GENERATOR_IS_MULTI_CONFIG GLOBAL PROPERTY GENERATOR_IS_MULTI_CONFIG) -if(GENERATOR_IS_MULTI_CONFIG) - string(REPLACE "\${BUILD_TYPE}" "$<CONFIG>" TEST_INSTALL_DIR ${CMAKE_INSTALL_PREFIX}) -else() - string(REPLACE "\${BUILD_TYPE}" "" TEST_INSTALL_DIR ${CMAKE_INSTALL_PREFIX}) -endif() +set(TEST_INSTALL_DIR ${CMAKE_INSTALL_PREFIX_WITH_CONFIG}) # Path to Blender and Python executables for all platforms. if(MSVC) diff --git a/tests/performance/api/graph.py b/tests/performance/api/graph.py index 6c9ba70141f..d95d386569e 100644 --- a/tests/performance/api/graph.py +++ b/tests/performance/api/graph.py @@ -74,7 +74,7 @@ class TestGraph: revisions[revision] = len(revisions) revision_dates[revision] = int(entry.date) - # Google Charts JSON data layout is like a spreadsheat table, with + # Google Charts JSON data layout is like a spreadsheet table, with # columns, rows, and cells. We create one column for revision labels, # and one column for each test. cols = [] diff --git a/tests/performance/tests/eevee.py b/tests/performance/tests/eevee.py new file mode 100644 index 00000000000..df7cac695da --- /dev/null +++ b/tests/performance/tests/eevee.py @@ -0,0 +1,129 @@ +# SPDX-License-Identifier: Apache-2.0 + +import os +import enum +import time + + +class RecordStage(enum.Enum): + INIT = 0, + WAIT_SHADERS = 1, + WARMUP = 2, + RECORD = 3, + FINISHED = 4 + + +WARMUP_SECONDS = 3 +WARMUP_FRAMES = 10 +SHADER_FALLBACK_SECONDS = 60 +RECORD_PLAYBACK_ITER = 3 +LOG_KEY = "ANIMATION_PERFORMANCE: " + + +def _run(args): + import bpy + + global record_stage + record_stage = RecordStage.INIT + + bpy.app.handlers.frame_change_post.append(frame_change_handler) + bpy.ops.screen.animation_play() + + +def frame_change_handler(scene): + import bpy + + global record_stage + global start_time + global start_record_time + global start_warmup_time + global warmup_frame + global stop_record_time + global playback_iteration + + if record_stage == RecordStage.INIT: + screen = bpy.context.window_manager.windows[0].screen + bpy.context.scene.sync_mode = 'NONE' + + for area in screen.areas: + if area.type == 'VIEW_3D': + space = area.spaces[0] + space.shading.type = 'RENDERED' + space.overlay.show_overlays = False + + start_time = time.perf_counter() + record_stage = RecordStage.WAIT_SHADERS + + elif record_stage == RecordStage.WAIT_SHADERS: + shaders_compiled = False + if hasattr(bpy.app, 'is_job_running'): + shaders_compiled = not bpy.app.is_job_running("SHADER_COMPILATION") + else: + # Fallback when is_job_running doesn't exists by waiting for a time. + shaders_compiled = time.perf_counter() - start_time > SHADER_FALLBACK_SECONDS + + if shaders_compiled: + start_warmup_time = time.perf_counter() + warmup_frame = 0 + record_stage = RecordStage.WARMUP + + elif record_stage == RecordStage.WARMUP: + warmup_frame += 1 + if time.perf_counter() - start_warmup_time > WARMUP_SECONDS and warmup_frame > WARMUP_FRAMES: + start_record_time = time.perf_counter() + playback_iteration = 0 + scene = bpy.context.scene + scene.frame_set(scene.frame_start) + record_stage = RecordStage.RECORD + + elif record_stage == RecordStage.RECORD: + current_time = time.perf_counter() + scene = bpy.context.scene + if scene.frame_current == scene.frame_end: + playback_iteration += 1 + + if playback_iteration >= RECORD_PLAYBACK_ITER: + stop_record_time = current_time + record_stage = RecordStage.FINISHED + + elif record_stage == RecordStage.FINISHED: + bpy.ops.screen.animation_cancel() + num_frames = RECORD_PLAYBACK_ITER * ((scene.frame_end - scene.frame_start) + 1) + elapse_seconds = stop_record_time - start_record_time + avg_frame_time = elapse_seconds / num_frames + fps = 1.0 / avg_frame_time + print(f"{LOG_KEY}{{'time': {avg_frame_time}, 'fps': {fps} }}") + bpy.app.handlers.frame_change_post.remove(frame_change_handler) + bpy.ops.wm.quit_blender() + + +if __name__ == '__main__': + _run(None) + +else: + import api + + class EeveeTest(api.Test): + def __init__(self, filepath): + self.filepath = filepath + + def name(self): + return self.filepath.stem + + def category(self): + return "eevee" + + def run(self, env, device_id): + args = {} + _, log = env.run_in_blender(_run, args, [self.filepath], foreground=True) + for line in log: + if line.startswith(LOG_KEY): + result_str = line[len(LOG_KEY):] + result = eval(result_str) + return result + + raise Exception("No playback performance result found in log.") + + def generate(env): + filepaths = env.find_blend_files('eevee/*') + return [EeveeTest(filepath) for filepath in filepaths] diff --git a/tests/python/CMakeLists.txt b/tests/python/CMakeLists.txt index 38c3fc4389a..ca3070b60ad 100644 --- a/tests/python/CMakeLists.txt +++ b/tests/python/CMakeLists.txt @@ -102,6 +102,11 @@ add_blender_test( ) add_blender_test( + script_pyapi_bpy_driver_secure_eval + --python ${CMAKE_CURRENT_LIST_DIR}/bl_pyapi_bpy_driver_secure_eval.py +) + +add_blender_test( script_pyapi_idprop --python ${CMAKE_CURRENT_LIST_DIR}/bl_pyapi_idprop.py ) @@ -630,8 +635,8 @@ if(WITH_CYCLES OR WITH_OPENGL_RENDER_TESTS) MESSAGE(WARNING "Disabling render tests because OIIO idiff does not exist") elseif(NOT EXISTS "${TEST_SRC_DIR}/render/shader") MESSAGE(WARNING "Disabling render tests because tests folder does not exist at ${TEST_SRC_DIR}") - elseif(NOT WITH_COMPOSITOR) - MESSAGE(WARNING "Disabling render tests because WITH_COMPOSITOR is disabled") + elseif(NOT WITH_COMPOSITOR_CPU) + MESSAGE(WARNING "Disabling render tests because WITH_COMPOSITOR_CPU is disabled") elseif(NOT WITH_OPENCOLORIO) MESSAGE(WARNING "Disabling render tests because WITH_OPENCOLORIO is disabled") else() @@ -730,7 +735,7 @@ if(WITH_CYCLES OR WITH_OPENGL_RENDER_TESTS) endif() endif() -if(WITH_COMPOSITOR) +if(WITH_COMPOSITOR_CPU) set(compositor_tests color converter diff --git a/tests/python/bl_blendfile_library_overrides.py b/tests/python/bl_blendfile_library_overrides.py index 1acc1e4d862..3ba99bd61e4 100644 --- a/tests/python/bl_blendfile_library_overrides.py +++ b/tests/python/bl_blendfile_library_overrides.py @@ -181,9 +181,125 @@ class TestLibraryTemplate(TestHelper, unittest.TestCase): assert(operation.operation == 'NOOP') +class TestLibraryOverridesResync(TestHelper, unittest.TestCase): + DATA_NAME_CONTAINER = "LibCollection" + DATA_NAME_RIGGED = "LibRigged" + DATA_NAME_RIG = "LibRig" + DATA_NAME_CONTROLLER_1 = "LibController1" + DATA_NAME_CONTROLLER_2 = "LibController2" + + def __init__(self, args): + self.args = args + + output_dir = pathlib.Path(self.args.output_dir) + self.ensure_path(str(output_dir)) + self.output_path = output_dir / "blendlib_overrides.blend" + self.test_output_path = output_dir / "blendlib_overrides_test.blend" + + bpy.ops.wm.read_homefile(use_empty=True, use_factory_startup=True) + + collection_container = bpy.data.collections.new(TestLibraryOverridesResync.DATA_NAME_CONTAINER) + bpy.context.collection.children.link(collection_container) + + mesh = bpy.data.meshes.new(TestLibraryOverridesResync.DATA_NAME_RIGGED) + obj_child = bpy.data.objects.new(TestLibraryOverridesResync.DATA_NAME_RIGGED, object_data=mesh) + collection_container.objects.link(obj_child) + armature = bpy.data.armatures.new(TestLibraryOverridesResync.DATA_NAME_RIG) + obj_armature = bpy.data.objects.new(TestLibraryOverridesResync.DATA_NAME_RIG, object_data=armature) + obj_child.parent = obj_armature + collection_container.objects.link(obj_armature) + + obj_child_modifier = obj_child.modifiers.new("", 'ARMATURE') + obj_child_modifier.object = obj_armature + + obj_ctrl1 = bpy.data.objects.new(TestLibraryOverridesResync.DATA_NAME_CONTROLLER_1, object_data=None) + collection_container.objects.link(obj_ctrl1) + + obj_armature_constraint = obj_armature.constraints.new('COPY_LOCATION') + obj_armature_constraint.target = obj_ctrl1 + + collection_sub = bpy.data.collections.new(TestLibraryOverridesResync.DATA_NAME_CONTROLLER_2) + collection_container.children.link(collection_sub) + obj_ctrl2 = bpy.data.objects.new(TestLibraryOverridesResync.DATA_NAME_CONTROLLER_2, object_data=None) + collection_sub.objects.link(obj_ctrl2) + + bpy.ops.wm.save_as_mainfile(filepath=str(self.output_path), check_existing=False, compress=False) + + def test_link_and_override_resync(self): + bpy.ops.wm.read_homefile(use_empty=True, use_factory_startup=True) + bpy.data.orphans_purge() + + link_dir = self.output_path / "Collection" + bpy.ops.wm.link( + directory=str(link_dir), + filename=TestLibraryOverridesResync.DATA_NAME_CONTAINER, + instance_collections=False, + ) + + linked_collection_container = bpy.data.collections[TestLibraryOverridesResync.DATA_NAME_CONTAINER] + assert(linked_collection_container.library is not None) + assert(linked_collection_container.override_library is None) + assert(len(bpy.data.collections) == 2) + assert(all(id_.library is not None for id_ in bpy.data.collections)) + assert(len(bpy.data.objects) == 4) + assert(all(id_.library is not None for id_ in bpy.data.objects)) + assert(len(bpy.data.meshes) == 1) + assert(all(id_.library is not None for id_ in bpy.data.meshes)) + assert(len(bpy.data.armatures) == 1) + assert(all(id_.library is not None for id_ in bpy.data.armatures)) + + override_collection_container = linked_collection_container.override_hierarchy_create( + bpy.context.scene, + bpy.context.view_layer, + ) + assert(override_collection_container.library is None) + assert(override_collection_container.override_library is not None) + # Objects and collections are duplicated as overrides, but meshes and armatures remain only linked data. + assert(len(bpy.data.collections) == 4) + assert(all((id_.library is None and id_.override_library is not None) for id_ in bpy.data.collections[:2])) + assert(len(bpy.data.objects) == 8) + assert(all((id_.library is None and id_.override_library is not None) for id_ in bpy.data.objects[:4])) + assert(len(bpy.data.meshes) == 1) + assert(len(bpy.data.armatures) == 1) + + bpy.ops.wm.save_as_mainfile(filepath=str(self.test_output_path), check_existing=False, compress=False) + + # Re-open the lib file, and change its ID relationships. + bpy.ops.wm.open_mainfile(filepath=str(self.output_path)) + + obj_armature = bpy.data.objects[TestLibraryOverridesResync.DATA_NAME_RIG] + obj_armature_constraint = obj_armature.constraints[0] + obj_ctrl2 = bpy.data.objects[TestLibraryOverridesResync.DATA_NAME_CONTROLLER_2] + obj_armature_constraint.target = obj_ctrl2 + + bpy.ops.wm.save_as_mainfile(filepath=str(self.output_path), check_existing=False, compress=False) + + # Re-open the main file, and check that automatic resync did its work correctly, remapping the target of the + # armature constraint to controller 2, without creating unexpected garbage IDs along the line. + bpy.ops.wm.open_mainfile(filepath=str(self.test_output_path)) + + override_collection_container = bpy.data.collections[TestLibraryOverridesResync.DATA_NAME_CONTAINER] + assert(override_collection_container.library is None) + assert(override_collection_container.override_library is not None) + # Objects and collections are duplicated as overrides, but meshes and armatures remain only linked data. + assert(len(bpy.data.collections) == 4) + assert(all((id_.library is None and id_.override_library is not None) for id_ in bpy.data.collections[:2])) + assert(len(bpy.data.objects) == 8) + assert(all((id_.library is None and id_.override_library is not None) for id_ in bpy.data.objects[:4])) + assert(len(bpy.data.meshes) == 1) + assert(len(bpy.data.armatures) == 1) + + obj_armature = bpy.data.objects[TestLibraryOverridesResync.DATA_NAME_RIG] + obj_ctrl2 = bpy.data.objects[TestLibraryOverridesResync.DATA_NAME_CONTROLLER_2] + assert(obj_armature.library is None and obj_armature.override_library is not None) + assert(obj_ctrl2.library is None and obj_ctrl2.override_library is not None) + assert(obj_armature.constraints[0].target == obj_ctrl2) + + TESTS = ( TestLibraryOverrides, TestLibraryTemplate, + TestLibraryOverridesResync, ) diff --git a/tests/python/bl_pyapi_bpy_driver_secure_eval.py b/tests/python/bl_pyapi_bpy_driver_secure_eval.py new file mode 100644 index 00000000000..953dbcd5381 --- /dev/null +++ b/tests/python/bl_pyapi_bpy_driver_secure_eval.py @@ -0,0 +1,220 @@ +# SPDX-License-Identifier: GPL-2.0-or-later + +# ./blender.bin --background -noaudio --python tests/python/bl_pyapi_bpy_driver_secure_eval.py -- --verbose +import bpy +import unittest +import builtins +from types import ModuleType + + +# ----------------------------------------------------------------------------- +# Mock Environment + + +expect_unreachable_msg = "This function should _NEVER_ run!" +# Internal check, to ensure this actually runs as expected. +expect_unreachable_count = 0 + + +def expect_os_unreachable(): + global expect_unreachable_count + expect_unreachable_count += 1 + raise Exception(expect_unreachable_msg) + + +__import__("os").expect_os_unreachable = expect_os_unreachable + + +expect_open_unreachable_count = 0 + + +def open_expect_unreachable(*args, **kwargs): + global expect_open_unreachable_count + expect_open_unreachable_count += 1 + raise Exception(expect_unreachable_msg) + + +mock_builtins = {**builtins.__dict__, **{"open": open_expect_unreachable}} + + +# ----------------------------------------------------------------------------- +# Utility Functions + + +def is_expression_secure(expr_str, verbose): + """ + Return (ok, code) where ok is true if expr_str is considered secure. + """ + # Internal function only for testing (not part of the public API). + from _bpy import _driver_secure_code_test + expr_code = compile(expr_str, "<is_expression_secure>", 'eval') + ok = _driver_secure_code_test(expr_code, verbose=verbose) + return ok, expr_code + + +# ----------------------------------------------------------------------------- +# Tests (Accept) + + +class _TestExprMixIn: + """ + Sub-classes must define: + - expressions_expect_secure: boolean, the expected secure state. + - expressions: A sequence of expressions that must evaluate in the driver name-space. + + Optionally: + - expressions_expect_unreachable: + A boolean, when true, it's expected each expression should call + ``expect_os_unreachable`` or ``expect_open_unreachable``. + """ + + # Sub-class may override. + expressions_expect_unreachable = False + + def assertSecure(self, expect_secure, expr_str): + is_secure, expr_code = is_expression_secure( + expr_str, + # Only verbose when secure as this is will result in an failure, + # in that case it's useful to know which op-codes caused the test to unexpectedly fail. + verbose=expect_secure, + ) + if is_secure != expect_secure: + raise self.failureException( + "Expression \"%s\" was expected to be %s" % + (expr_str, "secure" if expect_secure else "insecure")) + # NOTE: executing is not essential, it's just better to ensure the expressions make sense. + try: + exec( + expr_code, + {"__builtins__": mock_builtins}, + {**bpy.app.driver_namespace, **{"__builtins__": mock_builtins}}, + ) + # exec(expr_code, {}, bpy.app.driver_namespace) + ex = None + except BaseException as ex_test: + ex = ex_test + + if self.expressions_expect_unreachable: + if ex and ex.args == (expect_unreachable_msg,): + ex = None + elif not ex: + raise self.failureException("Expression \"%s\" failed to run `os.expect_os_unreachable`" % (expr_str,)) + else: + # An unknown exception was raised, use the exception below. + pass + + if ex: + raise self.failureException("Expression \"%s\" failed to evaluate with error: %r" % (expr_str, ex)) + + def test_expr(self): + expect_secure = self.expressions_expect_secure + for expr_str in self.expressions: + self.assertSecure(expect_secure, expr_str) + + +class TestExprMixIn_Accept(_TestExprMixIn): + expressions_expect_secure = True + + +class TestExprMixIn_Reject(_TestExprMixIn): + expressions_expect_secure = False + + +class TestAcceptLiteralNumbers(unittest.TestCase, TestExprMixIn_Accept): + expressions = ("1", "1_1", "1.1", "1j", "0x1", "0o1", "0b1") + + +class TestAcceptLiteralStrings(unittest.TestCase, TestExprMixIn_Accept): + expressions = ("''", "'_'", "r''", "r'_'", "'''_'''") + + +class TestAcceptSequencesEmpty(unittest.TestCase, TestExprMixIn_Accept): + expressions = ("()", "[]", "{}", "[[]]", "(())") + + +class TestAcceptSequencesSimple(unittest.TestCase, TestExprMixIn_Accept): + expressions = ("('', '')", "['', '_']", "{'', '_'}", "{'': '_'}") + + +class TestAcceptSequencesExpand(unittest.TestCase, TestExprMixIn_Accept): + expressions = ("(*('', '_'),)", "[*(), *[]]", "{*{1, 2}}") + + +class TestAcceptSequencesComplex(unittest.TestCase, TestExprMixIn_Accept): + expressions = ("[1, 2, 3][-1:0:-1][0]", "1 in (1, 2)", "False if 1 in {1, 2} else True") + + +class TestAcceptMathOperators(unittest.TestCase, TestExprMixIn_Accept): + expressions = ("4 / 4", "4 * 4", "4 // 4", "2 ** 2", "4 ^ -1", "4 & 1", "4 % 1") + + +class TestAcceptMathFunctionsSimple(unittest.TestCase, TestExprMixIn_Accept): + expressions = ("sin(pi)", "degrees(pi / 2)", "clamp(4, 0, 1)") + + +class TestAcceptMathFunctionsComplex(unittest.TestCase, TestExprMixIn_Accept): + expressions = ("-(sin(pi) ** 2) / 2", "floor(22 / 7)", "ceil(pi + 1)") + + +# ----------------------------------------------------------------------------- +# Tests (Reject) + +class TestRejectLiteralFStrings(unittest.TestCase, TestExprMixIn_Reject): + # F-String's are not supported as `BUILD_STRING` op-code is disabled, + # while it may be safe to enable that needs to be double-checked. + # Further it doesn't seem useful for typical math expressions used in drivers. + expressions = ("f''", "f'{1}'", "f'{\"_\"}'") + + +class TestRejectModuleAccess(unittest.TestCase, TestExprMixIn_Reject): + # Each of these commands _must_ run `expect_os_unreachable`, + # and must also be rejected as insecure - otherwise we have problems. + expressions_expect_unreachable = True + expressions = ( + "__import__('os').expect_os_unreachable()", + "exec(\"__import__('os').expect_os_unreachable()\")", + "(globals().update(__import__('os').__dict__), expect_os_unreachable())", + ) + + # Ensure the functions are actually called. + def setUp(self): + self._count = expect_unreachable_count + + def tearDown(self): + count_actual = expect_unreachable_count - self._count + count_expect = len(self.expressions) + if count_actual != count_expect: + raise Exception( + "Expected 'os.expect_os_unreachable' to be called %d times but was called %d times" % + (count_expect, count_actual), + ) + + +class TestRejectOpenAccess(unittest.TestCase, TestExprMixIn_Reject): + # Each of these commands _must_ run `expect_open_unreachable`, + # and must also be rejected as insecure - otherwise we have problems. + expressions_expect_unreachable = True + expressions = ( + "open('file.txt', 'r')", + "exec(\"open('file.txt', 'r')\")", + "(globals().update({'fake_open': __builtins__['open']}), fake_open())", + ) + + # Ensure the functions are actually called. + def setUp(self): + self._count = expect_open_unreachable_count + + def tearDown(self): + count_actual = expect_open_unreachable_count - self._count + count_expect = len(self.expressions) + if count_actual != count_expect: + raise Exception( + "Expected 'open' to be called %d times but was called %d times" % + (count_expect, count_actual), + ) + + +if __name__ == '__main__': + import sys + sys.argv = [__file__] + (sys.argv[sys.argv.index("--") + 1:] if "--" in sys.argv else []) + unittest.main() diff --git a/tests/python/bl_rigging_symmetrize.py b/tests/python/bl_rigging_symmetrize.py index 963be6d41d3..10ba99ac6e9 100644 --- a/tests/python/bl_rigging_symmetrize.py +++ b/tests/python/bl_rigging_symmetrize.py @@ -40,7 +40,7 @@ def check_loc_rot_scale(self, bone, exp_bone): def check_parent(self, bone, exp_bone): self.assertEqual(type(bone.parent), type(exp_bone.parent), - "Missmatching types in pose.bones[%s].parent" % (bone.name)) + "Mismatching types in pose.bones[%s].parent" % (bone.name)) self.assertTrue(bone.parent is None or bone.parent.name == exp_bone.parent.name, "Bone parent does not match on bone %s" % (bone.name)) @@ -56,17 +56,17 @@ def check_bendy_bones(self, bone, exp_bone): exp_value = getattr(exp_bone, var) self.assertEqual(type(value), type(exp_value), - "Missmatching types in pose.bones[%s].%s" % (bone.name, var)) + "Mismatching types in pose.bones[%s].%s" % (bone.name, var)) if isinstance(value, str): self.assertEqual(value, exp_value, - "Missmatching value in pose.bones[%s].%s" % (bone.name, var)) + "Mismatching value in pose.bones[%s].%s" % (bone.name, var)) elif hasattr(value, "name"): self.assertEqual(value.name, exp_value.name, - "Missmatching value in pose.bones[%s].%s" % (bone.name, var)) + "Mismatching value in pose.bones[%s].%s" % (bone.name, var)) else: self.assertAlmostEqual(value, exp_value, - "Missmatching value in pose.bones[%s].%s" % (bone.name, var)) + "Mismatching value in pose.bones[%s].%s" % (bone.name, var)) def check_ik(self, bone, exp_bone): @@ -81,7 +81,7 @@ def check_ik(self, bone, exp_bone): value = getattr(bone, var) exp_value = getattr(exp_bone, var) self.assertAlmostEqual(value, exp_value, - "Missmatching value in pose.bones[%s].%s" % (bone.name, var)) + "Mismatching value in pose.bones[%s].%s" % (bone.name, var)) def check_constraints(self, input_arm, expected_arm, bone, exp_bone): @@ -89,7 +89,7 @@ def check_constraints(self, input_arm, expected_arm, bone, exp_bone): expo_const_len = len(exp_bone.constraints) self.assertEqual(const_len, expo_const_len, - "Constraints missmatch on bone %s" % (bone.name)) + "Constraints mismatch on bone %s" % (bone.name)) for exp_constraint in exp_bone.constraints: const_name = exp_constraint.name @@ -111,28 +111,28 @@ def check_constraints(self, input_arm, expected_arm, bone, exp_bone): exp_value = getattr(exp_constraint, var) self.assertEqual(type(value), type(exp_value), - "Missmatching constraint value types in pose.bones[%s].constraints[%s].%s" % ( + "Mismatching constraint value types in pose.bones[%s].constraints[%s].%s" % ( bone.name, const_name, var)) if isinstance(value, str): self.assertEqual(value, exp_value, - "Missmatching constraint value in pose.bones[%s].constraints[%s].%s" % ( + "Mismatching constraint value in pose.bones[%s].constraints[%s].%s" % ( bone.name, const_name, var)) elif hasattr(value, "name"): - # Some constraints targets the armature itself, so the armature name should missmatch. + # Some constraints targets the armature itself, so the armature name should mismatch. if value.name == input_arm.name and exp_value.name == expected_arm.name: continue self.assertEqual(value.name, exp_value.name, - "Missmatching constraint value in pose.bones[%s].constraints[%s].%s" % ( + "Mismatching constraint value in pose.bones[%s].constraints[%s].%s" % ( bone.name, const_name, var)) elif isinstance(value, bool): self.assertEqual(value, exp_value, - "Missmatching constraint boolean in pose.bones[%s].constraints[%s].%s" % ( + "Mismatching constraint boolean in pose.bones[%s].constraints[%s].%s" % ( bone.name, const_name, var)) else: - msg = "Missmatching constraint value in pose.bones[%s].constraints[%s].%s" % ( + msg = "Mismatching constraint value in pose.bones[%s].constraints[%s].%s" % ( bone.name, const_name, var) self.assertAlmostEqual(value, exp_value, places=6, msg=msg) diff --git a/tests/python/bl_run_operators.py b/tests/python/bl_run_operators.py index a2478bd7547..ccb0814e5eb 100644 --- a/tests/python/bl_run_operators.py +++ b/tests/python/bl_run_operators.py @@ -317,7 +317,6 @@ def ctx_editmode_mesh_extra(): bpy.ops.object.shape_key_add(from_mix=False) bpy.ops.object.shape_key_add(from_mix=True) bpy.ops.mesh.uv_texture_add() - bpy.ops.mesh.vertex_color_add() bpy.ops.object.material_slot_add() # editmode last! bpy.ops.object.mode_set(mode='EDIT') diff --git a/tests/python/eevee_render_tests.py b/tests/python/eevee_render_tests.py index 8c6f08ae76e..9ed850fcb52 100644 --- a/tests/python/eevee_render_tests.py +++ b/tests/python/eevee_render_tests.py @@ -3,10 +3,12 @@ import argparse import os +import pathlib import shlex import shutil import subprocess import sys +from pathlib import Path def setup(): @@ -98,6 +100,26 @@ if inside_blender: sys.exit(1) +def get_gpu_device_type(blender): + command = [ + blender, + "-noaudio", + "--background" + "--factory-startup", + "--python", + str(pathlib.Path(__file__).parent / "gpu_info.py") + ] + try: + completed_process = subprocess.run(command, stdout=subprocess.PIPE) + for line in completed_process.stdout.read_text(): + if line.startswith("GPU_DEVICE_TYPE:"): + vendor = line.split(':')[1] + return vendor + except BaseException as e: + return None + return None + + def get_arguments(filepath, output_filepath): return [ "--background", @@ -133,11 +155,22 @@ def main(): idiff = args.idiff[0] output_dir = args.outdir[0] + gpu_device_type = get_gpu_device_type(blender) + reference_override_dir = None + if gpu_device_type == "AMD": + reference_override_dir = "eevee_renders/amd" + from modules import render_report report = render_report.Report("Eevee", output_dir, idiff) report.set_pixelated(True) report.set_reference_dir("eevee_renders") + report.set_reference_override_dir(reference_override_dir) report.set_compare_engine('cycles', 'CPU') + + test_dir_name = Path(test_dir).name + if test_dir_name.startswith('image'): + report.set_fail_threshold(0.051) + ok = report.run(test_dir, blender, get_arguments, batch=True) sys.exit(not ok) diff --git a/tests/python/gpu_info.py b/tests/python/gpu_info.py new file mode 100644 index 00000000000..426ce29e85d --- /dev/null +++ b/tests/python/gpu_info.py @@ -0,0 +1,26 @@ +# SPDX-License-Identifier: GPL-2.0-or-later + +""" +Prints GPU back-end information to the console and exits. + +Use this script as `blender --background --python gpu_info.py`. +""" +import bpy +import gpu +import sys + +# Render with workbench to initialize the GPU backend otherwise it would fail when running in +# background mode as the GPU backend won't be initialized. +scene = bpy.context.scene +scene.render.resolution_x = 1 +scene.render.resolution_y = 1 +scene.render.engine = "BLENDER_WORKBENCH" +bpy.ops.render.render(animation=False, write_still=False) + + +print('GPU_VENDOR:' + gpu.platform.vendor_get()) +print('GPU_RENDERER:' + gpu.platform.renderer_get()) +print('GPU_VERSION:' + gpu.platform.version_get()) +print('GPU_DEVICE_TYPE:' + gpu.platform.device_type_get()) + +sys.exit(0) diff --git a/tests/python/modules/render_report.py b/tests/python/modules/render_report.py index 15441918800..15d46d6d127 100755 --- a/tests/python/modules/render_report.py +++ b/tests/python/modules/render_report.py @@ -78,12 +78,18 @@ def test_get_name(filepath): return os.path.splitext(filename)[0] -def test_get_images(output_dir, filepath, reference_dir): +def test_get_images(output_dir, filepath, reference_dir, reference_override_dir): testname = test_get_name(filepath) dirpath = os.path.dirname(filepath) old_dirpath = os.path.join(dirpath, reference_dir) old_img = os.path.join(old_dirpath, testname + ".png") + if reference_override_dir: + override_dirpath = os.path.join(dirpath, reference_override_dir) + override_img = os.path.join(override_dirpath, testname + ".png") + if os.path.exists(override_img): + old_dirpath = override_dirpath + old_img = override_img ref_dirpath = os.path.join(output_dir, os.path.basename(dirpath), "ref") ref_img = os.path.join(ref_dirpath, testname + ".png") @@ -108,6 +114,7 @@ class Report: 'output_dir', 'global_dir', 'reference_dir', + 'reference_override_dir', 'idiff', 'pixelated', 'fail_threshold', @@ -127,6 +134,7 @@ class Report: self.output_dir = output_dir self.global_dir = os.path.dirname(output_dir) self.reference_dir = 'reference_renders' + self.reference_override_dir = None self.idiff = idiff self.compare_engine = None self.fail_threshold = 0.016 @@ -161,6 +169,9 @@ class Report: def set_reference_dir(self, reference_dir): self.reference_dir = reference_dir + def set_reference_override_dir(self, reference_override_dir): + self.reference_override_dir = reference_override_dir + def set_compare_engine(self, other_engine, other_device=None): self.compare_engine = (other_engine, other_device) @@ -343,7 +354,8 @@ class Report: name = test_get_name(filepath) name = name.replace('_', ' ') - old_img, ref_img, new_img, diff_img = test_get_images(self.output_dir, filepath, self.reference_dir) + old_img, ref_img, new_img, diff_img = test_get_images( + self.output_dir, filepath, self.reference_dir, self.reference_override_dir) status = error if error else "" tr_style = """ class="table-danger" """ if error else "" @@ -390,7 +402,8 @@ class Report: self.compare_tests += test_html def _diff_output(self, filepath, tmp_filepath): - old_img, ref_img, new_img, diff_img = test_get_images(self.output_dir, filepath, self.reference_dir) + old_img, ref_img, new_img, diff_img = test_get_images( + self.output_dir, filepath, self.reference_dir, self.reference_override_dir) # Create reference render directory. old_dirpath = os.path.dirname(old_img) diff --git a/tests/python/operators.py b/tests/python/operators.py index 548a2b50b05..fc2e8e39d4f 100644 --- a/tests/python/operators.py +++ b/tests/python/operators.py @@ -115,6 +115,18 @@ def main(): [OperatorSpecEditMode("dissolve_faces", {}, "VERT", {5, 34, 47, 49, 83, 91, 95})], ), + # dissolve limited + SpecMeshTest( + "SphereDissolveLimited", "testSphereDissolveLimited", "expectedSphereDissolveLimited", + [OperatorSpecEditMode("dissolve_limited", {"angle_limit": 0.610865}, "FACE", {20, 23, 26, 29, 32})], + ), + + # dissolve mode + SpecMeshTest( + "PlaneDissolveMode", "testPlaneDissolveMode", "expectedPlaneDissolveMode", + [OperatorSpecEditMode("dissolve_mode", {"use_verts": True}, "FACE", {0, 1, 2, 10, 12, 13})], + ), + # dissolve verts SpecMeshTest( "CubeDissolveVerts", "testCubeDissolveVerts", "expectedCubeDissolveVerts", @@ -332,6 +344,12 @@ def main(): [OperatorSpecEditMode("mark_seam", {}, "EDGE", {1})], ), + # merge normals + SpecMeshTest( + "CubeMergeNormals", "testCubeMergeNormals", "expectedCubeMergeNormals", + [OperatorSpecEditMode("merge_normals", {}, "FACE", {3, 5})], + ), + # select all SpecMeshTest( "CircleSelectAll", "testCircleSelectAll", "expectedCircleSelectAll", @@ -545,24 +563,6 @@ def main(): )], ), - # Vertex Colors - SpecMeshTest( - "VertexColorAdd", "testCubeColorAdd", "expectedCubeColorAdd", - [OperatorSpecEditMode("vertex_color_add", {}, "VERT", {})], - ), - SpecMeshTest( - "VertexColorRemove", "testCubeColorRemove", "expectedCubeColorRemove", - [OperatorSpecEditMode("vertex_color_remove", {}, "VERT", {})], - ), - SpecMeshTest( - "VertexColorSculptAdd", "testCubeSculptAdd", "expectedCubeSculptAdd", - [OperatorSpecEditMode("sculpt_vertex_color_add", {}, "VERT", {})], - ), - SpecMeshTest( - "VertexColorSculptRemove", "testCubeSculptRemove", "expectedCubeSculptRemove", - [OperatorSpecEditMode("sculpt_vertex_color_remove", {}, "VERT", {})], - ), - # Laplacian Smooth SpecMeshTest( "LaplacianSmoothDefault", "testSphereLaplacianSmoothDefault", "expectedSphereLaplacianSmoothDefault", diff --git a/tests/python/workbench_render_tests.py b/tests/python/workbench_render_tests.py index 3ceb0fb3226..e182b2a41e2 100644 --- a/tests/python/workbench_render_tests.py +++ b/tests/python/workbench_render_tests.py @@ -3,10 +3,12 @@ import argparse import os +import platform import shlex import shutil import subprocess import sys +from pathlib import Path def setup(): @@ -73,6 +75,11 @@ def main(): report.set_pixelated(True) report.set_reference_dir("workbench_renders") report.set_compare_engine('eevee') + + test_dir_name = Path(test_dir).name + if test_dir_name.startswith('hair') and platform.system() == "Darwin": + report.set_fail_threshold(0.050) + ok = report.run(test_dir, blender, get_arguments, batch=True) sys.exit(not ok) |