Welcome to mirror list, hosted at ThFree Co, Russian Federation.

git.blender.org/blender.git - Unnamed repository; edit this file 'description' to name the repository.
summaryrefslogtreecommitdiff
diff options
context:
space:
mode:
authorCampbell Barton <ideasman42@gmail.com>2021-09-22 07:48:01 +0300
committerCampbell Barton <ideasman42@gmail.com>2021-09-22 07:54:01 +0300
commit4d66cbd140b1648b79df0df695046cb718797b70 (patch)
tree945b1093ba250ad57611f44bda7ca0b8a31a0211
parent77061a5621015dfd0c9f89fd21cb23d706d0cec8 (diff)
Cleanup: spelling in comments
-rw-r--r--intern/cycles/blender/blender_gpu_display.cpp6
-rw-r--r--intern/cycles/blender/blender_gpu_display.h6
-rw-r--r--intern/cycles/device/cpu/kernel.cpp2
-rw-r--r--intern/cycles/device/cuda/kernel.cpp2
-rw-r--r--intern/cycles/device/device.h2
-rw-r--r--intern/cycles/device/device_denoise.h2
-rw-r--r--intern/cycles/device/optix/device_impl.cpp4
-rw-r--r--intern/cycles/device/optix/device_impl.h2
-rw-r--r--intern/cycles/integrator/denoiser.h10
-rw-r--r--intern/cycles/integrator/denoiser_device.cpp2
-rw-r--r--intern/cycles/integrator/denoiser_oidn.cpp4
-rw-r--r--intern/cycles/integrator/path_trace.cpp6
-rw-r--r--intern/cycles/integrator/path_trace.h16
-rw-r--r--intern/cycles/integrator/path_trace_work.h10
-rw-r--r--intern/cycles/integrator/path_trace_work_gpu.cpp6
-rw-r--r--intern/cycles/integrator/render_scheduler.cpp8
-rw-r--r--intern/cycles/integrator/render_scheduler.h10
-rw-r--r--intern/cycles/integrator/shader_eval.cpp2
-rw-r--r--intern/cycles/integrator/work_balancer.h2
-rw-r--r--intern/cycles/integrator/work_tile_scheduler.cpp2
-rw-r--r--intern/cycles/integrator/work_tile_scheduler.h4
-rw-r--r--intern/cycles/kernel/closure/bssrdf.h2
-rw-r--r--intern/cycles/kernel/device/cuda/compat.h2
-rw-r--r--intern/cycles/kernel/device/optix/compat.h2
-rw-r--r--intern/cycles/kernel/integrator/integrator_shade_volume.h8
-rw-r--r--intern/cycles/kernel/integrator/integrator_state.h4
-rw-r--r--intern/cycles/kernel/kernel_adaptive_sampling.h2
-rw-r--r--intern/cycles/kernel/kernel_film.h10
-rw-r--r--intern/cycles/kernel/kernel_random.h4
-rw-r--r--intern/cycles/kernel/kernel_types.h2
-rw-r--r--intern/cycles/kernel/svm/svm_bevel.h2
-rw-r--r--intern/cycles/render/buffers.cpp2
-rw-r--r--intern/cycles/render/buffers.h6
-rw-r--r--intern/cycles/render/gpu_display.h10
-rw-r--r--intern/cycles/render/session.cpp8
-rw-r--r--intern/cycles/render/session.h6
-rw-r--r--intern/cycles/render/tile.cpp4
-rw-r--r--intern/cycles/render/tile.h2
-rw-r--r--source/blender/blenkernel/intern/geometry_component_instances.cc4
-rw-r--r--source/blender/blenkernel/intern/gpencil_geom.cc2
-rw-r--r--source/blender/blenkernel/intern/mesh_convert.cc2
-rw-r--r--source/blender/editors/space_file/filesel.c2
-rw-r--r--source/blender/editors/space_sequencer/sequencer_draw.c2
-rw-r--r--source/blender/gpencil_modifiers/intern/MOD_gpencillength.c4
-rw-r--r--source/blender/nodes/geometry/nodes/node_geo_curve_sample.cc2
-rw-r--r--source/blender/windowmanager/intern/wm_dragdrop.c2
46 files changed, 102 insertions, 102 deletions
diff --git a/intern/cycles/blender/blender_gpu_display.cpp b/intern/cycles/blender/blender_gpu_display.cpp
index a79232af71f..aa2a9c17a8a 100644
--- a/intern/cycles/blender/blender_gpu_display.cpp
+++ b/intern/cycles/blender/blender_gpu_display.cpp
@@ -338,9 +338,9 @@ bool BlenderGPUDisplay::do_update_begin(const GPUDisplayParams &params,
* NOTE: Allocate the PBO for the the size which will fit the final render resolution (as in,
* at a resolution divider 1. This was we don't need to recreate graphics interoperability
* objects which are costly and which are tied to the specific underlying buffer size.
- * The downside of this approach is that when graphics interopeability is not used we are sending
- * too much data to GPU when resolution divider is not 1. */
- /* TODO(sergey): Investigate whether keeping the PBO exact size of the texute makes non-interop
+ * The downside of this approach is that when graphics interoperability is not used we are
+ * sending too much data to GPU when resolution divider is not 1. */
+ /* TODO(sergey): Investigate whether keeping the PBO exact size of the texture makes non-interop
* mode faster. */
const int buffer_width = params.full_size.x;
const int buffer_height = params.full_size.y;
diff --git a/intern/cycles/blender/blender_gpu_display.h b/intern/cycles/blender/blender_gpu_display.h
index b7eddf0afa7..1014c96cee4 100644
--- a/intern/cycles/blender/blender_gpu_display.h
+++ b/intern/cycles/blender/blender_gpu_display.h
@@ -134,7 +134,7 @@ class BlenderGPUDisplay : public GPUDisplay {
/* Make sure texture is allocated and its initial configuration is performed. */
bool gl_texture_resources_ensure();
- /* Ensure all runtime GPU resources needefd for drawing are allocated.
+ /* Ensure all runtime GPU resources needed for drawing are allocated.
* Returns true if all resources needed for drawing are available. */
bool gl_draw_resources_ensure();
@@ -146,7 +146,7 @@ class BlenderGPUDisplay : public GPUDisplay {
* NOTE: The texture needs to be bound. */
void texture_update_if_needed();
- /* Update vetrex buffer with new coordinates of vertex positions and texture coordinates.
+ /* Update vertex buffer with new coordinates of vertex positions and texture coordinates.
* This buffer is used to render texture in the viewport.
*
* NOTE: The buffer needs to be bound. */
@@ -200,7 +200,7 @@ class BlenderGPUDisplay : public GPUDisplay {
bool gl_draw_resource_creation_attempted_ = false;
bool gl_draw_resources_created_ = false;
- /* Vertex buffer which hold vertrices of a triangle fan which is textures with the texture
+ /* Vertex buffer which hold vertices of a triangle fan which is textures with the texture
* holding the render result. */
uint vertex_buffer_ = 0;
diff --git a/intern/cycles/device/cpu/kernel.cpp b/intern/cycles/device/cpu/kernel.cpp
index 0ab58ff8600..91282390e27 100644
--- a/intern/cycles/device/cpu/kernel.cpp
+++ b/intern/cycles/device/cpu/kernel.cpp
@@ -44,7 +44,7 @@ CPUKernels::CPUKernels()
/* Shader evaluation. */
REGISTER_KERNEL(shader_eval_displace),
REGISTER_KERNEL(shader_eval_background),
- /* Adaptive campling. */
+ /* Adaptive sampling. */
REGISTER_KERNEL(adaptive_sampling_convergence_check),
REGISTER_KERNEL(adaptive_sampling_filter_x),
REGISTER_KERNEL(adaptive_sampling_filter_y),
diff --git a/intern/cycles/device/cuda/kernel.cpp b/intern/cycles/device/cuda/kernel.cpp
index 0ed20ddf8e6..a4a7bfabce0 100644
--- a/intern/cycles/device/cuda/kernel.cpp
+++ b/intern/cycles/device/cuda/kernel.cpp
@@ -28,7 +28,7 @@ void CUDADeviceKernels::load(CUDADevice *device)
for (int i = 0; i < (int)DEVICE_KERNEL_NUM; i++) {
CUDADeviceKernel &kernel = kernels_[i];
- /* No megakernel used for GPU. */
+ /* No mega-kernel used for GPU. */
if (i == DEVICE_KERNEL_INTEGRATOR_MEGAKERNEL) {
continue;
}
diff --git a/intern/cycles/device/device.h b/intern/cycles/device/device.h
index 02b6edb56d0..399d5eb91df 100644
--- a/intern/cycles/device/device.h
+++ b/intern/cycles/device/device.h
@@ -213,7 +213,7 @@ class Device {
* The interoperability comes here by the meaning that the device is capable of computing result
* directly into an OpenGL (or other graphics library) buffer. */
- /* Check display si to be updated using graphics interoperability.
+ /* Check display is to be updated using graphics interoperability.
* The interoperability can not be used is it is not supported by the device. But the device
* might also force disable the interoperability if it detects that it will be slower than
* copying pixels from the render buffer. */
diff --git a/intern/cycles/device/device_denoise.h b/intern/cycles/device/device_denoise.h
index 02ee63fb0ad..dfdc7cc87b3 100644
--- a/intern/cycles/device/device_denoise.h
+++ b/intern/cycles/device/device_denoise.h
@@ -68,7 +68,7 @@ class DenoiseParams : public Node {
/* Viewport start sample. */
int start_sample = 0;
- /* Auxiliry passes. */
+ /* Auxiliary passes. */
bool use_pass_albedo = true;
bool use_pass_normal = true;
diff --git a/intern/cycles/device/optix/device_impl.cpp b/intern/cycles/device/optix/device_impl.cpp
index cd16b8c9f01..b54d423a183 100644
--- a/intern/cycles/device/optix/device_impl.cpp
+++ b/intern/cycles/device/optix/device_impl.cpp
@@ -571,7 +571,7 @@ class OptiXDevice::DenoiseContext {
int pass_stride = -1;
} guiding_params;
- /* Number of input passes. Including the color and extra auxillary passes. */
+ /* Number of input passes. Including the color and extra auxiliary passes. */
int num_input_passes = 0;
bool use_pass_albedo = false;
bool use_pass_normal = false;
@@ -956,7 +956,7 @@ bool OptiXDevice::denoise_run(DenoiseContext &context, const DenoisePass &pass)
/* Denoise in-place of the noisy input in the render buffers. */
output_layer = color_layer;
- /* Finally run denonising. */
+ /* Finally run denoising. */
OptixDenoiserParams params = {}; /* All parameters are disabled/zero. */
OptixDenoiserLayer image_layers = {};
image_layers.input = color_layer;
diff --git a/intern/cycles/device/optix/device_impl.h b/intern/cycles/device/optix/device_impl.h
index 742ae0f1bab..91ef52e0a5a 100644
--- a/intern/cycles/device/optix/device_impl.h
+++ b/intern/cycles/device/optix/device_impl.h
@@ -146,7 +146,7 @@ class OptiXDevice : public CUDADevice {
/* Read guiding passes from the render buffers, preprocess them in a way which is expected by
* OptiX and store in the guiding passes memory within the given context.
*
- * Pre=-processing of the guiding passes is to only hapopen once per context lifetime. DO not
+ * Pre=-processing of the guiding passes is to only happen once per context lifetime. DO not
* preprocess them for every pass which is being denoised. */
bool denoise_filter_guiding_preprocess(DenoiseContext &context);
diff --git a/intern/cycles/integrator/denoiser.h b/intern/cycles/integrator/denoiser.h
index 3101b45e31b..b02bcbeb046 100644
--- a/intern/cycles/integrator/denoiser.h
+++ b/intern/cycles/integrator/denoiser.h
@@ -33,7 +33,7 @@ class Progress;
/* Implementation of a specific denoising algorithm.
*
- * This class takes care of breaking down denosiing algorithm into a series of device calls or to
+ * This class takes care of breaking down denoising algorithm into a series of device calls or to
* calls of an external API to denoise given input.
*
* TODO(sergey): Are we better with device or a queue here? */
@@ -53,7 +53,7 @@ class Denoiser {
const DenoiseParams &get_params() const;
/* Create devices and load kernels needed for denoising.
- * The progress is used to communicate state when kenrels actually needs to be loaded.
+ * The progress is used to communicate state when kernels actually needs to be loaded.
*
* NOTE: The `progress` is an optional argument, can be nullptr. */
virtual bool load_kernels(Progress *progress);
@@ -64,7 +64,7 @@ class Denoiser {
* a lower resolution render into a bigger allocated buffer, which is used in viewport during
* navigation and non-unit pixel size. Use that instead of render_buffers->params.
*
- * The buffer might be copming from a "foreign" device from what this denoise is created for.
+ * The buffer might be coming from a "foreign" device from what this denoise is created for.
* This means that in general case the denoiser will make sure the input data is available on
* the denoiser device, perform denoising, and put data back to the device where the buffer
* came from.
@@ -95,8 +95,8 @@ class Denoiser {
* using OptiX denoiser and rendering on CPU.
*
* - No threading safety is ensured in this call. This means, that it is up to caller to ensure
- * that there is no threadingconflict between denoising task lazily initializing the device and
- * access to this device happen. */
+ * that there is no threading-conflict between denoising task lazily initializing the device
+ * and access to this device happen. */
Device *get_denoiser_device() const;
function<bool(void)> is_cancelled_cb;
diff --git a/intern/cycles/integrator/denoiser_device.cpp b/intern/cycles/integrator/denoiser_device.cpp
index 8088cfd7800..e8361c50f2f 100644
--- a/intern/cycles/integrator/denoiser_device.cpp
+++ b/intern/cycles/integrator/denoiser_device.cpp
@@ -77,7 +77,7 @@ bool DeviceDenoiser::denoise_buffer(const BufferParams &buffer_params,
local_render_buffers.reset(buffer_params);
/* NOTE: The local buffer is allocated for an exact size of the effective render size, while
- * the input render buffer is allcoated for the lowest resolution divider possible. So it is
+ * the input render buffer is allocated for the lowest resolution divider possible. So it is
* important to only copy actually needed part of the input buffer. */
memcpy(local_render_buffers.buffer.data(),
render_buffers->buffer.data(),
diff --git a/intern/cycles/integrator/denoiser_oidn.cpp b/intern/cycles/integrator/denoiser_oidn.cpp
index 1b5a012ec87..7fc2b2b1892 100644
--- a/intern/cycles/integrator/denoiser_oidn.cpp
+++ b/intern/cycles/integrator/denoiser_oidn.cpp
@@ -93,7 +93,7 @@ class OIDNPass {
* Is required for albedo and normal passes. The color pass OIDN will perform auto-exposure, so
* scaling is not needed for the color pass unless adaptive sampling is used.
*
- * NOTE: Do not scale the outout pass, as that requires to be a pointer in the original buffer.
+ * NOTE: Do not scale the output pass, as that requires to be a pointer in the original buffer.
* All the scaling on the output needed for integration with adaptive sampling will happen
* outside of generic pass handling. */
bool need_scale = false;
@@ -479,7 +479,7 @@ class OIDNDenoiseContext {
}
if (num_samples_ == 1) {
- /* If the avoid scaling if there is only one sample, to save up time (so we dont divide
+ /* If the avoid scaling if there is only one sample, to save up time (so we don't divide
* buffer by 1). */
return false;
}
diff --git a/intern/cycles/integrator/path_trace.cpp b/intern/cycles/integrator/path_trace.cpp
index 6c02316ac2b..bc43747718d 100644
--- a/intern/cycles/integrator/path_trace.cpp
+++ b/intern/cycles/integrator/path_trace.cpp
@@ -177,7 +177,7 @@ void PathTrace::render(const RenderWork &render_work)
void PathTrace::render_pipeline(RenderWork render_work)
{
- /* NOTE: Only check for "instant" cancel here. Ther user-requested cancel via progress is
+ /* NOTE: Only check for "instant" cancel here. The user-requested cancel via progress is
* checked in Session and the work in the event of cancel is to be finished here. */
render_scheduler_.set_need_schedule_cryptomatte(device_scene_->data.film.cryptomatte_passes !=
@@ -680,7 +680,7 @@ void PathTrace::write_tile_buffer(const RenderWork &render_work)
*
* Tiles are written to a file during rendering, and written to the software at the end
* of rendering (wither when all tiles are finished, or when rendering was requested to be
- * cancelled).
+ * canceled).
*
* Important thing is: tile should be written to the software via callback only once. */
if (!has_multiple_tiles) {
@@ -913,7 +913,7 @@ void PathTrace::process_full_buffer_from_disk(string_view filename)
* ensure proper denoiser is used. */
set_denoiser_params(denoise_params);
- /* Number of samples doesn't matter too much, since the sampels count pass will be used. */
+ /* Number of samples doesn't matter too much, since the samples count pass will be used. */
denoiser_->denoise_buffer(full_frame_buffers.params, &full_frame_buffers, 0, false);
render_state_.has_denoised_result = true;
diff --git a/intern/cycles/integrator/path_trace.h b/intern/cycles/integrator/path_trace.h
index 78ca68c1198..fc7713e6df9 100644
--- a/intern/cycles/integrator/path_trace.h
+++ b/intern/cycles/integrator/path_trace.h
@@ -83,7 +83,7 @@ class PathTrace {
void set_progress(Progress *progress);
/* NOTE: This is a blocking call. Meaning, it will not return until given number of samples are
- * rendered (or until rendering is requested to be cancelled). */
+ * rendered (or until rendering is requested to be canceled). */
void render(const RenderWork &render_work);
/* TODO(sergey): Decide whether denoiser is really a part of path tracer. Currently it is
@@ -110,7 +110,7 @@ class PathTrace {
/* Cancel rendering process as soon as possible, without waiting for full tile to be sampled.
* Used in cases like reset of render session.
*
- * This is a blockign call, which returns as soon as there is no running `render_samples()` call.
+ * This is a blocking call, which returns as soon as there is no running `render_samples()` call.
*/
void cancel();
@@ -120,11 +120,11 @@ class PathTrace {
* the data will be copied to the device of the given render buffers. */
void copy_to_render_buffers(RenderBuffers *render_buffers);
- /* Copy happens via CPU side buffer: data will be copied from the device of the given rendetr
+ /* Copy happens via CPU side buffer: data will be copied from the device of the given render
* buffers and will be copied to all devices of the path trace. */
void copy_from_render_buffers(RenderBuffers *render_buffers);
- /* Copy render buffers of the big tile from the device to hsot.
+ /* Copy render buffers of the big tile from the device to host.
* Return true if all copies are successful. */
bool copy_render_tile_from_device();
@@ -172,10 +172,10 @@ class PathTrace {
* Is called during path tracing to communicate work-in-progress state of the final buffer. */
function<void(void)> tile_buffer_update_cb;
- /* Callback which communicates final rendered buffer. Is called after pathtracing is done. */
+ /* Callback which communicates final rendered buffer. Is called after path-tracing is done. */
function<void(void)> tile_buffer_write_cb;
- /* Callback which initializes rendered buffer. Is called before pathtracing starts.
+ /* Callback which initializes rendered buffer. Is called before path-tracing starts.
*
* This is used for baking. */
function<bool(void)> tile_buffer_read_cb;
@@ -189,7 +189,7 @@ class PathTrace {
protected:
/* Actual implementation of the rendering pipeline.
- * Calls steps in order, checking for the cancel to be requested inbetween.
+ * Calls steps in order, checking for the cancel to be requested in between.
*
* Is separate from `render()` to simplify dealing with the early outputs and keeping
* `render_cancel_` in the consistent state. */
@@ -283,7 +283,7 @@ class PathTrace {
* affects both resolution and stride as visible by the integrator kernels. */
int resolution_divider = 0;
- /* Paramaters of the big tile with the current resolution divider applied. */
+ /* Parameters of the big tile with the current resolution divider applied. */
BufferParams effective_big_tile_params;
/* Denosier was run and there are denoised versions of the passes in the render buffers. */
diff --git a/intern/cycles/integrator/path_trace_work.h b/intern/cycles/integrator/path_trace_work.h
index 97b97f3d888..8c9c8811199 100644
--- a/intern/cycles/integrator/path_trace_work.h
+++ b/intern/cycles/integrator/path_trace_work.h
@@ -39,8 +39,8 @@ class PathTraceWork {
/* Create path trace work which fits best the device.
*
- * The cancel request flag is used for a cheap check whether cancel is to berformed as soon as
- * possible. This could be, for rexample, request to cancel rendering on camera navigation in
+ * The cancel request flag is used for a cheap check whether cancel is to be performed as soon as
+ * possible. This could be, for example, request to cancel rendering on camera navigation in
* viewport. */
static unique_ptr<PathTraceWork> create(Device *device,
Film *film,
@@ -107,7 +107,7 @@ class PathTraceWork {
/* Special version of the `copy_from_render_buffers()` which only copies denosied passes from the
* given render buffers, leaving rest of the passes.
*
- * Same notes about device copying aplies to this call as well. */
+ * Same notes about device copying applies to this call as well. */
void copy_from_denoised_render_buffers(const RenderBuffers *render_buffers);
/* Copy render buffers to/from device using an appropriate device queue when needed so that
@@ -119,7 +119,7 @@ class PathTraceWork {
* things are executed in order with the `render_samples()`. */
virtual bool zero_render_buffers() = 0;
- /* Access pixels rendered by this work and copy them to the coresponding location in the
+ /* Access pixels rendered by this work and copy them to the corresponding location in the
* destination.
*
* NOTE: Does not perform copy of buffers from the device. Use `copy_render_tile_from_device()`
@@ -182,7 +182,7 @@ class PathTraceWork {
unique_ptr<RenderBuffers> buffers_;
/* Effective parameters of the full, big tile, and current work render buffer.
- * The latter might be different from buffers_->params when there is a resolution divider
+ * The latter might be different from `buffers_->params` when there is a resolution divider
* involved. */
BufferParams effective_full_params_;
BufferParams effective_big_tile_params_;
diff --git a/intern/cycles/integrator/path_trace_work_gpu.cpp b/intern/cycles/integrator/path_trace_work_gpu.cpp
index 10baf869aa6..135466becc6 100644
--- a/intern/cycles/integrator/path_trace_work_gpu.cpp
+++ b/intern/cycles/integrator/path_trace_work_gpu.cpp
@@ -498,7 +498,7 @@ void PathTraceWorkGPU::compact_states(const int num_active_paths)
bool PathTraceWorkGPU::enqueue_work_tiles(bool &finished)
{
/* If there are existing paths wait them to go to intersect closest kernel, which will align the
- * wavefront of the existing and newely added paths. */
+ * wavefront of the existing and newly added paths. */
/* TODO: Check whether counting new intersection kernels here will have positive affect on the
* performance. */
const DeviceKernel kernel = get_most_queued_kernel();
@@ -508,7 +508,7 @@ bool PathTraceWorkGPU::enqueue_work_tiles(bool &finished)
int num_active_paths = get_num_active_paths();
- /* Don't schedule more work if cancelling. */
+ /* Don't schedule more work if canceling. */
if (is_cancel_requested()) {
if (num_active_paths == 0) {
finished = true;
@@ -729,7 +729,7 @@ void PathTraceWorkGPU::copy_to_gpu_display_naive(GPUDisplay *gpu_display,
gpu_display_rgba_half_.data_height != final_height) {
gpu_display_rgba_half_.alloc(final_width, final_height);
/* TODO(sergey): There should be a way to make sure device-side memory is allocated without
- * transfering zeroes to the device. */
+ * transferring zeroes to the device. */
queue_->zero_to_device(gpu_display_rgba_half_);
}
diff --git a/intern/cycles/integrator/render_scheduler.cpp b/intern/cycles/integrator/render_scheduler.cpp
index 4eb1dd941f9..3e5b3417a6a 100644
--- a/intern/cycles/integrator/render_scheduler.cpp
+++ b/intern/cycles/integrator/render_scheduler.cpp
@@ -233,7 +233,7 @@ void RenderScheduler::render_work_reschedule_on_cancel(RenderWork &render_work)
const bool has_rendered_samples = get_num_rendered_samples() != 0;
- /* Reset all fields of the previous work, canelling things like adaptive sampling filtering and
+ /* Reset all fields of the previous work, canceling things like adaptive sampling filtering and
* denoising.
* However, need to preserve write requests, since those will not be possible to recover and
* writes are only to happen once. */
@@ -246,7 +246,7 @@ void RenderScheduler::render_work_reschedule_on_cancel(RenderWork &render_work)
render_work.full.write = full_write;
/* Do not write tile if it has zero samples it it, treat it similarly to all other tiles which
- * got cancelled. */
+ * got canceled. */
if (!state_.tile_result_was_written && has_rendered_samples) {
render_work.tile.write = true;
}
@@ -817,7 +817,7 @@ int RenderScheduler::get_num_samples_to_path_trace() const
int num_samples_to_render = min(num_samples_pot, max_num_samples_to_render);
- /* When enough statistics is available and doing an offlien rendering prefer to keep device
+ /* When enough statistics is available and doing an offline rendering prefer to keep device
* occupied. */
if (state_.occupancy_num_samples && (background_ || headless_)) {
/* Keep occupancy at about 0.5 (this is more of an empirical figure which seems to match scenes
@@ -874,7 +874,7 @@ int RenderScheduler::get_num_samples_during_navigation(int resolution_divider) c
/* Always render 4 samples, even if scene is configured for less.
* The idea here is to have enough information on the screen. Resolution divider of 2 allows us
- * to have 4 time extra samples, so verall worst case timing is the same as the final resolution
+ * to have 4 time extra samples, so overall worst case timing is the same as the final resolution
* at one sample. */
return 4;
}
diff --git a/intern/cycles/integrator/render_scheduler.h b/intern/cycles/integrator/render_scheduler.h
index 9c2d107e46d..b7b598fb10c 100644
--- a/intern/cycles/integrator/render_scheduler.h
+++ b/intern/cycles/integrator/render_scheduler.h
@@ -83,7 +83,7 @@ class RenderWork {
} display;
/* Re-balance multi-device scheduling after rendering this work.
- * Note that the scheduler does not know anything abouce devices, so if there is only a single
+ * Note that the scheduler does not know anything about devices, so if there is only a single
* device used, then it is up for the PathTracer to ignore the balancing. */
bool rebalance = false;
@@ -203,7 +203,7 @@ class RenderScheduler {
* extra work needs to be scheduled to denoise and write final result. */
bool done() const;
- /* Update scheduling state for a newely scheduled work.
+ /* Update scheduling state for a newly scheduled work.
* Takes care of things like checking whether work was ever denoised, tile was written and states
* like that. */
void update_state_for_render_work(const RenderWork &render_work);
@@ -235,7 +235,7 @@ class RenderScheduler {
double guess_display_update_interval_in_seconds_for_num_samples_no_limit(
int num_rendered_samples) const;
- /* Calculate number of samples which can be rendered within current desred update interval which
+ /* Calculate number of samples which can be rendered within current desired update interval which
* is calculated by `guess_update_interval_in_seconds()`. */
int calculate_num_samples_per_update() const;
@@ -250,11 +250,11 @@ class RenderScheduler {
/* Whether adaptive sampling convergence check and filter is to happen. */
bool work_need_adaptive_filter() const;
- /* Calculate thretshold for adaptive sampling. */
+ /* Calculate threshold for adaptive sampling. */
float work_adaptive_threshold() const;
/* Check whether current work needs denoising.
- * Denoising is not needed if the denoiser is not configured, or when denosiing is happening too
+ * Denoising is not needed if the denoiser is not configured, or when denoising is happening too
* often.
*
* The delayed will be true when the denoiser is configured for use, but it was delayed for a
diff --git a/intern/cycles/integrator/shader_eval.cpp b/intern/cycles/integrator/shader_eval.cpp
index 465b4a8d4da..d35ff4cd03f 100644
--- a/intern/cycles/integrator/shader_eval.cpp
+++ b/intern/cycles/integrator/shader_eval.cpp
@@ -71,7 +71,7 @@ bool ShaderEval::eval(const ShaderEvalType type,
success = (device->info.type == DEVICE_CPU) ? eval_cpu(device, type, input, output) :
eval_gpu(device, type, input, output);
- /* Copy data back from device if not cancelled. */
+ /* Copy data back from device if not canceled. */
if (success) {
output.copy_from_device(0, 1, output.size());
read_output(output);
diff --git a/intern/cycles/integrator/work_balancer.h b/intern/cycles/integrator/work_balancer.h
index 94e20ecf054..fc5e561845e 100644
--- a/intern/cycles/integrator/work_balancer.h
+++ b/intern/cycles/integrator/work_balancer.h
@@ -32,7 +32,7 @@ struct WorkBalanceInfo {
double weight = 1.0;
};
-/* Balance work for an initial render interation, before any statistics is known. */
+/* Balance work for an initial render integration, before any statistics is known. */
void work_balance_do_initial(vector<WorkBalanceInfo> &work_balance_infos);
/* Rebalance work after statistics has been accumulated.
diff --git a/intern/cycles/integrator/work_tile_scheduler.cpp b/intern/cycles/integrator/work_tile_scheduler.cpp
index 3fc99d5b74d..e6ada2f46ee 100644
--- a/intern/cycles/integrator/work_tile_scheduler.cpp
+++ b/intern/cycles/integrator/work_tile_scheduler.cpp
@@ -81,7 +81,7 @@ void WorkTileScheduler::reset_scheduler_state()
bool WorkTileScheduler::get_work(KernelWorkTile *work_tile_, const int max_work_size)
{
/* Note that the `max_work_size` can be higher than the `max_num_path_states_`: this is because
- * the path trace work can decice to use smaller tile sizes and greedily schedule multiple tiles,
+ * the path trace work can decide to use smaller tile sizes and greedily schedule multiple tiles,
* improving overall device occupancy.
* So the `max_num_path_states_` is a "scheduling unit", and the `max_work_size` is a "scheduling
* limit". */
diff --git a/intern/cycles/integrator/work_tile_scheduler.h b/intern/cycles/integrator/work_tile_scheduler.h
index e4c8f701259..85f11b601c7 100644
--- a/intern/cycles/integrator/work_tile_scheduler.h
+++ b/intern/cycles/integrator/work_tile_scheduler.h
@@ -64,7 +64,7 @@ class WorkTileScheduler {
/* dimensions of the currently rendering image in pixels. */
int2 image_size_px_ = make_int2(0, 0);
- /* Offset and stride of the buffer within which scheduing is happenning.
+ /* Offset and stride of the buffer within which scheduling is happening.
* Will be passed over to the KernelWorkTile. */
int offset_, stride_;
@@ -87,7 +87,7 @@ class WorkTileScheduler {
* in the `get_work()`? */
int total_tiles_num_ = 0;
- /* In the case when the number of sam[les in the `tile_size_` is lower than samples_num_ denotes
+ /* In the case when the number of samples in the `tile_size_` is lower than samples_num_ denotes
* how many tiles are to be "stacked" to cover the entire requested range of samples. */
int num_tiles_per_sample_range_ = 0;
diff --git a/intern/cycles/kernel/closure/bssrdf.h b/intern/cycles/kernel/closure/bssrdf.h
index 0f9278bba89..e095314678a 100644
--- a/intern/cycles/kernel/closure/bssrdf.h
+++ b/intern/cycles/kernel/closure/bssrdf.h
@@ -76,7 +76,7 @@ ccl_device void bssrdf_setup_radius(Bssrdf *bssrdf, const ClosureType type, cons
const float inv_eta = 1.0f / eta;
const float F_dr = inv_eta * (-1.440f * inv_eta + 0.710f) + 0.668f + 0.0636f * eta;
const float fourthirdA = (4.0f / 3.0f) * (1.0f + F_dr) /
- (1.0f - F_dr); /* From Jensen's Fdr ratio formula. */
+ (1.0f - F_dr); /* From Jensen's `Fdr` ratio formula. */
const float3 alpha_prime = make_float3(
bssrdf_dipole_compute_alpha_prime(bssrdf->albedo.x, fourthirdA),
diff --git a/intern/cycles/kernel/device/cuda/compat.h b/intern/cycles/kernel/device/cuda/compat.h
index 665da43e1a1..3c85a8e7bd2 100644
--- a/intern/cycles/kernel/device/cuda/compat.h
+++ b/intern/cycles/kernel/device/cuda/compat.h
@@ -80,7 +80,7 @@ typedef unsigned long long uint64_t;
#define ccl_gpu_global_id_x() (ccl_gpu_block_idx_x * ccl_gpu_block_dim_x + ccl_gpu_thread_idx_x)
#define ccl_gpu_global_size_x() (ccl_gpu_grid_dim_x * ccl_gpu_block_dim_x)
-/* GPU warp synchronizaton */
+/* GPU warp synchronization. */
#define ccl_gpu_syncthreads() __syncthreads()
#define ccl_gpu_ballot(predicate) __ballot_sync(0xFFFFFFFF, predicate)
diff --git a/intern/cycles/kernel/device/optix/compat.h b/intern/cycles/kernel/device/optix/compat.h
index 4e255a135c6..fb9e094b535 100644
--- a/intern/cycles/kernel/device/optix/compat.h
+++ b/intern/cycles/kernel/device/optix/compat.h
@@ -81,7 +81,7 @@ typedef unsigned long long uint64_t;
#define ccl_gpu_global_id_x() (ccl_gpu_block_idx_x * ccl_gpu_block_dim_x + ccl_gpu_thread_idx_x)
#define ccl_gpu_global_size_x() (ccl_gpu_grid_dim_x * ccl_gpu_block_dim_x)
-/* GPU warp synchronizaton */
+/* GPU warp synchronization. */
#define ccl_gpu_syncthreads() __syncthreads()
#define ccl_gpu_ballot(predicate) __ballot_sync(0xFFFFFFFF, predicate)
diff --git a/intern/cycles/kernel/integrator/integrator_shade_volume.h b/intern/cycles/kernel/integrator/integrator_shade_volume.h
index 4a864b1e6ce..d44890f800e 100644
--- a/intern/cycles/kernel/integrator/integrator_shade_volume.h
+++ b/intern/cycles/kernel/integrator/integrator_shade_volume.h
@@ -30,7 +30,7 @@ CCL_NAMESPACE_BEGIN
#ifdef __VOLUME__
-/* Events for probalistic scattering */
+/* Events for probabilistic scattering. */
typedef enum VolumeIntegrateEvent {
VOLUME_PATH_SCATTERED = 0,
@@ -228,8 +228,8 @@ ccl_device void volume_shadow_heterogeneous(INTEGRATOR_STATE_ARGS,
/* compute attenuation over segment */
sd->P = new_P;
if (shadow_volume_shader_sample(INTEGRATOR_STATE_PASS, sd, &sigma_t)) {
- /* Compute expf() only for every Nth step, to save some calculations
- * because exp(a)*exp(b) = exp(a+b), also do a quick VOLUME_THROUGHPUT_EPSILON
+ /* Compute `expf()` only for every Nth step, to save some calculations
+ * because `exp(a)*exp(b) = exp(a+b)`, also do a quick #VOLUME_THROUGHPUT_EPSILON
* check then. */
sum += (-sigma_t * dt);
if ((i & 0x07) == 0) { /* ToDo: Other interval? */
@@ -648,7 +648,7 @@ ccl_device_forceinline void volume_integrate_heterogeneous(
}
}
- /* Write accumulated emisison. */
+ /* Write accumulated emission. */
if (!is_zero(accum_emission)) {
kernel_accum_emission(
INTEGRATOR_STATE_PASS, result.indirect_throughput, accum_emission, render_buffer);
diff --git a/intern/cycles/kernel/integrator/integrator_state.h b/intern/cycles/kernel/integrator/integrator_state.h
index 8cef9cf31e2..094446be02c 100644
--- a/intern/cycles/kernel/integrator/integrator_state.h
+++ b/intern/cycles/kernel/integrator/integrator_state.h
@@ -44,7 +44,7 @@
* INTEGRATOR_STATE_PASS_NULL: use to pass empty state to other functions.
*
* NOTE: if we end up with a device that passes no arguments, the leading comma will be a problem.
- * Can solve it with more macros if we encouter it, but rather ugly so postpone for now.
+ * Can solve it with more macros if we encounter it, but rather ugly so postpone for now.
*/
#include "kernel/kernel_types.h"
@@ -129,7 +129,7 @@ typedef struct IntegratorStateGPU {
*
* Note that there is a special access function for the shadow catcher state. This access is to
* happen from a kernel which operates on a "main" path. Attempt to use shadow catcher accessors
- * from a kernel which operates on a shadow catcher state will cause bad memory acces. */
+ * from a kernel which operates on a shadow catcher state will cause bad memory access. */
#ifdef __KERNEL_CPU__
diff --git a/intern/cycles/kernel/kernel_adaptive_sampling.h b/intern/cycles/kernel/kernel_adaptive_sampling.h
index 2bee12f0473..7d71907effe 100644
--- a/intern/cycles/kernel/kernel_adaptive_sampling.h
+++ b/intern/cycles/kernel/kernel_adaptive_sampling.h
@@ -60,7 +60,7 @@ ccl_device bool kernel_adaptive_sampling_convergence_check(const KernelGlobals *
const float4 A = kernel_read_pass_float4(buffer + kernel_data.film.pass_adaptive_aux_buffer);
if (!reset && A.w != 0.0f) {
- /* If the pixel was considered converged, its state will not change in this kernmel. Early
+ /* If the pixel was considered converged, its state will not change in this kernel. Early
* output before doing any math.
*
* TODO(sergey): On a GPU it might be better to keep thread alive for better coherency? */
diff --git a/intern/cycles/kernel/kernel_film.h b/intern/cycles/kernel/kernel_film.h
index fa93f4830d1..715d764fb31 100644
--- a/intern/cycles/kernel/kernel_film.h
+++ b/intern/cycles/kernel/kernel_film.h
@@ -393,7 +393,7 @@ film_calculate_shadow_catcher(const KernelFilmConvert *ccl_restrict kfilm_conver
ccl_global const float *in_catcher = buffer + kfilm_convert->pass_shadow_catcher;
/* NOTE: It is possible that the Shadow Catcher pass is requested as an output without actual
- * shadow catcher objects in the scene. In this case there will be no auxillary passes required
+ * shadow catcher objects in the scene. In this case there will be no auxiliary passes required
* for the devision (to save up memory). So delay the asserts to this point so that the number of
* samples check handles such configuration. */
kernel_assert(kfilm_convert->pass_offset != PASS_UNUSED);
@@ -404,14 +404,14 @@ film_calculate_shadow_catcher(const KernelFilmConvert *ccl_restrict kfilm_conver
ccl_global const float *in_matte = buffer + kfilm_convert->pass_shadow_catcher_matte;
/* No scaling needed. The integration works in way that number of samples in the combined and
- * shadow catcher passes are the same, and exposure is cancelled during the division. */
+ * shadow catcher passes are the same, and exposure is canceled during the division. */
const float3 color_catcher = make_float3(in_catcher[0], in_catcher[1], in_catcher[2]);
const float3 color_combined = make_float3(in_combined[0], in_combined[1], in_combined[2]);
const float3 color_matte = make_float3(in_matte[0], in_matte[1], in_matte[2]);
/* Need to ignore contribution of the matte object when doing division (otherwise there will be
* artifacts caused by anti-aliasing). Since combined pass is used for adaptive sampling and need
- * to contain matte objects, we subtrack matte objects contribution here. This is the same as if
+ * to contain matte objects, we subtract matte objects contribution here. This is the same as if
* the matte objects were not accumulated to the combined pass. */
const float3 combined_no_matte = color_combined - color_matte;
@@ -422,8 +422,8 @@ film_calculate_shadow_catcher(const KernelFilmConvert *ccl_restrict kfilm_conver
const float alpha = film_transparency_to_alpha(transparency);
/* Alpha-over on white using transparency of the combined pass. This allows to eliminate
- * artifacts which are happenning on an edge of a shadow catcher when using transparent film.
- * Note that we treat shadow catcher as straight alpha here because alpha got cancelled out
+ * artifacts which are happening on an edge of a shadow catcher when using transparent film.
+ * Note that we treat shadow catcher as straight alpha here because alpha got canceled out
* during the division. */
const float3 pixel = (1.0f - alpha) * one_float3() + alpha * shadow_catcher;
diff --git a/intern/cycles/kernel/kernel_random.h b/intern/cycles/kernel/kernel_random.h
index 41b7d76230a..240c92bf9d0 100644
--- a/intern/cycles/kernel/kernel_random.h
+++ b/intern/cycles/kernel/kernel_random.h
@@ -111,7 +111,7 @@ ccl_device_forceinline void path_rng_2D(
}
/**
- * 1D hash recomended from "Hash Functions for GPU Rendering" JCGT Vol. 9, No. 3, 2020
+ * 1D hash recommended from "Hash Functions for GPU Rendering" JCGT Vol. 9, No. 3, 2020
* See https://www.shadertoy.com/view/4tXyWN and https://www.shadertoy.com/view/XlGcRh
* http://www.jcgt.org/published/0009/03/02/paper.pdf
*/
@@ -124,7 +124,7 @@ ccl_device_inline uint hash_iqint1(uint n)
}
/**
- * 2D hash recomended from "Hash Functions for GPU Rendering" JCGT Vol. 9, No. 3, 2020
+ * 2D hash recommended from "Hash Functions for GPU Rendering" JCGT Vol. 9, No. 3, 2020
* See https://www.shadertoy.com/view/4tXyWN and https://www.shadertoy.com/view/XlGcRh
* http://www.jcgt.org/published/0009/03/02/paper.pdf
*/
diff --git a/intern/cycles/kernel/kernel_types.h b/intern/cycles/kernel/kernel_types.h
index 927e60e8729..66b7310ab65 100644
--- a/intern/cycles/kernel/kernel_types.h
+++ b/intern/cycles/kernel/kernel_types.h
@@ -380,7 +380,7 @@ typedef enum PassType {
* pass contains number of samples which contributed to the color components of the pass.
*
* PASS_SHADOW_CATCHER_SAMPLE_COUNT contains number of samples for which the path split
- * happenned.
+ * happened.
*
* PASS_SHADOW_CATCHER_MATTE contains pass which contains non-catcher objects. This pass is to be
* alpha-overed onto the backdrop (after multiplication). */
diff --git a/intern/cycles/kernel/svm/svm_bevel.h b/intern/cycles/kernel/svm/svm_bevel.h
index aab089d19ea..9d7ce202d49 100644
--- a/intern/cycles/kernel/svm/svm_bevel.h
+++ b/intern/cycles/kernel/svm/svm_bevel.h
@@ -255,7 +255,7 @@ ccl_device float3 svm_bevel(INTEGRATOR_STATE_CONST_ARGS,
/* Multiple importance sample between 3 axes, power heuristic
* found to be slightly better than balance heuristic. pdf_N
- * in the MIS weight and denominator cancelled out. */
+ * in the MIS weight and denominator canceled out. */
float w = pdf_N / (sqr(pdf_N) + sqr(pdf_T) + sqr(pdf_B));
if (isect.num_hits > LOCAL_MAX_HITS) {
w *= isect.num_hits / (float)LOCAL_MAX_HITS;
diff --git a/intern/cycles/render/buffers.cpp b/intern/cycles/render/buffers.cpp
index 1cdae3af7f5..186699596ac 100644
--- a/intern/cycles/render/buffers.cpp
+++ b/intern/cycles/render/buffers.cpp
@@ -300,7 +300,7 @@ void render_buffers_host_copy_denoised(RenderBuffers *dst,
DCHECK_EQ(dst_params.width, src_params.width);
/* TODO(sergey): More sanity checks to avoid buffer overrun. */
- /* Create a map of pass ofsets to be copied.
+ /* Create a map of pass offsets to be copied.
* Assume offsets are different to allow copying passes between buffers with different set of
* passes. */
diff --git a/intern/cycles/render/buffers.h b/intern/cycles/render/buffers.h
index c048234167d..a07e7289566 100644
--- a/intern/cycles/render/buffers.h
+++ b/intern/cycles/render/buffers.h
@@ -116,7 +116,7 @@ class BufferParams : public Node {
*
* When the scene passes are given, the buffer passes will be created from them and stored in
* this params, and then params are updated for those passes.
- * The `update_passes()` without parameters updates offsets and stries which are stored outside
+ * The `update_passes()` without parameters updates offsets and strides which are stored outside
* of the passes. */
void update_passes();
void update_passes(const vector<Pass *> &scene_passes);
@@ -140,7 +140,7 @@ class BufferParams : public Node {
protected:
void reset_pass_offset();
- /* Multipled by 2 to be able to store noisy and denoised pass types. */
+ /* Multiplied by 2 to be able to store noisy and denoised pass types. */
static constexpr int kNumPassOffsets = PASS_NUM * 2;
/* Indexed by an index derived from pass type and mode, indicates offset of the corresponding
@@ -171,7 +171,7 @@ class RenderBuffers {
/* Copy denoised passes form source to destination.
*
- * Buffer parameters are provided explicitly, allowing to copy pixelks between render buffers which
+ * Buffer parameters are provided explicitly, allowing to copy pixels between render buffers which
* content corresponds to a render result at a non-unit resolution divider.
*
* `src_offset` allows to offset source pixel index which is used when a fraction of the source
diff --git a/intern/cycles/render/gpu_display.h b/intern/cycles/render/gpu_display.h
index cbe347895a1..a01348d28d5 100644
--- a/intern/cycles/render/gpu_display.h
+++ b/intern/cycles/render/gpu_display.h
@@ -117,7 +117,7 @@ class GPUDisplay {
*
* NOTE: The GPUDisplay should be marked for an update being in process with `update_begin()`.
*
- * NOTE: Texture buffer can not be mapped while graphics interopeability is active. This means
+ * NOTE: Texture buffer can not be mapped while graphics interoperability is active. This means
* that `map_texture_buffer()` is not allowed between `graphics_interop_begin()` and
* `graphics_interop_end()` calls.
*/
@@ -125,7 +125,7 @@ class GPUDisplay {
/* Map pixels memory form texture to a buffer available for write from CPU. Width and height will
* define a requested size of the texture to write to.
* Upon success a non-null pointer is returned and the texture buffer is to be unmapped.
- * If an error happens during mapping, or if mapoping is not supported by this GPU display a
+ * If an error happens during mapping, or if mapping is not supported by this GPU display a
* null pointer is returned and the buffer is NOT to be unmapped.
*
* NOTE: Usually the implementation will rely on a GPU context of some sort, and the GPU context
@@ -149,7 +149,7 @@ class GPUDisplay {
* device API. */
DeviceGraphicsInteropDestination graphics_interop_get();
- /* (De)activate GPU display for graphics interoperability outside of regular display udpate
+ /* (De)activate GPU display for graphics interoperability outside of regular display update
* routines. */
virtual void graphics_interop_activate();
virtual void graphics_interop_deactivate();
@@ -206,8 +206,8 @@ class GPUDisplay {
GPUDisplayParams params_;
/* Mark texture as its content has been updated.
- * Used from places which knows that the texture content has been brough up-to-date, so that the
- * drawing knows whether it can be performed, and whether drawing happenned with an up-to-date
+ * Used from places which knows that the texture content has been brought up-to-date, so that the
+ * drawing knows whether it can be performed, and whether drawing happened with an up-to-date
* texture state. */
void mark_texture_updated();
diff --git a/intern/cycles/render/session.cpp b/intern/cycles/render/session.cpp
index 84407f8e6dd..c39232be2b0 100644
--- a/intern/cycles/render/session.cpp
+++ b/intern/cycles/render/session.cpp
@@ -116,7 +116,7 @@ Session::~Session()
}
#endif
- /* Make sure path tracer is destroyed before the deviec. This is needed because destruction might
+ /* Make sure path tracer is destroyed before the device. This is needed because destruction might
* need to access device for device memory free. */
/* TODO(sergey): Convert device to be unique_ptr, and rely on C++ to destruct objects in the
* pre-defined order. */
@@ -612,7 +612,7 @@ void Session::collect_statistics(RenderStats *render_stats)
}
/* --------------------------------------------------------------------
- * Tile and tile pixels aceess.
+ * Tile and tile pixels access.
*/
bool Session::has_multiple_render_tiles() const
@@ -650,7 +650,7 @@ bool Session::copy_render_tile_from_device()
bool Session::get_render_tile_pixels(const string &pass_name, int num_components, float *pixels)
{
/* NOTE: The code relies on a fact that session is fully update and no scene/buffer modification
- * is happenning while this function runs. */
+ * is happening while this function runs. */
const BufferParams &buffer_params = path_trace_->get_render_tile_params();
@@ -689,7 +689,7 @@ bool Session::set_render_tile_pixels(const string &pass_name,
const float *pixels)
{
/* NOTE: The code relies on a fact that session is fully update and no scene/buffer modification
- * is happenning while this function runs. */
+ * is happening while this function runs. */
const BufferPass *pass = buffer_params_.find_pass(pass_name);
if (!pass) {
diff --git a/intern/cycles/render/session.h b/intern/cycles/render/session.h
index 492cfdd1c09..5623604bfe8 100644
--- a/intern/cycles/render/session.h
+++ b/intern/cycles/render/session.h
@@ -120,7 +120,7 @@ class Session {
/* Callback is invoked by tile manager whenever on-dist tiles storage file is closed after
* writing. Allows an engine integration to keep track of those files without worry about
- * transfering the information when it needs to re-create session during rendering. */
+ * transferring the information when it needs to re-create session during rendering. */
function<void(string_view)> full_buffer_written_cb;
explicit Session(const SessionParams &params, const SceneParams &scene_params);
@@ -128,7 +128,7 @@ class Session {
void start();
- /* When quick cancel is requested path tracing is cancelles as soon as possible, without waiting
+ /* When quick cancel is requested path tracing is cancels as soon as possible, without waiting
* for the buffer to be uniformly sampled. */
void cancel(bool quick = false);
@@ -154,7 +154,7 @@ class Session {
void collect_statistics(RenderStats *stats);
/* --------------------------------------------------------------------
- * Tile and tile pixels aceess.
+ * Tile and tile pixels access.
*/
bool has_multiple_render_tiles() const;
diff --git a/intern/cycles/render/tile.cpp b/intern/cycles/render/tile.cpp
index eed75cc2372..28910bffa7b 100644
--- a/intern/cycles/render/tile.cpp
+++ b/intern/cycles/render/tile.cpp
@@ -282,7 +282,7 @@ static bool buffer_params_from_image_spec_atttributes(BufferParams *buffer_param
/* Configure image specification for the given buffer parameters and passes.
*
- * Image channels will ber strictly ordered to match content of corresponding buffer, and the
+ * Image channels will be strictly ordered to match content of corresponding buffer, and the
* metadata will be set so that the render buffers and passes can be reconstructed from it.
*
* If the tile size different from (0, 0) the image specification will be configured to use the
@@ -358,7 +358,7 @@ void TileManager::update(const BufferParams &params, const Scene *scene)
buffer_params_ = params;
- /* TODO(sergey): Proper Error handling, so that if configuration has failed we dont' attempt to
+ /* TODO(sergey): Proper Error handling, so that if configuration has failed we don't attempt to
* write to a partially configured file. */
configure_image_spec_from_buffer(&write_state_.image_spec, buffer_params_, tile_size_);
diff --git a/intern/cycles/render/tile.h b/intern/cycles/render/tile.h
index 124d0b3652c..71b9e966278 100644
--- a/intern/cycles/render/tile.h
+++ b/intern/cycles/render/tile.h
@@ -94,7 +94,7 @@ class TileManager {
* The file will be considered final, all handles to it will be closed. */
void finish_write_tiles();
- /* Check whether any tile ahs been written to disk. */
+ /* Check whether any tile has been written to disk. */
inline bool has_written_tiles() const
{
return write_state_.num_tiles_written != 0;
diff --git a/source/blender/blenkernel/intern/geometry_component_instances.cc b/source/blender/blenkernel/intern/geometry_component_instances.cc
index 4c10f5398b7..9479d012cb8 100644
--- a/source/blender/blenkernel/intern/geometry_component_instances.cc
+++ b/source/blender/blenkernel/intern/geometry_component_instances.cc
@@ -154,8 +154,8 @@ void InstancesComponent::ensure_geometry_instances()
/**
* With write access to the instances component, the data in the instanced geometry sets can be
- * changed. This is a function on the component rather than each reference to ensure const
- * correct-ness for that reason.
+ * changed. This is a function on the component rather than each reference to ensure `const`
+ * correctness for that reason.
*/
GeometrySet &InstancesComponent::geometry_set_from_reference(const int reference_index)
{
diff --git a/source/blender/blenkernel/intern/gpencil_geom.cc b/source/blender/blenkernel/intern/gpencil_geom.cc
index d7c906be18e..976b26a1f3a 100644
--- a/source/blender/blenkernel/intern/gpencil_geom.cc
+++ b/source/blender/blenkernel/intern/gpencil_geom.cc
@@ -600,7 +600,7 @@ static bool BKE_gpencil_stroke_extra_points(bGPDstroke *gps,
* \param dist: Length of the added section.
* \param overshoot_fac: Relative length of the curve which is used to determine the extension.
* \param mode: Affect to Start, End or Both extremes (0->Both, 1->Start, 2->End)
- * \param follow_curvature: True for appproximating curvature of given overshoot.
+ * \param follow_curvature: True for approximating curvature of given overshoot.
* \param extra_point_count: When follow_curvature is true, use this amount of extra points
*/
bool BKE_gpencil_stroke_stretch(bGPDstroke *gps,
diff --git a/source/blender/blenkernel/intern/mesh_convert.cc b/source/blender/blenkernel/intern/mesh_convert.cc
index 07dc6db05aa..467f7d4543e 100644
--- a/source/blender/blenkernel/intern/mesh_convert.cc
+++ b/source/blender/blenkernel/intern/mesh_convert.cc
@@ -1118,7 +1118,7 @@ static Mesh *mesh_new_from_mball_object(Object *object)
* balls and all evaluated child meta balls (since polygonization is only stored in the mother
* ball).
*
- * We create empty mesh so scripters don't run into None objects. */
+ * Create empty mesh so script-authors don't run into None objects. */
if (!DEG_is_evaluated_object(object) || object->runtime.curve_cache == nullptr ||
BLI_listbase_is_empty(&object->runtime.curve_cache->disp)) {
return (Mesh *)BKE_id_new_nomain(ID_ME, ((ID *)object->data)->name + 2);
diff --git a/source/blender/editors/space_file/filesel.c b/source/blender/editors/space_file/filesel.c
index 11b06d2b414..f7bdb4326a5 100644
--- a/source/blender/editors/space_file/filesel.c
+++ b/source/blender/editors/space_file/filesel.c
@@ -1271,7 +1271,7 @@ void file_params_rename_end(wmWindowManager *wm,
/* Ensure smooth-scroll timer is active, even if not needed, because that way rename state is
* handled properly. */
file_params_invoke_rename_postscroll(wm, win, sfile);
- /* Also always activate the rename file, even if renaming was cancelled. */
+ /* Also always activate the rename file, even if renaming was canceled. */
file_params_renamefile_activate(sfile, params);
}
diff --git a/source/blender/editors/space_sequencer/sequencer_draw.c b/source/blender/editors/space_sequencer/sequencer_draw.c
index 6f10a0f6c9e..53f1c35776c 100644
--- a/source/blender/editors/space_sequencer/sequencer_draw.c
+++ b/source/blender/editors/space_sequencer/sequencer_draw.c
@@ -1757,7 +1757,7 @@ static void draw_seq_strip_thumbnail(View2D *v2d,
}
/* Store recently rendered frames, so they can be reused when zooming. */
else if (!sequencer_thumbnail_v2d_is_navigating(C)) {
- /* Clear images in frame range occupied bynew thumbnail. */
+ /* Clear images in frame range occupied by new thumbnail. */
last_displayed_thumbnails_list_cleanup(
last_displayed_thumbnails, thumb_x_start, thumb_x_end);
/* Insert new thumbnail frame to list. */
diff --git a/source/blender/gpencil_modifiers/intern/MOD_gpencillength.c b/source/blender/gpencil_modifiers/intern/MOD_gpencillength.c
index e4a48925bf0..80b60547e92 100644
--- a/source/blender/gpencil_modifiers/intern/MOD_gpencillength.c
+++ b/source/blender/gpencil_modifiers/intern/MOD_gpencillength.c
@@ -140,8 +140,8 @@ static void applyLength(LengthGpencilModifierData *lmd, bGPdata *gpd, bGPDstroke
/* HACK: The second #overshoot_fac needs to be adjusted because it is not
* done in the same stretch call, because it can have a different length.
* The adjustment needs to be stable when
- * ceil(overshoot_fac*(gps->totpoints - 2)) is used in stretch and never
- * produce a result highter than totpoints - 2. */
+ * `ceil(overshoot_fac*(gps->totpoints - 2))` is used in stretch and never
+ * produce a result higher than `totpoints - 2`. */
const float second_overshoot_fac = lmd->overshoot_fac * (totpoints - 2) /
((float)gps->totpoints - 2) *
(1.0f - 0.1f / (totpoints - 1.0f));
diff --git a/source/blender/nodes/geometry/nodes/node_geo_curve_sample.cc b/source/blender/nodes/geometry/nodes/node_geo_curve_sample.cc
index 245394d3057..ac0cd510ffa 100644
--- a/source/blender/nodes/geometry/nodes/node_geo_curve_sample.cc
+++ b/source/blender/nodes/geometry/nodes/node_geo_curve_sample.cc
@@ -155,7 +155,7 @@ class SampleCurveFunction : public fn::MultiFunction {
spline_indices[i] = std::max(index, 0);
}
- /* Storing lookups in an array is unecessary but will simplify custom attribute transfer. */
+ /* Storing lookups in an array is unnecessary but will simplify custom attribute transfer. */
Array<Spline::LookupResult> lookups(mask.min_array_size());
for (const int i : mask) {
const float length_in_spline = lengths[i] - spline_lengths_[spline_indices[i]];
diff --git a/source/blender/windowmanager/intern/wm_dragdrop.c b/source/blender/windowmanager/intern/wm_dragdrop.c
index b6a04251ffb..6585349c83c 100644
--- a/source/blender/windowmanager/intern/wm_dragdrop.c
+++ b/source/blender/windowmanager/intern/wm_dragdrop.c
@@ -428,7 +428,7 @@ ID *WM_drag_get_local_ID_or_import_from_asset(const wmDrag *drag, int idcode)
}
/**
- * \brief Free asset ID imported for cancelled drop.
+ * \brief Free asset ID imported for canceled drop.
*
* If the asset was imported (linked/appended) using #WM_drag_get_local_ID_or_import_from_asset()`
* (typically via a #wmDropBox.copy() callback), we want the ID to be removed again if the drop