Cleanup: spelling

This commit is contained in:
Campbell Barton 2019-09-30 17:06:28 +10:00
parent 04133ee863
commit a6a0a09197
22 changed files with 36 additions and 36 deletions

View File

@ -27,7 +27,7 @@ namespace mv {
struct OperationListener {
// All hooks return true to continue or false to indicate the operation
// should abort. Hooks should be thread safe (reentrant).
// should abort. Hooks should be thread safe (re-entrant).
virtual bool Log(const string& message) = 0;
virtual bool Progress(double fraction) = 0;
virtual bool Cancelled() = 0;

View File

@ -31,7 +31,7 @@ struct ModifierUpdateDepsgraphContext;
struct Object;
struct Scene;
/* NOTE: bakeModifier() called from UI:
* needs to create new databloc-ks, hence the need for this. */
* needs to create new data-blocks, hence the need for this. */
struct bGPDframe;
struct bGPDlayer;
struct bGPDstroke;

View File

@ -140,7 +140,7 @@ typedef struct SubdivDisplacement {
/* This structure contains everything needed to construct subdivided surface.
* It does not specify storage, memory layout or anything else.
* It is possible to create different storages (like, grid based CPU side
* It is possible to create different storage's (like, grid based CPU side
* buffers, GPU subdivision mesh, CPU side fully qualified mesh) from the same
* Subdiv structure. */
typedef struct Subdiv {

View File

@ -150,7 +150,7 @@ typedef struct SubdivCCG {
struct CCGElem **edges;
int num_edges;
/* Loose vertices. Every element corresponds to a loose vertex from a coarse
* mesh, every coarse loose vertex corresponds to a single sundivided
* mesh, every coarse loose vertex corresponds to a single subdivided
* element. */
struct CCGElem *vertices;
int num_vertices;

View File

@ -110,7 +110,7 @@ const char *BKE_appdir_folder_default(void)
// #define PATH_DEBUG
/* returns a formatted representation of the specified version number. Non-reentrant! */
/* returns a formatted representation of the specified version number. Non-re-entrant! */
static char *blender_version_decimal(const int ver)
{
static char version_str[5];

View File

@ -3348,7 +3348,7 @@ void ntreeUpdateTree(Main *bmain, bNodeTree *ntree)
return;
}
/* avoid reentrant updates, can be caused by RNA update callbacks */
/* Avoid re-entrant updates, can be caused by RNA update callbacks. */
if (ntree->is_updating) {
return;
}
@ -3409,7 +3409,7 @@ void ntreeUpdateTree(Main *bmain, bNodeTree *ntree)
void nodeUpdate(bNodeTree *ntree, bNode *node)
{
/* avoid reentrant updates, can be caused by RNA update callbacks */
/* Avoid re-entrant updates, can be caused by RNA update callbacks. */
if (ntree->is_updating) {
return;
}
@ -3436,7 +3436,7 @@ bool nodeUpdateID(bNodeTree *ntree, ID *id)
return changed;
}
/* avoid reentrant updates, can be caused by RNA update callbacks */
/* Avoid re-entrant updates, can be caused by RNA update callbacks. */
if (ntree->is_updating) {
return changed;
}

View File

@ -241,7 +241,7 @@ static SeqCacheKey *seq_cache_choose_key(Scene *scene, SeqCacheKey *lkey, SeqCac
* This can happen because only FINAL_OUT item insertion will trigger recycling
* but that is also the point, where prefetch can be suspended.
*
* We could use temp cache as a shield and later untemp entry,
* We could use temp cache as a shield and later make it a non-temporary entry,
* but it is not worth of increasing system complexity.
*/
if (scene->ed->cache_flag & SEQ_CACHE_PREFETCH_ENABLE) {

View File

@ -31,7 +31,7 @@
# define BLI_qsort_r qsort_r
#endif
/* Quick sort reentrant */
/* Quick sort re-entrant */
typedef int (*BLI_sort_cmp_t)(const void *a, const void *b, void *ctx);
void BLI_qsort_r(void *a, size_t n, size_t es, BLI_sort_cmp_t cmp, void *thunk)

View File

@ -325,7 +325,7 @@ void BLI_linklist_apply(LinkNode *list, LinkNodeApplyFP applyfunc, void *userdat
#include "list_sort_impl.h"
#undef SORT_IMPL_FUNC
/* reentrant call */
/* re-entrant call */
#define SORT_IMPL_USE_THUNK
#define SORT_IMPL_FUNC linklist_sort_fn_r
#include "list_sort_impl.h"

View File

@ -312,7 +312,7 @@ static void listbase_double_from_single(Link *iter, ListBase *listbase)
#include "list_sort_impl.h"
#undef SORT_IMPL_FUNC
/* reentrant call */
/* re-entrant call */
#define SORT_IMPL_USE_THUNK
#define SORT_IMPL_FUNC listbase_sort_fn_r
#include "list_sort_impl.h"

View File

@ -1146,8 +1146,8 @@ bool invert_m4(float m[4][4])
* \return true on success (i.e. can always find a pivot) and false on failure.
* Mark Segal - 1992.
*
* \note this is less performant than #EIG_invert_m4_m4 (Eigen), but e.g.
* for non-invertible scale matrices, findinging a partial solution can
* \note this has worse performance than #EIG_invert_m4_m4 (Eigen), but e.g.
* for non-invertible scale matrices, finding a partial solution can
* be useful to have a valid local transform center, see T57767.
*/
bool invert_m4_m4_fallback(float inverse[4][4], const float mat[4][4])

View File

@ -90,7 +90,7 @@ BLI_INLINE char *med3(char *a, char *b, char *c, BLI_sort_cmp_t cmp, void *thunk
}
/**
* Quick sort reentrant.
* Quick sort re-entrant.
*/
void BLI_qsort_r(void *a, size_t n, size_t es, BLI_sort_cmp_t cmp, void *thunk)
{

View File

@ -239,7 +239,7 @@ void DEG_foreach_dependent_ID(const Depsgraph *depsgraph,
enum {
/* Ignore transform solvers which depends on multiple inputs and affects final transform.
* Is used for cases like snapping objects which are part of a rigid body simulation:
* without this there will be "false-positive" dependnecies between transform components of
* without this there will be "false-positive" dependencies between transform components of
* objects:
*
* object 1 transform before solver ---> solver ------> object 1 final transform

View File

@ -889,8 +889,8 @@ static void drw_sculpt_get_frustum_planes(Object *ob, float planes[6][4])
DRW_view_frustum_planes_get(DRW_view_default_get(), planes);
/* Transform clipping planes to object space. Transforming a plane with a
* 4x4 matrix is done by multiplying with the tranpose inverse. The inverse
* cancels out here since we transform by inverse(obmat). */
* 4x4 matrix is done by multiplying with the transpose inverse.
* The inverse cancels out here since we transform by inverse(obmat). */
float tmat[4][4];
transpose_m4_m4(tmat, ob->obmat);
for (int i = 0; i < 6; i++) {

View File

@ -1597,7 +1597,7 @@ void PE_update_object(Depsgraph *depsgraph, Scene *scene, Object *ob, int usefla
/* Only do this for emitter particles because drawing PE_FADE_TIME is not respected in 2.8 yet
* and flagging with PEK_HIDE will prevent selection. This might get restored once this is
* supported in drawing (but doesnt make much sense for hair anyways). */
* supported in drawing (but doesn't make much sense for hair anyways). */
if (edit->psys->part->type == PART_EMITTER) {
PE_hide_keys_time(scene, edit, CFRA);
}

View File

@ -3891,7 +3891,7 @@ static void nodelink_batch_draw(SpaceNode *snode)
void nodelink_batch_start(SpaceNode *UNUSED(snode))
{
/* TODO: partial workaround for NVIDIA driver bug on recent GTX/RTX cards,
* that breaks instancing when using indirect drawcal (see T70011). */
* that breaks instancing when using indirect draw-call (see T70011). */
g_batch_link.enabled = !GPU_type_matches(GPU_DEVICE_NVIDIA, GPU_OS_ANY, GPU_DRIVER_ANY);
}

View File

@ -97,7 +97,7 @@ typedef struct MovieClip {
struct MovieTracking tracking;
/**
* Context of tracking job used to synchronize data
* like framenumber in SpaceClip clip user.
* like frame-number in SpaceClip clip user.
*/
void *tracking_context;

View File

@ -433,7 +433,7 @@ typedef struct bNodeTree {
int flag;
/** Update flags. */
int update;
/** Flag to prevent reentrant update calls. */
/** Flag to prevent re-entrant update calls. */
short is_updating;
/** Generic temporary flag for recursion check (DFS/BFS). */
short done;

View File

@ -469,7 +469,7 @@ static void flatten_group_do(bNodeTree *ntree, bNode *gnode)
/* Flatten group to only have a simple single tree */
static void ntree_shader_groups_flatten(bNodeTree *localtree)
{
/* This is effectively recusive as the flattened groups will add
/* This is effectively recursive as the flattened groups will add
* nodes at the end of the list, which will also get evaluated. */
for (bNode *node = localtree->nodes.first, *node_next; node; node = node_next) {
if (ELEM(node->type, NODE_GROUP, NODE_CUSTOM_GROUP) && node->id != NULL) {

View File

@ -897,7 +897,7 @@ static PyObject *BPy_IDGroup_pop(BPy_IDProperty *self, PyObject *args)
pyform = BPy_IDGroup_MapDataToPy(idprop);
if (pyform == NULL) {
/* ok something bad happened with the pyobject,
/* ok something bad happened with the #PyObject,
* so don't remove the prop from the group. if pyform is
* NULL, then it already should have raised an exception.*/
return NULL;

View File

@ -1928,7 +1928,7 @@ static int pyrna_py_to_prop(
* layout.prop(self.properties, "filepath")
*
* we need to do this trick.
* if the prop is not an operator type and the pyobject is an operator,
* if the prop is not an operator type and the PyObject is an operator,
* use its properties in place of itself.
*
* This is so bad that it is almost a good reason to do away with fake
@ -3985,7 +3985,7 @@ static PyObject *pyrna_struct_type_recast(BPy_StructRNA *self)
}
/**
* \note Return value is borrowed, caller must incref.
* \note Return value is borrowed, caller must #Py_INCREF.
*/
static PyObject *pyrna_struct_bl_rna_find_subclass_recursive(PyObject *cls, const char *id)
{
@ -7192,7 +7192,7 @@ static PyObject *pyrna_srna_Subtype(StructRNA *srna)
#endif
/* Newclass will now have 2 ref's, ???,
* probably 1 is internal since decrefing here segfaults. */
* probably 1 is internal since #Py_DECREF here segfaults. */
/* PyC_ObSpit("new class ref", newclass); */

View File

@ -2351,26 +2351,26 @@ static PyObject *Vector_length_squared_get(VectorObject *self, void *UNUSED(clos
*
* axis_dict = {}
* axis_pos = {'x': 0, 'y': 1, 'z': 2, 'w': 3}
* axises = 'xyzw'
* while len(axises) >= 2:
* for axis_0 in axises:
* axis_chars = 'xyzw'
* while len(axis_chars) >= 2:
* for axis_0 in axis_chars:
* axis_0_pos = axis_pos[axis_0]
* for axis_1 in axises:
* for axis_1 in axis_chars:
* axis_1_pos = axis_pos[axis_1]
* axis_dict[axis_0 + axis_1] = (
* '((%s | SWIZZLE_VALID_AXIS) | '
* '((%s | SWIZZLE_VALID_AXIS) << SWIZZLE_BITS_PER_AXIS))' %
* (axis_0_pos, axis_1_pos))
* if len(axises) > 2:
* for axis_2 in axises:
* if len(axis_chars) > 2:
* for axis_2 in axis_chars:
* axis_2_pos = axis_pos[axis_2]
* axis_dict[axis_0 + axis_1 + axis_2] = (
* '((%s | SWIZZLE_VALID_AXIS) | '
* '((%s | SWIZZLE_VALID_AXIS) << SWIZZLE_BITS_PER_AXIS) | '
* '((%s | SWIZZLE_VALID_AXIS) << (SWIZZLE_BITS_PER_AXIS * 2)))' %
* (axis_0_pos, axis_1_pos, axis_2_pos))
* if len(axises) > 3:
* for axis_3 in axises:
* if len(axis_chars) > 3:
* for axis_3 in axis_chars:
* axis_3_pos = axis_pos[axis_3]
* axis_dict[axis_0 + axis_1 + axis_2 + axis_3] = (
* '((%s | SWIZZLE_VALID_AXIS) | '
@ -2380,7 +2380,7 @@ static PyObject *Vector_length_squared_get(VectorObject *self, void *UNUSED(clos
* %
* (axis_0_pos, axis_1_pos, axis_2_pos, axis_3_pos))
*
* axises = axises[:-1]
* axis_chars = axis_chars[:-1]
* items = list(axis_dict.items())
* items.sort(
* key=lambda a: a[0].replace('x', '0').replace('y', '1').replace('z', '2').replace('w', '3')