弹性波动力学笔记(一) 旋转矩阵简介
Introduction of Rotation Transformation And Rotation Matrix
1.1 Summary of vector analysis
A vector is defined as a directed line segment, having both magnitude and direction. The magnitude, or length, of a vector a will be represented by . The sum and the difference of two vectors, and the multiplication of a vector by a scalar (real number) are defined using geometric rules. Given two vectors a and b, two products between them have been defined.
Scalar, or dot, product:
where is the angle between the vectors.
Vector,or cross, product:
where is as before, and n is a unit vector (its length is equal to 1) perpendicular to both a and b such that the three vectors form a right-handed system.
An important property of the vector product, derived using geometric arguments, is the distributive law:
By introducing a rectangular Cartesian coordinate system it is possible to write a vector in terms of three components. Let ,and be the three unit vectors along the , and are axes of the coordinate. Then any vector can be written as
The components ,and are the orthogonal projections of in the directions of the three axes.
Before proceeding, a few words concerning the notation are necessary. A vector will be denoted by a bold-face letter, while its components will be denoted by the same letter in italics with subindex (literal or numerical). A bold-face letter with a subindex represents a vector, not a vector component. The three unit vectors defined above are examples of the latter. If we want to write the k-th component of the unit vector we will write . For example, ,,and . In addition , although vectors will usually be written in column form, when they are involved in matrix operations they should be considered as column vectors, as matrices of one column and three rows. For example, the matrix form of the scalar product is , where T indicates transposition.
When the scalar product is applied to the unit vectors we find
Equations(5) and (6) can be summarized as follows:
The symbol is known as Kronecker delta, which is an example of a second-order tensor, and will play an important role in this book. As an example of equation (7) , unless , in which case the scalar product is equal to 1.
Next we derive an alternative expression for a vector .Using equation (4), the scalar product of and is
Note that when applying equation(4) the subindex in the summation must be different from . To obtain equation(8) the following were used : the distributive law of the scalar product, the law of the product by a scalar and equation(7). Equation(8) shows that the component of can be written as
When equation (9) is introduced in equation (4) we find
This expression will be used in the discussion of dyadic.
In terms of its components the length of the vector is given by
Using purely geometric arguments it is found that the scalar and vector products can be written in component form as follows :
and
The last expression is based on the use of equation(2) and equation(3).
Vectors, and vector operations such as the scalar and vector products, among others, are defined independently of any coordinate system. Vector relations derived without recourse to vector components will be valid when written in component form regardless of the coordinate system used. Of course, the same vector may (and generally will) have different components in different coordinate systems, but they will represent the same geometric entity. This is true for Cartesian and more general coordinate systems, such as spherical and cylindrical ones, but in the following we will consider the former only.
Now suppose that we want to define new vector entities based on operations on the components of other vectors. In view of the comments in 1.1 it is reasonable to expect that not every arbitrary definition will represent a vector, an entity intrinsically independent of the coordinate system used to represent the space. To see this consider the following example, which for simplicity refers to vectors in two-dimensional (2-D) space.
1.2 Rotation of Cartesian coordinates. Definition of a vector
Let represent a Cartesian coordinate system and another system obtained from the previous one by a rotation about their common origin . Let and be the unit vectors along the three axes in the original and rotated systems. Finally, Let denote the cosine of the angle between and . The are known as direction cosines, and are related to and by
Given an arbitrary vector with components in the original system, we are interested in finding the new components in the rotated system. To find the relation between the sets of components we will consider first the relation between the corresponding unit vectors. Using (1.2.1) equation can be written as
Furthermore, in original and rotated systems can be written as
and
Now introduce (1.2.2) in (1.2.4)
Since(1.2.3) and (1.2.5) represent the same vector, and the three unit vector and are independent of each other, we conclude that
If we write the is in terms of the and replace them in equation (1.2.6) we find that
Note that equation(1.2.6) the sum is over the first subindex of , while in (1.2.7) equation the sum is over the second subindex of . This distinction is critical and must be respected.
Now we are ready to introduce the following definition of a vector:
Three scalars are the components of a vector if under a rotation of coordinate they transform according to equation(1.2.7).
What this definition means is that if we want to define a vector by some set of rules, we have to verify that the vector components satisfy the transformation equations.
Before proceeding we will introduce a summation convention (due to Einstein) that will simplify the mathematical manipulations significantly. The convention applies to monomial expressions (such as a single term in an equation) and consists of dropping the sum symbol and summing over repeated index. This convention requires that the same index should appear no more than twice in the same term.
Repeated indexes are known as dummy indices , while those that are not repeated. Using this convention, while those that are not repeated are called free indices. Using this convention, we will write, for example,
It is important to have a clear idea of the difference between free and dummy indices. A particular dummy index can be changed at will as long as it is replaced (in its two occurrences) by some other index not equal to any other existing indices in the same term. Free indices, on the other hand, are fixed and cannot be changed inside a single term. However, a free index can be replaced by another as long as the change is effected in all the terms in an equation, and the new index is different from all the other indices in the equation. In equation(1.2.9) is a dummy index and is a free index, while in equation (1.2.10) their role is reversed. The examples below show legal and illegal index manipulations.
The following relationship, derived from equation(1.2.9) , are true
because the repeated index was replaced by a different repeated index(equal to or ). However, it would not be correct to replace by because is already present in the equation. If were replaced by we would have
which would not be correct because the index has been appeared more than twice in the right-hand term, which is not allowed. Neither would it be correct to write
because the free index has been changed to only in the right-hand term. On the other hand, (1.3.9) can be written as
because the free index has been replaced by on both sides of the equation.
As equations (1.2.9) and (1.3.10) are of fundamental importance, it is necessary to pay attention to the fact that in the former the sum is over the first index of while in the latter the sum is over the second index of . Also note that (1.2.10) can be written as product of a matrix and a vector:
Where is the matrix with elements . It is clear that (1.2.9) can be written as
where the superscript indicates transposition.
Now we will derive an important property of . By introducing (1.2.10) in (1.2.9) we obtain
Note that it was necessary to change the dummy index in equation(1.2.10)to satisfy the summation convention. Equation(1.2.17) implies that any of the three components of is a combination of all three components. However, this cannot be generally true because is an arbitrary vector.
Therefore, the right-hand side of equation (1.2.17) must be equal to , which in turn implies that the product must be equal to unity when , and equal to zero when . This happens to be the definition of the Kronecker delta introduced in equation (1.2.7), so that
If equation (1.2.9) is introduced in equation (1.2.10) we obtain
Setting in equation (1.2.19) and writing in full gives
where the equality on the right-hand side follows from(1.2.2).
When ,equation(1.2.19) transform into
where the equality on the right-hand side also follows from (1.2.2). Therefore, (1.2.19) summarizes the fact that the are unit vectors orthogonal to each other, while (1.2.18) does the same thing for the s. Any set of vectors having these properties is known as an orthonormal set.
In matrix form, equation (1.2.18) and (1.2.19) can be written as
Where is the identity matrix.
Equation(1.2.22) can be written in the following useful way:
where the superscript-1 indicates matrix inversion. From (1.2.23) we also find
where vertical bars indicate the determinant of a matrix.
Linear transformations with a matrix such that its determinant squared is equal to 1 are known as orthogonal transformations. When , the transformation of one coordinate axis in a coordinate plane. An example of reflection is the transformation that leave the and axes unchanged and replaces the axis by . Reflections change the orientation of the space: if the original system is right-handed, then the new system is left-handed , and vice versa.
figure 1. Rotation of Axes. Primed and unprimed quantities refer to the original and rotated coordinate systems, respectively. Both
1.3 Rotation Matrices in two, three and n-dimensions space
A matrix is a representation of a linear transformation described in 1.2 chapter, which can be viewed as a machine that consumes a vector and spits out another vector. A rotation is a transformation with the property that the vector consumed by the machine and the vector spit out by the machine have the same length. That is, physically rotating a vector by an angle leaves the length of the vector unchanged. As a matrix equation, if is a rotation and is a vector ,then
A. Property of Two Dimension Rotation Matrices
In the case of . it is easy to work out the form of the most general real orthogonal matrix. set is 2D rotation matrix
According to property of rotation matrix , we can obtain
Thus, we can obtain three constraint equations, as expected from
Thus, we can obtain three constraint equations:
We now consider separately the cases of proper and improper rotations. In the case of proper rotations, we add a fourth equation (1.3.5)
Suppose that and . From equation(1.3.5) ,it follows that . Substituting this into equation (1.3.6) yields,
Combining equation. and(), we conclude that . Plugging this result back into yields . With these results, is automatically satisfied in light of . Hence,the most general real orthogonal matrix with determinant equal to 1 is given by,
Without loss of generality, we can set and ,since and satisfy same properties as and , namely, and , for . Indeed, the rotation matrix,
represent a proper counterclockwise rotation by an angle in the plane, as discussed .
In the case of improper rotations, we start with and impose , which again we can obtain follow equations:
Suppose that and . From equation (1.3.5) , it follows that . Substituting this into equation(1.3.5) :
we can conclude that . Plugging this result into equation (1.3.5) yields .With these results, eq(1.3.11) is automatically satisfied in light of .Hence, the most general real orthogonal matrix with determinant equal to -1 is given by,
we can again set and . Thus, an example of an improper rotation matrix is,
This matrix satisfies the property, . Note that we can express given equation (1.3.9) as the product of a proper rotation and a reflection,
B. Property of Three Dimension Rotation Matrices
It becomes quickly evident that the methods used in the previous section become much less practical for n=3. In this section I shall present the explicit form of the proper rotation matrix along with its most important properties. A derivation of the rotation matrix is given in an Appendix to these notes.
As previous noted, the most general three-dimensional rotation, which we henceforth by , can be specified by an-axis of rotation pointing in the direction of the unit vector , and a rotation angle . Conventionally, a positive rotation angle corresponds to a counterclockwise rotation. The direction of the axis is determined by the right hand rule. Namely, curl the fingers of your right hand around the axis of rotation, where your fingers point in the direction. Then, your thumb points perpendicular to the plane of rotation in the direction of . In general, rotation matrices do not commute under multiplication. However, if both rotations are taken with respect to the same fixed axis, then
Simple geometric considerations will convince you that the following relations are satisfied:
Combining these two results, it follows that
which implies that any three-dimensional rotation can be described by a counterclockwise rotation by an angle about an arbitrary axis , where . However, if we substitute , we conclude that
which means that for the special case of , and represent the same rotation. In particular, note that
Indeed for any choice of , the are the only non-trivial rotation matrices whose square is equal to the identity matrix. Finally, if then is the identity matrix (sometimes called the trivial rotation), independently of the direction of .
We now present an explicit form for . Since describes a rotation by an angle about an axis , the formula for will depend on the angle and on the coordinates of with respect to a fixed Cartesian coordinate system. Note that since is a unit vector, it follows that,
We can also express in terms of its polar and azimuthal angles ( and , respectively) with respect to a fixed . In particular,
One can check that equation (1.3.20) is indeed satisfied. Thus, depends on two independent parameters, and , which together with the rotation angle constitute the three independent parameters that describe a three dimensional rotation.
The explicit formula for the real orthogonal matrix with determinant equal to 1 is given by,
One can easily check that equations (1.3.15) and(1.3.16), the rotations represent the same rotation, and are writen as tensor form
where the Kronecker symbol was introduced in equation . Finally,as expected, , independently of the direction of . I leave it as an exercise to the student to verify explicitly that given in equation (1.3.22) satisfies the conditions and .
Although equation (1.3.22) looks complicated, one can present an elegant expression for the matrix elements of , denoted below by . Indeed, it is not difficult to check that the following expression for is equivalent to the matrix elements of exhibited in equation (1.3.22)
where is the Levi-Civita epsilon symbol, which is defined as follows,
Note that is the version of the Levi-Civita symbol introduced in the class handout entitled Determinant and the Adjugate.
It is instructive to check special cases of equation(1.3.22). For example, suppose we choose corresponding to a rotation axis that points along the positive . In this case, and , and equation(1.3.22) becomes
Of course, equation (1.3.26) is the expected result given the form of the two-dimension rotation matrix given in equation (1.3.9).
Likewise, one can choose either or corresponding to rotation axes that point along the positive (i.e. , ) or along the positive y-direction (i.e., and ), respectively. In these cases, equation (1.3.22) yields,
To learn more about the properties of a general three-dimensional rotation, consider the matrix representation , with respect to the standard basis . We can define a new coordinate system in which the unit vector points in the direction of the new ; the corresponding new basis will be denoted by . The matrix representation of the rotation with respect to is then given by equation(1.3.26);
Using the formalism developed in the class handout, Vector coordinates, matrix elements,changes of basis, and matrix diagonalization, there exists a real special orthogonal matrix such that
where is given by equation (1.3.26). In Appendix A, we will determine an explicit form of the matrix . However, the mere existence of the matrix in equation (1.3.29) is sufficient to provide a simple algorithm for determining the rotation axis (up to an overall sign) and the rotation angle that characterize a general three-dimensional rotation matrix.
C. Determining the rotation axis and the rotation angle
Given a general three-dimensional rotation matrix, , we can determine the angle of rotation and the axis of rotation . Using equation (1.3.22), the trace of is given by:
which coincides with our previous result obtain in equation (1.3.30). Thus, yields,
where is a consequence of the range of the rotation angle, .
To determine , we multiply equation (1.3.24) by and sum over and . Noting that
it follows that
If is a symmetric matrix (i.e. ), then automatically since is antisymmetric under the interchange of the indices and . In this case , and we must seek other means to determine . If , then one can divide both sides of equation (1.3.33) by . Using equation (1.3.33), we obtain:
More explicitly,
If we multiply equation (1.3.33) by and sum over , then
after using . This provides an additional check on the determination of the rotation angle.
Alternatively, we can define a matrix whose matrix elements are given by:
after using equation (1.3.24) for . Hence,
To determine up to an overall sign, we simply set (no sum) in equation (1.3.38), which fixes the value of
If , the overall sign of is fixed by equation (1.3.34) by . Note that equation (1.3.37) implies that . Summing over in equation (1.3.39) then yields
as required for a unit vector. As noted above, if is a symmetric matrix (i.e. ), then and cannot be determined from equation (1.3.34) . In this case, equation (1.3.30) determines whether or . For , equation (1.3.39) yields , which determines up to an overall sign. Equivalently, one can use equation (1.3.22) to derive
where the individual signs are determined up to an overall sign via
where the individual signs of is immaterial. Finally, in the case of (which corresponds to ), is the identity matrix, which is independent of the direction of .To summarize, equation(1.3.31), (1.3.3.35), (1.3.38) provide a simple algorithm for determining the unit vector and the rotation angle for any rotation matrix .
D. Euler angle representation of
An arbitrary rotation matrix can be written as:
where , and are called the Euler angles. The ranges of the Euler angles are: and . We shall prove these statements “by construction.” That is, we shall explicitly derive the relations between the Euler angles and the angles and that characterize the rotation , where and are the polar and azimuthal angle that define the axis of rotation ,
Multiplying out the three matrices on the right-hand side of equation (1.3.30), we obtain the Euler angle parameterization of the three-dimensional rotation matrix.
One can now make use of the results of section 4 to obtain and in terms of the Euler angles and .
Appendix A. An explicit formula for the Rotation Matrix
Suppose we wish to determine the explicit form of the rotation matrix . Here is one possible strategy. The matrix is specified with respect to the standard . Given that the explicit form for is known according to equation (1.3.29) suggests that we should transform to a new orthonormal basis, , in which new positive points the direction of . That is,
The new positive y-axis can be chosen to lie along
since by construction, is a unit vector orthogonal to . We complete the new right-handed coordinate system by choosing:
Following the class handout entitled, Vector coordinates, matrix elements, changes of basis, and matrix elements, changes of basis, and matrix diagonalization, we determine the matrix whose matrix elements are defined by
where the are the basis vectors of and the are the basis vectors of .The columns of are the coefficients of expansion of the new basis vectors in terms of the old basis vectors. Thus,
The inverse is easily computed since the columns of are orthonormal, which implies that is an orthogonal matrix, i.e. .
According to equation (16), of the class handout, Vector coordinates, matrix elements, changes of basis and matrix diagonalization,
where is the matrix with respect to the standard basis, and is the matrix with respect to the new basis (in which points along the new positive ). In particular,
Hence, equation (A-6) yields
where is given by equation (A-7) and .
For ease of notation, we define
Note that , since is a unit vector. Writing out the matrices :
Using and , the final matrix multiplication then yields,
which coincides with the result previously exhibited in the class handout entitled Rotation matrix.
【推荐】国内首个AI IDE,深度理解中文开发场景,立即下载体验Trae
【推荐】编程新体验,更懂你的AI,立即体验豆包MarsCode编程助手
【推荐】抖音旗下AI助手豆包,你的智能百科全书,全免费不限次数
【推荐】轻量又高性能的 SSH 工具 IShell:AI 加持,快人一步
· 25岁的心里话
· 闲置电脑爆改个人服务器(超详细) #公网映射 #Vmware虚拟网络编辑器
· 零经验选手,Compose 一天开发一款小游戏!
· 通过 API 将Deepseek响应流式内容输出到前端
· AI Agent开发,如何调用三方的API Function,是通过提示词来发起调用的吗