Proving that the angular velocity vector is equal to a limit involving the rotation vector

The angular velocity vector of a rigid body is defined as ω=r×v|r|2. But I'd like to show that that's equivalent to how most people intuitively think of angular velocity.

Euler's theorem of rotations states that any rigid body motion with one point fixed is equivalent to a rotation about some axis passing through the fixed-point. So let's consider a rigid body undergoing some motion with one point fixed, and for any times t1 and t2 let θ(t1,t2) denote the "rotation vector" of the rotation that's equivalent to the rigid body's motion between time t1 and time t2. For those who don't know, the rotation vector of a rotation is a vector whose magnitude is equal to the angle of the rotation and which points along the axis of the rotation; see this Wikipedia article.

Now my question is, how do we prove that the limit of θ(t1,t2)t2t1 as t2 goes to t1 exists, and that it's equal to the angular velocity vector?

This would all be much simpler if rotations were commutative since then the angular velocity would just equal the derivative of θ(t0,t) with respect to time. But since rotations are non-commutative, θ(t1,t2) does not equal θ(t0,t2)θ(t0,t1) and thus the relation between angular velocity and the time derivative of θ(t0,t) is considerably more complicated; see this journal paper for details.

Note: This is a follow-up to my question here.

EDIT: Note that what this journal paper calls α(t) would in my notation be written as θ(t0,t). The paper discusses the fact that the angular velocity vector ω(t) is not equal to the time derivative of α(t). This means that the limit of θ(t0,t2)θ(t0,t1)t2t1 does not equal ω(t1). But my question is about proving a slightly different statement, which is that the limit of θ(t1,t2)t2t1 as t2 goes to t1 DOES equal ω(t1). Note that the expressions θ(t0,t2)θ(t0,t1)t2t1 and θ(t1,t2)t2t1 are not equal, because θ(t0,t1)+θ(t1,t2) does not equal θ(t0,t2) due to the non-commutativity of rotations. So none of what I'm saying contradicts or seeks to disprove the journal paper.

@claudechuber I did provide a rigorous definition in my question. If we have a rigid body which undergoes some motion between time t1 and time t2 with one point fixed, then Euler's theorem of rotations states that there exists a rotation about some axis passing through that fixed point which is equivalent to that motion. I am defining θ(t1,t2) to be the rotation vector of that rotation, i.e. a vector which points in the direction of the axis of that rotation and whose magnitude is the angle of that rotation. What is unclear in that?
@claudechuber Wow, you're right, moving the point that you consider t0 does lead to the statement I want to prove through the equation proven in Peres' paper. If you want to post that as an answer (spelling out the details of course), I'm happy to accept it.
@claudechuber By the way, you may be interested in my follow-up question here: physics.stackexchange.com/q/287830/27396

Antworten (6)

If you take t0=t1 in the EDIT part of your question, with θ(t0,t1)=θ(t1,t1)=0, you are in the special case ω=α˙ from Asher Peres's paper you have mentioned, which then proves your statement according to your observation from the EDIT part, because in that case, you have

θ(t0,t2)θ(t0,t1)t2t1=θ(t1,t2)t2t1,
whose limit when t2t1 is equal to α˙(t1)=ω(t1).

Indeed, according to Asher Peres, we have:

w=α˙+1cosαα2(α×α˙)+αsinαα3(α×(α×α˙)),
which, for α(t)=θ(t1,t) and t=t1, and using α(t1)=θ(t1,t1)=0 (see above), reduces to
w(t1)=α˙(t1)+0.
Note that luckily 1cosα=O(α2) and αsinα=O(α3), hence there is not problem when passing to the limit when α0.

Thanks! Could you just add the equation relating α and the time derivative of ω from Peres' paper, that way people don't need to read the paper to make sense of your answer? And add a link to the paper.
Thanks for the edit! By the way, the fact that there's no problem in the limit as α goes to zero can also be seen in the earlier expression for ω in terms of an infinite series, because there α is only in the numerator, so when you set α equal to 0 only the time derivative of α survives.

The two concepts do seem to be similar. The key I think lies in the fact that one can express an infinitesimal angle as the arclength divided by the radius

δθ=δsr.
If the radius is expressed as a vector and the arclength as another vector indicating the direction of motion during rotation, then one can express the infinitesmial rotation as a cross-product
δθ=r×δs|r|2.
Now we just need to divide by the difference in time to make the connection
δθδt=r×δs|r|2δt.
In the appropriate limit, this then becomes
ω=dθdt=r|r|2×dsdt=r×v|r|2.

I'd like a more rigorous proof, using limits rather than using informal infinitesimal reasoning like this.

Consider a fixed point with location a or a rigid body.

To prove the rotation first establish that

da=dθ×a

This can be done with just geometry given that small angle approximations. For example the change in the x-direction is dax=azθyaydθz.

The expression can be written as

v=dadt=ω×a

The last part is to calculate

a×v=a×(ω×a)=ω(aa)a(rω)

Take the projection of the location perpendicular to the rotation r with rω=0 then

r×v=ωr2rω
ω=r×vr2


Edit 1

A more vigorous treatment involves creating a 3×3 rotation matrix, and applying small angle approximation to it. Use θ=(θx,θy,θz) as successive rotations

R=Rx(θx)Ry(θy)Rz(θz)=|cosθycosθzcosθysinθzsinθycosθxsinθz+sinθxsinθycosθzcosθxcosθzsinθxsinθysinθzsinθxcosθysinθxsinθzcosθxsinθycosθzsinθxcosθz+cosθxsinθysinθzcosθxcosθy|

All this now applied to a small angle to make da=dRaa such that sin()= and cos()=1

dR=Rx(dθx)Ry(dθy)Rz(dθz)=|1dθzdθydθz+dθxdθy1dθxdθydθzdθxdθy+dθxdθzdθx+dθydθz1|=|1dθzdθydθz1dθxdθydθx1|

So

da=dRaa=(dR1)a=dθ×a

[dθ×]=|0dθzdθydθz0dθxdθydθx0|

dθdt×=|0ωzωyωz0ωxωyωx0|

The last 3×3 matrix is called the vector cross product operator matrix. It is skew symmetric and it is used widely in computer graphics and in dynamics.

I'd like to see a more rigorous proof involving limits, not informal infinitesimal reasoning like this.
You can also create a 3×3 rotation matrix with 3 Euler angles and take small angle approximations to arrive at dθ× as a 3×3 operator.
Yeah, I'm aware of that. I just want to do things via axis-angle vectors. And I want a proof using limits, not infinitesimal reasoning
Like I said, I'd like an answer based on axis-angle vectors, not rotation matrices or Euler angles, and based on limits, not infinitesimals.
Infinitesimals and limits are the same thing you know. And if you notice I am using the angle vector θ.

I'll try to be as rigorous as possible but before diving in to the problem I want to explain my notation. I've dropped vector arrows if it is clear from the expression, which quantity is a vector and which is a scalar. Furthermore sometimes I've dropped the time dependence but one can deduce from the context which quantity time dependent. Note that differentiability is a local condition that is it doesn't care about what the function does far apart but just around the point that you are taking the derivative. This is apparent from the definition of a differentiable function. Let f:RRn be a function. We say that the function f is differentiable at a (fixed) point x0R if and only if the following limit exists:

limxx0f(x)f(x0)xx0=limh0f(x0+h)f(x0)h

if we substitute xx0=h. This is probably the usual definition of a derivative that you know. However for our purposes I want to use another equivalent definition, which might look weird at the first glance. The function f is differentiable at x0R if and only if there exists JRn and a function H:RRn, which is continuous at x0 with the value H(x0)=0Rn, such that the following holds for all xR:

f(x)f(x0)J(xx0)=H(x)|xx0|=:o(xx0)

There is an intuitive way to think about this definition, which basically tells you that the tangent line with "slope" J that you put at the function should have maximum a linear error. If you think about in terms of Taylor series the definition becomes more clear. Note again that the differentiability is a local condition. You can see this clearly from the second definition. There are absolutely no conditions on how the function H should behave like (except around x0, which we require it to be continuous since continuity also a local condition remember the ϵ,δ definition of continuity). Thus far away from x0 we just define H to be:

H(x):=f(x)f(x0)J(xx0)|xx0|

So with this intro let's return to our problem. We can only prove that the angular velocity vector is given by ω=θ˙(t0) locally around t0. Note that we will choose t0R to be a fixed but an arbitrary point. Thus you can show that for all t0R there exists a neighbourhood of ω(t0) such that θ˙(t0)=ω(t0). Of course I am assuming here that θ is a differentiable function over all R, which would be the case if you consider something "physical" since you can make the slope of θ arbitrarily large but in praxis you cannot make it not differentiable. I assume that you know that the Lie Algebra of SO(3) is so(3) which are basically all 3×3 skew-symmetric matrices. I'll choose t0=0 without loss of generality since you can translate the time axis and redefine your functions. Note that r(t)=R(t)r(0), where of course my origin is the fixed point and R:RSO(3) is a differentiable function with the property that R(0)=I. Thus the derivative of r at zero is given by:

r(t)r(0)vt=o(t)

for some vR3, which is the velocity vector. We want to figure out what v is in terms of r and R. Note that Rij(t)=Iijtϵijkωk+o(t) for some ωkR, where summation over k is implied. You might think that I'm doing something shady by calling these numbers ωk. Note that we are doing maths at this moment thus if you want you can call them wk. I'll explain later why physically ωk is the angular velocity but for now we have:

tϵijkωkrj(0)vit=o(t)

at this point I think it is quite obvious what you should choose as vi, namely you choose vi=ϵijkrj(0)ωk, which you can also write it as v(0)=ω×v(0). Assuming that ω is the angular velocity at the moment, let me calculate your identity. We just take the cross product with r from left:

r×v=r×(ω×v)=r2ω(rω)r=r2ω

where I used the fact that ωr.

Our main problem now is to show that the numbers ωk are in fact the angular velocity. Sadly there is no rigorous proof of this fact because know we are leaving the realm of maths and entering the realm of physics. I'll try to convince you of this fact by giving an example. You know that the position vector of a rotating point mass around z axis with angular velocity ω is given by:

r=(rcosωtrsinωt0)=R(t)(r00)=R(t)r(0)

where

R(t)=(cosωtsinωt0sinωtcosωt0001)

is the usual rotation matrix. By our definition the angular velocity vector has to be exactly the vector in z direction because I have chosen the wrong si(g)n (pun intended!) with length ω. Note that we have:

R(t)=(cos0sin00sin0cos00001)+ωt(sin0cos00cos0sin00000)+o(t)

comparing this with the equation Rij(t)=Iijtϵijkωk+o(t), we see that:

(0ωt0ωt00001)=(0ω3tω2tω3t0ω1tω2tω1t0)

and now you see ω1=ω2=0 and ω3=ω as promised. Now obviously this is not a proof of the fact that in general this holds but if you want you can also try this for a rotation in x and y axes and you get any combination thereof if you remember the vague written identity (s. wiki for the Baker–Campbell–Hausdorff formula):

exp[so(3)+so(3)]=SO(3)SO(3)

I have to gloss over some aspects like Lie algebra/group correspondence and locality of continuity of a function because I assumed that you are familiar with these ideas, if not feel free to say so in the comments and I'll edit my answer accordingly.

You say "We can only prove that the angular velocity vector is given by ω=θ˙(t0) locally around t0. Note that we will choose t0R to be a fixed but an arbitrary point. Thus you can show that for all t0R there exists a neighbourhood of ω(t0) such that θ˙(t0)=ω(t0)." But that is not true. The angular velocity vector ω is not the derivative of θ with respect to time, because of the non-commutative nature of rotations. See this journal paper for details.
@KeshavSrinivasan Now I understand what your confusion is (I think)! Note that my theta has the property that r=rθ, which is telling where the position vector is. Your θ is however some other vector and when we talk about the angular velocity we talk about the rate of change of my θ. Thus your equation is valid for my θ. In summary nobody is telling you that ω=θ˙ if we take your θ as you can also see in this journal paper that you have linked to. I think you need to refine your intuition about the equation ω=θ˙
OK, what is the definition of your θ? Is it r|r|? Are you saying that the derivative of r|r| with respect to time is equal to the angular velocity vector? I don't think that's true. And I'm pretty sure that there is no physically meaningful vector whose time derivative is the angular velocity vector.
Also, the purpose of my question is to show that (under my definition of θ) the limit of θ(t1,t2)t2t1 as t2 goes to t1 is equal to the angular velocity vector. I don't see how your answer addresses that.
@KeshavSrinivasan Well in the answer the θ vector is written as the matrix R and of course θ˙=ω (locally) in my answer. Furthermore as I mentioned in the comment above the if you take the limit of θ according to your definition you don't get the angular velocity vector. Who told you that and why do you believe that this should be the case? The article that you've linked shows that this is not the case??

(NOTE:Posted as new answer instead of edit to previous answer because we will look at this scenario slighly differently.


One important fact is that finite rotations do not commute in gerneral, however infinitesimal rotations always commute. Let's revisit this point later.

But first let's invoke some proper statements regarding theorems.

{{Euler's Theorum Formally Stated}} For {any} general proper, orthogonal operator R, there exists a fixed axis n^ and an angle Φ in the range 0Θπ such that R[Φn^]=R

{Source}: Analyitcal Mechanics for Relativity and Quantum Mechanics, Oliver Davis Johns, Oxford University Press, 2005)

({ {General Theorem :}}Angular Velocity of Parametized Operators)\ \emph{Any "`time-varying"'} rotated vector (fixed axis or not) may be written as V(t)=R[Φ(t)n^(t)]V ({Source}: Analyitcal Mechanics for Relativity and Quantum Mechanics, Oliver Davis Johns, Oxford University Press, 2005)

{{Consequence of second Theorem}}: If the operation of interest relates to the rotational velocity as ddtR=Ωikϵijkωj and the vector of interest is the position vector r then the derivative of these items is ddtV(t)=ω×V

then the explicit form of the rotational velocity in terms of a {general } rotational angle that solves these criteria (also see derivation of article) turns out to be ω(t)=Φ˙+sin(Φ)dn^dt+(1cos(Φ))n^×dn^dt

{{{But}}}... for a fixed axis, (as in the restrictions of Euler's theorem), ddtn^ turns out to be zero. Consequently, for a fixed axis of rotation

ω(t)|fixed axis=Φ˙n^

{{QED}}

Now let us return to the first point, which has not yet been demonstrated...

If ω(t)|=Φ˙n^ is valid for any fixed axis rotation, a simple deduction is that it should apply to a sequence of fixed axis rotations. This is somewhat of an intuitive connundrum however. One reason the truth of the initial statement is important.

Let us consider a sequence of two fixed axis rotations a and be

Allegedly ω(ta+tb)=ω(ta)n^a+ω(tb)nb^, and the operation should commute if the rotations are infinitesimal ...still EDITING.

Comments are not for extended discussion; this conversation has been moved to chat.

There are some differences in the notation of the article with those used in the original question. This answer is provided within the notation of the article.

Key point: The article states that α˙ω with α correponding to the angle of rotation.

Now we can equate r˙=ω×r with r˙=Ωr with Ω is an antisymetric matrix defined by Ωik=ϵijkωj. THESE are correct regarding positional changes of the object in terms of its angular velocity.

(I suspect what you are refering to as "lim Θ(t1,t2)t2t1..." is ACTUALLy Ω in the article, one reason I have migrated to the article's notation so as not to confuse this object with α˙ or Θ˙ in your notation.)

Your question as stated then reads to me "How do we proove THIS OBJECT (i.e. Ω exists and that it equates to the operation ω×)."

First of all, it has been defined in terms of the anti-simetric tensor ϵijkand the angular velocity ω ; ϵijkωjrk=ω×r by definition.

But my intuition is that you are intersted in how to arrive at this object starting from α instead. I believe this is demonstrated sufficiently in the first collum of the article.

To summarize: Letting the orthogonal matrix S=eα×, r=Sr0. it follows that

r˙=S˙S1r. Therefor Ω=S˙S1. If α exist, S exists and Ω may be expressed in term of it. It works from both angles. (Also the derivatives of S given in the article are calculable.)

To put another way equations of the same form have the same solution. On one hand we have r˙=S˙S1r on the other we have r˙=ω×r. Therby it is trival to make the interpretive association Ω=S˙S1


EDIT: In more explicit fashion

1) We have two items to consider initially.

A) The first is r˙=ω×r, the time derivitive of the position vector r as a function of itself. Let Ω be the matrix object that fulfills this relation. Ω BY DEFINITION is Ωϵijkωj where ωj are the components of the angular velocity.

B) The second is r˙=f(α) the time derivative of the position vector r as a function of the rotational angle vector α.

Logical connection: A=B

Since