The figure illustrates a the two first important properties of vector addition:
Addition is associative, i.e. u+v=v+u
Addition is commutative, i.e. (u+v)+w=u+(v+w)
More formally, we can introduce a vector space V. This can for example be the real (3-dimensional) world, R3. We can then define the full set of properties required for the vectors in V to be just vectors:
Vector summation
Associative addition (see above)
Commutative addition (see above)
There exists a unique zero vector, 0, such that 0+v=v,∀v∈V
There exists a unique vector −v, ∀v∈V, such that v+(−v)=0.
Multiplication by scalar Consider scalars a,b∈R, then
Associative multiplication by scalar, a(bv)=(ab)v
Commutative multiplication by scalar, av=v⋅a
Multiplication by zero and unity: 1v=v, 0v=0
Distributive multiplication by scalar wrt. vector addition: a(v+w)=av+aw
Distributive multiplication by scalar wrt. scalar addition: (a+b)v=av+bv
Assume that you have two different vectors in 2d and you want to construct a 3rd vector by scaling your vectors and adding them together. Since there are 2 dimensions, and you have two vectors, this should be possible, or?
However, what if your two vectors are pointing in the same direction (but have different length)? Then, you can only construct another vector in that direction. So even if you have two vectors, you are only spanning a line in your 2-dimensional space. I.e. you are only spanning a 1-dimensional space with your vectors. The reason is that these vectors are linearly dependent. Generally, we can formulate the definition of linear dependence as
Definition: With i=1,2,⋯,N, the vectors vi∈V are linearly dependent if there exists scalars ai∈R:aiai>0 (i.e. at least one ai=0) such that aivi=0). If no such scalars exists, the vectors vi are linearly independent.
v=aivi, which we just used, is called a linear combination of the vectors vi. (Recall the Einstein summation convention)
The scalar product, u⋅v, between two vectors, u and v, is defined as
u⋅v=∣∣u∣∣∣∣v∣∣cos(θ)
where ∣∣u∣∣ is the length of the vector u, and θ is the angle between vectors u and v. Note that we do not require a coordinate system to define this operation!
From this definition, it follows that ∣∣u∣∣=u⋅u as in this case, θ=0. Hence, u⋅u=0⇔u=0. Furthermore, if u and v are perpendicular vectors, then θ=π/2 and thus u⋅v=0.
When working with scalar products, the following algebraic rules apply
The cross (or vector) product, u×v, between two vectors, u and v, is defined as
u×v=∣∣u∣∣∣∣v∣∣sin(θ)n
with the same quantities as in Equation (1). The new vector, n, is the normal vector (with unit length) to the plane spanned by u and v, following the right-hand-rule. The following animation from Wikipedia illustrates this rule well
A couple of other properties, following from the definition in Equation (2), are also shown by this animation:
u×u=0
u×(−u)=0
∣∣u×v∣∣=∣∣u∣∣∣∣v∣∣ if u⊥v (i.e. u and v are perpendicular such that u⋅v=0)
When working with cross products, the following algebraic rules apply
Associative wrt. multiplication with scalar, a(u×v)=(au)×v
Here it is important to note the difference to scalar products, which are commutative, while cross products are anti-commutative (i.e. the order of multiplication matters!)
Above, we have seen that many properties of vectors can be defined without a basis system, simply by considering a vector as an object with a length and a direction. For these definitions, there is no need to introduce a coordinate system. However, if we want to perform actual calculations, everyting becomes easier if we define a base system from which we measure all quantities. Using a set of predefined so-called base vectors, we can express other vector quantities by using a linear combination of these base vectors. To ensure that this description is unique, we require that the base vectors are linearly independent. Otherwise, we can have the same vector described by different linear combinations.
Let's consider a basis system described by the linear independent vectors ei. A vector v, can then be described by the unique coefficients vi (considering the ei basis system) by v=viei. If we take the scalar product u⋅v, we get
u⋅v=uivjei⋅ej
resulting in the metric coefficients eij=ei⋅ej for the coordinate system. This coefficient complicates our calculations, but by choosing our basis vectors cleverly, we can make it become δij, resulting in u⋅v=uivi. But how can we choose ei, such that ei⋅ej=δij?
The solution is an orthonormal coordinate system (Cartesian coordinate system). The ortho comes from orthogonal, implying perpendicularity between the base vectors. If we consider the scalar product, this implies that ei⋅ej=0 if i=j. The normal comes from that the base vectors are normalized, i.e. have unit length (∣∣ei∣∣=1). This implies that ei⋅ej=1 if i=j. So with these two choices, we have exactly the definition of δij, for our orthonormal coordinate system with basis vectors ei:
As seen from the definition, the cross product is more complicated. Again, we consider two vectors, u=uiei and v=viei, in an orthonormal coordinate system, ei. The cross-product then becomes
u×v=(uiei)×(vjej)=uivjei×ej
Based on the definition of the cross product and the orthonormal coordinate system, we know the following in 3 dimensions
ei×ej={±ek0i=ji=jk=i,j
(In three dimensions, for i=j, only one choice remains for k=i,j) However, the direction of ±ek in the case that i=j is not properly defined. This depends on the order of our basis vectors. We therefore need to introduce this order in our definition of our coordinate system and we will choose a right-handed coordinate system (see the illustration used to define the cross product). For a coordinate system, this implies that e3=e1×e2. Using the Levi-Civita symbol, εijk, we then have
ei×ej=εijkek
in a right-handed orthonormal coordinate system. The cross product for two general vectors is then
u×v=uivjεijkek
Unless otherwise specified, coordinate systems discussed on this webpage are right-handed and orthonormal.