Point, Bi-point
A point r∈R2 is an ordered pair of real numbers, r=(x,y) with x∈R and y∈R. Here the first coordinate x stipulates the location on the horizontal axis and the second coordinate y stipulates the location on the vertical axis. Given two points r and r′ in R2 the directed line segment with departure point r and arrival point r′ is called the bi-point r, r′ and is denoted by [r,r′]. We say that r is the tail of the bi-point [r, r′] and that r′ is its head. The Euclidean length or norm of bi-point [a, b] is simply the distance between a and b and it is denoted by ∣∣[a,b]∣∣=(a1−b1)2+(a2−b2)2
Vector
A vector a∈R2 is a codification of movement of a bi-point: given the bi-point [r, r′], we associate to it the vector rr′=[x′−xy′−y] stipulating a movement of x′−x units from (x, y) in the horizontal axis and of y′−y units from the current position in the vertical axis.The zero vector 0=[00] indicates no movement in either direction.
Let u=0. Put Ru={λu:λ∈R} and let a∈R2, the affine line with direction vector u=[u1u2] and passing through a is the set of points on the plane.
a+Ru={(xy)∈R2:x=a1+tu1,y=a2+tu2,t∈R}
that is, the affine line is the Cartesian line with slope u1u2, Conversely, if y=mx+k is the equation of a Cartesian line, then
(xy)=[1m]t+(0k)
that is, every Cartesian line is also an affine line and one may take the vector [1m] as its direction vector.
Let x∈R2 and y∈R2. Their scalar product (dot product, inner product) is defined and denoted by x⋅y=x1y1+x2y2
Consider now two arbitrary vectors in R2,x,y , say. Under which conditions can we write an arbitrary vector v on the plane as a linear combination of x and y , that is, when can we find scalars a, b such that v=ax+by ?
v=ax+by⟺v1=ax1+by1,v2=ax2+by2⟺a=x1y2−x2y1v1y2−v2y1,b=x1y2−x2y1x1v2−x2v1
The above expressions for a and b make sense only if x1y2=x2y1.
So given two vectors in R2,x,y , an arbitrary vector v can be written as the linear combination v=ax+by,a∈R,b∈R if and only if x is not parallel to y.
Geometric Transformations in two dimensions
We now are interested in the following fundamental functions of sets on the plane: translations, scalings (stretching or shrinking) reflexions about the axes, and rotations about the origin.
Translate
A function Tv:R2→R2 is said to be a translation if it is of the form Tv(x)=x+v, where v is a fixed vector on the plane.A translation simply shifts an object on the plane rigidly (that is, it does not distort it shape or re-orient it), to a copy of itself a given amount of units from where it was.
Scale
A function Sa,b:R2→R2 is said to be a scaling if it is of the form Sa,b(r)=(axby),a,b∈R+
Reflexion
A function RH:R2→R2 is said to be a reflexion about the y-axis or horizontal reflexion if it is of the form RH(r)=(−xy)
A function RV:R2→R2 is said to be a reflexion about the x-axis or vertical reflexion if it is of the form RV(r)=(x−y)
A function RO:R2→R2 is said to be a reflexion about the orgin if it is of the form RO(r)=(x−−y)
Rotation
A function Rθ:R2→R2 is said to be a levogyrate rotation about the origin by the angle θ measured from the positive x-axis if Rθ(r)=(xcosθ−ysinθxsinθ+ycosθ)
linear transformation
A function L:R2→R2 is said to be a linear transformation from R2 to R2 if for all points a, b on the plane and every scalar λ, it is verified that L(a+b)=L(a)+L(b),L(λa)=λL(b)
affine transformation
A function A:R2→R2 is said to be an affine transformation from R2 to R2 if there exists a linear transformation L:R2→R2 and a fixed vector v∈R2 such that for all points x∈R2 it is verified that A(x)=L(x)+v
Let L:R2→R2 be a linear transformation. The matrix AL associated to L is the 2 × 2, (2rows, 2columns) array whose columns are in this order L((10)) and L((01))
Determinants in two dimensions
The determinant of the 2 × 2 matrix [abcd] is det[abcd]=ad−bc.
Consider now a simple quadrilateral with vertices r1=(x1,y1),r2=(x2,y2),r3=(x3,y3),r4=(x4,y4), listed in counterclockwise order. This quadrilateral is spanned by the vectors
Let [a;b]⊆R. A parametric curve representation r of a curve Γ is a function r:[a;b]→R , with
r(t)=(x(t)y(t))
and such that r([a;b])=Γ. r(a) is the initial point of the curve and r(b) its terminal point.
As the determinants in two dimensions corrensponding the area. SΔ=21det[xyx+Δxy+Δy]=21(xΔy−yΔx)
So for the parametric curves which can be seen as N(N→∞) points. So we have
Spc=21∮Γ(xdy−ydx)
Vectors in Space
The 3-dimensional Cartesian Space is defined and denoted by R3={r=(x,y,z):x∈R,y∈R,z∈R}
The dot product of two vectors a and b in R3 is a⋅b=a1b1+a2b2+a3b3
Let u and v be linearly independent vectors. The parametric equation of a plane containing the point a, and parallel to the vectors u and v is given by
r−a=pu+qv
in the other way, the equation of the plane in space can be written in the form ax+by+cz=d, which is the product of any v=(x+d1,y+d2,z+d3) and fixed p=(a,b,c) with condition ad1+bd2+cd3=0
Cross Product
We now define the standard cross product in R3 as a product satisfying the following properties.
Let x,y,z be vectors in R3 ,and let α∈R be a scalar. The cross product × is a closed binary operation satisfying
1: x⊥(x+y)andy⊥(x+y)
2: a×(b+c)=(a⋅c)b−(a⋅b)c
3: denote the convex angle between two vector x,y is θ, then ∣∣x×y∣∣=∣∣x∣∣∣y∣∣sinθ
corolary: Two non-zero vectors x,y , satisfy x×y=0 if and only if they are parallel.
Let a,b,c be linearly independent vectors in R3 . The signed volume of the parallelepiped spanned by them is (a×b)•c.
Matrices in three dimensions
We will briefly introduce 3 × 3 matrices. Most of the material will flow like that for 2 × 2 matrices.
A linear transformation T:R3→R3 is a function such that T(a+b)=T(a)+T(b),T(λa)=λT(a), for all points a, b in R3 and all scalars λ. Such a linear transformation has a 3×3 matrix representation whose columns are the vectors T(i),T(j),T(k).
Determinants in three dimensions
Since thanks to Theorem, the signed volume of the parallelepiped spanned by them is (a×b)•c, we define the determinant of A, det A, to be
Given a point (x, y, z) in Cartesian coordinates, its spherical coordinates are given by
x=ρcosθsinφ,y=ρsinθsinφ,z=ρcosφ
Here φ is the polar angle, measured from the positive z-axis, and θ is the azimuthal angle, measured from the positive x-axis. By convention, 0 ≤ θ ≤ 2π and 0 ≤ φ ≤ π.
Spherical coordinates are extremely useful when considering regions which are symmetric about a point.
Canonical Surfaces
In this section we consider various surfaces that we shall periodically encounter in subsequent sec- tions. Just like in one-variable Calculus it is important to identify the equation and the shape of a line, a parabola, a circle, etc., it will become important for us to be able to identify certain families of often-occurring surfaces. We shall explore both their Cartesian and their parametric form. We remark that in order to parametrise curves (“one-dimensional entities”) we needed one parameter, and that in order to parametrise surfaces we shall need to parameters.
Parametric Curves in Space
Let [a;b]⊆R. A parametric curve representation r of a curve Γ is a function r:[a;b]→R3
r(t)=⎝⎛x(t)y(t)z(t)⎠⎞
and such that r([a;b])=Γ. r(a) is the initial point of the curve and r(b) its terminal point. A curve is closed if its initial point and its final point coincide. The trace of the curve r is the set of all images of r, that is, Γ. The length of the curve is ∫Γ∣∣dr∣∣
Multidimensional Vectors
We briefly describe space in n-dimensions. The ideas expounded earlier about the plane and space carry almost without change.
Rn is the n-dimensional space, the collection Rn={⎝⎜⎜⎜⎛x1x2⋮xn⎠⎟⎟⎟⎞:xk∈R}
Given vectors a,b of Rn , their dot product is a⋅b=k=1∑nakbk
Cauchy-Bunyakovsky-Schwarz Inequality: Let x and y be any two vectors in Rn, then we have ∣x⋅y∣≤∣∣x∣∣y∣∣..
The form of the Cauchy-Bunyakovsky-Schwarz most useful to us will be ∣k=1∑nxkyk∣≤(k=1∑nxk2)21(k=1∑nyk2)21
Differentiation
Multivariable Functions
Let A⊆Rn . For most of this course, our concern will be functions of the form f:A→Rm
If m=1,we say that f is a scalar field. If m≥2,we say that f is a vector field.
Definition of the Derivative
Let A⊆Rn.A function f:A→Rm is said to be differentiable at a∈A if there is a linear transformation, called the derivative of f at a, Da(f):Rn→Rm such that
x→alim=∣∣x−a∣∣∣∣f(x)−f(a)−Da(f)(x−a)∣∣=0
The Jacobi Matrix
We now establish a way which simplifies the process of finding the derivative of a function at a given point.
Let A⊆Rn,f:A→Rm ,and put
To find partial derivatives with respect to the j-th variable, we simply keep the other variables fixed and differentiate with respect to the j-th variable.
Then each partial derivative ∂xj∂fi(x) exists, and the matrix representation of Dx(f) with respect to the standard bases of Rn and Rm is the Jacobi matrix
for eample, the function f(x,y)=(xy+yz,logexy), then jacobi matrix is f′(x,y)=[yx1x+zy1y0]
Gradients and Directional Derivatives
A function f:x∈Rn→f(x)∈Rm is called a vector field.
If m=1, it is called a scalar field.
Definition
Let f:Rnx→→Rf(x) be a scalar field.
The gradient of f is the vector defined and denoted by ∇f(x)=⎣⎢⎢⎢⎢⎡∂x1∂f(x)∂x2∂f(x)⋮∂xn∂f(x)⎦⎥⎥⎥⎥⎤
The gradient operator is the operator ∇=⎣⎢⎢⎢⎡∂x1∂∂x2∂⋮∂xn∂⎦⎥⎥⎥⎤
Definition
Let f:Rnx→→Rnf(x) be a vector field with f(x)=⎣⎢⎢⎢⎡f1(x)f2(x)⋮fn(x)⎦⎥⎥⎥⎤
The divergence of f is defined and denoted by divf(x)=∇⋅f(x)=∂x1∂f1(x)+∂x2∂f2(x)+⋯+∂xn∂fn(x)
Extrema (and Hessian matrix)
We now turn to the problem of finding maxima and minima for vector functions. As in the one-variable case, the derivative will provide us with information about the extrema, and the “second derivative” will provide us with information about the nature of these extreme points.
To define an analogue for the second derivative, let us consider the following. Let A⊂Rn and f:A→Rm be differentiable on A. We know that for fixed x0∈A, Dx0(f) is a linear transformation from Rn to Rm. This means that we have a function
T:Ax→→L(Rn,Rm)Dx(f)
where L(Rn,Rm) denotes the space of linear transformations from Rn to Rm. Hence, if we differentiate T at x0 again, we obtain a linear transformation Dx0(T)=Dx0(Dx0(f))=Dx02(f) from Rn to L(Rn,Rm). Hence, given x1∈Rn, we have Dx02(f)(x1)∈L(Rn,Rm). Again, this means that given $x_2 ∈ \R^n, D_{x_0}^2(f)(x_1))(x_2) ∈ \R^m $. Thus the function
is well defined, and linear in each variable x1 and x2, that is, it is a bilinear function.
Theorem:
Let A⊆Rn be an open set, and f:A→R be twice differentiable on A. Then the matrix of Dx2(f):Rn×Rn→R with respect to the standard basis is given by the Hessian matrix: