See Last Minute Notes on all subjects here.
A matrix represents a collection of numbers arranged in an order of rows and columns. It is necessary to enclose the elements of a matrix in parentheses or brackets.
A matrix with 9 elements is shown below.
This Matrix [M] has 3 rows and 3 columns. Each element of matrix [M] can be referred to by its row and column number. For example, a23=6
Order of a Matrix :
The order of a matrix is defined in terms of its number of rows and columns.
Order of a matrix = No. of rows ×No. of columns
Therefore Matrix [M] is a matrix of order 3 × 3.
Transpose of a Matrix :
The transpose [M]T of an m x n matrix [M] is the n x m matrix obtained by interchanging the rows and columns of [M].
if A= [aij] mxn , then AT = [bij] nxm where bij = aji
Properties of transpose of a matrix:
- (AT)T = A
- (A+B)T = AT + BT
- (AB)T = BTAT
Singular and Nonsingular Matrix:
- Singular Matrix: A square matrix is said to be singular matrix if its determinant is zero i.e. |A|=0
- Nonsingular Matrix: A square matrix is said to be non-singular matrix if its determinant is non-zero.
Square Matrix: A square Matrix has as many rows as it has columns. i.e. no of rows = no of columns.
Symmetric matrix: A square matrix is said to be symmetric if the transpose of original matrix is equal to its original matrix. i.e. (AT) = A.
Skew-symmetric: A skew-symmetric (or antisymmetric or antimetric) matrix is a square matrix whose transpose equals its negative.i.e. (AT) = -A.
Diagonal Matrix:A diagonal matrix is a matrix in which the entries outside the main diagonal are all zero. The term usually refers to square matrices.
Identity Matrix:A square matrix in which all the elements of the principal diagonal are ones and all other elements are zeros.Identity matrix is denoted as I.
Orthogonal Matrix: A matrix is said to be orthogonal if AAT = ATA = I
Idemponent Matrix: A matrix is said to be idemponent if A2 = A
Involutary Matrix: A matrix is said to be Involutary if A2 = I.
Adjoint of a square matrix:
Properties of Adjoint:
- A(Adj A) = (Adj A) A = |A| In
- Adj(AB) = (Adj B).(Adj A)
- |Adj A|= |A|n-1
- Adj(kA) = kn-1 Adj(A)
Inverse of a square matrix:
Here |A| should not be equal to zero, means matrix A should be non-singular.
Properties of inverse:
1. (A-1)-1 = A
2. (AB)-1 = B-1A-1
3. Only a non-singular square matrix can have an inverse.
Trace of a matrix :
Let A=[aij] nxn is a square matrix of order n, then the sum of diagonal elements is called the trace of a matrix which is denoted by tr(A). tr(A) = a11 + a22 + a33+ ……….+ ann. Remember trace of a matrix is also equal to the sum of eigen value of the matrix. For example:
Properties of trace of matrix:
Let A and B be any two square matrices of order n, then
- tr(kA) = k tr(A) where k is a scalar.
- tr(A+B) = tr(A)+tr(B)
- tr(A-B) = tr(A)-tr(B)
- tr(AB) = tr(BA)
Solution of a system of linear equations:
Linear equations can have three kind of possible solutions:
- No Solution
- Unique Solution
- Infinite Solution
Rank of a matrix: Rank of matrix is the number of non-zero rows in the row reduced form or the maximum number of independent rows or the maximum number of independent columns.
Let A be any mxn matrix and it has square sub-matrices of different orders. A matrix is said to be of rank r, if it satisfies the following properties:
- It has at least one square sub-matrices of order r who has non-zero determinant.
- All the determinants of square sub-matrices of order (r+1) or higher than r are zero.
Rank is denoted as P(A).
if A is a non-singular matrix of order n, then rank of A = n i.e. P(A) = n.
Properties of rank of a matrix:
- If A is a null matrix then P(A) = 0 i.e. Rank of null matrix is zero.
- If In is the nxn unit matrix then P(A) = n.
- Rank of a matrix A mxn , P(A) ≤ min(m,n). Thus P(A) ≤m and P(A) ≤ n.
- P(A nxn ) = n if |A| ≠ 0
- If P(A) = m and P(B)=n then P(AB) ≤ min(m,n).
- If A and B are square matrices of order n then P(AB) ? P(A) + P(B) – n.
- If Am×1 is a non zero column matrix and B1×n is a non zero row matrix then P(AB) = 1.
- The rank of a skew symmetric matrix cannot be equal to one.
System of homogeneous linear equations AX = 0.
- X = 0. is always a solution; means all the unknowns has same value as zero. (This is also called trivial solution)
- If P(A) = number of unknowns, unique solution.
- If P(A) < number of unknowns, infinite number of solutions.
System of non-homogenous linear equations AX = B.
- If P[A:B] ≠P(A), No solution.
- If P[A:B] = P(A) = the number of unknown variables, unique solution.
- If P[A:B] = P(A) ≠ number of unknown, infinite number of solutions.
Here P[A:B] is rank of gauss elimination representation of AX = B.
There are two states of the Linear equation system:
- Consistent State: A System of equations having one or more solutions is called a consistent system of equations.
- Inconsistent State: A System of equations having no solutions is called inconsistent system of equations.
Linear dependence and Linear independence of vector:
Linear Dependence: A set of vectors X1 ,X2 ….Xr is said to be linearly dependent if there exist r scalars k1 ,k2 …..kr such that: k1 X1 + k2X2 +……..kr Xr = 0.
Linear Independence: A set of vectors X1 ,X2….Xr is said to be linearly independent if for all r scalars k1,k2 …..krsuch that k1X1+ k2 X2+……..krXr = 0, then k1 = k2 =……. = kr = 0.
How to determine linear dependency and independency ?
Let X1, X2 ….Xr be the given vectors. Construct a matrix with the given vectors as its rows.
- If the rank of the matrix of the given vectors is less than the number of vectors, then the vectors are linearly dependent.
- If the rank of the matrix of the given vectors is equal to the number of vectors, then the vectors are linearly independent.
Eigen vector of a matrix A is a vector represented by a matrix X such that when X is multiplied with matrix A, then the direction of the resultant matrix remains the same as vector X.
Mathematically, above statement can be represented as:
AX = λX
where A is any arbitrary matrix, λ are eigen values and X is an eigen vector corresponding to each eigen value.
Here, we can see that AX is parallel to X. So, X is an eigen vector.
Method to find eigen vectors and eigen values of any square matrix A
We know that,
AX = λX
=> AX – λX = 0
=> (A – λI) X = 0 …..(1)
Above condition will be true only if (A – λI) is singular. That means,
|A – λI| = 0 …..(2)
(2) is known as characteristic equation of the matrix.
The roots of the characteristic equation are the eigen values of the matrix A.
Now, to find the eigen vectors, we simply put each eigen value into (1) and solve it by Gaussian elimination, that is, convert the augmented matrix (A – λI) = 0 to row echelon form and solve the linear system of equations thus obtained.
Some important properties of eigen values
Eigen values of real symmetric and hermitian matrices are real
Eigen values of real skew symmetric and skew hermitian matrices are either pure imaginary or zero
Eigen values of unitary and orthogonal matrices are of unit modulus |λ| = 1
If λ1, =λ2…….λn are the eigen values of A, then kλ1, kλ2…….kλn are eigen values of kA
If λ1, λ2…….λn are the eigen values of A, then 1/λ1, 1/λ2…….1/λn are eigen values of A-1
If λ1, λ2…….λn are the eigen values of A, then λ1k, λ2k…….λnk are eigen values of Ak
Eigen values of A = Eigen Values of AT (Transpose)
Sum of Eigen Values = Trace of A (Sum of diagonal elements of A)
Product of Eigen Values = |A|
Maximum number of distinct eigen values of A = Size of A
If A and B are two matrices of same order then, Eigen values of AB = Eigen values of BA
Probability refers to the extent of occurrence of events. When an event occurs like throwing a ball, picking a card from deck, etc ., then the must be some probability associated with that event.
- Random Event :- If the repetition of an experiment occurs several times under similar conditions, if it does not produce the same outcome everytime but the outcome in a trial is one of the several possible outcomes, then such an experiment is called random event or a probabilistic event.
- Elementary Event – The elementary event refers to the outcome of each random event performed. Whenever the random event is performed, each associated outcome is known as elementary event.
- Sample Space – Sample Space refers tho the set of all possible outcomes of a random event.Example, when a coin is tossed, the possible outcomes are head and tail.
- Event – An event refers to the subset of the sample space associated with a random event.
- Occurrence of an Event – An event associated with a random event is said to occur if any one of the elementary event belonging to it is an outcome.
- Sure Event – An event associated with a random event is said to be sure event if it always occurs whenever the random event is performed.
- Impossible Event – An event associated with a random event is said to be impossible event if it never occurs whenever the random event is performed.
- Compound Event – An event associated with a random event is said to be compound event if it is the disjoint union of two or more elementary events.
- Mutually Exclusive Events – Two or more events associated with a random event are said to be mutually exclusive events if any one of the event occurrs, it prevents the occurrence of all other events.This means that no two or more events can occur simultaneously at the same time.
- Exhaustive Events – Two or more events associated with a random event are said to be exhaustive events if their union is the sample space.
Probability of an Event – If there are total p possible outcomes associated with a random experiment and q of them are favourable outcomes to the event A, then the probability of event A is denoted by P(A) and is given by
P(A) = q/p
The probability of non occurrence of event A, i.e, P(A’) = 1 – P(A)
- If the value of P(A) = 1, then event A is called sure event .
- If the value of P(A) = 0, then event A is called impossible event.
- Also, P(A) + P(A’) = 1
- General – Let A, B, C are the events associated with a random experiment, then
- P(A∪B) = P(A) + P(B) – P(A∩B)
- P(A∪B) = P(A) + P(B) if A and B are mutually exclusive
- P(A∪B∪C) = P(A) + P(B) + P(C) – P(A∩B) – P(B∩C)- P(C∩A) + P(A∩B∩C)
- P(A∩B’) = P(A) – P(A∩B)
- P(A’∩B) = P(B) – P(A∩B)
- Extension of Multiplication Theorem – Let A1, A2, ….., An are n events associated with a random experiment, then
P(A1∩A2∩A3 ….. An) = P(A1)P(A2/A1)P(A3/A2∩A1) ….. P(An/A1∩A2∩A3∩ ….. ∩An-1)
Total Law of Probability – Let S be the sample space associated with a random experiment and E1, E2, …, En be n mutually exclusive and exhaustive events associated with the random experiment . If A is any event which occurs with E1 or E2 or … or En, then
P(A) = P(E1)P(A/E1) + P(E2)P(A/E2) + ... + P(En)P(A/En)
Conditional probability P(A | B) indicates the probability of event ‘A’ happening given that event B happened.
Derived from above definition of conditional probability by multiplying both sides with P(B)
P(A ∩ B) = P(B) * P(A|B)
A random variable is basically a function which maps from the set of sample space to set of real numbers. The purpose is to get an idea about the result of a particular situation where we are given probabilities of different outcomes.
Discrete Probability Distribution – If the probabilities are defined on a discrete random variable, one which can only take a discrete set of values, then the distribution is said to be a discrete probability distribution.
Continuous Probability Distribution – If the probabilities are defined on a continuous random variable, one which can take any value between two numbers, then the distribution is said to be a continuous probability distribution.
Cumulative Distribution Function –
Similar to the probability density function, the cumulative distribution function of a real-valued random variable X, or just distribution function of evaluated at , is the probability that will take a value less than or equal to .
For a discrete Random Variable,
For a continuous Random Variable,
The Uniform Distribution, also known as the Rectangular Distribution, is a type of Continuous Probability Distribution.
It has a Continuous Random Variable restricted to a finite interval and it’s probability function has a constant density over this interval.
The Uniform probability distribution function is defined as-
Expectation: The mean of the distribution, represented as E[x].
For uniform distribution,
For a positive real number the probability density function of a Exponentially distributed Random variable is given by-
where Rx is exponential random variables.
Existence of Limit – The limit of a function at exists only when its left hand limit and right hand limit exist and are equal i.e.
Some Common Limits –
L’Hospital Rule –
If the given limit is of the form or i.e. both and are 0 or both and are , then the limit can be solved by L’Hospital Rule.
If the limit is of the form described above, then the L’Hospital Rule says that –
where and obtained by differentiating and .
If after differentitating, the form still exists, then the rule can be applied continuously until the form is changed.
A function is said to be continuous over a range if it’s graph is a single unbroken curve.
A real valued function is said to be continuous at a point in the domain if –
exists and is equal to .
If a function is continuous at then-
Functions that are not continuous are said to be discontinuous.
The derivative of a real valued function wrt is the function and is defined as –
A function is said to be differentiable if the derivative of the function exists at all points of its domain. For checking the differentiability of a function at point ,
If a function is differentiable at a point, then it is also continuous at that point.
Note – If a function is continuous at a point does not imply that the function is also differentiable at that point. For example, is continuous at but it is not differentiable at that point.
Suppose be a function satisfying three conditions:
1) f(x) is continuous in the closed interval a ≤ x ≤ b
2) f(x) is differentiable in the open interval a < x < b
Then according to Lagrange’s Theorem, there exists at least one point ‘c’ in the open interval (a, b) such that:
Suppose f(x) be a function satisfying three conditions:
1) f(x) is continuous in the closed interval a ≤ x ≤ b
2) f(x) is differentiable in the open interval a < x < b
3) f(a) = f(b)
Then according to Rolle’s Theorem, there exists at least one point ‘c’ in the open interval (a, b) such that:
f ‘ (c) = 0
- Defination :Let f(x) be a function. Then the family of all ist antiderivatives is called the indefinite integral of a function f(x) and it is denoted by ∫f(x)dx.
The symbol ∫f(x)dx is read as the indefinite integral of f(x) with respect to x.
Thus ∫f(x)dx= ∅(x) + C.
Thus, the process of finding the indefinite integral of a function is called integration of the function.
Fundamental Integration Formulas –
- ∫xndx = (xn+1/(n+1))+C
- ∫(1/x)dx = (loge|x|)+C
- ∫exdx = (ex)+C
- ∫axdx = ((ax)/(logea))+C
- ∫sin(x)dx = -cos(x)+C
- ∫cos(x)dx = sin(x)+C
- ∫sec2(x)dx = tan(x)+C
- ∫cosec2(x)dx = -cot(x)+C
- ∫sec(x)tan(x)dx = sec(x)+C
- ∫cosec(x)cot(x)dx = -cosec(x)+C
- ∫cot(x)dx = log|sin(x)|+C
- ∫tan(x)dx = log|sec(x)|+C
- ∫sec(x)dx = log|sec(x)+tan(x)|+C
- ∫cosec(x)dx = log|cosec(x)-cot(x)|+C
Definite integrals are the extension after indefinite integrals, definite integrals have limits [a, b]. It gives the area of a curve bounded between given limits.
, It denotes the area of curve F(x) bounded between a and b, where a is the lower limit and b is the upper limit.
Note: If f is a continuous function defined on the closed interval [a, b] and F be an anti derivative of f. Then
Here, the function f needs to be well defined and continuous in [a, b].
Properties of definite integrals –
- Last Minute Notes - Computer Networks
- Mathematics | Set Operations (Set theory)
- Mathematics | The Pigeonhole Principle
- Mathematics | Matrix Introduction
- Mathematics | Introduction to Propositional Logic | Set 1
- Mathematics | Propositional Equivalences
- Mathematics | Predicates and Quantifiers | Set 1
- Mathematics | L U Decomposition of a System of Linear Equations
- Mathematics | Introduction of Set theory
- Mathematics | Mean, Variance and Standard Deviation
- Mathematics | Conditional Probability
- Mathematics | Sum of squares of even and odd natural numbers
- Mathematics | Eigen Values and Eigen Vectors
- Mathematics | Rolle's Mean Value Theorem
- Mathematics | Lagrange's Mean Value Theorem