Author: Dr. K. GURURAJAN: Class Notes of Engineering Mathematics Iv Subject Code: 06mat41
Author: Dr. K. GURURAJAN: Class Notes of Engineering Mathematics Iv Subject Code: 06mat41
Author: Dr. K. GURURAJAN: Class Notes of Engineering Mathematics Iv Subject Code: 06mat41
Another example is consider the experiment of finding how far this e – learning program
started by VTU yielding results. Usually the authorities visit the colleges and collect the
information about the programme from a sample of students. Later study will be done,
based on this sample study, suitable actions will be taken. Like this, one can give a
number of examples. With these in view, this chapter focuses a detailed discussion of
statistical inference.
This topic considers two different classes of problems
1. Hypothesis testing – we test a statement about the population parameter from which the
sample is drawn.
2. Estimation – A statistic obtained from the sample collected is used to estimate the
population parameter.
As we are studying population parameter based on some sample study, one can not do the
job with 100% accuracy since sample is drawn from the population and possible sample
may not represent the whole population. Therefore, usually we conduct analysis at
certain level of significance (lower than 100%. The possible choices include 99%, or
95% or 98% or 90%. Usually we conduct analysis at 99% or 95% level of
significance. The confidence with which a person rejects or accepts H0 depends upon
1 α = 1% 2.58
2 α = 2% 2.33
3 α = 5% 1.96
Critical values or Fiducial limit values for a single tailed test (right and test)
Setting a test criterion: The third step in hypothesis testing procedure is to construct a
test criterion. This involves selecting an appropriate probability distribution for the
particular test i.e. a proper probability distribution function to be chosen. Some of the
distribution functions used are t, F, when the sample size is small (size lower than 30).
However, for large samples, normal distribution function is preferred. Next step is the
computation of statistic using the sample items drawn from the population. Usually,
samples are drawn from the population by a procedure called random, where in each and
every data of the population has the same chance of being included in the sample. Then
the computed value of the test criterion is compared with the tabular value; as long the
calculated value is lower then or equal to tabulated value, we accept the null hypothesis,
otherwise, we reject null hypothesis and accept the alternate hypothesis. Decisions are
valid only at the particular level significance of level adopted.
During the course of analysis, there are two types of errors bound to occur. These are (i)
Type – I error and (ii) Type – II error.
(Dr. K. Gururajan, MCE, Hassan Page 3 - 3/5/2011)
Type – I error: This error usually occurs in a situation, when the null hypothesis is true,
but we reject it i.e. rejection of a correct/true hypothesis constitute type I error.
Type – II error: Here, null hypothesis is actually false, but we accept it. Equivalently,
accepting a hypothesis which is wrong results in a type – II error. The probability of
committing a type – I error is denoted by α where
α = Probability of making type I error = Probability [Rejecting H0 | H0 is true]
On the other hand, type – II error is committed by not rejecting a hypothesis when it is
false. The probability of committing this error is denoted by β . Note that
β = Probability of making type II error = Probability [Accepting H1 | H1 is false]
Critical region:
A region in a sample space S which amounts to Rejection of H0 is termed as critical
region.
One tailed test and two tailed test:
This depends upon the setting up of both null and alternative hypothesis.
A note on computed test criterion value:
biased one. Set up H1 : Yes, the coin is biased. As the coin is assumed be fair and it is
tossed 400 times, clearly we must expect 200 times heads occurring and 200 times tails.
Thus, expected number of heads is 200. But the observed result is 216. There is a
2500 - 2560
2500. Hence, zcal = = 1.7 which is lower than 1.96. Hence, we conclude
35.36
that die is a fair one.
(Dr. K. Gururajan, MCE, Hassan Page 5 - 3/5/2011)
3. A sample of 1000 days is taken from meteorological records of a certain district
and 120 of them are found to be foggy. What are the probable limits to the
percentage of foggy days in the district?
Solution: Let p denote the probability that a day is foggy in nature in a district as
120
reported by meteorological records. Clearly, p = = 0.12 and q = 0.88. With n =
1000
1000, the probable limits to the percentage of foggy days is given by p ± 3 pqn .
Using the data available in this problem, one obtains the answer as
4. A die was thrown 9000 times and a throw of 5 or 6 was obtained 3240 times. On
the assumption of random throwing, do the data indicate that die is biased? (Model
Question Paper Problem)
Solution: We set up the null hypothesis as H0 : Die is un - biased. Also,
1
×9000 = 3000 . Now the difference in these two results is 240. With p = 1/3, q = 2/3,
3
Difference
n = 9000, S.E . = npq = 44.72. Now consider the test criterion zcal = =
S.E.
240
= = 5.367 which is again more than the tabulated value. Therefore, we reject null
44.72
hypothesis and accept the alternate that die is highly biased.
(Dr. K. Gururajan, MCE, Hassan Page 6 - 3/5/2011)
Tests of significance for large samples:
In the previous section, we discussed problems pertaining to sampling of attributes. It is
time to think of sampling of other variables one may come across in a practical situation
such as height weight etc. We say that a sample is small when the size is usually lower
than 30, otherwise it is called a large one.
The study here is based on the following assumptions: (i) the random sampling
distribution of a statistic is approximately normal and (ii) values given by the samples are
sufficiently close to the population value and can be used in its place for calculating
σp
standard error. When the standard deviation of population is known, then S.E (X) =
n
where σ p denotes the standard deviation of population . When the standard deviation of
σ
the population is unknown, then S.E (X) = where σ is the standard deviation of
n
the sample.
Fiducial limits of population mean are:
σ
95% fiducial limits of population mean are X ± 1.96
n
σ .
99% fiducial limits of population mean are X ± 2.58
n
x -µ
Further, test criterion z cal =
S.E.
LLUSTRATIVE EXAMPLES
1. A sample of 100 tyres is taken from a lot. The mean life of tyres is found to be 39,
350 kilo meters with a standard deviation of 3, 260. Could the sample come from a
population with mean life of 40, 000 kilometers? Establish 99% confidence limits
within which the mean life of tyres is expected to lie.
Solution: First we shall set up null hypothesis, H0 : µ = 40, 000 , alternate hypothesis as
H1 : µ ≠ 40, 000 . We consider that the problem follows a two tailed test and
x -µ
expression for finding test criterion, zcal = . Here, µ =40, 000, x = 39, 350 and
S.E.
σ 3,260
σ = 3, 260 , n = 100. S.E. = = = 326 . Thus, z cal = 1.994. As this value is
n 100
slightly greater than 1.96, we reject the null hypothesis and conclude that sample has not
come from a population of 40, 000 kilometers.
The 99% confidence limits within which population mean is expected to lie is given as
as H1 : µ > 1, 600 hours . We consider that the problem follows a two tailed test and chose
α = 5% . Then corresponding to this, tabulated value is 1.96. Consider the expression
x -µ
for finding test criterion, zcal = . Here, µ =1, 600, x = 1, 570, n = 400 ,
S.E.
σ = 150 hours so that using all these values above, it can be seen that z cal = 4.0 which is
really greater than 1.96. Hence, we have to reject null hypothesis and to accept the
alternate hypothesis.
Test of significance of difference between the means of two samples
Consider two populations P1 and P2. Let S1 and S2 be two samples drawn at random
from these two different populations. Suppose we have the following data about these
two samples, say
S1 n1 x1 σ1
S2 n2 x2 σ2
then standard error of difference between the means of two samples S1 and S2
σ 12 σ 22 Difference of sample means
is S.E = + and the test criterion is Zcal = . The
n1 n2 Standard error
rest of the analysis is same as in the preceding sections.
When the two samples are drawn from the same population, then standard error is
When the standard deviations are un – known, then standard deviations of the two
s12 s22
samples must be replaced. Thus, S.E = + where s1 and s2 are standard
n1 n2
deviations of the two samples considered in the problem.
ILLUSTRATIVE EXAMPLES
1. Intelligence test on two groups of boys and girls gave the following data:
Girls 70 20 250
Is there a significant difference in the mean scores obtained by boys and girls?
2. A man buys 50 electric bulbs of “Philips” and 50 bulbs of “Surya”. He finds that
Philips bulbs give an average life of 1,500 hours with a standard deviation of 60
hours and Surya bulbs gave an average life of 1, 512 hours with a standard
deviation of 80 hours. Is there a significant difference in the mean life of the two
makes of bulbs?
relatively small. Also, we do see a number of problems involving small samples. With
these in view, here, we will initiate a detailed discussion on the same.
Here, too, the problem is about testing a statement about population parameter; i.e. in
ascertaining whether observed values could have arisen by sampling fluctuations from
some value given in advance. For example, if a sample of 15 gives a correlation
coefficient of +0.4, we shall be interested not so much in the value of the correlation in
the parent population, but more generally this value could have come from an un –
correlated population, i.e. whether it is significant in the parent population. It is widely
accepted that when we work with small samples, estimates will vary from sample to
sample.
Further, in the theory of small samples also, we begin study by making an assumption
that parent population is normally distributed unless otherwise stated. Strictly, whatever
the decision one takes in hypothesis testing problems is valid only for normal
populations.
Sir William Gosset and R. A. Fisher have contributed a lot to theory of small samples.
Sir W. Gosset published his findings in the year 1905 under the pen name “student”. He
gave a test popularly known as “t – test” and Fisher gave another test known as “z – test”.
These tests are based on “t distribution and “z – distribution”.
Gosset was employed by the Guinness and Son, Dublin bravery, Ireland which did not
permit employees to publish research work under their own names. So Gosset adopted
the pen name “student” and published his findings under this name. Thereafter, the t –
distribution commonly called student’s t – distribution or simply student’s distribution.
The t – distribution to be used in a situation when the sample drawn from a population is
of size lower than 30 and population standard deviation is un – known. The t – statistic,
i =n
∑( x )
2
x−µ −x
t cal is defined as t cal = ⋅ n where i
, x is the
γ = n − 1 = 12 S S= i =1
n −1
sample mean, n is the sample size, and x i are the data items.
The t – distribution function has been derived mathematically under the assumption of a
normally distributed population; it has the following form
The application of t – distribution includes (i) testing the significance of the mean of a
random sample i.e. determining whether the mean of a sample drawn from drawn from a
normal population deviates significantly from a stated value (i.e. hypothetical value of
the populations mean) and (ii) testing whether difference between means of two
independent samples is significant or not i.e. ascertaining whether the two samples comes
from the same normal population? (iii) Testing difference between means of two
dependent samples is significant? (iv) Testing the significance of on observed correlation
coefficient.
• Note down the sample size, n and the number of degrees of freedom,
1. The manufacturer of a certain make of electric bulbs claims that his bulbs have a
mean life of 25 months with a standard deviation of 5 months. Random samples of 6
such bulbs have the following values: Life of bulbs in months: 24, 20, 30, 20, 20, and
18. Can you regard the producer’s claim to valid at 1% level of significance? (Given
that t tab = 4.032 corresponding to γ = 5 ).
Solution: To solve the problem, we first set up the null hypothesis H0 : µ = 25 months ,
alternate hypothesis may be treated as H0 : µ < 25 months . To set up α = 1% , then
tabulated value corresponding to this level of significance is t tab |α =1% and γ =5 = 4.032
(4.032 value has been got by looking in the 5th row ) . The test criterion is given by
i =n
∑( x )
2
x−µ −x
t cal = ⋅ n where i
. Consider
S S= i =1
n −1
xi
(x )
2
x xi − x −x
i
24 1 1
26 3 9
30 7 49
20 23 -3 9
20 -3 9
18 -5 25
Solution: We shall set up H0 : µbefore = µafter i.e. there is no significant difference in the
blood pressure readings before and after the injection of the drug. The alternate
hypothesis is H0 : µbefore > µafter i.e. the stimulus resulted in an increase in the blood
pressure of the patients. Taking α = 1% and α = 5% , as n = 13, γ = n − 1 = 12 ,
respective tabulated values are t tab |α =1% and γ =12 = 3.055 and t tab |α =5% and γ =12 = 2.179 . Now,
we compute the value of test criterion. For this, consider
xi
(x )
2
x xi − x −x
i
5 2 4
2 -1 1
8 5 25
-1 -4 16
3 0 0
0 -3 9
-2 -5 25
1 -2 4
3
5 2 4
0 -3 9
4 1 1
6 3 9
8 5 25
∑( x )
2
−x 132 x−µ
Consider i
= = 11 = 3.317 . Therefore, t cal = ⋅ n may
S= i =1
12 S
n −1
0−3
be obtained as t cal = 13 = 3.2614 . As the calculated value 3.2614 is more than
3.317
the tabulated values of 3.055 and 2.179, we accept the alternate hypothesis that after the
drug is given to patients, there is an increase in the blood pressure level.
3. the life time of electric bulbs for a random sample of 10 from a large consignment
gave the following data: 4.2, 4.6, 3.9, 4.1, 5.2, 3.8, 3.9, 4.3, 4.4, 5.6 (in ‘000 hours).
Can we accept the hypothesis that the average life time of bulbs is 4, 000 hours?
Solution: Set up H0 : µ = 4, 000 hours , H1 : µ < 4, 000 hours . Let us choose that
α = 5% . Then tabulated value is t tab |α =5% and γ =9 = 2.262 . To find the test criterion,
consider
xi
(x )
2
x xi − x −x
i
∑( x )
2
−x 3.12 x−µ
Consider i
= = 0.589 . Therefore, t cal = ⋅ n is computed
S= i =1
9 S
n −1
4.4 − 4.0
as t cal = ⋅ 10 = 2.148. As the computed value is lower than the tabulated
0.589
value of 2.262, we conclude that mean life of time bulbs is about 4, 000 hours.
Recently, we have discussed t – distribution function (i.e. t – test). The study was based
on the assumption that the samples were drawn from normally distributed populations, or,
more accurately that the sample means were normally distributed. Since test required
such an assumption about population parameters. For this reason, A test of this kind is
called parametric test. There are situations in which it may not be possible to make any
rigid assumption about the distribution of population from which one has to draw a
sample.
Thus, there is a need to develop some non – parametric tests which does not require any
assumptions about the population parameters.
With this in view, now we shall consider a discussion on χ 2 distribution which does
not require any assumption with regard to the population. The test criterion
∑ ( Oi − Ei )
2
When Expected values are not given, one can calculate these by using the following
RT ⋅ CT
relation; E i : . Here, RT means the row total for the cell containing the row, CT
N
is for the column total for the cell containing columns, and N represent the total number
of observations in the problem.
The calculated χ2 value (i.e. test criterion value or calculated value) is compared
with the tabular value of χ 2 value for given degree of freedom at a certain prefixed
level of significance. Whenever the calculated value is lower than the tabular value, we
continue to accept the fact that there is not much significant difference between expected
and observed results.
On the other hand, if the calculated value is found to be more than the value suggested in
the table, then we have to conclude that there is a significant difference between observed
and expected frequencies.
This is an approximate test for relatively a large population. For the usage of test, the
following conditions must checked before employing the test. These are:
ILLUSTRATIVE EXAMPLES
1. From the data given below about the treatment of 250 patients suffering from a
disease, state whether new treatment is superior to the conventional test.
Conventional 60 20 80
Total 200 50 280
Solution: We set up null hypothesis as there is no significance in results due to the two
procedures adopted. The alternate hypothesis may be assumed as there could be some
difference in the results. Set up level of significance as
( 112 − 100 ) ( 71 − 50 ) ( 32 − 10 )
2 2 2
136 34 170
A 64 16 80
200 50 250
Oi Ei Oi − Ei ( Oi − Ei ) ( Oi − Ei )
2 2
Ei
140 136 4 16 0.118
60 64 -4 16 0.250
30 34 -4 16 0.471
20 16 4 16 1.000
Total 1.839
As the calculated value 1.839 is lower than the tabulated value χ |α = 0.05,γ =1 = 3.841 , we
2
accept the null hypothesis, namely, that there is not much significant difference between
the two procedures.
No. of heads 0 1 2 3 4 5
Frequency 6 27 72 112 71 32
Test the hypothesis that the data follow a binomial distribution function.
Solution: We shall set up the null hypothesis that data actually follows a binomial
distribution. Then alternate hypothesis is, namely, data does not follow binomial
distribution. Next, to set up a suitable level of significance, α = 5% , with n = 6, degrees
of freedom is γ = 5. Therefore, the tabulated value is χ |α = 0.05,γ =5 = 11.07 . Before
2
proceeding to finding test criterion, first we compute the various expected frequencies.
As the data is set to be following binomial distribution, clearly probability density
n k n −k
function is b ( n, p, k ) = p q . Here, n = 320, p = 0.5, q = 0.5 , and k
k
takes the values right from 0 up to 5. Hence, the expected frequencies of getting 0, 1, 2,
3, 4, 5 heads are the successive terms of the binomial expansion of 320 ⋅ ( p + q ) . Thus,
5
expected frequencies E i are 10, 50, 100, 100, 50, 10. Consider the test criterion given
∑( O − Ei )
2
i
by χ 2 | = i ;
cal
Ei
Here, observed values are: Oi : 6, 27, 72, 112, 71, 32
The expected values are: E i : 10, 50, 100, 100, 50, 10 . Consider
( 6 − 10 ) 2 ( 27 − 50 ) 2 ( 72 − 100 ) 2
χ |cal =
2
+ +
10 50 100
( 112 − 100 ) 2 ( 71 − 50 ) 2 ( 32 − 10 ) 2
+ + + = 78.68. As the calculated
100 50 10
value is very much higher than the tabulated value of 3.841, we reject the null hypothesis
and accept the alternate hypothesis that data does not follow the binomial distribution.
(Dr. K. Gururajan, MCE, Hassan - page 19, 23 – 05 – 2008)
Subject: Engineering Mathematics IV
Subject Code: 06MAT41
Author: Dr. K. Gururajan
Assistant Professor
Department of Mathematics
Malnad College of Engineering, Hassan – 573 201
UNIT 6:
Introduction: The previous chapter was devoted to a detailed study of basic probability
concepts. The following were taught to you:
• Random Experiment, Sample Space.
• Events in a Sample Space
• Definition of probability, Axioms
• Addition Law, Multiplication Rule
• Independent and Exclusive Events
• Bayes Rule
• Mainly computation of probability of events
From these analyses, it is clear that calculation of probability of events to be treated
individually. However, in a practical situation, one may be interested in finding the
probabilities of all the events and may wishes to have the results in a tabular form for any
future reference. Since for an experiment having n outcomes, totally, there are 2n totally
events; finding probabilities of each of these and keeping them in a tabular form may be
an interesting problem.
Let S denote the sample space of a random experiment. A random variable means it is a
rule which assigns a numerical value to each and every outcome of the experiment.
Thus, random variable may be viewed as a function from the sample space S to the set of
all real numbers; denoted as f :S →R. For example, consider the random
experiment of tossing three fair coins up. Then S = {HHH, HHT, HTH, THH, TTH,
THT, HTT, TTT}. Define f as the number of heads that appear. Hence, f ( HHH ) = 3 ,
f ( HHT ) = 2 , f ( HTH ) = 2 , f (THH ) = 2 , f ( HTT ) = 1 , f (THT ) = 1 , f (TTH ) = 1
and f (TTT ) = 0 . The same can be explained by means of a table as given below:
Note that all the outcomes of the experiment are associated with a unique number.
Therefore, f is an example of a random variable. Usually a random variable is denoted
by using upper case letters such as X, Y, Z etc. The image set of the random variable
may be written as f ( S ) = {0, 1, 2, 3} .
If the image set, X(S), is either finite or countable, then X is called as a discrete random
variable, otherwise, it is referred to as a continuous random variable i.e. if X is a CRV,
then X(S) is infinite and un – countable.
X = xi x1 x2 x3 . . . xn
f ( xi ) P (2 ≤ X < 5) f ( x1 ) f ( x2 ) f ( x3 ) . . . f ( xn )
• 0 ≤ f ( xi ) ≤ 1
• ∑ f ( x ) =1
i
i
On the other hand, X is a continuous random variable, then its probability function will
be usually given or has a closed form, given as f ( x ) = P ( X = x ) where x is defined over
a range of values., it is called as probability density function usually has some standard
form. This function too has the following properties:
• f ( x) ≥ 0
• 0 ≤ f ( x) ≤ 1
∞
• ∫
−∞
f ( x) = 1 .
To begin with we shall discuss in detail, discrete random variables and its distribution
functions. Consider a discrete random variable, X with the distribution function as
given below:
X = xi x1 x2 x3 . . . xn
f ( xi ) f ( x1 ) f ( x2 ) f ( x3 ) . . . f ( xn )
Using this table, one can find probability of various events associated with X. For
example,
• P ( xi ≤ X ≤ x j ) = P ( X = xi ) + P ( X = xi +1 ) + up to + P ( X = xj )
= f ( xi ) + f ( xi +1 ) + f ( xi + 2 ) + up to +f ( x j −1 ) +f ( xj )
• P ( xi < X < x j ) = P ( X = xi +1 ) + P ( X = xi + 2 ) +. . . + P ( X = x j −1 )
= f ( xi +1 ) + f ( xi + 2 ) + up to +f ( x j −1 )
{
= 1 − P ( X = x1 ) + P ( X = x2 ) up to +P ( X = x j -1 ) }
• 0 ≤ F ( x) ≤ 1
function.
• when x → ∞, F ( x ) approaches 1
• E ( X ) = ∑ x i ⋅ f ( xi )
i =1
i =n
• E ( X 2 ) = ∑ xi2 ⋅ f ( xi )
i =1
Var( X ) = E ( X 2 ) − { E ( X )}
2
•
ILLUSTRATIVE EXAMPLES:
X: 0 1 2 3 4 5 6
f ( xi ) : k 3k 5k 7k 9k 11k 13k
Find (i) k; (ii) F (4) ; (iii) P ( X ≥ 5) ; (iv) P (2 ≤ X < 5) (v) E(X) and (vi) Var (X).
Solution: To find the value of k, consider the sum of all the probabilities which equals to
1
49k. Equating this to 1, we obtain k = . Therefore, distribution of X may now be
49
written as
X: 0 1 2 3 4 5 6
1 3 5 7 9 11 13
f ( xi ) :
49 49 49 49 49 49 49
2
973 203
Var( X ) = E ( X 2 ) − { E ( X )}
2
= − .
49 49
Find (i) k, (ii) F (2) , (iii) P (−2 < X < 2) , (iv) P (−1 < X ≤ 2) , (v) E(X) , Variance.
Solution: Consider the result, namely, sum of all the probabilities equals 1,
0.1+ k + 0.2 + 2k + 0.3 + k = 1 Yields k = 0.1. In view of this, distribution function of
X may be formulated as
X: -2 -1 0 1 2 3
f ( xi ) : 0.1 0.1 0.2 0.2 0.3 0.1
∞
(i) f ( x ) ≥ 0, (ii) 0 ≤ f ( x ) ≤ 1 and ( iii ) ∫
−∞
f ( x ) dx = 1 .
• P (a ≤ X ≤ b) = ∫ f ( x ) dx , P (a < X ≤ b) = ∫ f ( x ) dx
a a
b x
• P (a < X < b) = ∫ f ( x ) dx , F ( x) = P ( X ≤ x) = ∫ f ( x ) dx
a −∞
∞ ∞
• E(X ) = ∫ x ⋅ f ( x ) dx , E( X2) = ∫x
2
⋅ f ( x ) dx
−∞ −∞
Var( X ) = E ( X 2 ) − { E ( X )}
2
•
• P (a < X < b) = F (b ) − F (a )
dF ( x )
• f ( x) = , if the derivative exists
dx
SOME STANDARD DISTRIBUTIONS OF A DISCRETE RANDOM VARIABLE:
Binomial distribution function: Consider a random experiment having only two
outcomes, say success (S) and failure (F). Suppose that trial is conducted, say, n number
of times. One might be interested in knowing how many number of times success was
achieved. Let p denotes the probability of obtaining a success in a single trial and q
stands for the chance of getting a failure in one attempt implying that p + q = 1. If the
experiment has the following characteristics;
We say that the problem is based on the binomial distribution. In a problem like this, we
define X as the random variable equals the number of successes obtained in n trials.
Then X takes the values 0, 1, 2, 3 . . . up to n. Therefore, one can view X as a discrete
random variable. Since number of ways of obtaining k successes in n trials may be
n n!
achieved in = , therefore, binomial probability function may be formulated
k k !( n − k ) !
n k n−k
as b(n, p, k ) = p q .
k
Illustrative examples:
1. It is known that among the 10 telephone lines available in an office, the chance
that any telephone is busy at an instant of time is 0.2. Find the probability that (i)
exactly 3 lines are busy, (ii) What is the most probable number of busy lines and
compute its probability, and (iii) What is the probability that all the telephones are
busy?
Solution:
Here, the experiment about finding the number of busy telephone lines at an instant of
time. Let X denotes the number of telephones which are active at a point of time, as there
are n = 10 telephones available; clearly X takes the values right from 0 up to 10. Let p
denotes the chance of a telephone being busy, then it is given that p = 0.2, a finite value.
The chance that a telephone line is free is q = 0.8. Since a telephone line being free or
working is independent of one another, and since this value being same for each and
every telephone line, we consider that this problem is based on binomial distribution.
Therefore, the required probability mass function is
10
b(10, 0.2,3) = ⋅ (0.2)3 ⋅ (0.8)7
3
(ii) With p = 0.2, most probable number of busy lines is n ⋅ p = 10 ⋅ 0.2 = 2 . The
10
probability of this number equals b(10, 0.2, 2) = ⋅ (0.2) ⋅ (0.8) .
2 8
2
(iii) The chance that all the telephone lines are busy = (0.2)10 .
2. The chance that a bomb dropped from an airplane will strike a target is 0.4. 6
bombs are dropped from the airplane. Find the probability that (i) exactly 2 bombs
strike the target? (ii) At least 1 strikes the target. (iii) None of the bombs hits the
target?
Solution: Here, the experiment about finding the number of bombs hitting a target. Let
X denotes the number of bombs hitting a target. As n = 6 bombs are dropped from an
airplane, clearly X takes the values right from 0 up to 6.
Let p denotes the chance that a bomb hits a target, then it is given that p = 0.4, a finite
value. The chance that a telephone line is free is q = 0.6. Since a bomb dropped from
airplane hitting a target or not is an independent event, and the probability of striking a
target is same for all the bombs dropped from the plane, therefore one may consider that
hat this problem is based on binomial distribution. Therefore, the required probability
10 6−k
mass function is b(10, 0.4, k ) = ⋅ (0.4) ⋅ (0.8) .
k
k
(i) To find the chance that exactly 2 bombs hits a target,
10
i.e. P[X = 2] = b(10, 0.4, 2) = ⋅ (0.4) ⋅ (0.8)
2 4
2
P[ X ≥ 1] = 1 − P[ X < 1] = 1 − P[ X = 0] = 1 − (0.6)6 .
(iii) The chance that none of the bombs are going to hit the target is P[X=0] = (0.6)6 .
n k n−k
probability mass function given by b(n, p, k ) = p q . Consider the expectation of
k
X, namely,
k =n
n
E ( X ) = ∑ k ⋅ p k q n− k
k =0 k
k =n
n!
= ∑k⋅ pk q n−k
k =0 k !( n − k ) !
k =n
n(n − 1)!
=∑ pp k −1q ( n−1+1− k )
k = 0 ( k − 1)!( n − 1 + 1 − k ) !
k =n
( n − 1)!
= np ⋅ ∑ p k −1q[( n−1− ( k −1)]
k = 0 ( k − 1)![( n − 1)!− ( k − 1)!]
k =n
( n − 1)!
= np ⋅ ∑ p k −1q[( n−1− ( k −1)]
k =1 ( k − 1)![( n − 1)!− ( k − 1)!]
k =n
n − 1 k −1 [( n−1− ( k −1)]
= np∑ p q
k = 2 k − 1
= np ⋅ ( p + q) n −1
= np as p + q = 1
Thus, expected value of binomial distribution function is np .
k =n
(n − 2)!
= n(n − 1) p 2 ∑ p k −2 q[( n −2 −( k −2)] + np
k = 0 ( k − 2)![( n − 2)!− ( k − 2)!]
k =n
(n − 2)!
= n(n − 1) p 2 ∑ p k −2 q[( n −2 −( k −2)] + np
k =2 ( k − 2)![( n − 2)!− ( k − 2)!]
k =n
n − 2 k −2 [( n− 2 −( k −2)]
= n(n − 1) p 2 ∑ p q + np
k =2 k − 2
= n(n − 1) p 2 + np
Therefore, Var( X ) = E ( X ) - { E ( X )}
2 2
= n(n − 1) p 2 + np − ( np )
2
= n2 p 2 − np2 + np − n2 p2
k =∞
e − λ λλ k −1
= ∑
k = 0 ( k − 1)!
k =∞
λ k −1 k =∞
λ k −1
= λ ⋅e −λ
⋅∑ . But ∑ = e λ , therefore it follows that for
k =1 ( k − 1)! k =1 ( k − 1)!
k =∞
e−λ λ k
= ∑ k2 ⋅
k =0 k!
k =∞
e−λ λ k
= ∑ k (k − 1 + 1) ⋅
k =0 k!
k =∞
e − λ λ k k =∞ e − λ λ k
= ∑ k (k − 1)
k =0 k!
+∑ k
k =0 k!
k =∞
e−λ λ k
= ∑ k (k − 1)
k =0 k!
+E ( X )
k =∞
λ k −2
= λ 2e − λ ∑ +λ
k = 2 ( k − 2)!
σ = Var( X ) = λ
Illustrative Examples:
It is known that the chance of an error in the transmission of a message through a
communication channel is 0.002. 1000 messages are sent through the channel; find
the probability that at least 3 messages will be received incorrectly.
Solution: Here, the random experiment consists of finding an error in the transmission
of a message. It is given that n = 1000 messages are sent, a very large number, if p
denote the probability of error in the transmission, we have p = 0.002, relatively a small
number, therefore, this problem may be viewed as Poisson oriented. Thus, average
number of messages with an error is λ = np = 2 . Therefore, required probability function
e −2 2 k
is .= p(2, k ) = , k = 0, 1, 2, 3, . . . ∞ . Here, the problem is about finding the
k!
probability of the event, namely,
P ( X ≥ 3) = 1 − P ( X < 3) = 1 − {P[ X = 0] + P[ X = 1] + P[ X = 2]}
k = 2 e −2 2k
= 1 − ∑
k =0 k !
= 1 − e −2 { 1 + 2 + 2} = 1 − 5e−2
2. A car hire –firm has two cars which it hires out on a day to day basis. The
number of demands for a car is known to be Poisson distributed with mean 1.5.
Find the proportion of days on which (i) There is no demand for the car and (ii) The
demand is rejected.
Solution: Here, let us consider that random variable X as the number of persons or
demands for a car to be hired. Then X assumes the values 0, 1, 2, 3. … . . . It is given
that problem follows a Poisson distribution with mean, λ = 1.5 . Thus, required probability
e −1.5 (1.5) k
mass function may be written as p(1.5, k ) = .
k!
(i) Solution to I problem consists of finding the probability of the event, namely
P[X = 0] = e −1.5 .
(ii) The demand for a car will have to be rejected, when 3 or more persons approaches the
firm seeking a car on hire. Thus, to find the probability of the event P[ X ≥ 3]. Hence,
(1.5) 2
P[ X ≥ 3] = 1 − P{ X < 3] = 1 − P[ X = 0,1, 2] = e −1.5 1 + 1.5 + .
2
x2
, −1 < x < 2
f ( x) = 3
0 elsewhere.
Solution: Consider F ( x ) = P ( X ≤ x ) = ∫
−∞
f ( t )dt
x x
Case (i) x ≤ -1 F ( x) = ∫
−∞
f ( t )dt = ∫ 0dt = 0
−∞
x −1 2 x
Case (iii) x = 2 F ( x ) = ∫
−∞
f ( t )dt = ∫
−∞
f ( t )dt + ∫ f ( t )dt + ∫ f ( t )dt
−1 2
2 2
t2 t3 8+1
= 0 + ∫ dt + 0 = = = 1 . Therefore,
−1
3 9 −1
9
0, x ≤ −1
3
x +1
F ( x) = , −1 < x < 2.
9
1, x ≥ 2.
Find (a) the value of k and (b) distribution function F(X) for X.
∞
∫ 2kxe
− x2
WKT dx = 1
0
∞
∫ ke
−t
⇒ dt = 1(put x 2 = t )
0
∞
⇒ ke − t =1
0
⇒ (0 + k ) = 1 ⇒ k = 1
0 x
= ∫
−∞
f ( t )dt + ∫ f ( t )dt , if x > 0
0
= 0 + ∫ 2te − t dt = ( − e − z )0x = (1 − e − x ) .
2 2 2
1 − e − x , for x ≥ 0
2
F ( x) =
0, otherwise.
3. Find the C.D.F of the R.V. whose P.D.F is given by
x
2 , for 0<x ≤ 1
1 , for 1<x ≤ 2
f ( x) = 2
3− x
, for 2<x ≤ 3
2
0, otherwise
x 0
0 1 x
1 x
t 1 2x − 1
= 0 + ∫ dt + ∫ dt =
0
2 1
2 4
0 1 2 x
0, if x ≤ 0
2
x , if 0<x ≤ 1
4
2 x − 1
F ( x) = , if 1<x ≤ 2
4
6x − x2 − 5
, if 2<x ≤ 3
4
1, if x>3
4. The trouble shooting of an I.C. is a R.V. X whose distribution
0, for x ≤ 3
function is given by F ( x) = 9
1− 2 , for x > 3.
x
If X denotes the number of years, find the probability that the I.C. will
work properly
(a) less than 8 years
(b) beyond 8 years
(c) anywhere from 5 to 7 years
(d) Anywhere from 2 to 5 years.
x 0, for x ≤ 3
Solution: We have F ( x ) = ∫ f ( t )dt = 9
0 1 − x 2 , for x > 3.
8
9
For (a): P ( x ≤ 8) = ∫ f ( t )dt = 1 − 2 = 0.8594
0
8
2
For Case (d): P (2≤ x≤ 5) = F (5) – F (2) = (1-9/5 ) – (0) = 0.64
d
Solution: We know that f ( x ) = [ F ( x)]
dx
0, x≤1
∴ f ( x ) = 4c( x − 1) , 1 ≤ x ≤ 3
3
0, x > 3.
4c( x − 1)3 , 1 ≤ x ≤ 3
∴ f ( x) =
0, otherwise
∞
3
− 3 dx =1 ⇒ c ( x −1)4 3 = 1
∫ 4 c ( x 1) 1
1
1
⇒16c =1 ∴c =
16
Using this, one can give the probability function just by substituting the value of c
above.
x
The probability distribution function may be written as F ( x ) = ∫ f ( x )dx which may be
−∞
1− e −λx , if 0 < x < ∞
computed as F ( x ) = .
0 , otherwise.
∞
−λx e −λ x
e 1 1
= λ x⋅ −1⋅ = - λ 0− =
−λ −λ 2 λ 2 λ
0
∞
( )
∞
Consider E X 2 = ∫ x 2 ⋅ f ( x )dx = ∫ x 2 ⋅λe −λx dx
−∞ 0
∞
−λx e −λx e −λx 2
= λ x2 e − 2 x + 2 =
−λ
−λ 2 −λ 3
0 λ2
( )
Var ( X ) = E X 2 - { E ( X )}
2= 1
λ2
.
1
The standard deviation is σ = Var( X ) = .
λ
1 1
an exponential distribution function, mean is known to be , so = 2 or λ =0.5 . The
λ λ
Among all the distribution of a continuous random variable, the most popular and widely
used one is normal distribution function. Most of the work in correlation and regression
analysis, testing of hypothesis, has been done based on the assumption that problem
follows a normal distribution function or just everything normal. Also, this distribution
is extremely important in statistical applications because of the central limit theorem,
which states that under very general assumptions, the mean of a sample of n mutually
The normal density function has well known bell shaped curve which will be shown on
2
−1 x − µ
1
the board and it may be given as f ( x) = e2 σ
, -∞ < x < ∞ where
σ 2π
−∞ < µ < ∞ and σ > 0 . It will be shown that µ and σ are respectively denotes mean
numerically and recorded in special table called normal distribution table. However, It
pertain to the standard normal distribution function by choosing µ and σ and their
z
1 −t 2
entries are values of the function, Fz ( z ) =
2π ∫
−∞
e 2
dt. Since the standard
−z
1
Fz ( − z ) =
2π
∫
−∞
f (t ) dt =1 - F ( z ) .
z
Thus, tabulations are done for positive values of z only. From this it is clear that
Normal distribution table, please refer table 3, in page number 591, Probability and
Statistics, Reliability, Queuing and Computer Science Applications” by K. S.
Trivedi, a PHI Publications. It is clear that
• P ( a ≤ X ≤ b ) = F (b ) - F ( a )
• P ( a < X < b ) = F (b ) - F ( a )
• P( a < X ) = 1 − P ( X ≤ a)= 1 − F (a )
Note: Let X be a normally distributed random variable taking a particular value, x, the
x−µ
corresponding value of the standardized variable is given by z = . Hence,
σ
F ( x ) = P ( X ≤ x ) = Fz x − µ .
σ
Therefore, number of electrical bulbs with life expectancy more than 2150 hours is
0.0336 × 2000 ≈ 67 .
Therefore, in a lot of 2000 bulbs, number of bulbs with life expectancy less than 1950
hours is 0.0668 * 2000 = 134 bulbs.
Finally, to find the probability of the event, namely,
P (1920 < X < 2060) = F (2060) − F (1920)
= F ( 0.3333) − F ( −2 )
z z
= F ( 0.3333) − 1 + F ( 2 )
z z
= 0.6293 − 1 + 0.9774 = 0.6065 .
Therefore, number of bulbs having life any where in between 1920 hours and 2060 hours
is 0.6065 * 2000 = 1213.
2. Assume that the reduction of a person’s oxygen consumption during
44.5 − 37.6
= 1 − Fz
4.6
= 1 − Fz ( 1.5 )
= 1 − 0.9332 = 0.0668 .
35.0 − 37.6
= Fz
4.6
= Fz ( −0.5652 )
= 1 − Fz ( 0.5652 )
= 1 − 0.7123 = 0.2877 .
(iii) Consider the probability of the event P[30.0 < X < 40.0]
= F (40) − F (30)
40 − 37.6 30 − 37.6
= Fz − Fz
4.6 4.6
= Fz (0.5217) − Fz ( −1.6522)
= 0.6985 − 1 + 0.9505 = 0.6490
= 1 − F (240)
240 − 200
= 1 − Fz
16
= 1 − Fz ( 2.5 )
= 1 – 0.9938
= 0.00621
240 − 200
1 − Fz
16 1 − Fz (2.5)
= =
210 − 200 1 − Fz (0.625)
1 − Fz
16
1 − 0.9939
= = 0.2335
1 − 0.73401
UNIT 8:
Introduction: So far the study was restricted to one dimensional random variable,
distribution function, and other related aspects. However, in real world, one come
across a number of situations where we do find the presence of two or correlated random
variables. For example, consider the experiment of finding how far e - learning
programme initiated by VTU Belgaum has become popular among the engineering
students? To find this say, authorities collect feed back from the students by visiting their
institutions. Let the problem be about finding the opinion of students regarding two
parameters; (i) quality of transmission from studio situated at Bangalore which we call it
as X and (ii) student’s interest in this kind of programs which we shall refer it to as Y.
For convenience of discussion, authorities visit seven colleges located at different parts of
the state and results are in the following table. We assume that these are given in terms of
percentages.
Y y1 y2 y3 y4 y5 y6 y7
In problems like this, we/authorities are certainly interested to learn the mood of the
students/teachers about the e – learning programme initiated by us of course with huge
cost. It is known to you that one satellite channel has been completely dedicated for this
purpose in India. Many people are involved in this programme to reach the un – reached
and needy.
One comes across many illustrative examples like that. Therefore, there is a necessity to
extend the study beyond single random variable.
Suppose that there exists a correlation between the random variables X and Y. Then X
and Y are jointly related/distributed variables Also, note that X and Y together assumes
values. The same can be shown by means of a matrix or a table.
Y y1 y2 y3 . . . . yn
X
x1 ( x1 , y1 ) ( x1 , y2 ) ( x1 , y3 ) .. . . ( x1 , yn )
x2 ( x2 , y1 ) ( x2 , y2 ) ( x 2 , y3 ) . . . . ( x 2 , yn )
x3 ( x3 , y1 ) ( x 3 , y2 ) ( x 3 , y3 ) . . . . ( x 3 , yn )
. . . . . . . . . . . . . .
. . . . . . . . . . . . . .
xm ( xm , y1 ) ( xm , y2 ) ( x m , y3 ) . . . . ( xm , yn )
1. h ( xi , y j ) ≥ 0
2. 0 ≤ h ( xi , y j ) ≤ 1
3. ∑∑ h ( x , y ) = 1
i j
i j
Y y1 y2 y3 . . . . yn
X
x1 h ( x1 , y1 ) h ( x1 , y2 ) h ( x1 , y3 ) .. . . h ( x1 , yn )
x2 h ( x2 , y1 ) h ( x2 , y2 ) h ( x2 , y3 ) . . . . h ( x 2 , yn )
x3 h ( x3 , y1 ) h ( x3 , y2 ) h ( x3 , y3 ) . . . . h ( x 3 , yn )
. . . . . . . . . . . . . .
. . . . . . . . . . . . . .
xm h ( xm , y1 ) h ( xm , y2 ) h ( xm , y3 ) . . . . h ( x m , yn )
Illustrative examples:
A fair coin is tossed 3 times. Let X denote the random variable equals to 0 or 1
accordingly as a head or a tail occurs on the first toss, and let Y be the random
variable representing the total number of heads that occurs. Find the joint
distribution function of (X, Y).
Solution: S = {HHT, HHT, HTH, THH, THT, TTH, HTT, TTT}. Thus, |S| = 8.
Here, X takes the values 0 or 1 accordingly as a H appears on the I toss or a Tail appears
on the I toss, while Y takes the values 0, 1, 2, 3 where these numbers represent the
number of heads appearing in the experiment. Observe that joint variables(X, Y) assume
eight values. Thus, there are (0, 0), (0, 1), (0, 2), (0, 3), (1, 0), (1, 1), (1, 2), (1, 3) –
totally 8 events. Thus, we need to find the probabilities of all these 8 events. First we
shall list the respective events corresponding to X and Y.
• [Y = 0] = {TTT},
• [Y = 3] = {HHH}
Y 0 1 2 3
X
0 h(0, 0) = 0 h(0, 1) = 1 h(0, 2) = 2 h(0, 3) = 1
8 8 8
1 h(1, 0) = 1 h(1, 1) = 2 h(1, 2) = 1 h(1, 3) = 0
8 8 8
Two cards are selected at random from a box which contains 5 cards numbered 1, 1,
2, 2, and 3. Let X denote the sum selected, and Y the maximum of two numbers
drawn. Determine the joint probability function of X and Y.
Solution: It is known that 2 cards can be selected from the available 5 in C (5, 2) = 10
ways. With cards being numbered as 1, 1, 2, 2, and 3, clearly, S = {(1, 1), (1, 2), (1, 2),
(1, 2), (1, 2), (1, 3), (1, 3), (2, 2), (2, 3), (2, 3)}. As X denote the sum of numbered
chosen, X takes the values 2, 3, 4, and 5. Y is defined as the maximum of the two
numbers, so Y takes the values 1, 2, 3. Therefore, (X, Y) assumes totally 12 values. It
can be shown as
(Dr. K. Gururajan, MCE, Hassan page 4)
1 2 3
2 (2, 1) (2, 2) (2, 3)
3 (3, 1) (3, 2) (3, 2)
X
4 (4, 1) (4, 2) (4, 3)
5 (5, 1) (5, 2) (5, 3)
Let h(x, y) denotes the joint probability function. Now, [X = 2] = {(1, 1)}, [Y = 1] =
{(1, 1)}, therefore, h(2, 1) = 1/10 = 0.1; Observe that [Y = 2] = {(1, 2), (1, 2), (1, 2), (1,
2), (2, 2)}, thus, [X = 2, Y = 2] = { }, so h(2, 2) = 0.0, a similar argument will show that
h(2, 3) = 0.0 since [y = 3] = {(1, 3), (1, 3), (2, 3), ( 2, 3)}. Next, [X = 3] = {(1, 2), (1, 2),
(1, 2), (1, 2)}, clearly, [X = 3, Y = 1] = { }, so h (3, 1) = 0.0, [X = 3, Y = 2] = {(1, 2), (1,
2), (1, 2), (1, 2)}, thus, h(3, 2) = 4/10 = 0.4. However, h (3, 3) = 0.0. This is because, [Y
= 3] = {(1, 3), (1, 3), (2, 3), (2, 3) and nothing is common between the events [X = 3] and
[Y = 3]. By following above arguments, it may be seen that h (4, 1) = 0, h (4, 2) = 0.1, h
(4, 3) = 0.2, h (5, 1) = 0.0, h (5, 2) = 0.0 and h (5, 3) = 0.2. Thus, the joint distribution
function of X and Y may be given as
Y 1 2 3
x = xt y = y u
H ( xt , yu ) = P[ X ≤ xt , Y ≤ yu ] = ∑ ∑ h( x, y ) .
x = x1 y = y1
The expectation of X and Y is
i =m j =n
E ( XY ) = ∑∑ ( xi ⋅ y j ) ⋅ h ( xi , y j ) . From the joint probability mass function, h(x, y) it
i =1 j =1
Y 0 1 2 3
X
0 h(0, 0) = 0 h(0, 1) = 1 h(0, 2) = 2 h(0, 3) = 1
8 8 8
1 h(1, 0) = 1 h(1, 1) = 2 h(1, 2) = 1 h(1, 3) = 0
8 8 8
The distribution function of X is got by adding the probabilities row wise. Thus,
xi : 0 1
1 1
f ( xi ) :
2 2
and the marginal distribution function of Y is to be obtained by the adding the
probabilities column wise. Therefore,
yi 0 1 2 3
g ( yj ) 1 3 3 1
8 8 8 8
ILLUSTRATIVE EXAMPLES:
Y -4 2 7
1 1/8 ¼ 1/8
X
5 2/8 1/8 1/8
Solution: First, we obtain the distribution functions of X and Y. To get the distribution
of X, it is sufficient to add the probabilities row wise. Thus,
(Dr. K. Gururajan, MCE, Hassan page 6)
X 1 5
f ( xi ) 1/2 1/2
E ( X ) = 1 ⋅ (1 / 2) + 5 ⋅ (1 / 2) = 3
E ( X 2 ) = 1 ⋅ (1 / 2) + 25 ⋅ (1 / 2) = 13
Var( X ) = E ( X 2 ) − { E ( X )} = 4 and σ x = 2 .
2
Similarly,
Y -4 2 7
E (Y ) = −4 ⋅ ( 3 / 8) + 2 ⋅ ( 3 / 8) + 7 ⋅ (2 / 8) = 1.0
E ( Y 2 ) = 16 ⋅ (3 / 8) + 4 ⋅ (3 / 8) + 49 ⋅ (2 / 8) = 39.5
Var(Y ) = E ( Y 2 ) − { E (Y )} = 38.5
2
σ y = 38.5 = 6.2048.
E ( XY ) = ∑∑ x ⋅ y ⋅ h( x , y )
Consider x y = (2)(1)(0.2) + (3)(2)(0.4) + (4)(2)(0.1) + (4)
(3)(0.2) + (5)(3)(0.2) = 8.8. With E(X) = 3, and E(Y) = 1, it follows that
Cov( X ,Y ) 5.8
Cov( X , Y ) = E ( XY ) − E ( X ) ⋅ E (Y ) = 5.8 and ρ ( x , y ) = σ x ⋅σ y
=
( 2.0)⋅( 6.2048)
= 0.4673 . As
It may be noted that so far the study on probability theory was concentrated on either a
single random variable or a two random variables (correlated variables) defined on the
same sample space, S of some random experiment. It is interesting to see that in a
practical situation, we come across numerous occasions involving a family or a collection
Let X and Y be two discrete random variables. One says that X and Y are independent
whenever the joint distribution function is the product of the respective marginal
( )
distribution functions. Equivalently, suppose that f ( xi ) , g y j and h xi , y j denote ( )
respectively, the distribution function of X, Y and that of X and Y, and
( ) ( )
h xi , y j = f ( xi ) ⋅ g y j , then we say that X and Y are independent random variables.
Note: If it is known that X and Y are independent random variables, then it is very easy
to compute the joint distribution function. This may be obtained by just multiplying the
respective probabilities in the corresponding row and column of the table. For example,
consider
X 0 1 2 3
f ( xi ) 0.2 0.2 0.5 0.1
y 3 5 6
( )
g yj 0.2 0.5 0.3
(
h xi , y j ) 3 5 6
0 0.04 0.10 0.06
1 0.04 0.10 0.06
2 0.10 0.25 0.15
3 0.02 0.05 0.03
1. If the state space I and index set T of a stochastic process is discrete, then it is called
Discrete – state – discrete – parameter (time) process.
2. On the other hand, if state space is continuous, and index set T is discrete, then we
have a continuous – state – discrete – parameter process.
3. Similarly, one have discrete state – continuous parameter process and
4 Continuous state – continuous parameter process.
P X (t ) ≤ x | X ( t n ) = x n , X ( tn −1 ) = xn −1 , . . . X ( t0 ) = xo
= P X (t ) ≤ x | X ( t n ) = x n
i.e. here, behavior of the stochastic process is such that probability distributions for its
future development depend only the present state and not on how the process arrived in
that state. Also, state space, I is discrete in nature.
Equivalently, in a Markov chain, (a Markov process where state space, I takes discrete
values), the past history is completely summarized in the current state, and, future is
independent of its past but depends only on present state.
First
ur we shall discuss few basic concepts pertaining to Markov chains. A vector
a = ( a1 , a2 , a3 . . . , an ) is called a probability vector if all the components are non–
negative and their sum is 1.
A matrix, A such that each row is a probability vector, then A is called stochastic matrix.
If A and B are stochastic matrices, then AB is a stochastic matrix. In fact, any power of
A is a a stochastic matrix.
A stochastic matrix A is said to be regular if all the entries of A are positive for some
power Am of A where m is a positive integer.
Result 1: Let A be a regular stochastic matrix. Then A has a unique fixed probability
vector, t. i.e. One can find a unique fixed vector t such that t = t ⋅ A .
Each outcome belongs to the state space, I = {a1 , a2 , a3 . . . a m } . The outcome of any
trial depends, at most, on the outcome of the trial and not on any other previous
outcomes. If the outcome in n th trial is ai then we say that the system is in ai state at
the n th stage. Thus, with each pair of states ( ai , a j ) , we associate a probability value
pij which indicate probability of system reaching the state ai from the state a j in n
steps. These probabilities form a matrix called transition probability matrix or just
transition matrix. This may be written as
Note: Here, i th row of M represents the probabilities of that system will change from
ai to a1 , a2 , a3 , . . . an . Equivalently, pij = P [ x n = j | xn −1 = i ] .
The probability that a Markov chain will move from state I to state j in exactly n – steps
is denoted by p ij = pij (n ) = P [ x m +n = j | xm = i ]
(n )
If M is the transition matrix of a Markov chain, then the n – step transition matrix may be
obtained by taking nth power of M. Suppose that a system at time t = 0 is in the state
a = ( a1 , a2 , a3 . . . a n ) where the process begins, then corresponding probability vector
(
a ( 0 ) = a1( 0 ) , a(20) , a3( 0) . . . a(0)
n
) denotes the initial probability distribution. Similarly, the
(n) (n ) (n) (n)
(
step transition probability matrix may be given as a = a1 , a 2 , a3 . . . a n . Now,
(n )
)
the (marginal) probability mass function of the random variable may be obtained
from the nth step transition probabilities and the initial distribution function as follows:
a ( n ) = a ( 0) ⋅ M ( n ) = a( 0) ⋅ M n . Thus, the probability distribution of a homogeneous Markov
chain is completely determined from the one step transition probability matrix M and the
initial probability distribution a (0) .
• a ( 1) = a ( 0 ) ⋅ M
• a ( 2 ) = a ( 1) ⋅ M = a ( 0 ) ⋅ M 2
• a ( 3 ) = a ( 2 ) ⋅ M = a( 0 ) ⋅ M 3
1 1 1 3 1 3 1
3 -
3 3 4 4 4 4
(i) (ii) (iii)
1 1 2 2 2 1
0
2 2 3 3 3 3
Solution: (i) is a stochastic matrix as all the entries are non – negative and the sum of all
values in both the rows equals 1. (ii) This is not a stochastic matrix, since the second row
is such that sum exceeding 1 (iii) this is again not a stochastic matrix, because one of the
entry is negative.
3. Find the unique fixed probability vector of the following stochastic matrices.
3 1
0 4 4 0 1 0
1 2
1 1 1 1 1
(i) A = 3 3 (ii) = 0
(iii) A =
2 2 6 2 3
1 0 2 1
0 1 0
0
3 3
(ii) We shall set up t = ( x , y , z ) as the unique fixed probability vector. Again this is a
probability vector, therefore, t = ( x , y , 1 − x − y ) . Now, consider the matrix equation,
3 1
0 4 4
1 1
(x, y , 1 − x − y ) = (x , y , 1 − x − y ) ⋅ 0
2 2
0 1 0
(iii) Here, the unique fixed probability vector may be obtained as t = (0.1, 0.6, 0.3) .
Solution: First we construct transition probability matrix. The problem here is about
whether the boy studies or not on consecutive nights. As his study is given to be
dependent on study in one night or not, therefore, it may be modeled in terms of Markov
chain. Thus, first we construct the required transition probability matrix. The following
table indicates the pattern of the study of the boy:
II night
Transition probability Boy Studying Boy not studying
0.3 0.7
Therefore, it is clear that required stochastic matrix is M = . To solve the
0.4 0.6
given problem, it is sufficient to compute a unique fixed probability vector, t = ( x , 1 − x )
such that t = t ⋅ M . Using the above, one obtains x = 0.3 x + 0.4(1 − x ) . From this
4
equation, we can calculate x = . Thus, a chance of boy studying in the long run is
11
36.7%.
6. A person’s playing chess game or tennis is as follows: If he plays chess game one
week then he switches to playing tennis the next week with a probability 0.2. On the
other hand, if he plays tennis one week, then there is a probability of 0.7 that he will
play tennis only in the next week as well. In the long run, how often does he play
chess game?
Solution: Like in the previous case, this one too is based on Markov process. Here,
parameters are about a player playing either chess or tennis game. Also, playing the
game in the next week is treated to be dependent on which game player plays in the
previous week. As usual, first we obtain transition probability matrix.
II Week
0.8 0.2
Clearly, the transition probability matrix is M = . Here, too, the problem is
0.3 0.7
to find a unique fixed probability vector, t = ( x , 1 − x ) such that t = t ⋅ M . Consider the
0.8 0.2
matrix equation; x 1 - x ) = (x 1- x) ⋅ . A simple multiplication work
0.3 0.7
results in x = 0.8 x + 0.7(1 − x ) from which we get x = 0.6 . Hence, we conclude that in
the long run, person playing chess game is about 60%.
7. A sales man S sells in only three cities, A, B, and C. Suppose that S never sells in
the same city on successive days. If S sells in city A, then the next day S sells in the
city B. However, if S sells in either B or C, then the next day S is twice likely to sell
in city A as in the other city. Find out how often, in the long run, S sells in each city.
Solution: First we shall obtain transition probability matrix which will have 3 rows and
3 columns. As a salesman does not sell in a particular city on consecutive days, clearly,
main diagonal is zero. It is given that if S sells in city A on first day, then next day he
will sell in city B, therefore, first row is 0, 1, 0. next, one can consider two cases: (i) S
selling city B or (ii) S selling in city C. It is also given that if S sells in city B with
probability p, then his chances of selling in city A, next day is 2p. There is no way, he
will sell in city C. Thus, we have p + 2p + 0 = 1 implying that p = 1/3. Therefore middle
row of the matrix is 2/3, 1/3, 0. Similarly, if S sells in city C with probability q, then his
chances of selling in city A, next day is 2q. Again, a chance of selling city B is 0. Thus,
last row is 2/3, 0 , 1/3. The probability transition matrix is
Next Day
A B C
A 0 1 0
I day
B 2 0 1
3 3
C 2 1 0
3 3
0 1 0
The matrix of the Markov process is M = 2 3 0 1 . As in the previous two
3
2 1 0
3 3
problems, we need to find a unique fixed probability vector, t = ( x , y , z ) such that
0 1 0
[ x y 1 - x - y ] = [ x y 1 - x - y ] ⋅ 2 3 0 13 Or
2 1 0
3 3
x = 2 y + 2 (1 − x − y ) and y = x + 1 (1 − x − y ) . Solving these two equations,
3 3 3
one obtains x = 0.4, y = 0.45 and z = 0.15 . Hence, chances of a sales man selling in
each of the cities is 40%, 45% and 15% respectively.
8. Mary’s gambling luck follows a pattern. If she wins a game, the probability of
winning the next game is 0.6. However, if she loses a game, the probability of losing
the next game is 0.7. There is an even chance that she wins the first game. (a) Find
the transition matrix M of the Markov process. (b) Find the probability that she
wins the second game? (c) Find the probability that she wins the third game? Find
out how often, in the long run, she wins?
10. Each year Rohith trades his car for a new car. If he has a Maruti, he trades it in
for a Santro. If he has a Santro car, then he trades it in for a Ford. However, is he
has a Ford car; he is just as likely to trade it in for a new Ford as to trade it in for a
Maruti or for a Santro. In 1995, he bought his first car which was a Ford.
(a) Find the probability that he has bought (i) a 1997 Buick, (ii) a 1998 Plymouth,
(iii) a 1998 Ford?
(b) Find out how often, in the long run, he will have a Ford?
Definition of various states:
Transient state: A state i is said to be transient (or non – recurrent) if and only if there is
a positive probability that process will not return to this state. For example, if we model
a program as a Markov chain all but the final state will be transient. Otherwise program
ends in an infinite loop.
Recurrent state: A state i is said to be recurrent if and only if process returns to this
state with probability one.
Periodic state: A recurrent state which returns to the state i at regular periods of time.
Absorbing state: A state i is said to be absorbing if and only if pii = 1 . Here, once a
Markov chain enters a state, and it remains there itself.
(Dr. K. Gururajan)
Unit 1: Numerical Methods
Numerical Solution of First Order and First Degree Ordinary
Differential Equations
Motivation:
The above laws define mechanism of change. When combined with continuity laws for
energy, mass or momentum, differential equation arises. The mathematical expression in
the above table is an example of the Conversion of a Fundamental law to an Ordinary
Differential Equation. Subsequent integration of these differential equations results in
mathematical functions that describe the spatial and temporal state of a system in terms of
energy, mass or velocity variations. In fact, such mathematical relationships are the basis
of the solution for a great number of engineering problems. But, many ordinary
differential equations arising in real-world applications and having lot of practical
significance cannot be solved exactly using the classical analytical methods. These ode
can be analyized qualitatively. However, qualitative analysis may not be able to give
accurate answers. A numerical method can be used to get an accurate approximate
solution to a differential equation. There are many programs and packages available for
solving these differential equations. With today's computer, an accurate solution can be
obtained rapidly. In this chapter we focus on basic numerical methods for solving initial
value problems.
Analytical methods, when available, generally enable to find the value of y for all
values of x. Numerical methods, on the other hand, lead to the values of y corresponding
only to some finite set of values of x. That is the solution is obtained as a table of values,
rather than as continuous function. Moreover, analytical solution, if it can be found, is
exact, whereas a numerical solution inevitably involves an error which should be small
but may, if it is not controlled, swamp the true solution. Therefore we must be concerned
with two aspects of numerical solutions of ODEs: both the method itself and its accuracy.
In this chapter some methods for the numerical solution of ODEs are described.
On finding the value of y1 for x = x1 from (1), y // ( x ), y /// ( x), y / V ( x ), etc . can be
evaluated at x = x1 from the differential equation
Problems:
1. Find by Taylor’s series method the value of y at x = 0.1 and 0.2 five places of decimals
dy
for the IVP = x 2 y −1, y ( 0) = 1 .
dx
Solution:
Given x 0 = 0, y 0 = 1 and f ( x, y ) = x 2 y −1
( x − 0) / ( x − 0) 2 // ( x − 0) 3 /// ( x − 0) 4 / V
y ( x) = y (0) + y (0) + y ( 0) + y ( 0) + y (0) +
1! 2! 3! 4!
x 2 // x 3 /// x 4 /V
i.e. y ( x) = y (0) + xy / (0) + y ( 0) + y ( 0) + y (0) +
2 6 24
(1)
y // ( x ) = 2 xy + x 2 y / ⇒⇒⇒ y // (0) = 0
Putting the values of y (0), y / (0), y // (0), y /// (0), y iv (0) in (1), we get
x2 x3 x4
y ( x ) = 1 + x (−1) + ( 0) + ( 2) + ( −6)
2 6 24
x3 x4
y ( x ) =1 − x + −
3 4
Hence y(0.1) = 0.90033 and y(0.2) = 0.80227.
2. Employ Taylor’s series method to obtain approximate value of y at x = 0.1 and 0.2 for
dy
the differential equation = 2 y + 3e x , y (0) = 0. Compare the numerical solution
dx
obtained with the exact solution.
Solution:
Given x0 = 0, y 0 = 0 and f ( x, y ) = 2 y + 3e x
Putting the values of y (0), y / (0), y // (0), y /// (0), y iv (0) in (2), we get
9 2 21 3 45 4
y ( x) = 0 + 3x + x + x + x
2 6 24
9 7 15 4
= 3x + x 2 + x 3 + x
2 2 8
Hence,
y(0.1) = 3(0.1)+4.5(0.1)2+3.5(0.1)3+1.875(0.1)4
= 0.3486875
and
y(0.2) = 3(0.2)+4.5(0.2)2+3.5(0.2)3+1.875(0.2)4
= 0.8110.
Exact Solution:
dy
The given differential equation can be written as − 2 y = 3e x which is Leibnitz’s linear
dx
differential equation.
y = −3e x + ce 2 x (3)
Solution:
Given x0 = 0, y 0 = 1 and f ( x, y ) = x 2 − y
Putting the values of y (0), y / (0), y // (0), y /// (0), y iv (0) in (4), we get
x2 x3 x4
y ( x) = 1 − x + + −
2 6 24
Hence,
= 0.9051625
= 0.8212667.
(0.3) 2 (0.3) 3 (0.3) 4
y (0.3) = 1 − (0.3) + + −
2 6 24
= 0.7491625.
= 0.6896.
Exact Solution:
dy
The given differential equation can be written as + y = x 2 a linear differential
dx
equation.
ye x = ∫ e x ( x 2 ) dx + c = ( x 2 − 2 x + 2)e x + c
y = ( x 2 − 2 x + 2) + ce − x (5)
y(0.1) = 0.9051625,
y(0.2) = 0.8212692,
y(0.3) = 0.7491817 and
y(0.4) = 0.6896799
The above solutions are tabulated as follows:
Assignments:
1. Use Taylor’s series method to find an approximate value of y at x = 0.1 of the IVP
dy
= x − y2, y ( 0) = 1 Answer : 0.9138
dx
2. Evaluate y(0.1) correct to six decimal places by Taylor’s series method if y(x) satisfies
y / = xy +1, y (0) = 1 Answer : 1.1053425
dy
3. Solve = x + y2, y = 1 at x = 0 using Taylor’s series method and compute y(0.1)
dx
and y(0.2) Answer : 1.1164 and 1.2725
dy
4. Using Taylor’s series method, compute the solution of = x + y, y (0) = 1 at the
dx
point x = 0.2 correct to three decimal places. Answer : 1.243
2. Modified Euler’s Method
dy
Consider the IVP = f ( x, y ), y ( x0 ) = y0 . The following two methods can be used to
dx
determine the solution at a point x = x n = x 0 + nh .
Euler’s Method :
y nE+1 = y n + hf ( x n , y n ), n = 0,1,2,3, (1)
y n +1 = y n +
h
2
[ ]
f ( x n , y n ) + f ( x n +1 , y nE+1 ) , n = 0,1,2,3, (2)
Remark:
1. The formulae (1) and (2) are also known as Euler’s Predictor – Corrector formula.
2. When Modified Euler’s method is applied to find the solution at a give point, we
first find the solution at that point by using Euler’s method and the same will be
used in the calculation of Modified Euler’s method. Also Modified Euler’s
method has to be applied repeatedly until the solution is stationary.
Problems :
dy y2
1. Solve =− , y (0) = 1 by Euler’s method by choosing h = 0.1 and h = 0.05.
dx 1+ x
Also solve the same problem by modified Euler’s method by choosing h = 0.05. Compare
the numerical solution with analytical solution.
Solution :
Analytical solution is : -
dy dx 1
=− ⇒ = log(1 + x) + c
y 2
1+ x y
(1) 2
y1 = y (0.05 ) = 1.0 − 0.05
1 + 0 = 0.95
(0.95 ) 2
y 2 = y (0.1) = 0.95 − 0.05
1 + 0.05 = 0.90702
(0.90702 ) 2
y 3 = y (0.15 ) = 0.90702 − 0.05
1 + 0 .1 = 0.86963
(0.86963 )
2
y 4 = y (0.2) = 0.86963 − 0.05
1 + 0.15 = 0.83675
y n2
Euler’s Formula is y n +1 = y n − 0.05 , n = 0,1,2 and 3
1 + xn
y2 y nE+1
Modified Euler Formula is y n +1 = y n − 0.025 n + , n = 0,1,2 and 3
1 + x n 1 + x n +1
(1) 2 (0.95 ) 2
y1(1) = y (0.05 ) = 1.0 − 0.025 1 + 0 1 + 0.05
+ = 0.95351
(1) 2
(0.95351 ) 2
y1( 2 ) = y (0.05 ) = 1.0 − 0.025
1 + 0 + = 0.95335
1 + 0.05
(1) 2 (0.95335 ) 2
y1( 3) = y (0.05 ) = 1.0 − 0.025 1 + 0 + 1 + 0.05
= 0.95336
(1) 2 (0.95336 ) 2
y1( 4 ) = y (0.05 ) = 1.0 − 0.025
1 + 0 +
= 0.95336
1 + 0.05
Hence y1 = y (0.05 ) = 0.95336
Stage – II: Finding y2 = y(0.1)
(0.95336 ) 2
y 2E = y (0.1) = 0.95336 − 0.05
1 + 0.05
= 0.91008
From Modified Euler’s formula, we have
(0.95336 ) 2 (0.91008 ) 2
y 2(1) = y (0.1) = 0.95336 − 0.025 1 + 0.05 + 1 + 0.1 = 0.91286
(0.95336 ) 2
(0.91286 )
2
y 2( 2 ) = y (0.1) = 0.95336 − 0.025
1 + 0.05 + 1 + 0.1 = 0.91278
(0.95336 ) 2 (0.91278 ) 2
y 2( 2 ) = y (0.1) = 0.95336 − 0.025
1 + 0.05 + 1 + 0.1
= 0.91278
(0.91278 ) 2
y 3E = y (0.15 ) = 0.91278 − 0.05
1 + 0.1
= 0.87491
From Modified Euler’s formula (for n = 2), we have
(0.91278 ) 2 (0.87491 ) 2
y 3(1) = y (0.15 ) = 0.91278 − 0.025 1 + 0.1 + 1 + 0.15 = 0.87720
(0.91278 ) 2
(0.87720 )
2
y 3( 2 ) = y (0.15 ) = 0.91278 − 0.025
1 + 0.1 + 1 + 0.15 = 0.87712
(0.91278 ) 2 (0.87712 ) 2
y 3( 3) = y (0.15 ) = 0.91278 − 0.025
1 + 0.1 + 1 + 0.15
= 0.87712
Hence y 3 = y (0.15 ) = 0.87712
Hence
y 4 = y (0.2) = 0.84550
2. Solve the following IVP by Euler’s modified method at 0.2 ≤ x ≤ 0.8 with h =
dy
0.2: = log 10 ( x + y ), y (0) = 2 .
dx
Solution:
Hence
y1 = y (0.2) = 2.0656
y 2(1) = y (0.4) = 2.0656 + 0.1[ log 10 (0.2 + 2.0656 ) + log 10 (0.4 + 2.1366 ] = 2.1415
y 2( 2 ) = y (0.4) = 2.0656 + 0.1[ log 10 (0.2 + 2.0656 ) + log 10 (0.4 + 2.1415 ] = 2.1416
y 2(3) = y (0.4) = 2.0656 + 0.1[ log 10 (0.2 + 2.0656 ) + log 10 (0.4 + 2.1416 ] = 2.1416
Hence
y 2 = y (0.4) = 2.1416
y 3(1) = y (0.6) = 2.1416 + 0.1[ log 10 (0.4 + 2.1416 ) + log 10 (0.6 + 2.2226 ] = 2.2272
y 3( 2 ) = y (0.6) = 2.1416 + 0.1[ log 10 (0.4 + 2.1416 ) + log 10 (0.6 + 2.2272 ] = 2.2272
Hence
y 3 = y (0.6) = 2.2272
Stage – IV: Finding y4 = y(0.8)
y 4(1) = y (0.8) = 2.2272 + 0.1[ log 10 (0.6 + 2.2272 ) + log 10 (0.8 + 2.3175 ] = 2.3217
y 4( 2 ) = y (0.8) = 2.2272 + 0.1[ log 10 (0.6 + 2.2272 ) + log 10 (0.8 + 2.3217 ] = 2.3217
Hence
y 4 = y (0.8) = 2.3217
xn yn
dy
3. Using modified Euler’s method solve the IVP = sin x + cos y, y ( 2.5) = 0
dx
at x = 3.5 in two steps, modifying the solution thrice at each stages. Here x is in
radians.
Solution:
y1(1) = y (3.0) = 0.0 + 0.25[ ( sin( 2.5) + cos( 0) ) + ( sin( 3.0) + cos( 0.7992 ) ) ] = 0.6092
y1( 2 ) = y (3.0) = 0.0 + 0.25[ ( sin( 2.5) + cos( 0) ) + ( sin( 3.0) + cos( 0.6092 ) ) ] = 0.6399
y1(3) = y (3.0) = 0.0 + 0.25[ ( sin( 2.5) + cos( 0) ) + ( sin( 3.0) + cos( 0.6399 ) ) ] = 0.6354
Hence
y1 = y (3.0) = 0.6354
Stage – II: Finding y2 = y(3.5)
y 2(1) = y (3.5) = 0.6354 + 0.25[ ( sin( 3.0) + cos( 0.6354 ) ) + ( sin( 3.5) + cos(1.10837 ) ) ] = 0.89572
y 2( 2) = y (3.5) = 0.6354 + 0.25[ ( sin( 3.0) + cos( 0.6354 ) ) + ( sin( 3.5) + cos( 0.89572 ) ) ] = 0.94043
y 2(3) = y (3.5) = 0.6354 + 0.25[ ( sin( 3.0) + cos( 0.6354 ) ) + ( sin( 3.5) + cos( 0.94043 ) ) ] = 0.93155
Hence
y 2 = y (3.5) = 0.93155
4. Using modified Euler’s method obtain the solution of the differential equation
dy
= x+ y with the initial condition y = 1 at x = 0 for the range 0 < x ≤ 0.6 in
dx
steps of 0.2.
Solution:
Given x0 =0, y 0 =1, h =0.2 and f ( x, y ) = x + y
To Find: y1 = y ( x1 ) = y (0.2), y 2 = y ( x 2 ) = y (0.4) and y 3 = y ( x3 ) = y (0.6)
Assignments :
The Taylor’s series method to solve IVPs is restricted by the difficulty in finding the
higher order derivatives. However, Runge – Kutta method do not require the calculations
of higher order derivatives. Euler’s method and modified Euler’s method are
Runge – Kutta methods of first and second order respectively.
dy
Consider the IVP = f ( x, y ), y ( x0 ) = y0 . Let us find the approximate value of y at
dx
x = x n +1 , n = 0,1,2,3,….. of this numerically, using Runge – Kutta method, as follows:
First let us calculate the quantities k1 , k 2 , k 3 and k 4 using the following formulae.
k1 = hf ( x n , y n )
h k
k 2 = hf x n + , y n + 1
2 2
h k
k 3 = hf x n + , y n + 2
2 2
k 4 = hf ( x n + h, y n + k 3 )
1
y n +1 = y n + ( k1 + 2k 2 + 2 k 3 + k 4 )
6
Problems:
1. Apply Runge – Kutta method, to find an approximate value of y when x = 0.2 given
dy
that = x + y, y (0) = 1 .
dx
Solution:
1
R – K method (for n = 0) is: y1 = y (0.2) = y 0 + ( k1 + 2 k 2 + 2 k 3 + k 4 ) -------- (1)
6
Now
k1 = hf ( x 0 , y 0 ) = 0.2 ×[0 +1] = 0.2
h k 0.2 0.2
k 2 = hf x0 + , y 0 + 1 = 0.2 × 0 + + 1 + = 0.2400
2 2 2 2
h k 0.2 0.24
k 3 = hf x0 + , y 0 + 2 = 0.2 × 0 + + 1 + = 0.2440
2 2 2 2
1
y1 = y (0.2) = 1 + ( 0.2 + 0.24 + 0.244 + 0.2888 ) = 1.2468
6
Hence the required approximate value of y is 1.2468.
dy y2 − x2
2. Using Runge – Kutta method of fourth order, solve = 2 , y (0) = 1 at x = 0.2
dx y + x2
& 0.4.
Solution:
y2 − x2
Given: x0 = 0, y 0 = 1, h = 0.2 and f ( x, y ) =
y2 + x2
1
y1 = y (0.2) = 1 + ( 0.2 + 2(0.19672 ) + 2(0.1967 ) + 0.1891 )
6
= 1+0.19599
= 1.19599
1
y 2 = y (0.4) = 1.19599 + ( 0.1891 + 2(0.1795 ) + 2(0.1793 ) + 0.1688 )
6
= 1.19599 + 0.1792
= 1.37519
3. Apply Runge – Kutta method to find an approximate value of y when x = 0.2 with
dy y
h = 0.1 for the IVP = 3x + , y (0) = 1 . Also find the Analytical solution and
dx 2
compare with the Numerical solution.
Solution:
y
Given: x 0 = 0, y 0 = 1, h = 0.1 and f ( x, y ) = 3 x +
2
Stage – I: Finding y1 = y (0.1)
1
R – K method (for n = 0) is: y1 = y (0.1) = y 0 + ( k1 + 2k 2 + 2k 3 + k 4 ) -------- (4)
6
= 1.0665242
= 1.0665242 + 0.1006976
= 1.1672218
Hence the required approximate value of y is 1.1672218.
Exact Solution
dy y
The given DE can be written as − = 3 x which is a linear equation whose solution is:
dx 2
x
y = −6 x −12 +13 e 2 .
The Exact solution at x = 0.1 is y(0.1) = 1.0665242 and at x = 0.2 is y(0.2) = 1.1672218
Both the solutions and the error between them are tabulated as follows:
xn yn yn Absolute
(Exact) (Numerical) Error
x y y
(Exact) (Numerical)
0.1 1.0665243 1.0665242
0.2 1.1672219 1.1672218
0.3 1.3038452 1.3038450
0.4 1.4782359 1.4782357
0.5 1.6923304 1.6923302
0.6 1.9481645 1.9481643
0.7 2.2478782 2.2478779
0.8 2.5937211 2.5937207
0.9 2.9880585 2.9880580
1.0 3.4333766 3.4333761
Assignments:
1. Apply Runge – Kutta method to compute an approximate value of y (0.2) and y(0.4), by
dy
taking h = 0.1 for the IVP 10 = x2 + y2 , y (0) = 1 . Answer: 1.0207,
dx
1.0438
2.Use Runge – Kutta method to find y when x = 0.2 in steps of 0.1 for the differential
dy y −x
equation dx = y + x , y (0) = 1 Answer: 1.0911278,
1.1678430
dy
3. Solve = ( y + x) , y (0.4) = 0.41 by Runge – Kutta method at x = 0.8 in two steps.
dx
Answer: 0.6103476,
0.8489914
4.Using Runge – Kutta method, find an approximate value of y for x = 0.2 in steps of 0.1
dy
if = x + y 2 , given that y = 1 when x = 0, Answer: 1.1165,
dx
1.2736
Multi-step Methods:
To solve a differential equation over an interval (xn, xn+1), using previous single-step
methods, only the values of y at the beginning of interval is required. However, in the
following methods, four prior values are needed for finding the value of yn+1 at a given
value of x. Also the solution at yn+1 is obtained in two stages. This method of refining an
initially crude estimate of yn+1 by means of a more accurate formula is known as
Predictor–Corrector method. A Predictor Formula is used to predict the value of yn+1
and then a Corrector Formula is applied to calculate a still better approximation of yn+1.
Now we study two such methods namely (i) Milne’s method and (ii) Adams – Bashforth
method.
4h
y 4( P ) = y 0 + [ 2 f1 − f 2 + 2 f 3 ]
3
Then we compute f 4 = f ( x 4 , y 4( P ) )
h
y 4(C ) = y 2 + f2 + 4 f3 +f
3 4
Then, an improved value of f4 is computed and again, corrector formula is applied to find
a better value of y4. We repeat the step until y4 remains unchanged.
dy
Given = f ( x, y ), y ( x 0 ) = y 0 . Using the given value of y ( x 0 ) = y 0 , we first
dx
compute
y ( x −1 ) = y ( x 0 − h) = y −1 , y ( x −2 ) = y ( x0 − 2h) = y −2 and y ( x −3 ) = y ( x 0 − 3h) = y −3 usin
g Taylor’s series method.
Next, we calculate f −1 = f ( x −1 , y −1 ), f −2 = f ( x −2 , y −2 ) and f −3 = f ( x −3 , y −3 ) .
Now, the value of y at x = x1 (or y1) can be determined in two stages:
h
y1( P ) = y 0 + [55 f 0 − 59 f −1 + 37 f −2 − 9 f −3 ]
24
h
y1( C ) = y 0 + [9 f1 + 19 f 0 − 5 f −1 + f −2 ]
24
Problems:
dy
1. Use Milne’s method to find y(0.3) for the IVP = x2 + y2 , y ( 0) = 1
dx
Solution:
First, let us find the values of y at the points x = -0.1, x = 0.1 and x = 0.2 by using
Taylor’s series method for the given IVP.
x 2 // x 3 ///
y ( x) = y (0) + xy / (0) + y ( 0) + y ( 0) -------- (1)
2 6
Given
y / ( x) = x 2 + y 2 ⇒⇒⇒ y / (0) = 0 + 1 = 1
y // ( x ) = 2 x + 2 yy / ⇒⇒⇒ y // (0) = 2 ×1 × 1 = 2
y /// ( x) = 2 + 2 yy // + 2( y / ) 2 ⇒⇒⇒ y /// (0) = 2 + 4 + 2 = 8
Using the values of y(0), y/(0), y//(0) and y///(0) in (1), we get
4x3
y ( x) = 1 + x + x 2 +
3
Putting x = -0.1, x = 0.1 and x = 0.2 in the above expression, we get
Given:
x 0 = −0.1, y 0 = 0.9087 and f 0 = 0.8357
x1 = 0, y1 = 1 and f 1 = 1
x 2 = 0.1, y 2 = 1.1113 and f 2 = 1.2449
x3 = 0.2, y 3 = 1.2507 and f 3 = 1.6043
To Find : y 4 = y ( x 4 ) = y (0.3)
4h
y 4( P ) = y (0.3) = y 0 + [ 2 f1 − f 2 + 2 f 3 ]
3
4(0.1)
= 0.9087 + [(2 ×1) −1.2449 + 2 ×1.6043 ]
3
= 1.4372
Now we compute f 4 = f (0.3,1.4372 ) = 2.1555
h
y 4( C ) = y (0.3) = y 2 + [ f2 + 4 f3 + f4 ]
3
0.1
y 4( C ) = y (0.3) = 1.1113 + [1.2449 + (4 ×1.6043 ) + 2.1555 ]
3
= 1.4386
Now, we compute f 4 = f (0.3,1.4386 ) = 2.1596
0. 1
y 4( C ,1) = y (0.3) = 1.1113 + [1.2449 + (4 ×1.6043 ) + 2.1596 ]
3
y(0.3) = 1.43869
dy
2. Given = x − y2, y (0) = 0 , y(0.2) = 0.02, y(0.4) = 0.0795 and y(0.6) = 0.1762.
dx
Compute y(1) using Milne’s Method.
Solution:
Stage - I : Finding y(0.8)
Given:
x 0 = 0, y 0 = 0 and f 0 = f ( x0 , y 0 ) = 0
x1 = 0.2, y1 = 0.02 and f 1 = f ( x1 , y1 ) = 0.1996
x 2 = 0.4, y 2 = 0.0795 and f 2 = f ( x 2 , y 2 ) = 0.3937
x3 = 0.6, y 3 = 0.1762 and f 3 = f ( x3 , y 3 ) = 0.56895
To Find : y 4 = y ( x 4 ) = y (0.8)
4h
y 4( P ) = y (0.8) = y 0 + [ 2 f1 − f 2 + 2 f 3 ]
3
4 ( 0 .2 )
=0+ [ (2 × 0.1996 ) − 0.3937 2 + 2 × 0.56895 ]
3
= 0.30491
Now we compute f 4 = f (0.8,0.30491 ) = 0.7070
h
y 4( C ) = y (0.8) = y 2 + f 2 + 4 f3 +f
3 4
0.2
y 4( C ) = y (0.8) = 0.0795 + [0.3937 + 4 × 0.56895 + 0.7070 ]
3
= 0.3046
0.2
y 4(C ,1) = y (0.8) = 0.0795 + [ 0.3937 + 4 × 0.56895 + 0.7072 ]
3
∴ y(0.8) = 0.3046
Given:
x1 = 0.2, y1 = 0.02 and f 1 = f ( x1 , y1 ) = 0.1996
x 2 = 0.4, y 2 = 0.0795 and f 2 = f ( x 2 , y 2 ) = 0.3937
x3 = 0.6, y 3 = 0.1762 and f 3 = f ( x3 , y 3 ) = 0.56895
x 4 = 0.8, y 4 = 0.3046 and f 4 = f ( x 4 , y 4 ) = 0.7072
To Find : y 5 = y ( x5 ) = y (1.0)
4h
y 5( P ) = y (1.0) = y1 + [2 f 2 − f3 + 2 f 4 ]
3
4(0.2)
= 0.02 + [(2 × 0.3937 ) − 0.56895 + 2 × 0.7072 ]
3
= 0.45544
Now we compute f 5 = f (1.0,0.45544 ) = 0.7926
h
y 5( C ) = y (1.0) = y 3 + [ f3 + 4 f4 + f5 ]
3
0.2
y 5( C ) = y (1.0) = 0.56895 + [0.56895 + 4 × 0.7072 + 0.7926 ]
3
= 0.4556
∴ y(1.0) = 0.4556
dy
3. Given = x 2 (1 + y ), y (1) = 1, y (1.1) = 1.233 , y (1.2) = 1.548 , y (1.3) = 1.979
dx
.Evaluate y(1.4) by Adam’s – Bashforth method.
Solution:
Given:
x −3 = 1, y −3 = 1 and f −3 = 2
x −2 = 1.1, y −2 = 1.233 and f −2 = 2.70193
x −1 = 1.2, y −1 = 1.548 and f −1 = 3.66912
x0 = 1.3, y 0 = 1.979 and f 0 = 5.03451
To Find : y1 = y ( x1 ) = y (1.4)
h
y1( P ) = y (1.4) = y 0 + [55 f 0 − 59 f −1 + 37 f −2 − 9 f −3 ]
24
(0.1)
= 1.979 + [ (55 × 5.03451 ) − (59 × 3.66912 ) + (37 × 2.70193 ) − (9 × 2)]
24
= 2.57229
Now we compute f 1 = f (1.4,2.57229 ) = 7.0017
h
y1(C ) = y (1.4) = y 0 + [9 f1 +19 f 0 − 5 f −1 + f −2 ]
24
0.1
= 1.979 + [(9 × 7.0017 ) + (19 × 5.03451 ) − (5 × 3.66912 ) + 2.70193 ]
24
y(1.4) = 2.57495
0.1
y1( C ,1) = 1.979 + [(9 × 7.0069 ) + (19 × 5.03451 ) − (5 × 3.66912 ) + 2.70193 ]
24
= 2.57514
Again f 1 = f (1.4, 2.57514 ) = 7.0073
0 .1
y1( C , 2 ) = 1.979 + [ (9 × 7.0073 ) + (19 × 5.03451 ) − (5 × 3.66912 ) + 2.70193 ]
24
y(1.4) = 2.57514
dy
4. Given = x 2 − y, y (0) = 1 . Find y(0.4) by Adam’s method.
dx
Solution:
First, let us find the values of y at the points x = 0.1, x = 0.2 and x = 0.3 by using
Taylor’s series method for the given IVP.
x 2 // x 3 ///
y ( x) = y (0) + xy / (0) + y ( 0) + y ( 0) -------- (2)
2 6
Given
y / ( x) = x 2 − y ⇒⇒⇒ y / (0) = 0 − 1 = −1
y // ( x) = 2 x − y / ⇒⇒⇒ y // (0) = 0 − (−1) = 1
y /// ( x) = 2 − 2 y // ⇒⇒⇒ y /// (0) = 2 − 1 = 1
Using the values of y(0), y/(0), y//(0) and y///(0) in (2), we get
x2 x3
y ( x) = 1 − x + +
2 3
Let x −3 = 0, y −3 = 1 and f −3 = −1
x −2 = 0.1, y −2 = 0.9051 and f −2 = 0.8951
x −1 = 0.2, y −1 = 0.8212 and f −1 = −0.7812
x 0 = 0.3, y 0 = 0.7492 and f 0 = −0.6592
To Find : y1 = y ( x1 ) = y (0.4)
I Stage : Predictor Method
h
y1( P ) = y (0.4) = y 0 + [55 f 0 − 59 f −1 + 37 f −2 − 9 f −3 ]
24
(0.1)
= 0.7492 + [55 × (−0.6592 ) − 59 × (−0.7812 ) + 37 × (−0.8951 ) − 9 × (−1)]
24
= 0.6896507
Now we compute f1 = f (0.4,0.6896507 ) = −0.5296507
h
y1( C ) = y (0.4) = y 0 + [9 f1 +19 f 0 − 5 f −1 + f −2 ]
24
0. 1
= 0.7492 + [9 × (−0.5297 ) +19 × (−0.6592 ) − 5 × (−0.7812 ) − 0.895125 ]
24
y(0.4) = 0.6896522
Assignments:
1. Using Adam’s method find y(1.4) for the IVP x2y/ + xy = 1; y(1) = 1, y(1.1) =
0.996, y(1.2) = 0.986, y(1.3) = 0.972.
dy
2. Given = 0.5 xy ; y (0) = 1 . Find y(0.4) using Adam’s – Bashforth predictor –
dx
corrector formulae.
3. Solve by Milne’s predictor – corrector method, the differential equation
dy
= y − x 2 with the following starting values : y(0) = 1, y(0.2) = 1.12186,
dx
y(0.4) = 1.4682, y(0.6) = 1.7379 to find the value of y when x = 0.8.
4. Using Milne’s method, obtain the solution of the equation
y = (1 + x 2 ) y 2 ; y (0) = 1 at
/ 1
x = 0.4.
2
Answers : (1) 0.94934, (2) 1.0408, (3) 2.0111 & (4) 1.2797
LECTURE NOTES OF
ENGINEERING MATHEMATICS–IV (Sub Code: 06 MAT41)
Text Book:
Higher Engineering Mathematics by
Dr. B.S.Grewal (36th Edition – 2002)
Khanna Publishers,New Delhi
Reference Book:
Advanced Engineering Mathematics by
E. Kreyszig (8th Edition – 2001)
John Wiley & Sons, INC. New York
SPECIAL FUNCTIONS
Prepared by
Dr. M. SANKAR
Professor and Head
Department of Mathematics
Sapthagiri College of Engineering
Bangalore – 560 057
Introduction
Many Differential equations arising from physical problems are linear but have variable
coefficients and do not permit a general analytical solution in terms of known functions.
Such equations can be solved by numerical methods (Unit – I), but in many cases it is
easier to find a solution in the form of an infinite convergent series. The series solution of
certain differential equations give rise to special functions such as Bessel’s function,
Legendre’s polynomial. These special functions have many applications in engineering.
d2y dy
x2 2
+x + ( x2 − n2 ) y = 0 (i)
dx dx
where n is a non negative real constant or parameter.
dx 2
= ∑
r =0
a r ( k + r )( k + r − 1) x k +r −2
( )∑ a x
∞ ∞ ∞
x 2 ∑ a r (k + r )( k + r − 1) x k +r −2 +x ∑ a r ( k + r ) x k +r −1 + x 2 − n 2 r
k +r
=0
r =0 r =0 r =0
∞ ∞ ∞ ∞
i.e., ∑ a r (k + r )( k + r − 1) x +∑ a r ( k + r ) x k +r +∑ a r x k +r +2 − n 2 ∑ a r x k +r = 0
k +r
r =0 r =0 r =0 r =0
Grouping the like powers, we get
[ ]
∞ ∞
∑ a r (k + r )(k + r − 1) + (k + r ) − n 2 x k +r + ∑ a r x k +r +2 = 0
r =0 r =0
∑ a [( k + r ) ]
∞ ∞
r
2
− n 2 x k + r + ∑ a r x k +r +2 = 0 (iii)
r =0 r =0
Equating the coefficient of xk from the first term and equating it to zero, we get
[ ]
a 0 k 2 − n 2 = 0. Since a 0 ≠ 0, we get k 2 − n 2 = 0, ∴ k = ±n
Coefficient of xk+1 is got by putting r = 1 in the first term and equating it to zero, we get
[ ]
i.e., a1 (k + 1) 2 − n 2 = 0. This gives a1 = 0, since (k + 1) 2 − n 2 = 0 gives, k + 1 = ± n
which is a contradiction to k = ± n.
Let us consider the coefficient of xk+r from (iii) and equate it to zero.
[ ]
i.e, a r (k + r ) 2 − n 2 + a r − 2 = 0.
− a r −2
∴ ar =
(k + r ) 2 − n 2 [ ] (iv)
i.e., a1 = a5 = a7 = …… = 0
∞
y = ∑ a r x k + r = x k ( a 0 + a 1 x + a 2 x 2 + a 3 x 3 + a 4 x 4 + )
r =0
a0 a0
∴ y1 = x n a 0 − x2 + x 4 −
4( n + 1) 32( n + 1)( n + 2)
x 2
x 4
i.e., y1 = a 0 x n 1 − 2 + 5 − (v)
2 (n + 1) 2 (n + 1)( n + 2)
This is a solution of the Bessel’s equation.
x2 x4
y 2 = a 0 x −n 1 − 2 + 5 − (vi)
2 ( −n + 1) 2 ( −n + 1)( −n + 2)
The complete or general solution of the Bessel’s differential equation is y = c1y1 + c2y2,
where c1, c2 are arbitrary constants.
Now we will proceed to find the solution in terms of Bessel’s function by choosing
1
a0 = and let us denote it as Y1.
2 n
( n +1)
xn x 2 1 x
4
1
i.e., Y1 = 1 − + −
2n 2 (n +1) 2 (n +1)( n + 2) ⋅ 2
(n +1)
x
n 1 x
2
1 x
4
1
= − + −
2 (n +1) (n +1) (n +1) ( n +1)( n + 2) (n +1) ⋅ 2
2 2
We have the result Γ (n) = (n – 1) Γ (n – 1) from Gamma function
Hence, Γ (n + 2) = (n + 1) Γ (n + 1) and
Γ (n + 3) = (n + 2) Γ (n + 2) = (n + 2) (n + 1) Γ (n + 1)
x
n 1 x
2
1 x
4
1
Y1 = − + −
2 (n +1) (n + 2) ( n + 3) ⋅ 2
2 2
which can be further put in the following form
x
n ( −1) 0 x
0
( −1)1 x
2
( −1) 2 x
4
Y1 = + + +
2 (n +1) ⋅ 0!
(n + 2) ⋅1! 2 (n + 3) ⋅ 2! 2
2
n ∞ 2r
x ( −1) r x
=
2
∑ (n + r +1) ⋅ r! 2
r =0
∞ n +2 r
x 1
= ∑(−1) r ⋅ ⋅
r =0 2 ( n + r +1) ⋅ r!
This function is called the Bessel function of the first kind of order n and is denoted by
Jn(x).
∞ n +2 r
x 1
Thus J n ( x ) = ∑( −1) r ⋅ ⋅
r =0 2 (n + r +1) ⋅ r!
∞ n +2 r
x 1
Proof : By definition, J n ( x ) = ∑(−1) r ⋅ ⋅
r =0 2 ( n + r +1) ⋅ r!
n +2 r
∞ x 1
∴ J n ( −x ) = ∑( −1 ) r ⋅− ⋅
r =0 2 ( n +r +1 ) ⋅ r!
n +2 r
∞ x 1
i.e., = ∑( −1 ) r ⋅ ( −1) n +2 r ⋅
r =0 2 ( n +r +1 ) ⋅ r!
n +2 r
∞ x 1
= ( −1) n ∑( −1 ) r ⋅ ⋅
r =0 2 ( n +r +1 ) ⋅ r!
n
Thus, J n (-x) = ( −1) J n (x)
Recurrence Relations:
Recurrence Relations 1:
d
dx
[ ]
x n J n ( x ) = x n J n −1 ( x )
From definition,
n +2 r 2( n +r )
∞ x 1 ∞ x 1
x n J n ( x ) = x n ∑( −1 ) r ⋅ ⋅ = ∑( −1 ) r ⋅ ⋅
r =0 2 ( n +r +1 ) ⋅ r! r =0 2 ( n +r +1 ) ⋅ r!
[ ]
2( n +r )−1
d ∞ 2( n +r ) x
∴ x n J n ( x ) = ∑( −1 ) r ⋅
dx r =0 2 n +2 r ( n +r +1 ) ⋅ r!
∞ ( n +r ) x n +2 r −1
= x n ∑( −1 ) r ⋅
r =0 2 n +2 r −1 ( n +r ) ( n +r ) ⋅ r!
∞ ( x / 2 ) ( n −1) +2 r
= x n ∑( −1 ) r ⋅ = x n J n −1 ( x )
r =0 ( n −1 +r +1 ) ⋅ r!
d
Thus,
dx
[ ]
x n J n ( x ) = x n J n −1 ( x ) --------(1)
Recurrence Relations 2:
d
dx
[ ]
x −n J n ( x ) = −x −n J n +1 ( x )
From definition,
n +2 r
∞ x 1
x −n J n ( x ) = x −n ∑( −1 ) r ⋅ ⋅
r =0 2 ( n +r +1 ) ⋅ r!
2r
∞ x 1
= ∑( −1 ) r ⋅ ⋅
r =0 2 ( n +r +1 ) ⋅ r!
∴
d
[ ∞
]
x −n J n ( x ) = ∑( −1 ) r ⋅
2 r x 2 r −1
dx r =0 2 n +2 r ( n +r +1 ) ⋅ r!
∞ x n +1+2( r −1 )
= −x −n ∑( −1 ) r −1 ⋅
r =1 2 n +1+2( r −1 ) ( n +r +1 ) ⋅( r −1 )!
Let k = r – 1
∞ x n +1+2 k
= −x −n ∑( −1 ) k ⋅ = −x −n J n +1 ( x )
k =0 2 n +1+2 k ( n +1 +k +1 ) ⋅ k !
d
Thus,
dx
[ ]
x −n J n ( x ) = −x −n J n +1 ( x ) --------(2)
x
Recurrence Relations 3: J n ( x ) = [ J n −1 ( x ) + J n +1 ( x )]
2n
We know that
d
dx
[ ]
x n J n ( x ) = x n J n −1 ( x )
x −n J n/ ( x ) − nx −n −1 J n ( x ) = −x −n J n +1 ( x )
Dividing by –x–n we get − J n/ ( x ) + ( n / x )J n ( x ) = J n +1 ( x ) --------(4)
Adding (3) and (4), we obtain 2nJ n ( x ) = x[ J n −1 ( x ) + J n +1 ( x )]
x
i.e., [ J n −1 ( x ) + J n +1 ( x )]
Jn( x ) =
2n
1
Recurrence Relations 4: J n/ ( x ) = [ J n −1 ( x ) − J n +1 ( x )]
2
Subtracting (4) from (3), we obtain 2 J n/ ( x ) = [ J n −1 ( x ) − J n +1 ( x )]
1
i.e., [ J n −1 ( x ) − J n +1 ( x )]
J n/ ( x ) =
2
n
Recurrence Relations 5: J n/ ( x ) = J n ( x ) − J n +1 ( x )
x
This recurrence relation is another way of writing the Recurrence relation 2.
n
Recurrence Relations 6: J n/ ( x ) = J n −1 ( x ) − J n ( x )
x
This recurrence relation is another way of writing the Recurrence relation 1.
2n
Recurrence Relations 7: J n +1 ( x ) = J n ( x ) − J n −1 ( x )
x
This recurrence relation is another way of writing the Recurrence relation 3.
Problems:
2 2
Prove that ( a ) J 1 / 2 ( x ) = sin x ( b ) J −1 / 2 ( x ) = cos x
πx πx
By definition,
n +2 r
∞ x 1
J n ( x ) = ∑( −1 ) r ⋅ ⋅
r =0 2 ( n +r +1 ) ⋅ r!
Putting n = ½, we get
1 / 2 +2 r
∞ x 1
J 1 / 2 ( x ) = ∑( −1 ) r ⋅ ⋅
r =0 2 ( r + 3 / 2 ) ⋅ r!
x
2 4
1 x 1 x 1
J1/ 2( x ) = − + − --------(1)
2 Γ ( 3 / 2 ) 2 Γ ( 5 / 2 )1! 2 Γ ( 7 / 2 )2!
Using the results Γ (1/2) = √ π and Γ (n) = (n – 1) Γ (n–1), we get
π 3 π 15 π
Γ(3 / 2 ) = ,Γ ( 5 / 2 ) = ,Γ (7 / 2 ) = and so on.
2 4 8
Using these values in (1), we get
x 2 x2 4 x 4
8
J1/ 2( x ) = − + −
2 π 4 3 π 16 15 π .2
x 2 x3 x 5 2 x 3 x 5
= ⋅ x − + − = x − + −
2π x
6 120
xπ
3! 5!
2
J1/ 2( x ) = sin x
πx
Putting n = - 1/2, we get
−1 / 2 +2 r
∞ x 1
J −1 / 2 ( x ) = ∑( −1 ) r ⋅ ⋅
r =0 2 ( r +1 / 2 ) ⋅ r!
x
2 4
1 x 1 x 1
J −1 / 2 ( x ) = − + − --------(2)
2 Γ ( 1 / 2 ) 2 Γ ( 3 / 2 )1! 2 Γ ( 5 / 2 )2!
Using the results Γ (1/2) = √ π and Γ (n) = (n – 1) Γ (n–1) in (2), we get
2 1 x2 2 x 4 4
J −1 / 2 ( x ) = − + −
x π 4 π 16 3 π .2
2 x 2 x 4
= 1 − + −
xπ
2! 4!
2
J −1 / 2 ( x ) = cos x
πx
2. Prove the following results :
2 3 − x 2 3
(a) J5/ 2( x ) = sin x − cos x and
πx x 2 x
2 3 − x 2 3
(b ) J −5 / 2 ( x ) = cos x + sin x
πx x 2 x
Solution :
x
We prove this result using the recurrence relation J n ( x ) = [ J n −1 ( x ) + J n +1 ( x )] ------
2n
(1).
3
Putting n = 3/2 in (1), we get J 1 / 2 ( x ) + J 5 / 2 ( x ) = J3/ 2( x )
x
3
∴ J5/ 2( x ) = J3 / 2( x )− J1/ 2( x )
x
3 2 sin x − x cos x 2
i .e., J5/ 2( x ) = − πx sin x
x πx x
2 3 sin x − 3 x cos x − x 2 sin x 2 ( 3 − x 2 ) 3
J5/ 2( x ) = = sin x − cos x
πx x2 πx x 2 x
3
Also putting n = - 3/2 in (1), we get J −5 / 2 ( x ) + J −1 / 2 ( x ) = − J −3 / 2 ( x )
x
3 − 3 2 x sin x + cos x 2
∴ J −5 / 2 ( x ) = − J −3 / 2 ( x ) − J −1 / 2 ( x ) =
−
− cos x
x x πx x πx
2 3 x sin x + 3 cos x − x 2 cos x 2 3 3 −x2
i .e., J −5 / 2 ( x ) = = sin x + cos x
π x x2 π x x x 2
3. Show that
d
dx
[ ]
J n2 ( x ) + J n2+1 ( x ) =
2
x
nJ[ 2
n( x ) −( n +1 ) J n2+1 ( x ) ]
Solution:
L.H.S =
d
dx
[ ]
J n2 ( x ) + J n2+1 ( x ) = 2 J n ( x )J n/ ( x ) + 2 J n +1 ( x )J n/ +1 ( x ) ------- (1)
Hence,
d
dx
[ 2
] [
J n2 ( x ) + J n2+1 ( x ) = nJ n2 ( x ) −( n + 1 )J n2+1 ( x )
x
]
1
4. Prove that J 0// ( x ) = [ J 2 ( x ) − J 0 ( x )]
2
Solution :
1
We have the recurrence relation J n/ ( x ) = [ J n −1 ( x ) − J n +1 ( x )] -------(1)
2
1 1
Putting n = 0 in (1), we get J 0/ ( x ) = [ J −1 ( x ) − J 1 ( x )] = [ − J 1 ( x ) − J 1 ( x )] = −J 1 ( x )
2 2
Thus, J 0/ ( x ) = −J 1 ( x ) . Differentiating this w.r.t. x we get, J 0// ( x ) = −J 1/ ( x ) ----- (2)
1
Now, from (1), for n = 1, we get J 1/ ( x ) = [ J 0 ( x ) − J 2 ( x )] .
2
Using (2), the above equation becomes
1
[ J 0 ( x ) − J 2 ( x )]orJ 0// ( x ) = 1 [ J 2 ( x ) − J 0 ( x )] .
− J 0// ( x ) =
2 2
1
Thus we have proved that, J 0 ( x ) = [ J 2 ( x ) − J 0 ( x )]
//
2
2
5. Show that (a) ∫ J 3 ( x )dx = c − J 2 ( x ) − J 1( x )
x
(b) ∫ xJ 02 ( x )dx =
1 2 2
2
[
x J 0 ( x ) + J 12 ( x ) ]
Solution :
(a) We know that
d
dx
[ ]
x −n J n ( x ) = −x −n J n +1 ( x ) or ∫ x −n J n +1 ( x )dx = −x −n J n ( x )
------ (1)
Now,
2 −2 2 −2 −2
∫ J 3 ( x )dx = ∫ x ⋅ x J 3 ( x )dx + c = x ⋅ ∫ x J 3 ( x )dx − ∫ 2 x ∫ x J 3 ( x )dx dx + c [ ]
[ ] [ ]
= x 2 ⋅ − x −2 J 2 ( x ) − ∫ 2 x − x −2 J 2 ( x ) dx + c ( from (1) when n = 2)
2 2
= c − J 2 ( x ) −∫ J 2 ( x )dx = c − J 2 ( x ) − J 1 ( x ) ( from (1) when n = 1)
x x
2
Hence, ∫ J 3 ( x )dx = c − J 2 ( x ) − J 1 ( x )
x
1 1
(b) ∫ xJ 02 ( x )dx = J 02 ( x ) ⋅ x 2 − ∫ 2 J 0 ( x ) ⋅ J 0/ ( x ). x 2 dx (Integrate by parts)
2 2
1
= x 2 J 02 ( x ) + ∫ x 2 J 0 ( x ) ⋅ J 1 ( x )dx (From (1) for n = 0)
2
1 d
= x 2 J 02 ( x ) + ∫ xJ 1 ( x ) ⋅ [ xJ 1 ( x )]dx
2 dx
d
[ xJ 1 ( x )] = xJ 0 ( x ) from recurrence relation (1)
dx
=
1 2 2
2
1 1
[
x J 0 ( x ) + [ xJ 1 ( x )] 2 = x 2 J 02 ( x ) + J 12 ( x )
2 2
]
Generating Function for Jn(x)
x
( t −1 / t ) ∞
To prove that e 2 = ∑t n J n ( x )
n =−∞
or
x
If n is an integer then Jn(x) is the coefficient of tn in the expansion of e 2 ( t −1 / t ) .
Proof:
x
We have e 2 ( t −1 / t ) = e xt / 2 × e − x / 2t
n n +2 n +4 n +2 r
1 x 1 x 1 x ∞ x 1
= − + − = ∑( −1 )r = Jn( x )
n! 2 ( n + 1 )! 1! 2 ( n + 2 )! 2! 2 r =0 2 Γ ( n + r + 1 )r!
Similarly, if we collect the coefficients of t–n in the product, we get J–n(x).
x
( t −1 / t ) ∞
Thus, e 2 = ∑t n J n ( x )
n =−∞
[ ]
x
( t −1 / t ) ∞
Result: e2 = J 0 ( x ) + ∑ t n + ( −1 ) n t −n J n ( x )
n =1
Proof :
x
( t −1 / t ) ∞ −1 ∞
e 2 = ∑t n J n ( x ) = ∑t n J n ( x ) + ∑t n J n ( x )
n =−∞ n =−∞ n =0
∞ ∞ ∞ ∞
= ∑ t − n J − n ( x ) + J 0 ( x ) + ∑ t n J n ( x ) = J 0 ( x ) + ∑ t − n ( − 1 ) n J n ( x ) + ∑ t n J n ( x ) { J − n ( x ) = ( − 1 ) n J n ( x
n =1 n =1 n =1 n =1
[ ]
x
( t −1 / t ) ∞
Thus, e 2 = J 0 ( x ) + ∑ t n + ( −1 ) n t −n J n ( x )
n =1
Problem 6: Show that
1π
(a) J n ( x ) = ∫ cos( nθ − x sin θ )dθ , n being an integer
π 0
1π
(b) J 0 ( x ) = ∫ cos( x cos θ )dθ
π 0
(c) J 02 + 2 J 12 + 2 J 22 + J 32 + = 1
Solution :
[ ]
x
( t −1 / t ) ∞
We know that e 2 = J 0 ( x ) + ∑ t n + ( −1 ) n t −n J n ( x )
n =1
= J 0 ( x ) + tJ 1 ( x ) + t 2 J 2 ( x ) + t 3 J 3 ( x ) + + t −1 J −1 ( x ) + t −2 J −2 ( x ) + t −3 J −3 ( x ) +
Since J −n ( x ) = ( −1 )n J n ( x ) , we have
x
e2
( t −1 / t )
( ) ( )
= J 0 ( x ) + J 1 ( x )( t − 1 / t ) + J 2 ( x ) t 2 + 1 / t 2 + J 3 ( x ) t 3 − 1 / t 3 + -----
(1)
Let t = cosθ + i sinθ so that tp = cospθ + i sinpθ and 1/tp = cospθ - i sinpθ .
From this we get, tp + 1/tp = 2cospθ and tp – 1/tp = 2i sinpθ
Using these results in (1), we get
x
( 2 i sin θ )
e 2 = e ix sin θ = J 0 ( x ) + 2[ J 2 ( x ) cos 2θ + J 4 ( x ) cos 4θ + ] + 2i[ J 1 ( x ) sin θ + J 3 ( x ) sin 3θ + ]
-----(2)
ixsinθ
Since e = cos(xsinθ ) + i sin(xsinθ ), equating real and imaginary parts in (2) we get,
cos( x sin θ ) = J 0 ( x ) + 2[ J 2 ( x ) cos 2θ + J 4 ( x ) cos 4θ + ] ----- (3)
sin( x sin θ ) = 2[ J 1 ( x ) sin θ + J 3 ( x ) sin 3θ + ] ----- (4)
These series are known as Jacobi Series.
Now multiplying both sides of (3) by cos nθ and both sides of (4) by sin nθ and
integrating each of the resulting expression between 0 and π , we obtain
1π J n ( x ), n is even or zero
∫ cos( x sin θ ) cos nθdθ =
π 0 0 , n is odd
1π 0, n is even
and ∫ sin( x sin θ ) sin nθdθ =
π 0 J n ( x ), n is odd
π π π
, if p = q
Here we used the standard result ∫ cos pθ cos qθdθ = ∫ sin pθ sin qθdθ = 2
0 0 0 , if p ≠ q
From the above two expression, in general, if n is a positive integer, we get
1π 1π
Jn( x ) = ∫ [ cos( x sin θ ) cos nθ + sin( x sin θ ) sin nθ ] dθ = ∫ cos( nθ − x sin θ )dθ
π 0 π 0
(b) Changing θ to (π /2) θ in (3), we get
cos( x cos θ ) = J 0 ( x ) + 2[ J 2 ( x ) cos( π − 2θ ) + J 4 ( x ) cos( π − 4θ ) + ]
cos( x cos θ ) = J 0 ( x ) − 2 J 2 ( x ) cos 2θ + 2 J 4 ( x ) cos 4θ −
Integrating the above equation w.r.t θ from 0 to π , we get
π π
∫ cos( x cos θ )dθ = ∫ [ J 0 ( x ) − 2 J 2 ( x ) cos 2θ + 2 J 4 ( x ) cos 4θ − ]
0 0
π
π sin 2θ sin 4θ
∫ cos( x cos θ )dθ = J 0 ( x ) ⋅θ − 2 J 2 ( x ) + 2J 4 ( x ) − = J 0 ( x ) ⋅π
0 2 4 0
1π
Thus, J 0 ( x ) = ∫ cos( x cos θ )dθ
π 0
(c) Squaring (3) and (4) and integrating w.r.t. θ from 0 to π and noting that m and n
being integers
π π π
∫ cos ( x sin θ )dθ = [ J 0 ( x )] ⋅ π + 4[ J 2 ( x )] + 4[ J 4 ( x )] 2 +
2 2 2
0 2 2
2 π π
π
∫ sin ( x sin θ )dθ = 4[ J 1 ( x )] + 4[ J 3 ( x )] 2 +
2
0 2 2
π
2
[ 2 2 2
Adding, ∫ dθ = π = π J 0 ( x ) + 2 J 1 ( x ) + 2 J 2 ( x ) + J 3 ( x ) +
0
]
Hence, J 02 + 2 J 12 + 2 J 22 + J 32 + = 1
Proof:
When β = α , the RHS of (4) takes 0/0 form. Its value can be found by
considering α as a root of Jn(x) = 0 and β as a variable approaching to α . Then (4)
gives
1 αJ n/ ( α )J n ( β )
Lt ∫ xJ n ( αx )J n ( βx )dx = Lt
β →α 0 β →α β 2 −α 2
Applying L’Hospital rule, we get
αJ n/ ( α ) J n/ ( β ) 1 /
{ }
1 2
Lt ∫ xJ n ( αx )J n ( β x )dx = Lt = J n ( α ) --------(5)
β →α 0 β →α 2β 2
n
We have the recurrence relation J n/ ( x ) = J n ( x ) − J n +1 ( x ) .
x
n
∴ J n/ ( α ) = J n ( α ) − J n +1 ( α ). Since J n ( α ) = 0 , we have J n/ ( α ) = −J n +1 ( α )
α
1
Thus, (5) becomes Lt ∫ xJ n ( α x )J n ( β x )dx =
β →α 0
1 /
2
{ 1
}
J n ( α ) = { J n + 1 ( α )} 2
2
2
LECTURE NOTES OF
ENGINEERING MATHEMATICS–IV (Sub Code: 06 MAT41)
Text Book:
Higher Engineering Mathematics by
Dr. B.S.Grewal (36th Edition – 2002)
Khanna Publishers,New Delhi
Reference Book:
Advanced Engineering Mathematics by
E. Kreyszig (8th Edition – 2001)
John Wiley & Sons, INC. New York
SPECIAL FUNCTIONS
Prepared by
Dr. M. SANKAR
Professor and Head
Department of Mathematics
Sapthagiri College of Engineering
Bangalore – 560 057
Introduction
Many Differential equations arising from physical problems are linear but have variable
coefficients and do not permit a general analytical solution in terms of known functions.
Such equations can be solved by numerical methods (Unit – I), but in many cases it is
easier to find a solution in the form of an infinite convergent series. The series solution of
certain differential equations give rise to special functions such as Bessel’s function,
Legendre’s polynomial. These special functions have many applications in engineering.
r =0
/ //
2. Find the derivatives y and y from the assumed solution and substitute in to the
given DE which results in an infinite series with various powers of x equal to
zero.
3. Now equate the coefficients of various powers of x to zero and try to obtaina
recurrence relation from which the constants a0, a1, a2,…. can be determined.
4. When substituted the values of a0, a1, a2,…. in to the assumed solution, we get the
power series solution of the given DE in the form y = Ay1(x) + By2(x), where A
and B are arbitrary constants.
In general, The above type of DE can be solved by the following two methods:
r =0
constants to be determined. Here all the constants ar’s will be expressed in terms of a0
and a1 only.
Problems :
d2y
1. Obtain the series solution of the equation + y =0 ----- (1)
dx 2
∞
Let y = ∑ a r x
r
----- (2) be the series solution of (1).
r =0
∞ ∞
Hence, y = ∑ a r r x , y // = ∑ a r r (r − 1) x r − 2 ,
/ r −1
r =0 r =0
Now (1) becomes
∞ ∞
∑ ar r (r − 1) x r −2 +∑ ar x r = 0
r =2 r =1
− a0 − a1 − a2 a0 − a3 a
a2 = ; a3 = ; a4 = = ; a5 = = 1 ;
2 6 12 24 20 120
− a 4 − a0 − a5 − a1
a6 = = ; a7 = = ; and so on.
30 720 42 5040
Substituting these values in the expanded form of (1), we get,
y = a 0 + a1 x + a 2 x 2 + a3 x 3 + a 4 x 4 +
x2 x4 x6 x3 x5 x7
i.e., y = a 0 1 − + − + + a1 x − + − +
2 24 720 6 120 5040
x2 x4 x6 x3 x5 x7
Hence y = a 0 1 − + − + + a1 x − + − + is the required
2! 4! 6! 3! 5! 7!
solution of the given DE.
d2y dy
x2 2
+x + (x 2 − n2 ) y = 0 (i)
dx dx
where n is a non negative real constant or parameter.
∞
dy
Hence, = ∑a r ( k + r ) x k +r −1
dx r =0
d2y ∞
dx 2
= ∑
r =0
a r ( k + r )( k + r − 1) x k +r −2
+( x )∑ a x
∞ ∞ ∞
x 2
∑a
r =0
r (k + r )( k + r − 1) x k + r −2
+x ∑ a r ( k + r ) x
r =0
k + r −1 2
−n 2
r =0
r
k +r
=0
∞ ∞ ∞ ∞
i.e., ∑ a r (k + r )( k + r − 1) x +∑ a r ( k + r ) x k +r +∑ a r x k +r +2 − n 2 ∑ a r x k +r = 0
k +r
r =0 r =0 r =0 r =0
Grouping the like powers, we get
[ ]
∞ ∞
∑ a r (k + r )(k + r − 1) + (k + r ) − n 2 x k +r + ∑ a r x k +r +2 = 0
r =0 r =0
[ ]
∞ ∞
∑ a r (k + r ) 2 − n 2 x k + r + ∑ a r x k +r +2 = 0
r =0 r =0
(iii)
Now we shall equate the coefficient of various powers of x to zero
Equating the coefficient of xk from the first term and equating it to zero, we get
[ ]
a 0 k 2 − n 2 = 0. Since a 0 ≠ 0, we get k 2 − n 2 = 0, ∴ k = ±n
Coefficient of xk+1 is got by putting r = 1 in the first term and equating it to zero, we get
[ ]
i.e., a1 (k + 1) 2 − n 2 = 0. This gives a1 = 0, since (k + 1) 2 − n 2 = 0 gives, k + 1 = ± n
which is a contradiction to k = ± n.
Let us consider the coefficient of xk+r from (iii) and equate it to zero.
[ ]
i.e, a r (k + r ) 2 − n 2 + a r − 2 = 0.
− a r −2
∴ ar =
[
(k + r ) 2 − n 2 ] (iv)
−a 1
a3 = = 0 , a1 = 0
6n + 9
i.e., a1 = a5 = a7 = …… = 0
∞
y = ∑ a r x k + r = x k ( a 0 + a 1 x + a 2 x 2 + a 3 x 3 + a 4 x 4 + )
r =0
x2 x4
y 2 = a 0 x −n 1 − 2 + 5 − (vi)
2 ( −n + 1) 2 ( −n + 1)( −n + 2)
The complete or general solution of the Bessel’s differential equation is y = c1y1 + c2y2,
where c1, c2 are arbitrary constants.
Now we will proceed to find the solution in terms of Bessel’s function by choosing
1
a0 = and let us denote it as Y1.
2 n
( n +1)
xn x 2 1 x
4
1
i.e., Y1 = 1 − + −
2n 2 (n +1) 2 (n +1)( n + 2) ⋅ 2
(n +1)
x
n 1 x
2
1 x
4
1
= − + −
2 (n +1) (n +1) (n +1) ( n +1)( n + 2) (n +1) ⋅ 2
2 2
We have the result Γ (n) = (n – 1) Γ (n – 1) from Gamma function
Hence, Γ (n + 2) = (n + 1) Γ (n + 1) and
Γ (n + 3) = (n + 2) Γ (n + 2) = (n + 2) (n + 1) Γ (n + 1)
x
n 1 x
2
1 x
4
1
Y1 = − + −
2 (n +1) (n + 2) ( n + 3) ⋅ 2
2 2
which can be further put in the following form
x
n ( −1) 0 x
0
( −1)1 x
2
( −1) 2 x
4
Y1 = + + +
2 (n +1) ⋅ 0! (n + 2) ⋅1! 2 (n + 3) ⋅ 2! 2
2
n 2r
x ∞
( −1) r x
=
2
∑
r =0
( n + r +1) ⋅ r! 2
∞ n +2 r
x 1
= ∑(−1) r ⋅ ⋅
r =0 2 ( n + r +1) ⋅ r!
This function is called the Bessel function of the first kind of order n and is denoted by
Jn(x).
∞ n +2 r
x 1
Thus J n ( x ) = ∑( −1) r ⋅ ⋅
r =0 2 (n + r +1) ⋅ r!
∞ n +2 r
x 1
Proof : By definition, J n ( x ) = ∑(−1) r ⋅ ⋅
r =0 2 ( n + r +1) ⋅ r!
n +2 r
∞ x 1
∴ J n ( −x ) = ∑( −1 ) r ⋅− ⋅
r =0 2 ( n +r +1 ) ⋅ r!
n +2 r
∞ x 1
i.e., = ∑( −1 ) r ⋅ ( −1) n +2 r ⋅
r =0 2 ( n +r +1 ) ⋅ r!
n +2 r
∞ x 1
= ( −1) n ∑( −1 ) r ⋅ ⋅
r =0 2 ( n +r +1 ) ⋅ r!
Recurrence Relations:
Recurrence Relations 1:
d
dx
[ ]
x n J n ( x ) = x n J n −1 ( x )
From definition,
n +2 r 2( n +r )
∞ x 1 ∞ x 1
x n J n ( x ) = x n ∑( −1 ) r ⋅ ⋅ = ∑( −1 ) r ⋅ ⋅
r =0 2 ( n +r +1 ) ⋅ r! r =0 2 ( n +r +1 ) ⋅ r!
∴
d
[ ]
∞
x n J n ( x ) = ∑( −1 ) r ⋅
2( n +r ) x 2( n +r )−1
dx r =0 2 n +2 r ( n +r +1 ) ⋅ r!
∞ ( n +r ) x n +2 r −1
= x n ∑( −1 ) r ⋅
r =0 2 n +2 r −1 ( n +r ) ( n +r ) ⋅ r!
∞ ( x / 2 ) ( n −1) +2 r
= x n ∑( −1 ) r ⋅ = x n J n −1 ( x )
r =0 ( n −1 +r +1 ) ⋅ r!
d
Thus,
dx
[ ]
x n J n ( x ) = x n J n −1 ( x ) --------(1)
Recurrence Relations 2:
d
dx
[
x −n J n ( x ) = −x −n J n +1 ( x ) ]
From definition,
n +2 r
∞ x 1
x −n J n ( x ) = x −n ∑( −1 ) r ⋅ ⋅
r =0 2 ( n +r +1 ) ⋅ r!
2r
∞ x 1
= ∑( −1 ) r ⋅ ⋅
r =0 2 ( n +r +1 ) ⋅ r!
∴
d
[ ∞
]
x −n J n ( x ) = ∑( −1 ) r ⋅
2 r x 2 r −1
dx r =0 2 n +2 r ( n +r +1 ) ⋅ r!
∞ x n +1+2( r −1 )
= −x −n ∑( −1 ) r −1 ⋅
r =1 2 n +1+2( r −1 ) ( n +r +1 ) ⋅( r −1 )!
Let k = r – 1
∞ x n +1+2 k
= −x −n ∑( −1 ) k ⋅ = −x −n J n +1 ( x )
n +1+2 k
k =0 2 ( n +1 +k +1 ) ⋅ k !
Thus,
d
dx
[ ]
x −n J n ( x ) = −x −n J n +1 ( x ) --------(2)
x
Recurrence Relations 3: J n ( x ) = [ J n −1 ( x ) + J n +1 ( x )]
2n
We know that
d
dx
[ ]
x n J n ( x ) = x n J n −1 ( x )
x −n J n/ ( x ) − nx −n −1 J n ( x ) = −x −n J n +1 ( x )
Dividing by –x–n we get − J n/ ( x ) + ( n / x )J n ( x ) = J n +1 ( x ) --------(4)
Adding (3) and (4), we obtain 2nJ n ( x ) = x[ J n −1 ( x ) + J n +1 ( x )]
x
i.e., [ J n −1 ( x ) + J n +1 ( x )]
Jn( x ) =
2n
1
Recurrence Relations 4: J n/ ( x ) = [ J n −1 ( x ) − J n +1 ( x )]
2
Subtracting (4) from (3), we obtain 2 J n/ ( x ) = [ J n −1 ( x ) − J n +1 ( x )]
1
i.e., [ J n −1 ( x ) − J n +1 ( x )]
J n/ ( x ) =
2
n
Recurrence Relations 5: J n/ ( x ) = J n ( x ) − J n +1 ( x )
x
This recurrence relation is another way of writing the Recurrence relation 2.
n
Recurrence Relations 6: J n/ ( x ) = J n −1 ( x ) − J n ( x )
x
This recurrence relation is another way of writing the Recurrence relation 1.
2n
Recurrence Relations 7: J n +1 ( x ) = J n ( x ) − J n −1 ( x )
x
This recurrence relation is another way of writing the Recurrence relation 3.
Problems:
2 2
Prove that ( a ) J 1 / 2 ( x ) = sin x (b ) J −1 / 2 ( x ) = cos x
πx πx
By definition,
n +2 r
∞ x 1
J n ( x ) = ∑( −1 ) r ⋅ ⋅
r =0 2 ( n +r +1 ) ⋅ r!
Putting n = ½, we get
1 / 2 +2 r
∞ x 1
J 1 / 2 ( x ) = ∑( −1 ) r ⋅ ⋅
r =0 2 ( r + 3 / 2 ) ⋅ r!
x
2 4
1 x 1 x 1
J1/ 2( x ) = − + − --------(1)
2 Γ ( 3 / 2 ) 2 Γ ( 5 / 2 )1! 2 Γ ( 7 / 2 )2!
Using the results Γ (1/2) = √ π and Γ (n) = (n – 1) Γ (n–1), we get
π 3 π 15 π
Γ(3 / 2 ) = ,Γ ( 5 / 2 ) = ,Γ (7 / 2 ) = and so on.
2 4 8
Using these values in (1), we get
x 2 x2 4 x 4
8
J1/ 2( x ) = − + −
2 π 4 3 π 16 15 π .2
x 2 x3 x 5 2 x 3 x 5
= ⋅ x − + − = x − + −
2π x
6 120
xπ
3! 5!
2
J1/ 2( x ) = sin x
πx
Putting n = - 1/2, we get
−1 / 2 +2 r
∞ x 1
J −1 / 2 ( x ) = ∑( −1 ) r ⋅ ⋅
r =0 2 ( r +1 / 2 ) ⋅ r!
x
2 4
1 x 1 x 1
J −1 / 2 ( x ) = − + − --------(2)
2 Γ ( 1 / 2 ) 2 Γ ( 3 / 2 )1! 2 Γ ( 5 / 2 )2!
Using the results Γ (1/2) = √ π and Γ (n) = (n – 1) Γ (n–1) in (2), we get
2 1 x2 2 x 4
4
J −1 / 2 ( x ) = − + −
x π 4 π 16 3 π .2
2 x 2 x 4
= 1 − + −
xπ
2! 4!
2
J −1 / 2 ( x ) = cos x
πx
2. Prove the following results :
2 3 − x 2 3
(a) J5/ 2( x ) = sin x − cos x and
πx x 2 x
2 3 − x 2 3
(b ) J −5 / 2 ( x ) = cos x + sin x
πx x 2 x
Solution :
x
We prove this result using the recurrence relation J n ( x ) = [ J n −1 ( x ) + J n +1 ( x )] ------
2n
(1).
3
Putting n = 3/2 in (1), we get J 1 / 2 ( x ) + J 5 / 2 ( x ) = J3/ 2( x )
x
3
∴ J5/ 2( x ) = J3 / 2( x )− J1/ 2( x )
x
3 2 sin x − x cos x 2
i .e., J5/ 2( x ) = − sin x
x πx x πx
2 3 sin x − 3 x cos x − x 2 sin x 2 ( 3 − x 2 ) 3
J5/ 2( x ) = = sin x − cos x
πx x2 πx x 2 x
3
Also putting n = - 3/2 in (1), we get J −5 / 2 ( x ) + J −1 / 2 ( x ) = − J −3 / 2 ( x )
x
3 − 3 2 x sin x + cos x 2
∴ J −5 / 2 ( x ) = − J −3 / 2 ( x ) − J −1 / 2 ( x ) =
−
− π x cos x
x x π x x
2 3 x sin x + 3 cos x − x 2 cos x 2 3 3 −x2
i .e., J −5 / 2 ( x ) = = sin x + cos x
π x x2 π x x x 2
3. Show that
d
dx
[ ]
J n2 ( x ) + J n2+1 ( x ) =
2
x
nJ [ 2
n( x ) −( n +1 ) J n2+1 ( x ) ]
Solution:
L.H.S =
d
dx
[ ]
J n2 ( x ) + J n2+1 ( x ) = 2 J n ( x )J n/ ( x ) + 2 J n +1 ( x )J n/ +1 ( x ) ------- (1)
Hence,
d
dx
[ 2
] [
J n2 ( x ) + J n2+1 ( x ) = nJ n2 ( x ) −( n + 1 )J n2+1 ( x )
x
]
1
4. Prove that J 0// ( x ) = [ J 2 ( x ) − J 0 ( x )]
2
Solution :
1
We have the recurrence relation J n/ ( x ) = [ J n −1 ( x ) − J n +1 ( x )] -------(1)
2
1 1
Putting n = 0 in (1), we get J 0/ ( x ) = [ J −1 ( x ) − J 1 ( x )] = [ − J 1 ( x ) − J 1 ( x )] = −J 1 ( x )
2 2
Thus, J 0 ( x ) = −J 1 ( x ) . Differentiating this w.r.t. x we get, J 0// ( x ) = −J 1/ ( x ) ----- (2)
/
1
Now, from (1), for n = 1, we get J 1/ ( x ) = [ J 0 ( x ) − J 2 ( x )] .
2
Using (2), the above equation becomes
1
[ J 0 ( x ) − J 2 ( x )]orJ 0// ( x ) = 1 [ J 2 ( x ) − J 0 ( x )] .
− J 0// ( x ) =
2 2
1
Thus we have proved that, J 0// ( x ) = [ J 2 ( x ) − J 0 ( x )]
2
2
5. Show that (a) ∫ J 3 ( x )dx = c − J 2 ( x ) − J 1 ( x )
x
(b) ∫ xJ 02 ( x )dx =
1 2 2
2
[
x J 0 ( x ) + J 12 ( x ) ]
Solution :
(a) We know that
d
dx
[ ]
x −n J n ( x ) = −x −n J n +1 ( x ) or ∫ x −n J n +1 ( x )dx = −x −n J n ( x )
------ (1)
Now,
2 −2 2 −2 −2
[
∫ J 3 ( x )dx = ∫ x ⋅ x J 3 ( x )dx + c = x ⋅ ∫ x J 3 ( x )dx − ∫ 2 x ∫ x J 3 ( x )dx dx + c ]
[ ] [ ]
= x 2 ⋅ − x −2 J 2 ( x ) − ∫ 2 x − x −2 J 2 ( x ) dx + c ( from (1) when n = 2)
2 2
= c − J 2 ( x ) −∫ J 2 ( x )dx = c − J 2 ( x ) − J 1 ( x ) ( from (1) when n = 1)
x x
2
Hence, ∫ J 3 ( x )dx = c − J 2 ( x ) − J 1 ( x )
x
1 1
(b) ∫ xJ 02 ( x )dx = J 02 ( x ) ⋅ x 2 − ∫ 2 J 0 ( x ) ⋅ J 0/ ( x ). x 2 dx (Integrate by parts)
2 2
1 2 2
= x J 0 ( x ) + ∫ x 2 J 0 ( x ) ⋅ J 1 ( x )dx (From (1) for n = 0)
2
1 d
= x 2 J 02 ( x ) + ∫ xJ 1 ( x ) ⋅ [ xJ 1 ( x )]dx
2 dx
d
[ xJ 1 ( x )] = xJ 0 ( x ) from recurrence relation (1)
dx
=
1 2 2
2
1 1
[
x J 0 ( x ) + [ xJ 1 ( x )] 2 = x 2 J 02 ( x ) + J 12 ( x )
2 2
]
Generating Function for Jn(x)
x
( t −1 / t ) ∞
To prove that e 2 = ∑t n J n ( x )
n =−∞
or
x
If n is an integer then Jn(x) is the coefficient of tn in the expansion of e 2 ( t −1 / t ) .
Proof:
x
We have e 2 ( t −1 / t ) = e xt / 2 × e − x / 2t
[ ]
x
( t −1 / t ) ∞
Result: e2 = J 0 ( x ) + ∑ t n + ( −1 ) n t −n J n ( x )
n =1
Proof :
x
( t −1 / t ) ∞ −1 ∞
e 2 = ∑t n J n ( x ) = ∑t n J n ( x ) + ∑t n J n ( x )
n =−∞ n =−∞ n =0
∞ ∞ ∞ ∞
= ∑ t − n J − n ( x ) + J 0 ( x ) + ∑ t n J n ( x ) = J 0 ( x ) + ∑ t − n ( − 1 ) n J n ( x ) + ∑ t n J n ( x ) { J − n ( x ) = ( − 1 ) n J n ( x
n =1 n =1 n =1 n =1
[ ]
x
( t −1 / t ) ∞
Thus, e2 = J 0 ( x ) + ∑ t n + ( −1 ) n t −n J n ( x )
n =1
Problem 6: Show that
1π
(a) J n ( x ) = ∫ cos( nθ − x sin θ )dθ , n being an integer
π 0
1π
(b) J 0 ( x ) = ∫ cos( x cos θ )dθ
π 0
(c) J 02 + 2 J 12 + 2 J 22 + J 32 + = 1
Solution :
[ ]
x
( t −1 / t ) ∞
We know that e 2 = J 0 ( x ) + ∑ t n + ( −1 ) n t −n J n ( x )
n =1
= J 0 ( x ) + tJ 1 ( x ) + t 2 J 2 ( x ) + t 3 J 3 ( x ) + + t −1 J −1 ( x ) + t −2 J −2 ( x ) + t −3 J −3 ( x ) +
Since J −n ( x ) = ( −1 )n J n ( x ) , we have
x
e 2
( t −1 / t )
( ) (
= J 0 ( x ) + J 1 ( x )( t − 1 / t ) + J 2 ( x ) t 2 + 1 / t 2 + J 3 ( x ) t 3 − 1 / t 3 + ) -----
(1)
Let t = cosθ + i sinθ so that tp = cospθ + i sinpθ and 1/tp = cospθ - i sinpθ .
From this we get, tp + 1/tp = 2cospθ and tp – 1/tp = 2i sinpθ
Using these results in (1), we get
x
( 2 i sin θ )
e2 = e ix sin θ = J 0 ( x ) + 2[ J 2 ( x ) cos 2θ + J 4 ( x ) cos 4θ + ] + 2i[ J 1 ( x ) sin θ + J 3 ( x ) sin 3θ + ]
-----(2)
ixsinθ
Since e = cos(xsinθ ) + i sin(xsinθ ), equating real and imaginary parts in (2) we get,
cos( x sin θ ) = J 0 ( x ) + 2[ J 2 ( x ) cos 2θ + J 4 ( x ) cos 4θ + ] ----- (3)
sin( x sin θ ) = 2[ J 1 ( x ) sin θ + J 3 ( x ) sin 3θ + ] ----- (4)
These series are known as Jacobi Series.
Now multiplying both sides of (3) by cos nθ and both sides of (4) by sin nθ and
integrating each of the resulting expression between 0 and π , we obtain
1π J n ( x ), n is even or zero
∫ cos( x sin θ ) cos nθdθ =
π 0 0 , n is odd
1π 0, n is even
and ∫ sin( x sin θ ) sin nθdθ =
π 0 J n ( x ), n is odd
π π π
, if p = q
Here we used the standard result ∫ cos p θ cos q θd θ = ∫ sin pθ sin q θd θ = 2
0 0 0 , if p ≠ q
From the above two expression, in general, if n is a positive integer, we get
1π 1π
J n ( x ) = ∫ [ cos( x sin θ ) cos nθ + sin( x sin θ ) sin nθ ] dθ = ∫ cos( nθ − x sin θ )dθ
π 0 π 0
(b) Changing θ to (π /2) θ in (3), we get
cos( x cos θ ) = J 0 ( x ) + 2[ J 2 ( x ) cos( π − 2θ ) + J 4 ( x ) cos( π − 4θ ) + ]
cos( x cos θ ) = J 0 ( x ) − 2 J 2 ( x ) cos 2θ + 2 J 4 ( x ) cos 4θ −
Integrating the above equation w.r.t θ from 0 to π , we get
π π
∫ cos( x cos θ )dθ = ∫ [ J 0 ( x ) − 2 J 2 ( x ) cos 2θ + 2 J 4 ( x ) cos 4θ − ]
0 0
π
π sin 2θ sin 4θ
∫ cos( x cos θ )dθ = J 0 ( x ) ⋅θ − 2 J 2 ( x ) + 2J 4 ( x ) − = J 0 ( x ) ⋅π
0 2 4 0
1π
Thus, J 0 ( x ) = ∫ cos( x cos θ )dθ
π 0
(c) Squaring (3) and (4) and integrating w.r.t. θ from 0 to π and noting that m and n
being integers
π π π
∫ cos ( x sin θ )dθ = [ J 0 ( x )] ⋅ π + 4[ J 2 ( x )] + 4[ J 4 ( x )] 2 +
2 2 2
0 2 2
2 π π
π
∫ sin ( x sin θ )dθ = 4[ J 1 ( x )] + 4[ J 3 ( x )] 2 +
2
0 2 2
π
2
[ 2 2 2
Adding, ∫ dθ = π = π J 0 ( x ) + 2 J 1 ( x ) + 2 J 2 ( x ) + J 3 ( x ) + ]
0
Hence, J 02 + 2 J 12 + 2 J 22 + J 32 + = 1
Proof:
d
/
Similarly v = Jn(β x) gives v = [ J n ( βx )] = βJ n/ ( βx ) . Substituting these values in (3),
dx
we get
1 αJ n/ ( α ) J n ( β ) − βJ n ( α ) J n/ ( β )
∫ xJ n ( αx ) J n ( βx )dx = ------- (4)
0 β 2 −α 2
If α and β are the two distinct roots of Jn(x) = 0, then Jn(α ) = 0 and Jn(β ) = 0, and
π
hence (4) reduces to ∫ xJ n ( αx )J n ( βx )dx = 0 .
0
This is known as Orthogonality relation of Bessel functions.
When β = α , the RHS of (4) takes 0/0 form. Its value can be found by
considering α as a root of Jn(x) = 0 and β as a variable approaching to α . Then (4)
gives
1 αJ n/ ( α )J n ( β )
Lt ∫ xJ n ( αx )J n ( βx )dx = Lt
β →α 0 β →α β 2 −α 2
Applying L’Hospital rule, we get
1
Lt ∫ xJ n ( αx )J n ( β x )dx = Lt
β →α 0 β →α
αJ n/ ( α ) J n/ ( β ) 1 /
2β
{2
}
= J n ( α ) --------(5)
2
n
We have the recurrence relation J n/ ( x ) = J n ( x ) − J n +1 ( x ) .
x
n
∴ J n/ ( α ) = J n ( α ) − J n +1 ( α ). Since J n ( α ) = 0 , we have J n/ ( α ) = −J n +1 ( α )
α
{ }
1 1 / 1
J n ( α ) = { J n + 1 ( α )} 2
2
Thus, (5) becomes Lt ∫ xJ n ( α x )J n ( β x )dx =
β →α 0 2 2