Professional Documents
Culture Documents
Lecture note 1
Dr.Narasimha H Ayachit1 and Dr.G.Neeraja Rani2
What we observe is not nature itself, but nature exposed to our method of questioning -
---- Heisenberg
Classical mechanics and classical physics which are studied to understand the law of
nature in terms of what we see and feel about our surroundings is basically depends
upon our common sense. The overall performance of all these classical phenomena
depends upon the nature of basic constituents of these surroundings that is on atomic
and molecular behavior. The phenomena occur at these microstructures do not seems
to work on common sense. This is because of the concept of quantization of energy
which led to the branch of Quantum Mechanics further which led to the new
beginning of Modern Physics.
In this lecture note 1, a brief introduction to this concept has been given trying to drive
home the ‘duality of nature’ as given by de-Broglie hypothesis. This note also includes
some of the references to the work carried out earlier to de-Broglie’s hypothesis, which
are the basis of this hypothesis, like interference, photo-electric effect etc.
The study of any science and in particular Engineering deals with the acquiring
knowledge about the energy transfer from one place to another place in its original
form or in a form in which it is needed by converting the original energy in the most
efficient manner. The technology we say is better if the efficiency increases i.e. the
energy for efficient transfer, the mode of its propagation and knowledge about it are
1
essential features and hence it is always the rate equations of propagation of energy
becomes important, which has the knowledge of energy and medium in which it is
happening.
The initial study was on light energy and its propagation. Initially it was
assumed that light travels in a straight line (rectilinear propagation of light) which
was explained through Pin-Hole camera (Appendix1.1) , which could only explain
the formation of shadow and images (due to lenses and mirrors) through reflections,
refractions and total internal reflections. The Young’s double slit experiment
(Appendix1.2) lead to the wave theory of light through Hygen’s Wave theory and
wave theory. The polarization (Appendix1.3) of light lead to show that light is a
transverse Electro Magnetic radiation with electrical vector and magnetic vector
perpendicular to each other. The wave theory could also explain all those, which
through Plank’s radiation concepts lead to show that, the light radiation when
interacts with some alkali metals rather behave with particle nature i.e. energy in
the form of quanta (photons)( Appendix1.5), i.e. when more than one come together
also show that they act individually, contrary to the wave nature ( the wave nature
as per the principle of superposition act collectively when more than one wave act at
point and also its energy acts on surroundings). This was evident from the fact that
2
no electrons are liberated even when enormous number of photons are made to fall
on alkali metals, the frequency of whose are less than the threshold frequency and
only increase in the kinetic energy but no increase in the number of electrons when
the frequency of the photons is increased by keeping the density of photons same.
hν = W + (1/2)mv2
the electron and v is the velocity of the ejected electrons. Thus, one can conclude
that the photoelectric effect can be explained by considering that, the incident
energy behaves as small energy packets leading to one to one interaction between
photons and electrons. This lead to show that light some times behaving like waves
(under interference, diffraction and polarization) and some times as particle (Photo
effect. In this X-rays are made to collide with the electrons of a system and during
collision one can see that the energy exchange between two happens as if it is a
in this particular experiment, they behave like a particle which is inferred through
the shift in wavelength observed. Further, the explanation of Planck’s black body
radiation using Planck’s equation, that energy is in quanta, namely E = hν show that
3
WAVE – PARTICLE DUALISM (de-Broglie’s hypothesis)
that light behaves as waves. While the Photoelectric effect infers that the
light at the time of interaction show that they behave as quanta( particle) .
2. The X- rays through the process of diffraction show that they are waves,
while the same X-rays show that they are scattered as if they are particles
circumstances, then one can even expect that entities which ordinarily behaves as
circumstances”.
and p is momentum of the particle/ wave. λ is the wave length, known as de-
Broglie wavelength.
4
de-Broglie hypothesis opened up a new thinking in almost all the
fields of Physics. In fact it can be treated as the new beginning of the Modern
think that the electron just do not follow the path of the circle but has to go
whizzing around this path in the form of a wave. But this is not sufficient.
energy takes place and if this has to happen there is no other way of thinking
that the number of waves around the orbit should be an integer? Of course
this has been proved later leading to the selection rules of transitions.
5
Appendix I
Brewster.
presented below.
When light falls on an object, each point on its surface reflects light
rays in all the directions. If these rays are made to pass through a hole they
produce a inverted image of the object on the projection plane and this image
6
being at all parts proportional to the object clearly indicates the receiving of
because it does not involve aberrations and the one which works perfectly as
(same frequency, nearly the same amplitude and are always in phase with
7
The coherent waves spreading from the two pin holes S1 and S2
superimpose on each other and produces alternatively bright and dark bands
on the screen CD. When crest falls on crest or trough falls on trough
produces a bright band (maximum energy) and when crest falls on trough or
amplitude and hence the division of energy from the same source S and
1.3 POLARIZATION
conditioned compared to the natural state. Is it not right to wonder under the
state of polarization the system is under strain!!!? The greatest inference the
polarization of light has given is not only that light is a transverse wave, it is
an electromagnetic wave.
polarization etc are possible through the combination of plane polarized light
which can be achieved through different methods like use of Nicole prism,
polaroiders etc.,
8
Figure3. Plane polarization
Let the unpolarized light waves vibrating in all the planes perpendicular to
the direction of propagation are passed through the polarizers 1 and 2, when
both the slits are parallel to each other and are perpendicular to the direction
9
of propagation the plane polarization is observed. In any other direction the
frequency falls on a certain alkali metal surfaces it has been observed that the
metals in a vaccum tube and noting the current created in the vaccum tube to avoid
the influence of the surrounding on the experiment achieved the following results.
10
Here in are also presented the explanation how these results lead to the concept of
quantization of the energy associated with waves, here after called as photon and
their behavior while interacting with the matter as if they are particles but not
waves.
that below a certain frequency not a single electron was liberated and above this
frequency the electrons were liberated. This frequency was named as threshold
frequency. This threshold frequency is different for different metals. It is also found
that increase in the frequency of the incident light increases the energy of the ejected
electrons while their number remained the same. This indicates the transfer of light
energy to the electrons in one to one corresponding manner leading to think that the
energy associated with each light wave as quanta of energy referred earlier as
photon. This one to one interaction is obviously similar to the transfer of momentum
from particle to particle involving elastic collision. No release of electron below the
threshold frequency indicates the quanta of energy associated with photon is not
The experiment was also conducted by varying the intensity of light below
the threshold frequency. Under, no intensity of light it is found that the electrons are
liberated showing that they are not collectively been able to create an energy which
is sufficient to knock out a single electron from the metal surface. This non collective
11
effect of light waves clearly indicates the energy transfer is not in favor of wave
Above the threshold frequency it has been found that the number of
one interaction between photon and electrons which have interacted with each other
The above inferences drawn form the results made Einstein to propose the
hν = W+ (½)mv max2
Where ν is the frequency associated with the incident light, h being the
Planck’s constant and hence hν is the energy associated with the photon. W is
the work function given by hν0 with ν0 being threshold frequency. The last term
in the above equation is associated kinetic energy with the liberated electrons.
statements made in the literature stating that the energy hν is partly used to
12
to the electron through a perfect elastic collision and this electron which gets
excited converts this energy into the mechanical energy which is used for the
pocket which can not be broken in to parts and hence this energy associated
with each transition is quantized and this energy is known as a quanta. These
quanta of energy which gets transferred to the material after interaction in full,
otherwise where the momentum transfer does not take place is physically
13
MODERN PHYSICS
Lecture note 2
Dr.Narasimha H Ayachit1 and Dr.G.Neeraja Rani2
opening of scientific gold mine consisting of great research work. This is evident from
the scientific out put that led in between 1925-1928, for example;
Periodic table,
statistical laws. The statistical nature of the universe and the principle of
a free electron can not exist with in a nucleus. This inference drawn
14
explained the nuclear structure and later made us to stay in this world
elements were lying in the dust and nobody knew how they resulted from the atoms or
molecules and what information they convey i.e., this whole data was as if coming from a
black box. The understanding of this and its impact in the later developments through
quantum physics formed on the basis of de-broglie hypothesis stand to the testimony of
There is no end to make the list of implications. We can add a few if you want to
this non ending list like invention like LASER, concepts of symmetry and identity,
probability concept of wave function and very recent development of Nanoscience and
Nanotechnology.
15
Davisson Germer Experiment
The de-Broglie’s hypothesis and in fact the whole of the quantum theory which
units of hf and the concept that only this unit can interact in toto with matter was not well
accepted by all probably because of not able to visualize the association of a wave with a
particle. The two important experiments which removed all the doubts regarding this are
one Compton effect and the other most important one is Davisson Germer experiment.
In line with Bragg’s law one may infer that if diffraction has to happen there
should be a variation of intensity up and down as the angle of scattering increases. This
indicates the wave nature, while gradual decrease in the intensity with the scattering
angle indicates Rayleigh’s scattering in consistent with the particle nature. As the
electrons are well known as particles it is expected them to show the Rayleigh’s
scattering but as per the de-Broglie’s hypothesis if these electrons are made to move fast
they should behave like waves after scattering. As per the de-Broglie wave length
expression for electrons having momentum generated by applying a voltage of 54V the
Or λ = 1.66 X 10-10m.
Thus, as per condition at which diffraction has to take place we should have a
material having set of planes with interplanar spacing approximately around 1.66A 0.
16
Requirement of this value will slightly vary depending upon the applied voltage used to
accelerate electrons.
the dimensions mentioned above and to carry out the experiment in such a way that the
motion of electrons is not affected before scattering and after scattering. Davisson
Germer were successful in achieving this with nickel metal as a target by carrying out the
carried out in vaccum has been discussed in detailed along with the inferences drawn
below.
17
Experimental set up
It consists of a tungsten filament (F), variable voltage source, Nickel metal (N)
and an electron detector (D). When the filament is heated the electrons are liberated from
it by thermionic emission. These electrons are passed through narrow slits in order to get
a fine beam of electrons. This electron beam was accelerated and directed at the nickel
target which was mounted on a support by the applied voltage. When the electrons hit the
target metal they get scattered around the metal. The electron detector was mounted on an
arc so that it could be rotated to different angles (ф) to observe the scattered electrons
from the metal surface. Keeping the accelerating potential constant the intensity of
electrons was measured by varying the angle ф. Polar graphs (figure6) below show the
18
It was a great surprise to find that an intense peak at certain angle and increase in
the intensity with the applied accelerating potentials up to 54V and then decreases. The
maximum intensity for this peak was observed for 54V and ф=500.
If the electrons are considered to behave like particles then inconsistent with
Rayleigh’s scattering, in the above experiment the graph should not contain any peak
with intensity gradually increasing. Thus, the peak indicates the occurrence the
phenomena of diffraction and hence strongly indicate the wave nature of electrons, which
The interplanar spacing value is 0.91A0 for a set of planes under study in nickel
metal. With this interplanar spacing and with applied voltage of 54V Davisson Germer
found the diffraction observed at an angle of 65 0. With these values and using the Bragg’s
law (2dsinӨ = nλ) with n = 1, λ turns out to be 1.65X10 -10m. Comparing this value with
the one calculated using de-Broglie equation infer that electrons do behave like waves. It
also infers the duality of matter. As acceleration of electrons before the scattering stand
with the particle nature while diffraction observed after scattering stands for wave
nature.
19
Concept of Group velocity and Phase velocity.
Undoubtedly in the form of a wave which is a resultant of many waves almost having
nearly the same amplitude determined through the de-Broglie wave equation λ = h/p, the
value of which is being very small but having slightly different frequencies. That is what
one is trying to understand through an experiment which involves the motion of energy
through a wave, which is resultant of group of waves. Thus, any concept to study
involving de-Broglie hypothesis has to have its basis on group wave concept. However,
when momentum transfers take place, these individual waves are looked as quanta of
energy having a concept of particle. However, for an individual wave it is the change in
phase that becomes important. In view of these concepts and the fact that any energy
associated with motion of the system is visualized through its velocity, the present study
demands defining of three types of velocities, namely Group velocity, Particle velocity
Take some amount of water in a non rigid container in which each molecule is
under motion with respect to other molecules. When this container is moving with certain
velocity v the individual molecules will have velocity of motion very close to the velocity
of the container. Thus the individual velocities of the water molecules can be associated
with phase velocity and that of a container which is the mean of all the phase velocities to
be forming a group velocity. When this container transfers its momentum after impact to
some other body the body will be feeling the velocity of the container as if it is one
particle. This shows that the particle velocity and the group velocities are one and the
same. This can be further visualized in better manner assuming that this type of
20
containers are received continuously at equal intervals of time by the body which gets
impacted will only be able to sense the velocities of the individual containers rather than
feel concept of particle velocity equivalent to group velocity. Thus the group velocity of
Phase velocity
Describes how a peak (crest and troughs) of a wave moves along the direction of
propagation and the velocity with which it moves is known as phase velocity.
To understand the movement of the peak of a wave, let the weights are suspended
on the ends of the springs and all them bounce with same frequency but all of them start
slightly at different times. Then these weights will form a traveling sine wave as shown in
21
Consider a point on the sine wave at a particular instant of time the wave equation
After certain time say t0 if the point is displaced by a distance x the equation is
given by
Y = Asin[ω(t-t0)].
Where k is defined as the propagation constant or the wave number which is also
defined as k= 2Π/ λ.
‘v’ mentioned above as stated earlier is associated with the individual wave and
hence it is nothing but phase velocity. Thus phase velocity vphase = ω/k. ---------1.1
Group Velocity
When a group of two or more waves of nearly same frequencies are superimposed
on each other the resultant wave will have different properties from those of the
individual waves. Variation in the amplitude is found in the resultant wave which
To obtain an expression for group velocity one has to consider the superposition of
discussed the superposition of two waves which are a part of a group wave that is the
waves which have slightly different angular frequency which are superimposed along a
line. For simplicity only two waves are taken, whose amplitudes(say A) have been
22
If Y1 is the displacement of one wave at a particular instant of time(t) and
Considering another wave with a displacement Y2 at the same instant (t) and
position (x) at that of Y1 but having slightly different frequency (ω+Δω) and hence also
The resultant displacement due to the superposition of the above two waves is,
A −B A+B
Since, sinA + sinB = 2cos ( ) sin ( ),
2 2
∆ω ∆k 2ω + ∆ω 2k + ∆k
Y = 2Acos {( )t ( ) x} sin {( )t–( ) x}
2 2 2 2
As the difference in frequency of the two waves is very small because Y 1 and Y2
form part of a group Δω and Δk terms in the sin part can be neglected as they appear in
combination with 2ω and 2k which are quite large compared to these quantities. The
same can not be done in cosine part as these terms do not appear in combination with 2ω
∆ω ∆k
Y = 2Acos {( )t ( ) x} sin (ωt-kx) ------- (1.5)
2 2
23
From a look at the equations (1.2) and (1.5) one can conclude that under the
approximation mentioned above, the wave component of the original reference wave and
the resultant wave remains the same while the amplitude changes from
∆ω ∆k
‘A’ to ‘2Acos {( )t ( ) x}’
2 2
Figure 9.
From the figure it is evident that the amplitude of the resulting wave is as per the
cosine component of equation (1.5) and it reaches its maximum when cosine component
∆ω ∆k ∆ω ∆k
cos {( )t ( ) x} = 1 i.e. {( )t ( ) x} = 0
2 2 2 2
∆ω
x/t = ( ) , x being the position at a given instant t of the resultant
∆k
wave i.e. group wave, obviously x/t represents the group velocity (Vg). Under the
MODERN PHYSICS
24
Lecture note 3
Dr.Narasimha H Ayachit1 and Dr.G.Neeraja Rani2
de-Broglie in consistent with his hypothesis had given an expression for the
wavelength associated with the matter wave namely, λ = h/p where h is a plank’s
constant and p is the momentum associated with the matter wave (in some of the books
the above equation has been derived by equating, Plank’s quantized energy E = hν and
mass energy relation E = mc2, obtained through Einstein’s theory of relativity. That is by
putting, E = hν = hc/ λ = mc2 i.e. λ = h/ (mc). Treating m as the mass and c as the
velocity and hence mc as momentum, one may write λ = h/p. But the question here is m
is of what? and c is of what? Certainly both do not belong to the same quantity and hence
although the approach seems to give the right equation but it is not the right way to do it).
It has been proved that the de-Broglie wave length is in consistent with the
wavelength associated with matter waves, undoubtedly arising out of the duality and
hence out of group wave and particle concepts. Below is presented the derivation of
expression for de-Broglie wavelength with this concept. As per the Plank’s quantization
E = (1/2)mv2 + V -------(1.10)
25
By equating the above two equations,
hν = (1/2)mv2 + V --------(1.12)
The left hand side is synonymous with duality of energy as per the photo-electric
effect and the right hand side of the above equation stands for particle nature as the two
terms are associated with the motion of electrons liberated in the photoelectric effect.
Differentiating the equation for photo electric effect i.e. (1.12) we get
∴ (dν/v) = (m/h)dv
Integrating we get,
For a particle having a velocity zero the wavelength can be assumed to be infinite
i.e. λ = ∞ for v = 0. Substituting this in the above equation we obtain the constant of
1/λ = p/h
Please note v is the velocity of matter waves which are associated with those
26
Relation between group velocity and particle velocity
∴ dω = (2Π/h)dE ------(1.18)
By relativistic approach we can also obtain the same results as shown below.
27
2
v particle
by substituting, p= m0vparticle/ 1 − in (1.25) we get
c2
It can be seen that this result is in consistent with our argument put forth earlier
with an example.
(ii) They represent, the wave associated with a particle if the constituents are
electromagnetic radiation.
(iii) The wavelength associated with the particles λ = h/p where h is Plank’s
is the velocity of light and E is the energy associated with the photon.
(iv) The amplitude of these matter waves being very small and being the resultant
given time.
(v) Under the limiting case the relation between group velocity and the phase
matter waves have the velocity more than that of light. This observation compel
28
one to think that matter waves are not the waves which can be physically felt but
(vi) The velocity associated with the electromagnetic radiation remains constant
for all the wave lengths while the velocity of matter wave differs under different
29
In the sharp formulation of the law of causality-- "if we know the present
exactly, we can calculate the future"-it is not the conclusion that is wrong
but the premise.
--Heisenberg, in uncertainty principle paper, 1927
We are treating motion of a particle in the form of group. Thus any thing we want to
measure physically, what we see is because of the parameters connected with a group.
When we want to measure we try to individualize them and hence the deviation in the
measurement which exist between that of an individual parameter and a group
parameters. Thus to this extent we have always an uncertainty built up.
The effect seen even in an individual wave is because of the interaction of matter waves
with the system under study. As per the Planck’s quantum theory the energy hν cannot be
broken i.e. when an interaction takes place what part of the wave has come in interaction
is uncertain. Apart from this the proof can be done in Relativistic and Non-Relativistic.
According to “classical physics” It is easy to measure both the position and velocity of a
macroscopic system i.e. the future motion of a particle could be exactly predicted, or
"determined," from knowledge of its present position and momentum and all of the forces
acting upon it. But when it comes to microscopic system (subatomic particles) this is not
true as explained by the Heisenberg's uncertainty principle. According to Heisenberg,
because one cannot know the precise position and momentum of a particle at a given
instant, so its future cannot be determined. One cannot calculate the precise future motion
of a particle, but only a range of possibilities for the future motion of the particle.
Because a moving particle is associated with a group wave hence according to Born’s
probability interpretation the particle may be present any where with in the group wave.
If we consider the group wave is to be narrow, then the position of the particle can be
found accurately where as the inaccuracy in calculation of its velocity or momentum
increases. On the other hand if we consider the group wave to be wide, then its velocity
or momentum can be found accurately whereas there is lot of inaccuracy in locating the
particle. That is
30
Following Heisenberg's derivation of the uncertainty relations, one starts with a particle
moving all by itself through empty space. To describe thus, one would refer to certain
measured properties of the particle. Four of these measured properties which are
important for the uncertainty principle, are the position, its momentum, its energy, and
the time. These properties appear as "parameters" in equations that describe the particles
motion.
The uncertainty relations may be expressed as.
∆ x∆ p > h/2π −−− (1.27)
∆ E∆ t > h/2π −−− (1.28)
∆ L∆ Ө> h/2π −−− (1.29)
Where,
1 ∆ x is the uncertainty or imprecision (standard deviation) of the position
measurement.
2 ∆ p is the uncertainty of the momentum measurement in the x direction at the
same time as the x measurement.
3 ∆ E is the uncertainty in the energy measurement.
4 ∆ t is the uncertainty in the time measurement at the same time as the energy is
measured.
5 ∆ L is the uncertainty in the angular momentum measurement
6 ∆ Ө is the uncertainty in the angle measurement at the same time as the angular
momentum is measured
7 h is a constant from quantum theory known as Planck's constant, a very tiny
number.
8 Π is pi from the geometry of circles.
If we measure the position of a moving atomic particle with a great accuracy that is ∆ x
is very small almost tending to zero then, what happens to the measurement of the
momentum ∆ p, which we measure at the same instant? From the equation
∆ p>h/2π ∆ x
we can see that the uncertainty in the momentum measurement, ∆ p is very large, that is
it approaches infinity since ∆ x in the denominator is very small.
When two or more waves superimpose on each other forming a group wave the resultant
we get is
∆ω ∆k
Y = 2Acos {( )t ( ) x} sin (ωt-kx)
2 2
31
∆ω ∆k
Of the above equation the term 2Acos {( )t ( ) x} determines the amplitude of
2 2
the group wave. Thus the certainty lies in between two nodes of the wave shown in the
Fig.10.
The distance between these two consecutive nodes(x1,x2) can be obtained as follows. First
∆ω ∆k ∆ω
node is found at ( )t ( ) x1 = (2n+1) π / 2 and the second node at ( )t (
2 2 2
∆k
) x2 = {(2n+1) π / 2}+π . Therefore by finding the difference between these two one
2
∆k
can write ( ) (x1-x2) = π . By taking (x1-x2) = Δx, as the uncertainty in position one
2
can write Δx = 2π /Δk. But k = 2π/λ = 2πp/h.
Therefore Δk = (2πΔp)/h.
Δx Δp = h.
Extending this argument to two and three dimensions respectively one can prove,
Δx Δp = h/2π, Δx Δp = h/4π
Gamma-Ray Microscope
32
The one of the best experiment which can illustrate the uncertainty in the measurement of
two conjugate quantities simultaneously is the gamma ray microscope through resolving
power of a very high resolution microscope (Fig.11).
Fig.11
Δp=(2hν sinθ)/(c).
=h
This is along a point, if it is taken over an area in 2 dimensions it becomes h/2π and
over the space h/4π .
Applications of Uncertainty Principle
1. The non-existence of free electron in the nucleus.
The diameter of nucleus of any atom is of the order of 10-14m. If any electron is
confined within the nucleus then the uncertainty in its position (∆ x) must not be greater
than 10-14m.
33
According to Heisenberg’s uncertainty principle, equation (1.27)
∆ x ∆ p > h / 2π
This is the uncertainty in the momentum of electron and then the momentum of
the electron must be in the same order of magnitude. The energy of the electron
can be found in two ways one is by non relativistic method and the other is by
relativistic method.
Non-Relativistic method:
E = p2/ 2m
∴E = (1.055X10-20)2/ (2X9.11X10-31) J
= 0.0610X10-9J
= 3.8X108eV
The above value for the kinetic energy indicates that an electron with a
momentum of 1.055X10-20 kg.m /s and mass of 9.11X10-31kg to exist with in the
nucleus, it must have energy equal to or greater than this value. But the
experimental results on β decay show that the maximum kinetic an electron can
have when it is confined with in the nucleus is of the order of 3 – 4 Mev.
Therefore the free electrons cannot exist within the nucleus.
Relativistic method:
34
Where m0 is the particle’s rest mass and m is the mass of the particle with velocity
v.
Squaring the above equation we get,
= (m02v2c2)/ (c2-v2)
= (m02c4)
= c2(p2 + m02c2)
Substituting the value of momentum from eq(1.30) and the rest mass as =
9.11X10-31kg we get the kinetic energy of the electron as
The second term in the above equation being very small and may be neglected
then we get
E > 1.5X10-12 J
The above value for the kinetic energy indicates that an electron with a
momentum of 1.055X10-20 kg.m /s and mass of 9.11X10-31kg to exist with in the
nucleus it must have energy equal to or greater than this value. But the
experimental results on β decay show that the maximum kinetic an electron can
have when it is confined with in the nucleus is of the order of 3 - 4 Mev.
Therefore the electrons cannot exist within the nucleus.
35
Whenever the energy interacts with the matter the atoms get excited and the
excited atom gives up its excess energy by emitting a photon of certain frequency
which leads to the spectrum. Whatever may be the resolving power of the
spectrometer is used to record these spectral lines no spectral line is observed
perfectly sharp. The broadening in the spectral lines is observed due to the
indeterminateness in the atomic energies. To measure the energies accurately the
time (∆ t) required is more that is ∆ t tends to infinity or otherwise. According to
Heisenberg’s uncertainty relation
∆ E = h / 2π ∆ t
36
How does the uncertainty measured? It is through the probability that means de-
Broglie wave which is basically a group wave defined by group velocity, a
corresponding frequency wave number and de-Broglie wavelength has to be
represented by probability equation. Since it is being a wave Therefore any
mathematical model representing this shall be a combination of sine and cosine
terms which are basically complex in nature. This model shall represent the
probabilistic value of finding a particle at a given instant at a given point. Such a
representation is what is known as a wave function. This wave function does
physically means a representation of probabilistic value and since it involves sine
and cosine terms it has to be complex in its nature. This type of a wave function is
written as
ψ = Ae-i(ωt-kx) ------(1.35)
This wave function as mentioned earlier is complex and to obtain a real part
which gives a physical meaning is only through multiplying with by its complex
conjugate. This shows that the probability is measured through (ψψ*)1/2 where
ψ* complex conjugate of ψ.
The value of amplitude of ψ is determined through the following equations. For
an existing particle the probability of finding it is under the limits of boundary
conditions imposed
∞
1. Integral of ∫
−∞
ψ ψ* dψ is equal to one.
∞
2. For a non existing particle ∫
−∞
ψ ψ* dψ is equal to zero..
The above two equations strongly advocate towards quantized energy states for an
existing particle leading to its physical significance.
From the above discussion the characteristics of a waveform representing a de-
Broglie wave can be listed as below
1. Wave function is complex
2. It is Probabilistic in nature
∞
3. Its value is such that ∫ −∞
ψ ψ* dψ = 1
4. Since it is to find a probability through integration it means the function has to
be differentiable.
5. Since it is differentiable in the existing space it should be continuous.
6. It is not only continuous it should be single valued over the space.
1. Time dependent →These are the wave equations associated with de-Broglie
waves in which amplitude decreases with time. For example the motion of the
electron/hole across the pn junction of a semiconductor diode. The equation of
motion for this type of de-Broglie waves are known as Time Dependent
Schrödinger Wave equation
37
2. Time independent → these are the wave equations associated with de-Broglie
waves in which amplitude remains constant at all times. This is associated
with those systems where the energy is conserved in which the de-Broglie
wave forms a stationary wave. Example being the motion of an electron inside
a metal (topic discussed later in the form of a quantum model of a free
electron) in which the electron get confined to the metal as the potential in
side the metal is less than the potential outside the metal. However these
electrons move from one place to another place and form de-Broglie waves of
stationary type. The equation of motion written for such particles is known as
Time Independent Schrödinger wave equation.
Let us consider a particle having a mass ‘m’ associated with a de-Broglie wave of
wave length ‘λ’ and moving with a group velocity ‘v’ then λ = h/mv. Let this
wave equation be a stationary wave that is time independent type then the wave
function associated with this wave ψ can be written as
ψ = Ae-i(ωt-kx)
Since it forms a stationary wave the value of ψ changes with position rather than
time and hence to obtain the realistic differential equation the above wave
function is only differentiable with ‘x’ and ‘t’ is taken as constant,
d2ψ/dx2 = -i2k2Ae-i(ωt-kx)
= -k2Ae-i(ωt-kx)
= -k2 ψ
38
If E is the total energy of the particle and V is the potential energy associated with
the particle then KE = E –V
2
Therefore k2 = {8π m(E-V)}/h2
The solution of a equation of motion written for a de-Broglie wave has seen
earlier is trigonometrical in nature(complex). Thus any trigonometrical function
will have a most general solution and hence they are not unique. All these
possible solutions written in the form of wave functions are known as Eigen
functions. This can be seen from the fact that a particle which forms a stationary
wave in a given potential can have more than one wavelength and hence
frequency.
In view of the fact the particle can exist with different frequencies corresponding
to different Eigen functions naturally corresponds to different energies that is a
particle can have more than one energy at different timings satisfying the
Schrödinger wave equation are known as Eigen values.
Process of Normalization
Let us consider a particle of mass ‘m’ inside a potential well moving in a direction
say along the breadth of the potential well. Let us assume that this as X-axis and
let the breadth of the potential well be ‘L’ (Fig.12). Let this particle move along
X-axis with out doing any work in other words at any energy of the particle it is
moving in an equipotential space. Since always potential difference is a measure
of the difference between the potentials the potential difference inside the box can
39
be considered as zero. Thus the Schrödinger wave equation of time independent
type for this particle can be written as
Fig.12
2
d2ψ/dx2 + [(8π mE)/h2] ψ = 0.
This is taken as time independent type because the particle is confined to the
potential well and hence can be taken as a case where the energy dissipation does
not take place. This can only happen when formation of stationary waves takes
place.
As already known the solution of this type of Schrödinger wave equation will
have a solution in its complex form. That is it contains sine and cosine terms.
Thus, the most general solution can be for the wave function ψ associated with
particle can be written as
Note that here‘t’ is considered as zero because the solution is for the time
independent type.
ψ = 0 at x≤ 0 and also at x≥ L.
Using these boundary conditions in the solution of the wave function can be
written as
1. at x = 0, ψ = 0
Then B = 0
2. at x = L, ψ = 0
40
Therefore kL = nπ
Or k = (nπ )/ L
Since ‘n’ can take all possible integral values that means there is more than one
wave function. These all values of sin are called Eigen functions. For each Eigen
function we associate an Eigen value i.e. the allowed energies associated with that
particle.
We know that k = nπ / L
2
Then k2 = n2π /L2
2 2
Or 4π / λ 2 = n2π /L2 ( k = 2π / λ )
En are called Eigen values. ‘n’ can take any value i.e. energy can be any value
since it is an infinite potential well that is the energy gap between the
surroundings and the well is infinite. Below is represented a schematic diagram of
wave function associated with particle having n = 1,2,3…. (the position of a
particle shown at any point should not misunderstood as its exact position as it
41
only indicates the most probabilistic value). Infact, dealing with quantum
mechanics one has to discourage the interpretation of anything through a diagram.
However for understanding the concept and visualizing the things in the
probabilistic manner these pictures are represented.
For n = 1; E1 = h2/8mL2
For n = 2; E2 = 4h2/8mL2
And Ψ2 = (2/L)1/2sin(2π x)/L
with ψ2 = 0 at x = 0
ψ2 = (2/L)1/2 at x = L/4
ψ2 = 0 at x = L/2
ψ2 = - (2/L)1/2 at x = 3L/4
ψ2 = 0 at x = L
Fig.13
42
It may please be noted that
1. The wave formed are stationary waves
2. Energies are discrete
3. The gap between two energy levels is not uniform.
43
Traditionally light is defined as a radiation which when falls on objects and comes to our
eyes (reflection) gives a sense of vision. The modulation in the reflected light gives a lot
of information. However it will not be able to acquire the full knowledge as this light is
most unsystematic in nature. More scientifically light is defined as an electromagnetic
radiation which carries an information after reflection through the process of modulation
and which can be sensed by the use of proper detectors. Thus, basically light is the energy
in the form of electromagnetic radiation. This can happen only through the process of
emission, that is, a system coming from a higher energy state to a lower energy state. Of
course, all emissions need not give us light. The one which gives us light is said to
perform radiative transition and while the others are called non-radiative transitions
(example, cooling of water). The process involved in which one is interested is the
emission of radiation in the form of light. This light although have enormous good
characteristics has the following negative characteristics.
2. Not coherent
Amplification of any parameter of a system is defined as achieving the out put more than
the normal out put that is an efficient out put compare to its in put in a parameter in
which we are interested at a given instant of time. This can be achieved by the following
processes.
44
In case of light, for example, a small converging lens can produce a large amount of
heating effect. This is because of an amplification achieved through the process of
reduction of loss. The working of a table lamp with a reflector increases the intensity per
unit area on the table is an example of negative feed back. Increasing the intensity of light
in Phosphorescence is an example of external energy being fed. Thus, for any
amplification to happen the process is to be controlled from out side. This is known as
process of stimulation. If there is no control from out side the system behaves on its own
as per its spontaneous nature and hence a process that happens without control is
spontaneous process and the performance will be more erratic. In case of an ordinary
emission of light the properties as listed above exist because the process involved is a
spontaneous emission. Some how if this light is amplified that is the process is
controlled externally one would get an amplified light. This light expected to give us
other than the characteristics mentioned above through the process of amplification
achieved through stimulation as explained by Einstein and this light is called LASER an
acronym for Light Amplification by Stimulated Emission of Radiation.
1. Highly concentrated energy that is the energy per unit cross sectional area
should be very high.
3. Highly monochromatic. If the light contains more than one frequency naturally
they can not be in phase and hence the reduction of energy takes place. Thus it is
expected that to bring about coherence the source should be highly
monochromatic
4. Phase coherence, it is not just sufficient that light is monochromatic but every
ray in a beam should be in phase with respect to each other so that intensity do not
reduce. This is known as phase coherence
45
THEORY OF LASERS
To achieve Laser the basic requirement is to achieve amplification so as to get the
characteristics mentioned earlier. Thus one needs to achieve controlled emission against
spontaneous emission. Before achieving emission the system has to be excited to the
higher state that is in a typical two energy level system the atoms or molecules have to be
put to a higher energy level through the process of absorption of radiation. Let us say at
any given instant of time N1 be the number of atoms/molecules in the ground state and N2
be the number of atoms/molecules in the excited state corresponding to energy of energy
levels E1 and E2. The number of atoms/molecules are determined through the equations E1
– E2 = hν and according to Boltzman’s equation N2 = N1Exp (-h ν /kT), where h is the
Planck’s constant, ν is the frequency of the radiation, k is the Boltzman’s constant and T
is the temperature at which absorption is taking place.
Fig.1
The negative sign indicates that the N1 decreases as the absorption takes place. This is
also written as
B12 is known as Einstein’s coefficient of absorption. Once a system gets excited the
emission starts to takes place. The most probable is the one which happens on its own
that is spontaneous emission (Fig.2). Thus the rate of spontaneous emission (dN2/dt)sp is
only proportional to the number of atoms in the excited state i.e, N2.
46
Fig.2
Therefore dN2/dt α - N2
The negative sign indicates that the N2 decreases as the emission takes place. This is also
written as
Fig.3
(dN2/dt)st α - ρN2
The negative sign indicates that the N2 decreases as the emission takes place. This is also
written as
47
(dN2/dt)st = - B21ρN2 (1.3)
N2 = N1Exp(-h ν /kT) or N1/N2 = exp((h ν /kT) substituting this in the above equation
Since the above equation is for an equilibrium state the radiation density ‘ρ’ is the same
as given for a black body by Planck.
Where ‘ν’ is the frequency of the radiation, ‘c’ velocity of light, ‘k’ Boltzman’s constant
and T is the temperature of the system
Equation (1.9) gives the relation between Einstein’s coefficients. It may be here noted
that a system which has good absorption power that is higher value of Einstein’s
coefficient of absorption has also good ability to give stimulated emission.
48
Therefore
The equations (1.7) and (1.10) can be taken as the expression for energy density.
The direct implication of requirement of population inversion is that the laser can not be
achieved through a system which has two energy levels as it is a relation between only
those two energy levels. (However in the recent times people have achieved laser
through two energy level systems, the basic principle being that the rate of absorption is
made faster than the rate of emission. This can be achieved only when one type of laser
is used as the source to achieve another type of laser).
(i). Stable (Ground) state – It is that state in which if an atom/molecule exists for ever
unless forced upon to leave, that is, its life time is infinite.
(iii). Metastable state – The atom/molecule excited to this state spends time of about
milli second, that is, nearly 1010 times more than the unstable state, such states are known
as Metastable state. The emission from such states is known as delayed emissions or
49
Phosphorescence. Since life time is higher these states can be utilized to bring about
population inversion and made to interact with the external control so as to get a
stimulated emission.
Let an energy ‘hν/’ be made to fall on a system at energy state E1 (Fig.4) so as to get
excited to E2. E2 being unstable state part of the atoms/molecules comes back to E1 i.e.
some of the atoms /molecules come directly to E1 by emitting radiation while few gets in
to the state E3. The excitation from E1 to E2 is continuous and hence with time the
population in E3 goes on increasing as E3 is metastable state. A continuous absorption
between E1 and E2 and transitions between E2 and E3 the population in E3 goes on building
up till a stage reaches where population in E3 level becomes more than that in E1. After a
milli second when transition happens from metastable state to stable state i.e. E3 to E1 that
will be the amplified signal as the population in E3 is greater than the population in E1.
Note that the Boltzman equation has not been violated because absorption is between E 1
and E2 while the population inversion is achieved between E3 and E1. This amplified
signal (Phosphorescence) which comes due to the transition from E 3 to E1 say with a
frequency ‘ν’ is made to fall back on the system. This signal interacts with the
atoms/molecules in the metastable state resulting stimulated emission of radiation. These
emitted photons will have the same direction, same frequency and same phase with
respect to each other. This process is continuously repeated and hence an enough amount
of photons are emitted. A part of this is taken out as a source and the remaining is sent
back into the system for further continuous process. The part of the light that is coming
out is highly monochromatic, highly directional and in phase and hence it is coherence in
all respects and hence it is a Laser.
Apart from the above theoretical phenomena, the laser action is supported through
experimental setups and construction. The laser active material is kept in a cavity which
is long in length and smaller in breadth with two mirrors along its cross section with a
small hole at one side through which part of the laser beam emerges. A typical diagram is
presented as below (Fig.5). The mirrors used are dielectric mirrors with 99.9% reflection.
The energy to the system is given along its length of the cavity. The distance between the
two mirrors is so adjusted that it is exactly equal to integral multiples of amplified
frequency. This creates a resonance and helps in absorption of the signal leading to the
better stimulated emission. Since the length of the tube is long and two parallel mirrors
with a small hole in one of the mirrors are used, the light that is emerging through this
small hole becomes unidirectional.
50
Fig.4
Fig.5
51
Some of the terminologies used are as below
(i) Active material. The material which is responsible for laser action that is a
material in which amplified signal is attained through population inversion.
(ii) Resonator. The cavity in which the active material is placed and in which a
resonance is brought about by adjusting the two mirrors such that the distance
between them is integral multiples of wavelength of the amplified signal.
(iii) Pumping technique. The method of exciting the system to a higher state by
supplying external energy in the above example from E1 to E2 is what is
known as pumping. There are different methods of pumping. If the light
energy is used to excite the system then the pumping is called as optical
pumping, when the electrical energy is used it is called as electrical pumping,
chemical energy brings the chemical pumping and so on.
As can be understood from the above discussion the material to act as a laser has to meet
lot of conditions like B21 to be large, existence of metastable state etc. and hence the
choice of material becomes very limited.
In a similar manner the discussion can be extended to more than three energy level
systems. However, in any system transition from a metastable state to lower energy state
is essential. In case of four energy level system the transition is from metastable state to
another state which is mostly unstable from which nonradiative transition takes place to
the lower state. In view of this building of population inversion becomes much faster and
hence are found to be more efficient than three energy level systems.
52
Lecture notes 8
RUBY LASER
Ruby laser was fabricated by Maiman in 1960 and was the first laser to be
operated successfully. It consists of abundant amount of aluminum oxide with some of
the alumina ions, replaced by chromium ions. The concentration of chromium ions is
approximately 0.05% by weight. The aluminum oxide is known as host material which
absorbs the incident energy and transfers it to chromium ions, which take part in lasing
action. The energy level diagram is as shown in Figure 6. From the figure, it can be
observed that ruby laser is a three level laser.
Figure 7 shows a typical setup with a flash lamp (e.g., Xenon or Krypton) pumped
pulsed laser. As the absorption bands of the ruby laser and the emission spectra of
available flash lamps matches very well with each other makes the pumping of atoms to
higher energy levels more efficient. The helical flash lamp is surrounded by a cylindrical
reflector to direct the pump light onto the ruby efficiently. A typical ruby was in the form
of a rod of length about 2 to 20cm and diameter about 0.1 to 2cm. The ends of the rod
was grounded and polished to make them parallel and flat. One end of the rod was
completely silvered and the other end was partially silvered through which the laser beam
emerges.
When the light from the flash lamp falls on the ruby rod the chromium ions get
excited to E1 and E2 levels from the ground level. These excited chromium ions relax
rapidly through a non radiative transition to the pair of levels M, which is the upper laser
level. The life time of the level M which is metastable state is of the order of 3ms from
where the lasing action takes place by emitting a photon of wavelength 6943Ao.
One of the interesting things about ruby laser is that it is a pulsed laser. This is
because of the flash lamp used which in itself is pulsating type so in a given pulse the
population inversion gets built up as it reaches the threshold value till the next pulse is
absorbed. The population depletion takes place below the threshold value of inversion
and hence lasing action stops for a very small time. This leads to pulsating laser to come
out. A very close look at the spectrum of the out put, thus, gives a spike like structure.
In spite of ruby being an inefficient laser because it is a 3 energy level system still
it finds a large amount of applications because of the laser characteristics.
1. The absorption band of ruby laser matches very well with the emission
spectra of flash lamp in it self can be used for pumping.
2. It has long life time.
3. It has narrow line width.
4. Because it falls in the visible region it is used in photographic emulsion,
photodetectors and in pulsed holography.
53
Figure 6
Figure 7
He-Ne LASER
He-Ne laser was the first gas laser to be operated successfully and is the most
widely used with continuous power out put in the range of a fraction of mW to tens of
54
mW. In gases the atoms are characterized by sharp energy levels compared to those in
solids. Therefore an electrical discharge is used to pump the atoms instead of a flash lamp
or continuous high power lamp, which is used in a system having broad absorption band.
In case of external mirror type, to avoid the reflection from the ends of discharge
tube, the Brewster windows are attached at the two ends. These windows have almost
100% transmittance excluding very small absorption and scattering losses. The laser
beam gets polarized in the plane of incidence and transmits through the windows and this
polarized laser will be perpendicular to the plane of incidence gets reflected back. This
leads to linearly polarized beam as an out put. The orientation of the laser beam can be
varied either by rotating the tube or the laser itself.
Figure 8
55
Figure 9
When an electrical discharge is passed through the gas, the electrons, which are
accelerated collides with the He atoms in the discharge tube and excites them to a higher
energy levels i.e. to F2 and F3 (Figure 10). As the life time of these atoms being more (10-4
&5X10-6 s respectively) many He atoms get collected there and these atoms can not go to
directly to the ground energy level by emitting a photon as this transition is not optically
allowed. The energy levels F2 and F3 act as metastable levels and the energies of these
levels almost coincide with the E4 and E6 energy levels of Ne. The excited He atoms
collide (perfectly elastic collision) ground Ne atoms and transfers whole of energy to Ne
atom, excites them to E4 and E6 energy levels. As the gas mixture contains He and Ne in
the ratio 10:1, the Ne atoms are selectively populated to E4 and E6 levels.
The Ne atoms in E6 level take a transition to E3 level by emitting the very popular
radiation of He-Ne laser of wavelength 6328A0. In Ne atom the life time of E6 energy
level is very much greater than the life time of E3 level and hence the population
inversion is achieved between E6 and E3 levels. From E3 level the atom goes to E2
56
spontaneously as the life time of E3 is too small (~10-8s). The E2 energy level is a
metastable energy level and thus the Ne atoms get collected there. These atoms instead of
going to ground state by colliding with the walls of the tube may absorb the
spontaneously emitted radiation and there by get excited back to E3 level which effects
the achievement of population inversion and hence the efficiency of the laser. To avoid
the absorption of the radiation in E2 level the diameter of the tube is decreased to as
narrow as possible.
Two more possible transitions E4 to E3 and E6 to E5 can be observed in the energy
level diagram, which emits the radiations of wavelength 1.15µm and 3.39µm
respectively.
Figure 10
57
Lecture notes 9
APPLICATIONS OF LASERS
Lasers find variety of applications in various fields due to their properties which
are much different from the ordinary light. Few applications like laser welding, cutting,
drilling, measurement of pollutants in the atmosphere and holography are discussed here.
The major property of laser is its high power, high directionality. The character of high
directionality means the energy is concentrated with in the beam and hence the
precession in space is highly uniform. The major problem of a laser is its high energy a
major part of which goes as reflected light. So to find any application of laser the
remaining energy after reflection should be capable of interacting with the material. In
this part also a small amount of energy also gets wasted as transmission energy. In view
of all these one need to select a laser source that has very high energy and such lasers are
known as industrial laser. The examples of the industrial laser are CO2 laser and Nd-YAG
lasers. The combination of these lasers with robotics and optical fibers has made these
mechanical applications more diverse in nature.
LASER DRILLING
In case of a conventional drilling a drill bit is held to the specimen with large
amount of force between them apart from the force used to hold the drilling machine and
the specimen. This brings about large amount of mechanical friction and hence most of
the energy gets wasted in producing of heat, sound and other energies. The heat energy
produced creates an uneven distribution and in the process either the specimen gets
distorted or spoiled hence, proper lubricant is essential. Apart from this the specimen also
puts the force on the drill bit resulting in to deformation of drill bit. This leads to non-
uniform drills created by the same drill bit. Apart from this if the drill is needed to be
very small in diameter then having a drill bit with high mechanical strength of that size
becomes impossible but in the today’s world of miniaturization thus becomes essential.
These practical problems can be overcome by using high power lasers as there is no
mechanical contact between the laser and the specimen and the highly directional laser
can also be concentrated on a narrow space. Drilling can be done at any angle and
through very hard and brittle materials. However the drilling through lasers needs a high
precession both in space and time. Typical drilling through laser is as described below.
High energetic pulses of 10-4 to 10-3 s duration are made to fall on the material.
The heat generated due to these pulses evaporates the material locally, thus a hole will be
been formed. Nd-YAG laser is used for metals and the CO 2 laser is used for metallic and
non-metallic materials. The pulsed laser exposure is controlled through a time controller,
as any variation in this will result in not only the variation in depth but also in width.
Further, to avoid accumulation of molten metal in the drill a high-pressure inert gas
preferably argon gas is passed. A typical schematic diagram is shown in figure 11. The
drills created through Lasers are highly uniform. Since lasers can be concentrated on a
small space very small drills can be created.
58
Figure 11
LASER WELDING
Laser welding plays very important role to reduce the roughness of the melted
surface, eliminate mechanical effects and to limit the heat effected area in the material.
Laser welding are generally used for welding multilayer in which the thermal properties
differ at interfaces. Compared to the other types of welding such as arc welding and
electron beam welding the laser welding is more efficient because it is contact less
process, does not destruct the material and the structure remains homogeneous. The laser
welding process is as explained below.
59
Figure 12
LASER CUTTING
Laser beam is made to concentrate at a point and the specimen is moved with a
uniform speed. This creates molten state in the metal. Repeating this movement to and fro
for number of times the depth is made to go on increasing and ultimately the whole of the
specimen can be cut. Since laser can be concentrated on a very small space and the
surface energy is very smooth the cutting becomes uniform. This cutting through laser
light is facilitated using high pressure oxygen so that the efficiency becomes better. This
also helps in high uniform cutting. The combination of this system with the programmed
robot will give a better control over the whole process. Coupling the laser with NC
machine will make the three dimensional cutting possible. Surface after cutting will be
very smooth and hence no polishing is required. A typical schematic laser cutting
diagram is presented in figure13.
Figure 13
Of all the pollutions the air pollution in the atmosphere is considered to be the
most hazardous which basically results from the creation of dust, smoke, flash etc. These
pollutants are in the size of nano meters. The challenge here is not only the size of the
particle but also the very less quantity of pollutant. The conventional method of Chemical
analysis fails because of the less quantity and the practical limitations in the measurement
of quantity accurately. The accurate measurement becomes essential at lesser quantities
because even small variation of air pollutants can create drastic effect on a life cycle.
60
The other method of quantitative analysis is through the spectroscopic analysis
which depends on Kirchhoff’s law which states that a material which has a capability to
emit certain wavelengths at higher temperatures also has the capacity to absorb the same
wavelength at lower temperatures. These wavelengths are the characteristic of that
particular material and the amount of absorption that happens at lower temperature is
directly proportional to the concentration of the material. This is achieved usually by
making a continuous light to fall on the substance known as Absorption technique. By
measuring the absorption and the wavelength the quantitative and qualitative analysis of
the elements present are done.
But in case of air pollution the quantity is so small that even this method finds it
difficult to identify using an ordinary light source. Thus to achieve the accuracy, lasers
are used as a source of light because of their important properties like monochromatic,
high radiation density and highly concentrated energy so that it can even interact with the
smallest percentage of element present in the atmosphere. For such type of experiments
Dye laser or tunable lasers which gives continuous source of wavelengths, of course, one
wavelength at a given instant time are used. Further, to increase the concentration of the
pollutant the sample is collected at atmospheric pressure is compressed to high pressure,
so that the concentration of the pollutant per unit volume is increased so that the detection
becomes easier.
A laser light is split into two parts one is made to fall on the sample and reflected
back to the detector with the help of a concave dielectric mirror. The other beam is used
as reference beam and made to fall on the detector such that it is out of phase with the
one which comes through the sample. By comparing these two the variation in the
intensity at different wavelengths of Dye laser can be recorded. By knowing the
wavelengths the elements can be identified and by knowing the variation in the intensity
with reference to Nitrogen or Carbon dioxide or Oxygen, the concentration of the
pollutants can be measured. With these values the corresponding concentration of the
pollutants at atmospheric pressure can be calculated.
61
Lecture notes 10
HOLOGRAPHY
When we see around us we can feel that the objects are in three dimensions that
is, we have the ability to analyze the intensity distribution and the depth of the different
points of an object. In other words we have the ability to analyze the amplitude variation
and path difference between the two points. When a photograph of an object is taken all
that we see is the amplitude variation but no path difference. This leads to a two
dimensional view. In view of this, there is a necessity that a technique has to be
developed, such that when photograph is taken it should have the information about
amplitude variation and path difference between the two points. This is what is known as
Holography. The information about the path difference between the two points is nothing
but the information regarding the phase difference between the two rays coming from
two different points, That is, to have a holography what that we need is the information
regarding amplitude variation and phase variation. The only experiment which can give
us both these information is an interference experiment and hence the techniques of
holography has to essentially depend upon the interference phenomena.
Some times some sketches gives us a three dimensional feel and also with the
help of specially prepared spectacle, one can get three dimensions feeling. This should
not be understood as holography, they are just illusions in reality.
In principle holography can be had using multiple wave lengths of source but a
single color will give a better information regarding the phase difference and hence use of
monochromatic source of light is preferable. Further, whenever a difference is measured
if the reference is uniform, the measurement becomes clearer. Hence, a laser as a source
becomes very important in holography because not only it is monochromatic but also it
has both the phase and temporal coherence.
PRINCIPLE
Consider two coherent waves P and Q are made to fall on the photographic plate
with different angles of incidence as shown in the Figure14. Superposition of these two
waves results in interference fringes on the photographic plate which can be observed
after developing the film. These fringes carry the information regarding phase and
62
intensity of the beam. If once again the coherent waves P and Q are made to fall on the
film at the same positions then these waves pass through fringe pattern and the waves P/
and Q/ can be observed from the transmission side these waves appear as if they are the
continuation of the waves P and Q respectively (Figure 15). If one of the incident waves
say Q is blocked and only one wave P is made to fall on the film. This wave gets
diffracted by the fringes that are present on the film and the emergent wave gets split in
to two parts. One of the part gets deviated from its path and emerges as if it is the
continuation wave Q (Q/) which is not present there and the other part emerges as P/
Figure 16. In this way in holography, the arrangement is done to get one of the coherent
waves reflected from the object called object beam on to the photographic plate when the
reference beam is made to incident directly on the photographic plate in the same position
as explained above. The reference beam gets diffracted by the hologram and produces
secondary wavelets. These secondary wavelets superimpose on each other and produce
maxima in certain directions and generate a real and virtual images of the object. On the
transmission side one can view the real image and on the other side virtual image of a
three dimensional object (Figure 17).
Figure 14
63
Figure 15
Figure 16
64
Figure 17
In both the techniques part of the incident laser beam is made to fall on the object and
the other part on to the photographic plate which has been placed in front of the object.
65
Figure 18
Figure 19
Applications
1. Holographic Interferometry.
The one of the important testing process is non destructive testing of specimen
which basically depends upon interferometric methods. The nature of holographic
process to give objective analysis of a specimen when illuminated with a
reconstruction wave allows us to have interference leading to a method called double
exposure holographic interferometry. This uses the photographic process between the
object wave and a reference wave of a specimen which is under stress alternatively to
create a hologram. The holograms so created give us two object waves when
illuminated with reconstructive wave, the first one for an unstressed object and the
other one for the stressed object. When these are overlapped again they gives us the
interference pattern leading to the information about the qualitative and quantitative
stress distribution. Since interference is involved in the whole process the slightest
variation in the stress distribution can be analyzed.
66
reconstructed object of a strained specimen. If the object undergoes any strain, fringes
will be formed and this can be studied as function of time. Such type of studies is
used in study of vibrating objects like musical instruments etc.
2. Microscopy
3. Character recognition
Two laser beams one as reference and another as a sample having a constant
phase difference are superimposed such that they produce interference. The laser
source used will be plane polarized and hence alternatively bright and dark lines of
uniform thickness will be formed. Using this, holograms are produced which are
nothing but diffraction gratings. Such a uniform formation of gratings is impossible
as all other methods involve mechanical approach which needs a very high control.
5. Information coding
In the present era of quantum computing which involves optical coding requires
the information to be stored in a compact space and in more secured manner. This is
achieved through virtual storage of information through a hologram. The information
can be reconstructed and acquired through this hologram.
67
OPTICAL FIBERS
Two of the greatest developments of recent past are in communication and
Miniaturization. The contribution to this has come through different developments in
Science and Engineering. Polymer Science is one of such field, which has resulted into
the realization of an optical fiber.
Optical fiber, basically, does a job of efficient transfer of light signal from one
point to another. In past also the light was used for communication system so also at
present, for example traffic lights. The basic problem with light is that it has the property
of scattering and hence the distance through which it can be communicated. In any solid
medium if this light is made to confine, then also it is impossible to make the light to
travel from one point to another point with out much scattering. This is basically because
of the large number of scattering points (scattering points should not be considered as
some other particles in side the material but it is the change in the refractive index of the
material because of change in density at various points). Polymers are able to overcome
this, of course, in very low dimensions.
Optical fiber with such polymers works on the principle of Total Internal
Reflection and hence the material has to have more than one density, which is to be well
controlled. Although Total Internal Reflection was known long back, development in
transmission of light in a confined space could not be achieved because of non-
availability of technology described above. With the development of optical fiber
practically we find its applications in every field of Communication, Medicine, Industry
and sensors. The combination of optical fibers with lasers has revolutionized the Science
and Technology. In the following paragraphs the development in communication as one
of the examples is discussed.
Our conventional communication system was earlier through cable and later
through microwaves although they gave us a large advantage, have their own limitations.
For example through the cables, the communication is achieved by the modulation of
68
electrical signals, which are governed by the flow of electrons. As every one knows
electrons have mass and hence inertia. This in itself creates a problem in the speed of
communication. This speed not only depends upon the electrons but also on the material
of the cable, which results in to resistance. Apart from this, speed gets affected because
the electrons have charge hence they interact with the surroundings leading to loss of
energy. To overcome this, regular amplification is required. Because electrons have
inertia, whenever a cable bends the electrons are likely to penetrate from one cable to
another and hence cross hearing is possible. This is a severe problem in rainy season.
In a bundle of a wire through which electrons are moving, a part of that can be cut
and the information can be tapped and hence affects the security. For the reasons
mentioned above we have to use a copper wire or an alloy of it, which comes as a natural
limited resource. With increasing demand and decrease in natural resources it has put a
lot of problem in communication. There are several other drawbacks like weight of the
cable, durability, corrosion etc. But the greatest drawback of this cable communication
system is its limited bandwidth because each can transmit a very limited number signal
through it simultaneously.
The microwave signal, which belongs to wire less communication, has a problem
in its production basically because it is an electronic gadget and hence the speed of
production of such signals is slow. Earlier these signals were reflected through
ionosphere but at present through the satellite from one point to another. This distance
itself creates a lot of delay in the signal processing. More than any thing these signals are
left in space and whoever is capable of collecting it will be able to do so and hence there
is no security. Especially in an era in which we are talking of SKUD missiles and HIGH
END machines in defense the delay in signal transmission can create havoc and also the
insecurity. There are many more problems associated with the microwave technology one
of them again being low band width. Signal distortion is also very high in this system.
69
principle of Total Internal Reflection all the disadvantages observed in cable
communication and microwave communication system can be overcome. For example
light being a photon has no inertia and its speed is enormous. Since they are
electromagnetic radiation they do not interact with each other and hence minimum losses.
You have an infinite number of visible wavelengths and hence bandwidth is enormously
high. Since there is only one fiber for one modulated signal tapping is impossible as the
communication gets cut down. The optical fiber, which is used to send a signal, can also
be used to receive the signal. Since the material of an optical fiber is a polymer it can be
synthesized in abundance and it can have high mechanical strength to weight ratio etc.
Total Internal Reflection
Consider a ray S1 of light made to travel from a denser medium of refractive index
n1 to rarer medium of refractive index n2 (n1 > n2 ) which fall on the interface between two
mediums at an angle of incidence Ө1 with the normal. Due to the change in medium at
point ‘P’ the ray refracts at an angle r1. The angle of refraction will be more than the
angle of incidence as the ray passing from a denser to a rarer medium. If the angle of
incidence is increased from Ө1 to Ө2 the angle of refraction also increases from r1 to r2. As
the angle of incidence increased the angle of refraction also increases and at a particular
angle of incidence the angle of refraction becomes 900, that is, the refracted ray moves
along the interface between the two mediums. This angle of incidence is known as
critical angle and represented by Өc. Further increase in the angle of incidence increases
the angle of refraction greater than 900, which makes the ray to get back to the same
medium, which is known as total internal reflection (fig. 1a).
Figure 1a
70
According to Snell’s law n1 sinӨ1 = n2 sinӨ2 and When Ө1 = Өc , Ө2 = 900
Figure 1b
71
Angle of acceptance and Numerical Aperture
The angle of acceptance is defined as the maximum angle with which a ray can be
sent in to the optical fiber such that it suffers total internal reflection and reaches the other
end of the fiber.
Consider a ray SO is made to enter in to the optical fiber with an angle ‘Ө0’ with
the fiber axis. Let angle of refraction at ‘O’ is Ө1 = θ 0. After refraction the ray falls on
the interface between core and cladding mediums at point ‘A’ at an angle Ө2 [( 90- Ө1)].
If Ө2 is equal to the critical angle Өc then the ray travels along the interface between core
and the cladding mediums as shown in figure 2. Consider another ray PO is made to enter
the optical fiber with an angle of incidence greater than Ө0 with the fiber axis. The angle
of refraction at ‘O’ will be greater than Ө1 and further the ray falls at point ‘B’ on the
interface with an angle of incidence less than Ө2 and hence the ray escapes through the
cladding in to the surrounding medium. Now, Consider another ray QO is made to enter
the optical fiber with an angle of incidence less than Ө0 with the fiber axis. The angle of
refraction at O will be lesser than Ө1 and further the ray falls at point ‘C’ on the interface
with an angle of incidence greater than Ө2 that is greater than Өc and hence the ray suffers
total internal reflection and propagates through the core medium and reaches the other
end. Thus any ray that enters the optical fiber with an angle either equal to Ө0 or lesser
than Ө0 at ‘O’ propagates through the optical fiber and the other rays escapes through the
cladding. The angle Ө0 is called the angle of acceptance. If SO is rotated around the fiber
axis keeping Ө0 it produces a conical surface. Any ray sent in to the fiber with in this
conical surface is accepted by the fiber and hence Ө0 is also known as the acceptance
cone half angle.
Figure 2
72
The Numerical Aperture (NA) is defined as the light gathering capacity of an
optical fiber which is expressed as sin of angle of acceptance that is,
NA = sin Ө0
n1 cosӨ1 = n2 ,
The ray will be able to propagate through the fiber only when
sin Ө0 ≤ N.A.
73
Fractional Index Change (∆)
Is defined as the ratio of the refractive index difference between the core and
cladding to the refractive index of the core.
From the above equation it can be observed that increasing the value of ‘∆’, one
can increase the light gathering capacity of an optical fiber. But a large value of ‘∆’ leads
to intermodal dispersion, which causes signal distortion.
This optical fiber is made up of core of constant refractive index n1 with lesser
diameter and cladding of slightly lower refractive index n2. The name step index is
74
given because the refractive index profile (Fig.3) for this fiber makes a step change at
the interface between the core and the cladding medium. Single mode is because of
smaller diameter because of which the optical fiber can allow for propagation of only
few modes through it at a time. As the signal transmitted is less, the signal broadening
is not practically found which makes these fibers highly advantages. The signal
propagates almost along the axis of the core as shown in the figure. The source used
for coupling the signal in to the fiber can be only lasers because of its diameter. These
are widely used in sub marine cable systems.
Figure 3
The refractive index profile remains same as the step index single mode fiber
except for the core diameter which is almost five times greater than the single mode
fiber. Because of the larger core diameter this fiber allows many modes to propagate
through it. The refractive index profile and propagation of modes is as illustrated in
the figure 4. Each mode is made to enter in to the optical fiber with different angles of
incidents which makes these modes to propagate different distances through the fiber,
due to which broadening of the signal (intermodal dispersion) takes place at the out
put end. This in turn restricts the bandwidth of the optical fiber. These fibers can use
the spatially along with incoherent optical sources such as LED’s. In addition to
lasers can be easily coupled because of larger core diameter and larger numerical
75
apertures. These are basically used in DATA links, which don’t require very large
bandwidths.
Figure 4
In this fiber the refractive index of the core is varied as a function of radial
distance from the center of the fiber and the refractive index of the cladding is
maintained constant. The refractive index profile and the propagation of modes is
as shown in the figure 5. Though the signals take different distances through the
fiber the signal distortion is not seen as they travel through the medium with
different refractive indices. The one that covers more distance travels through low
refractive index than the one that covers less distance. These are used for
illumination purposes and where the total variation in the light matters as in the
case of sensors.
76
Figure 5
Attenuation
In any communication system the major concern is the loss of information, which
basically comes because of attenuation. That is attenuation is the measure of loss in the
energy in its transformation from one point to another point. The effort of technology is
to always see that this attenuation is minimum. To achieve this one has to have the
knowledge of different parameters leading to attenuation that is, attenuation basically
comes because of the interaction of the energy or radiation with the material of the wave-
guide. Naturally the characteristic of attenuation will thus be dependent on both the
power loss and the length of the wave-guide. In view of the radiative properties always
the loss of power is logarithmic in nature and this loss being proportional to the length,
the coefficient of attenuation is defined as
∝ = {[10/L) (log (Pin /Pout)]}
The factor 10 comes because of the unit used is decibels. In an ideal wave guide Pin =
Pout is the one. Here Pin, Pout and L are in put power, out put power and the length of the
fiber respectively. The attenuation or the fiber loss (the loss of signal or power) because
of different parameters are as listed below,
1. Absorption loss
2. Scattering loss
3. Bending loss
4. Core and cladding loss
5. Dispersion loss etc.
1. Absorption loss:
Since optical fibers are made up of glass polymers and hence the absorption is
resulted due to
1. Defects in the composition
2. Impurities in the material (extrinsic absorption)
3. Characteristic absorption of the material of the fiber (intrinsic absorption)
77
Apart from the above factors the amount of attenuation happens because of OH ions
present and the existence of transition metals. The presence of OH ions cannot be
avoided because the hydrolysis process involved in the preparation of optical fibers.
Thus, in the transmission of signal through the optical fibers, care has to be taken not
to use the wavelength of light where absorption occurs due to OH ions and the
characteristic peaks of the material of the fiber appear.
2. Scattering loss
However best the material may be, it is impossible to have its density uniform
through out. The variation in the density leads to the variation in the refractive index,
of course to a very little extent; even this little variation causes refraction resulting
into attenuation. The complexity increases because of the variation in the refractive
index with wavelength also. Thus, the scattering induced attenuation, which is a
Rayleigh’s type of scattering of light, is very complex in nature.
Structural inhomogenities and defects created in the fiber fabrication can also
cause scattering. This scattering being dependent on value of wavelength and it
decreases with increase in wavelength. Thus, using higher wavelengths one can
reduce this type of losses.
4. Bending losses
Because of the bending of the fiber arising out of its installation from one point to
another which are known as macroscopic bending, that is bending because of the
curvatures in the fiber also leads to attenuation. Another type of bending arising out
of uneven forces on the interface leading to change in the refractive indices of core
and clad resulting in to microscopic losses. These can be reduced by having high
radius of curvatures at the time of cabling and by using a high ordered clad which has
a good resistance for small variation in external forces.
78
5. Core and cladding losses:
Since core and cladding have to have different refractive indices and their
multiplicity being high in step index multimode and graded index fibers, where
more than one material is used, the attenuation coefficients are going to be
different. This in itself creates scattering and hence attenuation when ray of light
tries to travel from one medium to another.
6. Dispersion loss:
The dispersion in the light is obvious because of the refractive indices of the
different wavelengths being different in the transmission. This arises out of the
characteristic of the material. Because of this the angle of incidence for different
wavelengths will be different when a ray tries to travel towards the clad resulting
in to uneven total internal reflection and some times transmission of light from
core to clad. This type of attenuation thus arises because of the wavelength itself.
79
Application of optical fibers for communication
The optical fibers are widely used in various fields like communications, Medical,
Domestic etc. One of its applications in point to point communication is discussed in
detail here (Figure 6).
In optical fiber communication the signal can be transmitted through the fiber only in the
form of optical signal. For which first the audio signal is converted into electrical analog
signals with the help of transmitter in the telephone and these analog signal are fed in to a
coder where they get converted in to binary form. This binary data is made to enter in to a
optical transmitter where these get converted in to optical signals by modulating the light
signals. These optical signals are fed in to the optical fiber with in the angle of
acceptance. At the out put end the optical signals are fed in to a photo detector which
converts them in to Binary form and then in to a decoder where the signal converted in to
analog form and finally gets converted in to voice at the receiver end. As the signal
propagates through the optical fiber there may be loss of signal due to various factors like
as discussed in attenuation and also the signal may be distorted due to the spreading of
pulses with time which is mainly because of the variation in the velocity of the different
spectral components through the optical fiber. With distance the loss of the signal
80
becomes more. Hence before the entire gets lost it is necessary to have repeaters which
amplifies the signal but there is no such device which can directly amplify the optical
signal and therefore at each repeater the optical signal needs to be converted in to
electrical signal then amplify and again convert it back in to optical signal and then fed in
to the fiber. This process has to be followed at every repeater, which restricts the speed of
the signal transmission.
81
3. Dielectrics
Introduction
One way of studying a material is to characterize it by study of the influence of the
applied electrical field. One of the way in which, the materials are classified is by looking
into their electrical conductivity under the influence of external electrical field. The
classification has lead into naming of materials on this effect, namely as conductors, semi
conductor and insulators.
Insulators are those materials, which do not conduct electricity. With this statement one
may conclude that these are the materials which do not get influenced by electric field,
which is however not true. The insulators, which are also known as Dielectrics have been
found to have an influence of external electrical field. Faraday first observed this, through
the working of capacitors. An insulator is defined as that material which does not allow
the conduction of electrons when an electrical field is applied due to the absence of free
electrons. The dielectric being a non-conductor has the ability to hold the electrostatic
field and hence can store the electrical energy. This is possible due to the phenomenon of
polarization of what are known as electric dipoles (permanent and induced).
The other way of looking dielectric & dielectric constant through the experimental
observations in a parallel plate capacitor experiment when a voltage V is applied to
parallel plate capacitor, a change (Q) which is proportional to applied voltage (V)
accumulates on the two plates, with the relation Q=CV, Here, C is called capacitance of
parallel plate capacitor. The capacitance of the capacitor has been found to be directly
82
proportional to area of the plates & inversely proportional to the distance between the
plates. It is found that the insertion of a dielectric material between the plates increases
the capacitance of the capacitor. Further the value of C has been found to be proportional
to the thickness of the dielectrics. This increase in capacitance can be only explained by
considering that when a dielectric material is inserted between the plates should decrease
the electrical field inside the dielectrics. The capacitance of a capacitor with a dielectric
and without a dielectric i.e. vacuum are compared, it will give dielectric constant i.e.
relative permittivity of the dielectric material as explained above.
∈o , the permittivity of vacuum has no much of physical meaning other than that it gives
the fundamental conversion factor. On the other hand the dielectric constant ∈r , gives the
insight into atomic structure of the material as ∈r is equal to 1 for vacuum & for all other
materials it is greater than 1.
A dipole consists of two equal & opposite charges, which are separated by a distance. If
Q is the charges separated by distance‘d’ then, the dipole – moment P of the dipole is
given by P =Q d. The moment acts towards the positive charge.
In a system, the centers of negative charges & positive charges act at the same point and
then dipole moment does not exist. However, under the influence of an external field, if
the centers are displaced, then a dipole will come into existence and the dipole is said to
be an induced dipole. This is the case as shown in the figure (1). The dipole moment of
such a system is said to be an induced dipole –moment and will exist as long as external
field exists.
In certain materials like ionic crystals the charges are displaced and hence have
permanent dipoles and are said to have permanent dipole moment. Under the influence of
83
external electric field, these will have additional effects along with the permanent effect
i.e. both induced dipole moment and permanent dipole moments.
In case of molecules depending upon, whether they are polar or non-polar in nature, their
exist a permanent dipole moment or not respectively. However, under influence of
electric field the non-polar molecules show induced dipole moments because of change in
electrical field distribution within a molecule. This also depends upon symmetry of the
molecules. The examples of polar dielectric materials are CH3Cl, H2O while that of non-
polar are H2, CO2, CH4, etc.
When no electrical field is applied, for these systems where permanent dipole moments
exist, will have their resultant equal to zero. They arrange themselves in such a way that
the net effects because of individual dipoles are nullified as shown in the figure (2).
When an external electrical field is applied, these dipoles get partially aligned along
applied electric field. The extent to which they are polarized depends upon the applied
electric field & the permittivity of the material. This process of re-orientation is known as
polarization.
In case of system where no permanent dipole moments exist, under the influence of
external electric field, the dipoles are induced & these dipole moments so induced will
align themselves in a direction opposite to the direction of the applied field in accordance
with Lenz’s law. This process in which a net induced dipole moment is results is also a
process of polarization.
The amount of polarization depends upon the resultant dipole –moment. The resultant
dipole moment creates an electrical field inside the dielectrics, so as to reduce the effect
of external electrical field. This explains why the capacitance of a capacitor increases
with the increases with the introduction of dielectric material inside a capacitor. The
amount of polarization is measured by a quantity known as polarization density P. The
polarization density P is defined as resultant dipole moment induced per unit volume of a
material. The polarization is thus naturally is a measure of change in capacitance with
84
vacuum and a dielectric material within the parallel plates of a capacitor. It is further, can
be shown that E = - P(in SI units) where Ep is the average macroscopic electric field
within a dielectric material due to induced polarization. At atomic level or molecular
level, the dipole moment induced is also proportional to applied electric field. If p is the
dipole moment induced per atom, p α E i.e. p = α E, where α is called polarizability
(defined as electric dipole moment induced in the atom by an electric field of unit
strength). i.e. P = Np= N α E. Thus under polarization state D is given by D = E + P (in
SI units), D is also known as displacement field vector. But D = εE i.e. D= ∈o ∈rE = E
+ P.
As, the polarization density is a measure of the change in the capacitance of the capacitor,
we have
P = Cm - C o
Where, Co and Cm are the capacitances of the capacitor with vacuum and dielectric
material as medium respectively.
.: P= [Evac- Em] = [(Evac/ Em) - 1] Em = [ε-1] Em
P/E = ( ∈ – 1)
From the above equation we get
∈ =1 +P/E
The polarization is basically brought about because of three aspects, which contribute
towards polarization density. They are
i) The Electronic polarization
ii) The anomic polarization
iii) Orientational polarization
iv) Space Charge polarization
85
1) The electronic polarization:
In some systems of dielectrics where the permanent dipole moment of an atom/molecule
involved is zero, i.e. in which the centers of negative and positive charges meet in such a
way to give net dipole moment zero, an induced dipole-moment comes into
effect under the influence of external electrical field due to the separation of charge
centers. The polarization brought about by such cases is known as electronic polarization
(figure 1) & the polarizability is known as electronic polarizability and denoted by αe.
2) Ionic Polarization:
Ionic crystalline materials like KCl, NaCl, etc. show ionic polarization. The external
electric field shifts the positive ions relative to negative ones (figure3a &b). This shift
produces polarization. The electronic polarizability measures the shift in the centre of
electric clouds while the ionic polarizability measures the shift of the ions relative to each
other. The ionic polarization is demanded by αi .
86
3) Orientation polarization :
When an external field is applied to systems of molecules like water, phenol etc carrying
permanent dipole moment Po, they tend to align themselves along the field i.e. the
orientation of these dipoles are changed leading to polarization effect (figure 4). This
polarization is known as orientation polarization and the corresponding polarizability as
induced polarizability. The polarizability of this type is found to be proportional to square
of the dipole moment ‘p’ and inversely proportional to T (temperature). This orientation
polarizability is the only polarizability, which depends upon temperature.
Thus if N is number of dipoles created per unit volume, then polarization density is given
by
P2
P=Np = N ( α e + α i + α 0 ) = N (α e + α i + )
3KT
Let us consider a dielectric system in case of liquids and solids in particular in this
section. In earlier discussions the effect of external field was discussed on the system.
The effect produced by this external field at atomic level depending upon the nature and
the surrounding is going to be different. One of the major effect is the formation of
induced dipoles at each atomic centers of the system. These induced dipoles create their
own electric field around that and creates local fields much different from the external
87
electrical fields. This effect can be neglected in case of gaseous dielectrics in view of
large distance between molecules however, which is not the case in solids and liquids. In
view of this the local field in liquids and solids differ from the microscopic field as
considered in earlier discussion.
Thus the study of effect at a particular location where atom is situated has to be studied
because
a) Influence of external electric field.
b) Field arising from the manual interactions of other, atoms, known as induced
field.
The equation for induced dipole moment is
Pi = αe E i
Where Ei is total induced field called local field. Thus the potential at any point in
space, at a distance r and angle θ from the mid-point of a dipole is given by
1 Ri co sθ
V (r, θ) =
4Π ∈ 0 r2
Thus, the field around the dipole will have two component with respect r & θ
∂ ∂
v(r , θ ) = 1 Pi c o θs ( r − 2 )
.: Er =
or 4Π ∈ o ∂r
2 Pi cosθ
=
4Π ∈ o r 3
1 2 Pi
the problem reduces to Er = here cos θ = 1 & sin θ =0
4Π ∈ o r3
88
The model with dimensional array is schematically is pictured as below. If Ao is a
molecule under study,
The effect due to the adjust atom A+1, A+2,------------------, and A-1, A-2, have to be taken
into effect. Assuming systematic arrangement of atoms, at a distance ‘a’ from each other,
the effect of A+1& A-1, on Ao are respectively, (1/4Пε0).(2pi/a3) because of A+1, it is
(1/4Пε0).[2(-pi)/(-a)3] because of A-1. The total effect because of both is then (1/Пε0).
( pi/a3) similarly, because of atoms A+2 and A-2 it will be (1/Пε0).[pi/(2a)3] and so on.
Because of nth atom on both sides it will be (1/Пε0).[pi/(na)3]. Thus, the electrical field,
due to the other atoms in general will be,
89
for infinite array of atoms
Therefore, the local electrical field at an atom is given by
1
∑
α
Ei=E + pi/ Пε0a3 n =0
n3
Where E is the external field,
1 1 1
∑
α
.: Ei=E + 1.2 pi/ Пε0a3 since n =0 3 =1+ 3
+ 3 + − − − − − − − − − − − = 1.2
n 2 3
2
NPi
interception Y axis will give N ( α e + α i ) while the slope . Thus, holds good only
3k
in gaseous state. From the above it is clear that, the effect of temperature on those
materials, which do not have permanent dipole moment is equal to zero. i.e. only those
gaseous systems whose molecules contain permanent dipole moment show a temperature
effect. In consistent with the above observations the groups of CH 4, CCl3, CH3Cl2 and
CH3Cl the value of slope goes on increasing in order of their dipole moment values.
A further, general equation has been given, Clausius – Mosotti, which is
90
∈ r − 1 Nα Pi
= Where, α = α e + α i + as explained in section (3.3.)
∈ r + 2 3∈o 3kT
NM A
If MA is the molecular weight of the molecule involved then = N A , then the above
α
∈ r −1 N A N A Pi
equation can be written as = (α e + α i + )
∈
r+ 2 α 3 ∈ o 3 K T
Therefore
Er− 1 MA N a NA P 2 i
= (α e + α i ) +
∈ r+ 2 P 3 ∈ o 9k ∈ o
2
∈r −1 MA 1 Pi
The plot of against with gives straight by with slope as .
∈N +2 P T
9k ∈ o
These results also go along with discussion as done in case of gases system above. The
above equation is known as Debye’s equation, and it helps in the determination of
permanent dipole moment of a molecule. To measure the dipole moment; the dielectric
constants at different temperatures will have to be measured using different methods.
∈o A
the medium. If C1 is the capacitance of the capacitor in air is equal to where A and
d
d are the area of the plates and d is the distance between them. The capacitance, is
91
determined with a material whose dielectric measurement is to be carried out. If C m is the
capacitance of the capacitor with dielectric medium, with dielectric constant ∈r then
∈o ∈r A
Cm =
d
Cm
.: =∈r
C air
The variation of dielectric constant depends upon the temperature in general on the state
of the material and nature of molecules like its shape, size and mass etc. In general, in
liquid state the dipoles are comparatively free to rotate in an external electrical field. At
freezing point these dipoles also freeze & at this stage only contributions from ionic &
electronic part come into effect. Thus, the variation of ∈r with temperature changes
depending upon which type of polarizability is more predominant.
92
Since D is having the components dielectric will also have two components one
Do
corresponding to cosine component namely ∈' ( w) = cosδ and sine component as,
Eo
D
o
∈
" (W ) =
( ) sinδ . These values of dielectric
E
o
constants have been found to be frequency dependent. The two dielectric constants in a
complex notation can be, thus written as ε/(ω) – i ε//(ω) and is usually denoted by ε*(ω)
i.e. ε*(ω) = ε/(ω) – i ε//(ω). ε(ω ) being a cosine component has been taken as real part and
ε//(ω) being a sine component as imaginary part. Then, the displacement vector can be
written in its complex notation as
D = ∈* E 0e iωt
" (ω
∈ )
=tan δ
' (ω
From above, it is clear that
∈ )
tan δ is known as loss tangent or dielectric loss. In fact ∈" is a measure of energy loss
during the application of the electrical field, which is used up in the form of the work
carried out in changing polarization & hence appearing as Joule heating. This heating has
to be thus appeared as loss and hence ∈" appears as an imaginary part of the dielectric
constant.
The energy loss i.e. dissipation of energy per unit time is given by
2Π
W= ω
ω
2Π ∫ JE dt
0
93
2Π
ω
W= ω
2Π ∫ ω(−D sin ωt + D
0
1 2 cos ωt ). E0 cos ωt dt
2Π 2Π
∫0 ∫ (cos ωt ) dt
2
2Π 2Π 0
2Π 2Π
ω ω
= ω ( −D1 Eo ) sin 2ωtdt + ω D2 ∈o
2 2
∫ ∫(
1+cos 2ωt )
2 dt
2Π 0
2Π 0
2Π 2Π
ω ω
since
∫ sin 2ωtdt = ∫ cos 2Πωtdt = 0,
0 0
ω2 D2 E 0 t 2Π
ω
ω2 D2 Eo 2Π = ω D E
ω= 2 =
ω
2 o
2Π 0 2Π.2 2
ω D2 ω 2
.: ω = E o2 = Eo ∈ 1 1 and also
2 Eo 2
ω
ω= D 0 Eo sin δ
2
The equation for ‘ω’ thus, indicates that the dielectric loss is directly proportional to sinδ.
At low δ values sin δ ~ tanδ is known as loss tangent. δ is known as loss angle.
94
electrical polarizability is responsible for dielectric constant at optical frequencies. The
dielectric associated with ionic polarizations fall in the region of infrared radiation. For
orientation polarization or permanent dipole moments, the range of frequency under
study will be radio frequency and microwave ranges. The figure 5 indicates a schematic
diagram of variation of α with frequency. The variation in the dielectric constant will be
of the same type except for the numerical value along Y-axis as the variation in dielectric
constant is directly proportional to polarizability.
Figure 5
Dipoles tend to align itself along the direction of the applied electrical field and in
varying field it tries to follow the field. In the process the dipoles try to interact with each
other and this results into dielectric loss and appears as heat. This energy loss is asociated
with imaginary dielectric constant. This is also evident in case of an LCR circuit, with
power is being given by
Power = VR IR Cos φ
1
With tan δ = . Thus, it may be concluded that the dielectric losses are because of
ωCR
ionization, leakage current, polarization and structural inhomogeneity.
95
These are the materials which are similar to ferromagnetic materials under the influence
of electric field i.e. the materials in which the polarization is not linear with applied
electrical fields. Naturally, such materials exhibit hysterisis(Figure 6). After applying the
electrical fields, the dielectrics get polarized and even after the removal of the electrical
field they process the spontaneous polarization Ps. The value of Ps is found to be the
function of temperature. Every ferroelectric has a critical temperature above which they
c
behave as Para-electric as per the Curie –Weiss-law, namely γ = is in case
T − Tc
ferromagnetic material ψ is the electrical susceptibility C α Tc are constants. Tc is
known as Curie temperature. It is found that, while going from ferroelectric to Para –
electrics the dielectrics involved go into transitions, transition being occurring from order
to disorder in crystal structure or displacement of ions. Some of the examples of
ferroelectric are BaTio3 (curie temp2-3800k, Ps ≈ 0.26 C/m2), Rochelle salt KNbO3(Curie
temp2-7080k; Ps ≈ 0.026667 C/m2), etc. Another important feature of ferroelectric crystals
is they do not have centre of symmetry.
Figure 6
3.8 Piezoelectricity:
96
Polarization can also be changed in dielectrics by application of stress or temperature. If
polarization changes because of the application of stress such materials are as known as
Piezoelectric. This type of change in polarization is seen in all the crystals where no
centre of symmetry exists. Nearly twenty classes of crystals have been found to exhibit
the change in polarization under stress. BaTiO3 has been found to be piezoelectric. The
change in symmetry an account of change in temper is known piezoelectricity BaTiO 3
has found to be pyroelectric also(Tourmaline is an example of pyroelectric). However,
every material, which is piezoelectric, is not essentially a pyroelectric. The piezoelectric
materials are characterized by piezoelectric strain co-efficient and electromechanical
coupling K. Quartz is most widely used piezoelectric.
Variation of polarized in all these materials have produced large application as
transducer.
+
- +
- -
Electric Field
Figure 1
Figure 2
97
Figure3a E=0
98
Electric field
Figure3b
99
Electric field
+e
-e
Figure 4
100
Eng Physics Unit VII
CRYSTAL STRUCTURES
Professor B V Sadashivmurty, Sri Jayachamarajendra College of Engineering,
Mysore
Narasimha Ayachit and G Neeraja Rani, SDM College of Engineering and
Technology, Dharwad
Introduction
One has to understand the crystal structure (arrangement of atoms) thoroughly because
solids have specific properties which depend on the structure. The materials are
generally classified into (i) solids, (ii) liquids and (iii) gases. Solids are further classified
into Crystalline and Non-Crystalline (Amorphous) solids depending on the arrangement
of atoms. If the atoms are arranged periodically through out the solid then it is said to be
crystalline other wise it is said to be non crystalline solid.
The atoms / molecules are electrically neutral. But when atoms or molecules are brought
closer together, a repulsive force operates between the similar charges in the atoms or
molecules. An attractive force operates between the dissimilar charges. The ultimate
force, holding the particles together in solids, is the resultant of attractive and repulsive
forces (figure 1)
Repulsive force
Attractive force
r0
101
together either by (1) Ionic bonds, (2) Covalent bonds, (3) Metallic bonds or (4)
Molecular bonds.
Figure 2
Basis: A set of an identical atom/s which is correlated to lattice points is called basis.
Crystal structure: A crystal structure is formed when the basis is substituted in
the space lattice i.e
Lattice + Basis = Crystal structure (figure3).
+ =
102
Bravais Lattice: The Bravais Lattice has infinite number of lattice points in it. If a
set of identical atoms / molecules are substituted in the space lattice then the lattice is said
to be Bravais lattice. The surroundings of any atom/molecule is same as any other
atom/molecule in the lattice. Otherwise it is said to be non-Bravais lattice. Below are the
some figures 4(a) to 4(e) representing both Bravais lattice and non- Bravais lattices.
+ =
Fig 4(a)
Bravais Lattice
+ / =
Fig. 4(b)
Non-Bravais Lattice
Fig.4(c)
Examples of Bravais Lattice & Non-Bravais Structure (Two Dimensional)Fig4(d)
103
Fig. 4(e)
4 Primitive
5 Base Centered a≠b≠c
3 Orthorhombic
6 Body Centered α = β = γ = 90°
7 Face Centered
8 Primitive a=b≠c
4 Tetragonal
9 Body Centered α = β = γ = 90°
a=b=c
5 Trigonal 10 Primitive
α = β = γ < 120°, ≠ 90°
a=b≠c
6 Hexagonal 11 Primitive
α = β = 90°, γ = 120°
7 Cubic 12 Primitive a=b≠c
α = β = γ = 90°
13 Body Centered
104
14 Face Centered
Crystal Systems
Symbols used
• P - Primitive - simple unit cell
• F - Face-centered - additional point in the center of each face
• I - Body-centered - additional point in the center of the cell
• C - Centered - additional point in the center of each end
• R - Rhombohedral - Hexagonal class only
Table2
BASIC VECTORS:
A coordinate system is used to represent the position of lattice points in space lattice. The
periodically repeating arrangement of all lattice points in space can be described by the
operation of parallel displacement called a translation vector R .
105
Figure 5
Let a and b be two vectors having equal magnitudes and oriented along AB and AC
respectively as shown in the figure 5. With a and b as coordinate vectors, the position
vector R of any lattice point can be written as,
R = n1 a + n2 b
Where, n1 and n2 are integers, whose values depend upon the location of lattice points
with respect to the origin.
A crystal lattice is a space lattice in which the lattice sites are occupied by atoms or
clusters of atoms. Each lattice point is associated with atom or group atoms, are called the
basis.
Figure 6
The Basis must be identified in composition, arrangement and orientation such that the
crystal appears exactly the same at one point as it does at other equivalent points. The
figure 6 shows the basis consisting of a group of two atoms. When the basis is repeated
with correct periodicity in all direction the crystal structure is obtained. Thus,
106
UNIT CELL
The entire lattice structure of a crystal can be generated by identical blocks known as unit
cell. The unit cell may be a group of ions, atoms or molecules. The unit cell is the
smallest building block or geometric figure from which the crystal is built up by
repeating it in three dimensions figure 7.
Figure 7
A unit cell can be chosen in a number of ways. The figure8 shows two ways of choosing
a unit cell in two dimensions. In the first way the unit cell contains lattice points only at
the corners. This type of unit cells is called as Primitive Cells. In the second way the unit
cell contains lattice points in addition to the corners. These are known as Non Primitive
Cells.
Figure8
Note: The unit cells differ from the primitive cell in that it is not restricted to being the
equivalent of one lattice point. Thus, unit cells may also be primitive cells, but all the
primitive cells need not be Unit Cells.
LATTICE PARAMETERS
To completely illustrate the crystal structure, the basic minimum parameters required are:
(1) The inter atomic or intermolecular molecular distance in x-, y- and z- direction
(a b and c)
107
(2) The angles ( α , β and γ ) between the x and y, y and z and z and x axes.
The angle between the x and y axes is taken as α , the angle between the y and z axes is
taken as β, angle between the Z and x axes is taken as γ are called the crystal
parameters(figure 7 represents cubic structure).
While dealing with a crystal system, it is necessary to refer to the crystal planes, and
directions of the straight lines joining the lattice points in a space lattice. A notation
system which uses a set of three integers (n1, n2 & n3) is adopted to describe both the
positions of planes or directions within the lattice.
Figure9
A resultant vector R which joins Lattice points A and B (figure9) can be represented by
an Equation.
108
Figure10
If x, y and Z are the starting co-ordinates for a plane then the Miller indices of the
plane are obtained by the following procedure:
(1) Consider the x-m y- and z co-ordinates of the lattice points of the plane lying on
the x-, y- and the z- directions of a reference frame
(3) Convert these fractions into whole numbers by multiplying all the numbers by a
common multiplier.
These simplified numbers, derived from the x-, y- and z- co-ordinates of the lattice
points on the plane are named as the h, k and l values of the plane and is called the
Miller index of the plane. All the planes parallel to this plane will have the same
indices. So the hkl values for a plane also represent a family of all the parallel
planes. The miller index a set of parallel planes is written as <hkl).
Example: Given that,
6 3 3
X= a Y= b Z= c
5 4 2
109
x y z 6 3 3
, , = , ,
a b c 5 4 2
Taking reciprocals of the three fractions
5 4 2
6 , 3 , 3
Multiplying throughout by least common multiple ‘4’ for the denominator, we have
the Miller indices
(5 8 4)
If a plane is oriented parallel to a coordinate axis, its intercept with the coordinate is
taken as infinity, since the reciprocal of infinity is zero, the corresponding Miller
indices value will also be zero.
Thus the Miller indices is a set of 3 lowest possible integers whose ratio taken in
order is the same as that of the reciprocals of the Miller integers of the planes on the
corresponding axes in the same order. Similar to the case of representation of
directions in the space lattice, any given Miller indices set represents all parallel
equidistant crystal planes for a given space lattice. Owing the rotational symmetry,
certain planes which are not parallel to each other become in distinguishable from the
crystallographic point of view. In such cases, Miller indices are enclosed in braces {}
instead of parenthesis or brackets, which represents all the equivalent planes. For
example in the case of cubic lattice, the 6 planes referring to 6 faces of a unit cell are
represented by the Miller indices as (1 0 0) (0 0 1) (010) (0 1 0) (0 0 1) collectively
designated as {1 0 0 }
Figure11
110
To get the interplanar distance, consider a plane ABC with Miller indices (hkl). In
the reference frame, draw a normal to the plane from the origin. Let OP is the normal
to the plane ABC. Let angle POA = α ’, and POB = β’ and POC =γ ’ be the angles
made by the normal to the plane with the x, y and the z directions. OP = d is the
interplanar distance
OP d
Cos β = =
OB b / k
OP d
Cos γ = =
OC c / l
h2d 2 k 2d 2 l 2d 2
+ 2 + 2 =1
a2 b c
h2 k 2 l 2
d2 2 + 2 + 2 =1
a b c
111
1
d2hkl = h + k + l
2
2 2
2
a
b2 c2
1
1/ 2
dhkl = h 2 k2 l2
2 + 2 + 2
a b c
Density is a macroscopic property. Basically it is the mass per unit volume. In case of
crystals, mass of atoms packed in a conventional unit cell per unit volume of the cell
gives the density of the crystal.
nM
m=
Na
From equation ( i ), mass of atoms in the unit cell in terms of the density of the crystal
is given by,
nM
ρ=
N AV
112
nM
ρ= kg / m 3
N AV
The number of atoms at equal and least distance from a given atom in the structure is
the coordination number can be taken as the first nearest neighbors of an atom in
the structure.
Atomic packing factor is the ratio of the volume of the unit cell occupied by atoms
to the net volume of the unit cell
The following are the possible structures found in a cubic system of crystals,
resulting due to the kinds of packing of atoms,
In a simple cubic structure, the space lattice is cubic. Atoms are placed at all lattice
locations. Therefore, an atom at a lattice location will be in contact with all its six
nearest neighbors (figure12). Thus the co-ordination number for a simple cubic
structure is N=6. In a simple cubic structure, the lattice constant is the cube edge of
the conventional unit cell. Thus the lattice constant is the distance between the centers
of two neighboring atoms. That is the lattice constant a = 2r, where r is the atomic
radius. The atoms at the lattice locations are shared by eight unit cells. Each atom at
the lattice location of the unit cell contributes (1/8) to the unit cell.
The number of atoms contained in a unit cell in the structure is given by the
number of lattice locations in the unit cell multiplied by the contribution of the atom
at each location. That is the number of atoms in a unit cell of the simple cubic
structure is given by
113
Figure12
v
=
( n ) X[4 / 3πr 3 ] = (n )( 4)( πr 3 ) = nπ
APF =
V a3 3( 2r )3 6
nπ
APF = = 0.523 = (52 %)
6
In body centered cubic structure, the space lattice is cubic. Atoms are placed at all
lattice locations and there will be an additional atom at the body center of the cubic
unit cell. In this structure, the atoms stacked along the body diagonal of the cubic unit
cell are in contact. That is, the atom at the body center of the unit cell will be in
contact with all other atoms at the cubic lattice locations of the unit cell. Each atom in
the structure will have eight nearest neighbors (figure 13). Thus, the co-ordination
114
number for a body centered cubic structure is N = 8. In a body centered cubic
structure, the lattice constant (a), is the cube edge of the conventional unit cell.
Figure13
Calculation of lattice constant in terms of atomic radius
Let d be diagonal distance between the atoms in a plane of the unit cell. Then,
d 2 = (a 2 + a 2 ) = 2a 2 − − − − − − − (i)
If D is the diagonal of the cubic unit cell, then
D 2 = (d 2 + a 2 ) = ( 2a 2 + a 2 ) = 3a 2 − (ii )
But, the atoms along the body diagonal, are in contact, Therefore D = 4r
And D2 = (4r)2 - - - - - - - - (iii)
From equations (i) and ( iii) ,
D2 = (4r)2 = 3a2,
Thus, the relation between the lattice constant (a ) and atomic radius is,
4r
a=
3
The atoms at the lattice locations are shared by eight unit cell. Therefore,
each atom at the lattice location of the unit cell contributes ( 1/8 ) to the unit cell. The
number of atoms contained in the unit cell of the structure is given by the number of
lattice locations in the unit cell multiplied by the contribution of the atom at each
location plus the atom at the body center of the unit cell. That is the number of atoms
in a unit cell the simple cubic structure is given by
The parameters describing the structure of a body centered cubic structure are :
115
6 Volume of an atom a = (4/3) π
r3
v
=
( n )X[4 / 3πr 3 ] = (n )( 4)( πr 3 ) = (n )( 4 / 3)( πr 2 )
APF = V a3 3(4 / 3r )3 [64 r 3 /( 3 3 )]
( n )( 4 / 3)( πr 3 ) 3 π(n )
APF =
[64 r 3
/( 3 3) ] =
16
= 0.68 (68 %)
In a face centered cubic structure, the space lattice is cubic. The conventional unit
cell consists of eight small cubelets. Atoms are placed at alternate lattice locations.
The atoms stacked along the face diagonal of the cubic unit cell are in contact. The
atom at any lattice location of the unit cell will have twelve nearest neighbors
(figure14)
Figure14
Let d be the diagonal distance between the atoms in a plane of the conventional
unit cell. Then,
d2= (a2 + a2 ) = 2a2 - - - - - - - ( i )
But, the atoms along the face diagonal of the conventional unit cell are in contact.
Therefore,
d = 4r - - - - - - - ( ii )
2 2 2 2
And d = ( a + a ) = 2a - - - - ( iii )
From equations ( i ) and ( iii ) ,
d2 = ( 4r)2 = 2a2 ,
Thus, the relation between the lattice constant ( a ) and atomic radius is,
116
4r
A=
2
The atoms at the lattice locations of the conventional unit cell are shared
by eight such cells. Therefore each atom at the lattice location of the unit cell
contributes ( 1 /8 ) to the unit cell. The number of atoms belonging to the unit cell in
the structure is given by the number of lattice locations in the unit cell multiplied by
the contribution of the atom at each location plus the contributions of the atoms at the
face centered locations. The number of atoms in a unit cell of the face centered cubic
structure is given by
n = ( No. of lattice locations X contribution of atoms at these location )
( No. of atoms at the face centers X their contribution to the unit cell )
n = [ 8 X (1 /8 )] + [ 6 X ( 1 / 2)] = ( 1 +3 ) = 4
The parameters describing the structure of a face centered cubic structure are :
Atomic packing factor is the ratio of the volume of the unit cell occupied by
atoms to the net volume of the unit cell
Volume of the unit cell occupied by the atom is given by
v = (number of atoms in a unit cell ) X [volume of an atom]
v = (n ) X [ 4 / 3 π3 ]
( n )( 4)( πr 3 ) 2 π(n )
APF =
[ 3
3 64 r /( 2 2) ] =
24
= 0.74 (74 %)
117
Sodium chloride is an ionic crystal. It is a compound of the alkali halide family. Due
to the proximity of the sodium atom with a chlorine atom, the valence electron from
the sodium atom is transferred to the chlorine atom. With this transfer of an electron,
the sodium atom becomes a cation and the chlorine is converted to an anion. Then,
these ions are held by the ionic force and the bond that holds the atoms together is
ionic.
A conventional unit cell of the structure consists of eight cubic primitive cells
(figure 18). In the conventional unit cell of the crystal, sodium atoms occupy the FCC
positions of a cubic structure. Chlorine atoms are placed at the intermediate positions
between the sodium atoms. Chlorine atoms independently from an FCC structure
similarly sodium atoms also from an independent FCC structure. The lattice constant
of both the FCC structures is a = 5.63Au. A unit cell of a sodium chloride crystal can
be regarded as the structure formed by the inter-penetration of sodium FCC lattice
with the Chlorine FCC lattice through one half of the lattice constant. This can be
treated as the graphical representation of the crystal structure.
The crystal structure can also be illustrated in terms of the co-ordinates of the
atoms in the structure. For this, an orthogonal co-ordinate system is considered. The
origins of the co-ordinate system can either be at the location of a chlorine atom or at
the location of a sodium atom. With the origin at the location of the sodium atom, the
co-ordinates of the sodium and chlorine atoms in the crystal structure are shown
below. Sodium chloride is an ionic crystal formed by the transfer of the valence
electron from sodium atom to the chlorine atom. Due to this transfer the sodium atom
is converted to a cation and the chlorine is converted to an anion. These ions are held
by the ionic force and the bond that holds the atoms together is ionic bond. In the
unit cell of the compound, chlorine atoms occupy the FCC positions of a cubic
structure. Sodium atoms occupy edge center positions.
Figure18
Atoms Co-ordinates
118
000. 100, 101 , 001 ,
½0½
Sodium ½½0,1½½,½½1,
0½½,
110 , 111 , 011 , 010 , ½
1½
½ 0, 10 ½, ½ 01 , 00 ½ ,
1 ½ 0, 1 ½ 1, 0 ½ 1, 0 ½
Chlorine
0, ½ ½ ½
½ 1 0, 1 ½ ½ , ½ 1 1, 0
1½
nA
ρ= kg / m 3
NAV
N = [ n (Na) ] + [ n ( CI ) ]
[ n ( CI ) ] = 4 ( 1/4 )+ 4 ( 1/4 )+ 1 = 4
119
A = A (Na ) + A ( CI ) = ( 23 X 10-3 ) + ( 33.45 X 10-3 ) kg
A = ( 58.43 X 10-3 ) Kg
nA 4(58.43X10−3 )
ρ= = 23 10) 3
= (2.174Ξ103 )kg / m3
N A V (6.023X10 )(5.63X10 −
Diamond is the crystalline form of carbon. The structure consists of the periodic
stacking of carbon atoms in its crystalline state.
A conventional unit cell of structure consists of eight cubic primitive cells ( figure
16 ). On the basis of the positions of the carbon atoms in the unit cell they are
classified as first kind and second kind only for the sake of convenience. Otherwise
they are the same carbon atoms. The carbon atoms placed at the FCC positions of the
conventional unit cell are the first kind of carbon atoms. Another set of carbon atoms
occupy the body centered positions in the primitive cubic cells in the structure. The
second type of carbon atoms are placed at the body centers of the alternate cubic
primitive cells. These carbon atoms independently form another FCC structure.
Therefore, diamond crystal can be regarded as a structure produced by the penetration
of one type of carbon atoms forming FCC structure in to the other FCC structure,
formed by the second kind of carbon atoms, through a distance equal to three fourth
the body diagonal (figure 19).
If the locations of the atoms in a diamond structure is seen in the primitive part of
the unit cell (one of the eight cubelets), they look like a tetrahedra. In these tetrahedra,
one atom at the center is bonded to three others in the primitive cell. Periodic stacking
of the tetrahedral structure also produces the diamond crystal.
The distance between the two carbon atoms place on the cube edge of the
conventional unit cell is the lattice constant and it is, a = 3.567 Au for diamond
crystal. This can be treated as the graphical representation of the crystal structure.
The crystal structure can also be illustrated in terms of the co-ordinates of the
atoms in the structure. To write the co-ordinates of the carbon atoms in the structure,
an orthogonal co-ordinate system is considered. The origin of the co-ordinate system
is fixed at one of the carbon atoms. With this co-ordinate system the co-ordinates of
the atoms in the structure are shown below.
120
Figure19
Atoms Co-ordinates
000. 100, 101 , 001 , ½ 0
½
Carbon ½½0,1½½,½½1,0
atoms of ½½,
first kind
110 , 111 , 011 , 010 , ½ 1
½
Carbon ¼ ¼ ¼, 3/4 3/4 1/4,
atoms of
second 1/4 3/4, 3/4 ,3/4 1/4 3/4
kind
nA
ρ= kg / m 3
NAV
In diamond structure, the number of carbon atoms in the conventional unit cell is
n=8
121
Atomic weight of carbon is A = 12.01 X 10-3 kg.
Lattice constant a = 3.567 Au
Volume of the vunit cell V= a3 = ( 3.567 X 10-10)3
And the Avogadro number is NA = ( 6.023 X 1023)
122
Notes by Prof. B.N.Meera, Bangalore University
1. The valence electron of the atom is loosely bound to their respective atom. In
metals they become the free electrons. These electrons are free to move
throughout the metal and are hence termed as free electrons. Since electrical
conductivity is due to these free electrons, these are also termed as conduction
electrons.
2. These free electrons move in the metal according to Kinetic theory of gases.
3. Like the molecules of gas in a container, free electrons in these metals move
randomly (in the absence of electric field) with a velocity, which is the
average velocity or rms (root mean square) velocity. In the absence of
externally applied potential difference there are on an average as many
electrons wandering through a given cross section of the conductor in a given
direction as there are in the opposite direction. Hence the net current us zero.
123
Consider a conductor of uniform cross section. If a voltage V is applied to a metal
of resistance R, and I is the current in the conductor, we know by ohm’s law,
V = IR 1)
J =σ E 2)
where σ is the electrical conductivity and A is the cross sectional area of the conductor.
(Note 3).
Let us find an expression for J. Let vd be the drift velocity of the electrons. Hence in one
second the electrons travel a distance of vd. Let n be the number density of the electrons
in the metal. Consider unit volume of the metal. This contains n free electrons, each of
charge e. The total charge is en. The total charge crossing unit area per unit time is nevd.
This is the definition of current density.
∴J = nev d .
The applied voltage generates an electric field E. The force F acting on the electrons due
to the field is -eE. It is important to note that free electrons obey Newton’s Laws. For
example, electrons moving in the opposite direction to the field will be speeded up and
those moving in the same direction as E will be slowed down. Those moving
perpendicular to E will acquire a sideways component of velocity and so on. The
equation of motion for an electron is
F = −eE = ma
Integrating we get
vd = −(eE / m)τ
where τ is the time for which the electron moves before undergoing collision. This
average time between collisions is known as the scattering time or relaxation time
124
J = ne (eE / m)τ
= ne 2τE / m
σ =J / E
= ne 2τ / m
Note electrical conductivity depends on n – free electron number density. If the carrier
density is however not known, then the conductive properties of the materials cannot be
fully determined, for σ = nτ and only the product of the two quantities can be obtained.
Mean time between collisions and the mean free path
An important concept that is of importance in Drudes theory (which is a important
concept in Kinetic theory of gases) is the concept of mean time between collision and
mean free path. As discussed above, the electrons travels in the presence of electric field
for a time τ (mean free time) before it undergoes a collision. The electron travels with
the rms velocity in the absence of electric field. In the presence of applied electric field,
the drift velocity imparted to the electron is very small compared to the rms velocity. For
ex, in copper metal the rms velocity is equal to where as the drift velocity is . The time
taken by the electrons in traversing during time τ is hence decided not by the drift
velocity but by the much greater velocity – the rms velcity (average velocity).
τ = λ / v rms
where λ is the average distance traveled by the electron between collisions. Here v is
the average electron speed, which Drude estimated from the classical partition of energy:
mv 2 3k BT
=
2 2
3k B T
i.e. v=
m
It is also important to note that drift velocity is a concept which is of relevance in the
Drude free electron theory but not in Kinetic theory of gases (Note 4).
Note:
1. Drude put forth this model in the year 1900. J.J. Thomson discovered electron in
the year 1897. However, a complete understanding of atom was established (in
the year 1914) by Rutherford. He used the results of scattering of alpha particles
by gold foil to arrive at the model. It is interesting to note that the details of the
125
nature of the atom (as understood today) was not available to Drude at the time of
formulating of his theory- a fine example of insight and intuition.
V = IR
El = ( JA ) R .
3k B T
v=
m
Substituting we get,
126
Note that the rms velocity is determined by temperature alone.
v d = −(eE / m)τ
We get 3.67 x 10-4 m/s. Note the vast difference in the magnitudes of drift velocity and
rms velocity.
Note 4: Considering the extremely simplistic assumptions made in modeling the atomic
processes in the Drude free electron theory, it is natural to expect that the model cannot give
a good agreement with experimental results. However, the model quite successfully explains
the electrical behavior of metals and is even at present considered as a workable theory as a
starting point in understanding electrical processes in metals.
The main objections to Drude theory are the following. In the model proposed by Drude,
electrons are treated as “non-interacting” gas molecules. Unlike gas molecules electrons are
charged. There is a strong electrostatic interaction between these electrons and this
interaction is assumed to be non-existent in the Drude theory. Drude also did not dwell in any
detail the exact cause and mechanism of collision in his theory (remember the theory was
proposed in 1900 when there existed no clear picture of an atom!!). Also the expression
derived for conductivity of metal does not explicitly show any temperature dependence of
conductivity. Electrical conductivity was known to depend on temperature.
127