0% found this document useful (0 votes)
220 views693 pages

Full

Uploaded by

Im print
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
220 views693 pages

Full

Uploaded by

Im print
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 693

PHYSICAL METHODS IN

CHEMISTRY AND
NANO SCIENCE

Pavan M. V. Raja & Andrew R. Barron


Rice University
Rice University
Physical Methods in Chemistry and Nano
Science

Pavan M. V. Raja & Andrew R. Barron


This text is disseminated via the Open Education Resource (OER) LibreTexts Project (https://LibreTexts.org) and like the hundreds
of other texts available within this powerful platform, it is freely available for reading, printing and "consuming." Most, but not all,
pages in the library have licenses that may allow individuals to make changes, save, and print this book. Carefully
consult the applicable license(s) before pursuing such effects.
Instructors can adopt existing LibreTexts texts or Remix them to quickly build course-specific resources to meet the needs of their
students. Unlike traditional textbooks, LibreTexts’ web based origins allow powerful integration of advanced features and new
technologies to support learning.

The LibreTexts mission is to unite students, faculty and scholars in a cooperative effort to develop an easy-to-use online platform
for the construction, customization, and dissemination of OER content to reduce the burdens of unreasonable textbook costs to our
students and society. The LibreTexts project is a multi-institutional collaborative venture to develop the next generation of open-
access texts to improve postsecondary education at all levels of higher learning by developing an Open Access Resource
environment. The project currently consists of 14 independently operating and interconnected libraries that are constantly being
optimized by students, faculty, and outside experts to supplant conventional paper-based books. These free textbook alternatives are
organized within a central environment that is both vertically (from advance to basic level) and horizontally (across different fields)
integrated.
The LibreTexts libraries are Powered by MindTouch® and are supported by the Department of Education Open Textbook Pilot
Project, the UC Davis Office of the Provost, the UC Davis Library, the California State University Affordable Learning Solutions
Program, and Merlot. This material is based upon work supported by the National Science Foundation under Grant No. 1246120,
1525057, and 1413739. Unless otherwise noted, LibreTexts content is licensed by CC BY-NC-SA 3.0.
Any opinions, findings, and conclusions or recommendations expressed in this material are those of the author(s) and do not
necessarily reflect the views of the National Science Foundation nor the US Department of Education.
Have questions or comments? For information about adoptions or adaptions contact [email protected]. More information on our
activities can be found via Facebook (https://facebook.com/Libretexts), Twitter (https://twitter.com/libretexts), or our blog
(http://Blog.Libretexts.org).

This text was compiled on 06/03/2022


TABLE OF CONTENTS
This book is intended as a survey of research techniques used in modern chemistry, materials science, and nano science. The topics
are grouped, not be method per se, but with regard to the type of information that can be obtained.

1: Elemental Analysis
1.1: Introduction to Elemental Analysis
1.2: Spot Tests
1.3: Introduction to Combustion Analysis
1.4: Introduction to Atomic Absorption Spectroscopy
1.5: ICP-AES Analysis of Nanoparticles
1.6: ICP-MS for Trace Metal Analysis
1.7: Ion Selective Electrode Analysis
1.8: A Practical Introduction to X-ray Absorption Spectroscopy
1.9: Neutron Activation Analysis (NAA)
1.10: Total Carbon Analysis
1.11: Fluorescence Spectroscopy
1.12: An Introduction to Energy Dispersive X-ray Spectroscopy
1.13: X-ray Photoelectron Spectroscopy
1.14: Auger Electron Spectroscopy
1.15: Rutherford Backscattering of Thin Films
1.16: An Accuracy Assessment of the Refinement of Crystallographic Positional Metal Disorder in Molecular Solid Solutions
1.17: Principles of Gamma-ray Spectroscopy and Applications in Nuclear Forensics

2: Physical and Thermal Analysis


2.1: Melting Point Analysis
2.2: Molecular Weight Determination
2.3: BET Surface Area Analysis of Nanoparticles
2.4: Dynamic Light Scattering
2.5: Zeta Potential Analysis
2.6: Viscosity
2.7: Electrochemistry
2.8: Thermal Analysis
2.9: Electrical Permittivity Characterization of Aqueous Solutions
2.10: Dynamic Mechanical Analysis
2.11: Finding a Representative Lithology

3: Principles of Gas Chromatography


3.1: Principles of Gas Chromatography
3.2: High Performance Liquid chromatography
3.3: Basic Principles of Supercritical Fluid Chromatography and Supercrtical Fluid Extraction
3.4: Supercritical Fluid Chromatography
3.5: Ion Chromatography
3.6: Capillary Electrophoresis

1
4: Chemical Speciation
4.1: Magnetism
4.2: IR Spectroscopy
4.3: Raman Spectroscopy
4.4: UV-Visible Spectroscopy
4.5: Photoluminescence, Phosphorescence, and Fluorescence Spectroscopy
4.6: Mössbauer Spectroscopy
4.7: NMR Spectroscopy
4.8: EPR Spectroscopy
4.9: X-ray Photoelectron Spectroscopy
4.10: ESI-QTOF-MS Coupled to HPLC and its Application for Food Safety
4.11: Mass Spectrometry

5: Reactions Kinetics and Pathways


5.1: Dynamic Headspace Gas Chromatography Analysis
5.2: Gas Chromatography Analysis of the Hydrodechlorination Reaction of Trichloroethene
5.3: Temperature-Programmed Desorption Mass Spectroscopy Applied in Surface Chemistry

6: Dynamic Processes
6.1: NMR of Dynamic Systems- An Overview
6.2: Determination of Energetics of Fluxional Molecules by NMR
6.3: Rolling Molecules on Surfaces Under STM Imaging

7: Molecular and Solid State Structure


7.1: Crystal Structure
7.2: Structures of Element and Compound Semiconductors
7.3: X-ray Crystallography
7.4: Low Energy Electron Diffraction
7.5: Neutron Diffraction
7.6: XAFS
7.7: Circular Dichroism Spectroscopy and its Application for Determination of Secondary Structure of Optically Active Species
7.8: Protein Analysis using Electrospray Ionization Mass Spectroscopy
7.9: The Analysis of Liquid Crystal Phases using Polarized Optical Microscopy

8: Structure at the Nano Scale


8.1: Microparticle Characterization via Confocal Microscopy
8.2: Transmission Electron Microscopy
8.3: Scanning Tunneling Microscopy
8.4: Spectroscopic Characterization of Nanoparticles
8.5: Using UV-Vis for the detection and characterization of silicon quantum dots
8.6: Characterization of Graphene by Raman Spectroscopy
8.7: Characterization of Graphene by Raman Spectroscopy
8.8: Characterization of Bionanoparticles by Electrospray-Differential Mobility Analysis
8.9: Characterization of Bionanoparticles by Electrospray-Differential Mobility Analysis
Index

2
9: Surface Morphology and Structure
9.1: Interferometry
9.2: Atomic Force Microscopy (AFM)
9.3: SEM and its Applications for Polymer Science
9.4: Catalyst Characterization Using Thermal Conductivity Detector
9.5: Nanoparticle Deposition Studies Using a Quartz Crystal Microbalance

10: Device Performance


10.1: A Simple Test Apparatus to Verify the Photoresponse of Experimental Photovoltaic Materials and Prototype Solar Cells
10.2: Measuring Key Transport Properties of FET Devices

Index

Index

Glossary

Physical Methods in Chemistry and Nano Science (Barron) is shared under a CC BY 4.0 license and was authored, remixed, and/or curated by
Pavan M. V. Raja & Andrew R. Barron via source content that was edited to conform to the style and standards of the LibreTexts platform; a
detailed edit history is available upon request.

3
CHAPTER OVERVIEW
1: Elemental Analysis
The purpose of elemental analysis is to determine the quantity of a particular element within a molecule or material.
1.1: Introduction to Elemental Analysis
1.2: Spot Tests
1.3: Introduction to Combustion Analysis
1.4: Introduction to Atomic Absorption Spectroscopy
1.5: ICP-AES Analysis of Nanoparticles
1.6: ICP-MS for Trace Metal Analysis
1.7: Ion Selective Electrode Analysis
1.8: A Practical Introduction to X-ray Absorption Spectroscopy
1.9: Neutron Activation Analysis (NAA)
1.10: Total Carbon Analysis
1.11: Fluorescence Spectroscopy
1.12: An Introduction to Energy Dispersive X-ray Spectroscopy
1.13: X-ray Photoelectron Spectroscopy
1.14: Auger Electron Spectroscopy
1.15: Rutherford Backscattering of Thin Films
1.16: An Accuracy Assessment of the Refinement of Crystallographic Positional Metal Disorder in Molecular Solid Solutions
1.17: Principles of Gamma-ray Spectroscopy and Applications in Nuclear Forensics

1: Elemental Analysis is shared under a CC BY 4.0 license and was authored, remixed, and/or curated by Pavan M. V. Raja & Andrew R. Barron
via source content that was edited to conform to the style and standards of the LibreTexts platform; a detailed edit history is available upon
request.

1
1.1: Introduction to Elemental Analysis
The purpose of elemental analysis is to determine the quantity of a particular element within a molecule or material. Elemental
analysis can be subdivided in two ways:
Qualitative: determining what elements are present or the presence of a particular element.
Quantitative: determining how much of a particular or each element is present.
In either case elemental analysis is independent of structure unit or functional group, i.e., the determination of carbon content in
toluene (C H CH ) does not differentiate between the aromatic sp carbon atoms and the methyl sp carbon.
6 5 3
2 3

Elemental analysis can be performed on a solid, liquid, or gas. However, depending on the technique employed the sample may
have to be pre-reacted, e.g., by combustion or acid digestion. The amounts required for elemental analysis range from a few gram
(g) to a few milligram (mg) or less.
Elemental analysis can also be subdivided into general categories related to the approach involved in determining quantities.
Classical analysis relies on stoichiometry through a chemical reaction or by comparison with known reference sample.
Modern methods rely on nuclear structure or size (mass) of a particular element and are generally limited to solid samples.
Many classical methods they can be further classified into the following categories:
Gravimetric in which a sample is separated from solution as a solid as a precipitate and weighed. This is generally used for
alloys, ceramics, and minerals.
Volumetric is the most frequently employed involves determination of the volume of a substance that combines with another
substance in known proportions. This is also called titrimetric analysis and is frequently employed using a visual end point or
potentiometric measurement.
Colorimetric (spectroscopic) analysis requires the addition of an organic complex agent. This is commonly used in medical
laboratories as well as in the analysis of industrial wastewater treatment.
The biggest limitation in classical methods is most often due to sample manipulation rather than equipment error, i.e., operator
error in weighing a sample or observing an end point. In contrast, the errors in modern analytical methods are almost entirely
computer sourced and inherent in the software that analyzes and fits the data.

1.1: Introduction to Elemental Analysis is shared under a CC BY 4.0 license and was authored, remixed, and/or curated by Pavan M. V. Raja &
Andrew R. Barron via source content that was edited to conform to the style and standards of the LibreTexts platform; a detailed edit history is
available upon request.

1.1.1 https://chem.libretexts.org/@go/page/55810
1.2: Spot Tests
Spot tests (spot analysis) are simple chemical procedures that uniquely identify a substance. They can be performed on small
samples, even microscopic samples of matter with no preliminary separation. The first report of a spot test was in 1859 by Hugo
Shiff for the detection of uric acid. In a typical spot test, a drop of chemical reagent is added to a drop of an unknown mixture. If
the substance under study is present, it produces a chemical reaction characterized by one or more unique observables, e.g., a color
change.

Detection of Chlorine
A typical example of a spot test is the detection of chlorine in the gas phase by the exposure to paper impregnated with 0.1% 4-
4'bis-dimethylamino-thiobenzophenone (thio-Michler's ketone) dissolved in benzene. In the presence of chlorine the paper will
change from yellow to blue. The mechanism involves the zwitterionic form of the thioketone

This, in turn, undergoes an oxidation reaction and subsequent disulfide coupling

Bibliography
L. Ben-Dor and E. Jungreis, Microchimica Acta, 1964, 52, 100.
F. Feigl, Spot Tests in Organic Analysis, 7th Ed. Elsevier, New York, 2012
N. MacInnes, A. R. Barron, R. S. Soman, and T. R. Gilbert, J. Am. Ceram. Soc., 1990, 73, 3696.
H. Schi , Ann. Chim. Acta, 1859, 109, 67.

1.2: Spot Tests is shared under a CC BY 4.0 license and was authored, remixed, and/or curated by Pavan M. V. Raja & Andrew R. Barron via
source content that was edited to conform to the style and standards of the LibreTexts platform; a detailed edit history is available upon request.

1.2.1 https://chem.libretexts.org/@go/page/55811
1.3: Introduction to Combustion Analysis
Applications of Combustion Analysis
Combustion, or burning as it is more commonly known, is simply the mixing and exothermic reaction of a fuel and an oxidizer. It
has been used since prehistoric times in a variety of ways, such as a source of direct heat, as in furnaces, boilers, stoves, and metal
forming, or in piston engines, gas turbines, jet engines, rocket engines, guns, and explosives. Automobile engines use internal
combustion in order to convert chemical into mechanical energy. Combustion is currently utilized in the production of large
quantities of H . Coal or coke is combusted at 1000 ◦ C in the presence of water in a two-step reaction. The first step shown in
2

involved the partial oxidation of carbon to carbon monoxide.


C(g) + H O(g) ⟶ CO(g) + H (g) (1.3.1)
2 2

The second step involves a mixture of produced carbon monoxide with water to produce hydrogen and is commonly known as the
water gas shift reaction.

CO(g) + H O(g) → CO (g) + H (g) (1.3.2)


2 2 2

Although combustion provides a multitude of uses, it was not employed as a scientific analytical tool until the late 18th century.

History of Combustion
In the 1780's, Antoine Lavoisier (figure 1.3.1 ) was the first to analyze organic compounds with combustion using an extremely
large and expensive apparatus (figure 1.3.2 ) that required over 50 g of the organic sample and a team of operators.

Figure 1.3.1 : French chemist and renowned "father of modern Chemistry" Antoine Lavoisier (1743-1794).

Figure 1.3.2 : Lavoisier's combustion apparatus. A. Lavoisier, Traité Élémentaire de Chimie, 1789, 2, 493-501.
The method was simplified and optimized throughout the 19th and 20th centuries, first by Joseph Gay- Lussac (Figure 1.3.3), who
began to use copper oxide in 1815, which is still used as the standard catalyst.

1.3.1 https://chem.libretexts.org/@go/page/55812
Figure 1.3.3 : French chemist Joseph Gay-Lussac (1778-1850).
William Prout (Figure 1.3.4) invented a new method of combustion analysis in 1827 by heating a mixture of the sample and CuO
using a multiple-flame alcohol lamp (Figure 1.3.5) and measuring the change in gaseous volume.

Figure 1.3.4 : English chemist, physician, and natural theologian William Prout (1785-1850).

Figure 1.3.5 : Prout's combustion apparatus. W. Prout, Philos. T. R. Soc. Lond., 1827, 117, 355.
In 1831, Justus von Liebig (Figure 1.3.6)) simplified the method of combustion analysis into a "combustion train" system (Figure
1.3.7) and Figure 1.3.8)) that linearly heated the sample using coal, absorbed water using calcium chloride, and absorbed carbon

dioxide using potash (KOH). This new method only required 0.5 g of sample and a single operator, and Liebig moved the sample
through the apparatus by sucking on an opening at the far right end of the apparatus.

1.3.2 https://chem.libretexts.org/@go/page/55812
Figure 1.3.6 : German chemist Justus von Liebig (1803-1873).

Figure 1.3.7 : Print of von Liebig's "combustion train" apparatus for determining carbon and hydrogen composition. J. Von Liebig,
Annalen der Physik und Chemie, 1831, 21.

Figure 1.3.8 : Photo of von Liebig's "combustion train apparatus" for determining carbon and hydrogen composition. The Oesper
Collections in the History of Chemistry, Apparatus Museum, University of Cincinnati, Case 10, Combustion Analysis. For a 360o
view of this apparatus, click here.
Jean-Baptiste André Dumas (Figure 1.3.9)) used a similar combustion train to Liebig. However, he added a U-shaped aspirator that
prevented atmospheric moisture from entering the apparatus (Figure 1.3.10)).

1.3.3 https://chem.libretexts.org/@go/page/55812
Figure 1.3.9 : French chemist Jean-Baptiste André Dumas (1800-1844).

Figure 1.3.10 : Dumas' apparatus; note the aspirator at 8. Sourced from J. A. Dumas, Ann. der Chem. and Pharm., 1841, 38, 141.
In 1923, Fritz Pregl (Figure 1.3.11)) received the Nobel Prize for inventing a micro-analysis method of combustion. This method
required only 5 mg or less, which is 0.01% of the amount required in Lavoisier's apparatus.

Figure 1.3.11 : Austrian chemist and physician Fritz Pregl (1869-1930).


Today, combustion analysis of an organic or organometallic compound only requires about 2 mg of sample. Although this method
of analysis destroys the sample and is not as sensitive as other techniques, it is still considered a necessity for characterizing an
organic compound.

Categories of combustion

1.3.4 https://chem.libretexts.org/@go/page/55812
Basic flame types
There are several categories of combustion, which can be identified by their flame types (Table 1.3.1). At some point in the
combustion process, the fuel and oxidant must be mixed together. If these are mixed before being burned, the flame type is referred
to as a premixed flame, and if they are mixed simultaneously with combustion, it is referred to as a nonpremixed flame. In addition,
the ow of the flame can be categorized as either laminar (streamlined) or turbulent (Figure 1.3.12).
Table 1.3.1 : Types of combustion systems with examples. Adapted from J. Warnatz, U. Maas, and R. W. Dibble, Combustion: Physical and
Chemical Fundamentals, Modeling and Simulation, Experiments, Pollutant Formation, 3rd Ed., Springer, Berlin (2001).
Fuel/oxidizer mixing Fluid motion Examples

Spark-ignited gasoline engine, low NOx


Premixed Turbulent
stationary gas turbine
Flat flame, Bunsen flame (followed by a
Premixed Laminar
nonpremixed candle for Φ>1)
Pulverized coal combustion, aircraft turbine,
Nonpremixed Turbulent
diesel engine, H2/O2 rocket motor

Nonpremixed Laminar Wood fire, radiant burners for heating, candle

Figure 1.3.12 : Schematic representation of (a) laminar flow and (b) turbulent flow.
The amount of oxygen in the combustion system can alter the ow of the flame and the appearance. As illustrated in Figure 1.3.13, a
flame with no oxygen tends to have a very turbulent flow, while a flame with an excess of oxygen tends to have a laminar flow.

1.3.5 https://chem.libretexts.org/@go/page/55812
Figure 1.3.13 : Bunsen burner flames with varying amounts of oxygen and constant amount of fuel. (1) air valve completely closed,
(2) air valve slightly open, (3) air valve half open, (4) air valve completely open.

Stoichiometric combustion and calculations


A combustion system is referred to as stoichiometric when all of the fuel and oxidizer are consumed and only carbon dioxide and
water are formed. On the other hand, a fuel-rich system has an excess of fuel, and a fuel-lean system has an excess of oxygen
(Table 1.3.2).
Table 1.3.2 : Examples of stoichiometric, fuel-rich, and fuel-lean systems.
Combustion type Reaction example

Stoichiometric 2H
2
+O
2
⟶ 2H O
2

Fuel-rich (H left over)


2
3H
2
+O
2
⟶ 2H O
2
+
H
2

Fuel-lean (O left over)


2
CH
4
+3O
2
⟶ 2H O
2
+
CO
2
+O
2

If the reaction of a stoichiometric mixture is written to describe the reaction of exactly 1 mol of fuel (H in this case), then the mole 2

fraction of the fuel content can be easily calculated as follows, where ν denotes the mole number of O in the combustion reaction 2

equation for a complete reaction to H O and CO ,


2 2

1
xfuel, stoich = (1.3.3)
1 +v

For example, in the reaction


1
H + O → H O +H
2 2 2 2 2 2

we have v = 1

2
, so the stoichiometry is calculated as
1
xH ,stoich = = 2/3
2
1 + 0.5

However, as calculated this reaction would be for the reaction in an environment of pure oxygen. On the other hand, air has only
21% oxygen (78% nitrogen, 1% noble gases). Therefore, if air is used as the oxidizer, this must be taken into account in the
calculations, i.e.
xN = 3.762(xO ) (1.3.4)
2 2

1.3.6 https://chem.libretexts.org/@go/page/55812
The mole fractions for a stoichiometric mixture in air are therefore calculated in following way:
1
xfuel, stoich = (1.3.5)
1 + v(4.762)

xO ,stoich = v(xfuel, stoich) (1.3.6)


2

xN ,stoich = 3.762(xO ,stoich) (1.3.7)


2 2

Example 1.3.1:

Calculate the fuel mole fraction (x fuel


) for the stoichiometric reaction:

CH +2 O + (2 × 3.762)N → CO + 2 H O + (2 × 3.762)N
4 2 2 2 2 2

Solution
In this reaction ν = 2, as 2 moles of oxygen are needed to fully oxidize methane into H 2
O and
CO .
2

1
xfuel, stoich = = 0.09502 = 9.502 mol%
1 + 2 × 4.762

Exercise 1.3.1

Calculate the fuel mole fraction for the stoichiometric reaction:

C H +5 O + (5 × 3.762)N → 3 CO + 4 H O + (5 × 3.762)N
3 8 2 2 2 2 2

Answer
The fuel mole fraction is 4.03%

Premixed combustion reactions can also be characterized by the air equivalence ratio, λ :
xair / xfuel
λ = (1.3.8)
xair, stoich/ xfuel,stoich

The fuel equivalence ratio, Φ, is the reciprocal of this value


Φ = 1/λ (1.3.9)

Rewriting 1.3.5 in terms of the fuel equivalence ratio gives:


1
xfuel = (1.3.10)
1 + v(4.672/Φ)

xair = 1 − xfuel (1.3.11)

xO = xair /4.762 (1.3.12)


2

xN = 3.762(xO ) (1.3.13)
2 2

The premixed combustion processes can also be identified by their air and fuel equivalence ratios (Table 1.3.3 ).
Table 1.3.3 : Identification of combustion type by Φ and λ values.
Type of combustion Φ λ

Rich >1 <1

Stoichiometric =1 =1

Lean <1 >1

1.3.7 https://chem.libretexts.org/@go/page/55812
With a premixed type of combustion, there is much greater control over the reaction. If performed at lean conditions, then high
temperatures, the pollutant nitric oxide, and the production of soot can be minimized or even avoided, allowing the system to
combust efficiently. However, a premixed system requires large volumes of premixed reactants, which pose a fire hazard. As a
result, nonpremixed combusted, while not being efficient, is more commonly used.

Instrumentation
Though the instrumentation of combustion analysis has greatly improved, the basic components of the apparatus (Figure 1.14) have
not changed much since the late 18th century.

Figure 1.3.14 : Combustion apparatus from the 19th century. The Oesper Collections in the History of Chemistry Apparatus
Museum, University of Cincinnati, Case 10, Combustion Analysis. For a 360o view of this apparatus, click here.
The sample of an organic compound, such as a hydrocarbon, is contained within a furnace or exposed to a ame and burned in the
presence of oxygen, creating water vapor and carbon dioxide gas (Figure 1.3.15). The sample moves first through the apparatus to
a chamber in whichH O is absorbed by a hydrophilic substance and second through a chamber in which CO is absorbed. The
2 2

change in weight of each chamber is determined to calculate the weight of H O and CO . After the masses of H O and CO have
2 2 2 2

been determined, they can be used to characterize and calculate the composition of the original sample.

Figure 1.3.15 : Typical modern combustion apparatus with a furnace.

Calculations and determining chemical formulas


Hydrocarbons
Combustion analysis is a standard method of determining a chemical formula of a substance that contains hydrogen and carbon.
First, a sample is weighed and then burned in a furnace in the presence of excess oxygen. All of the carbon is converted to carbon
dioxide, and the hydrogen is converted to water in this way. Each of these are absorbed in separate compartments, which are
weighed before and after the reaction. From these measurements, the chemical formula can be determined.
Generally, the following reaction takes place in combustion analysis:

1.3.8 https://chem.libretexts.org/@go/page/55812
Ca H + O (xs) → aCO + b/2 H O (1.3.14)
b 2 2 2

Example 1.3.2:

After burning 1.333 g of a hydrocarbon in a combustion analysis apparatus, 1.410 g of H O and 4.305 g of CO were 2 2

produced. Separately, the molar mass of this hydrocarbon was found to be 204.35 g/mol. Calculate the empirical and molecular
formulas of this hydrocarbon.
Step 1: Using the molar masses of water and carbon dioxide, determine the moles of hydrogen and carbon that were produced.
1 mol H O 2 mol H
2
1.410 g H O × × = 0.1565 mol H
2
18.015 g H O 1 mol H O
2 2

1 mol CO 1 mol C
2
4.3051 g CO × × = 0.09782 mol C
2
44.010 g CO 1 mol CO
2 2

Step 2: Divide the larger molar amount by the smaller molar amount. In some cases, the ratio is not made up of two integers.
Convert the numerator of the ratio to an improper fraction and rewrite the ratio in whole numbers as shown

0.1565 mol H 1.600 mol H 16/10 mol H 8/5 mol H 8 mol H


= = = =
0.09782 mol C 1 mol C 1 mol C 1 mol C 5 mol C

Therefore, the empirical formula is C 5


H
8
.
Step 3: To get the molecular formula, divide the experimental molar mass of the unknown hydrocarbon by the empirical
formula weight.
 Molar mass  204.35 g/mol
= =3
 Empirical formula weight  68.114 g/mol

Therefore, the molecular formula is (C 5


H )
8 3
or C 15
H
24
.

Exercise 1.3.2

After burning 1.082 g of a hydrocarbon in a combustion analysis apparatus, 1.583 g of H O and 3.315 g of CO were 2 2

produced. Separately, the molar mass of this hydrocarbon was found to be 258.52 g/mol. Calculate the empirical and molecular
formulas of this hydrocarbon.

Answer
The empirical formula is C 3
H
7
, and the molecular formula is (C 3
H )
7 6
orC 18
H
42
.

Compounds containing carbon, hydrogen, and oxygen


Combustion analysis can also be utilized to determine the empiric and molecular formulas of compounds containing carbon,
hydrogen, and oxygen. However, as the reaction is performed in an environment of excess oxygen, the amount of oxygen in the
sample can be determined from the sample mass, rather than the combustion data

Example 1.3.3:

A 2.0714 g sample containing carbon, hydrogen, and oxygen was burned in a combustion analysis apparatus; 1.928 g of H O 2

and 4.709 g of CO were produced. Separately, the molar mass of the sample was found to be 116.16 g/mol. Determine the
2

empirical formula, molecular formula, and identity of the sample.


Step 1: Using the molar masses of water and carbon dioxide, determine the moles of hydrogen and carbon that were produced.
1 mol H O 2 mol H
2
1.928 g H O × × = 0.2140 mol H
2
18.015 g H O 1 mol H O
2 2

1.3.9 https://chem.libretexts.org/@go/page/55812
1 mol CO 1 mol C
2
4.709 g CO × × = 0.1070 mol C
2
44.010 g CO 1 mol CO
2 2

Step 2: Using the molar amounts of carbon and hydrogen, calculate the masses of each in the original sample.
1.008 g H
0.2140 mol H × = 0.2157 g H
1 mol H

12.011 g C
0.1070 mol C × = 1.285 g C
1 mol C

Step 3: Subtract the masses of carbon and hydrogen from the sample mass. Now that the mass of oxygen is known, use this to
calculate the molar amount of oxygen in the sample.
2.0714g sample − 0.2157 g H − 1.285 g C = 0.5707 g O

1 mol O
0.5707 mol O × = 0.03567 g O
16.00 g O

Step 4: Divide each molar amount by the smallest molar amount in order to determine the ratio between the three elements.
0.03567 mol O
= 1.00 mol O = 1 mol O
0.03567

0.1070 mol C
= 3.00mol C = 3 mol C
0.03567

0.2140 mol H
= 5.999 mol H = 6 mol H
0.03567

Therefore, the empirical formula is C


3
H O
6
.
Step 5: To get the molecular formula, divide the experimental molar mass of the unknown hydrocarbon by the empirical
formula weight.
 Molar mass  116.16 g/mol
= =2
 Empirical formula weight  58.08 g/mol

Therefore, the molecular formula is (C 3


H O)
6 2
or C
6
H
12
O
2
.

1.3.10 https://chem.libretexts.org/@go/page/55812
Structure of possible compounds with the molecular formula C H O : (a) butylacetate, (b) sec-butyl acetate, (c) tert-butyl
6 12 2

acetate, (d) ethyl butyrate, (e) haxanoic acid, (f) isobutyl acetate, (g) methyl pentanoate, and (h) propyl proponoate.

Exercise 1.3.3

A 4.846 g sample containing carbon, hydrogen, and oxygen was burned in a combustion analysis apparatus; 4.843 g of \
(\ce{H2O}\) and 11.83 g of \(\ce{CO2}\) were produced. Separately, the molar mass of the sample was found to be 144.22
g/mol. Determine the empirical formula, molecular formula, and identity of the sample.

Answer
The empirical formula is \(\ce{C4H8O}\), and the molecular formula is (\(\ce{C4H8O)2}\) or \(\ce{C8H16O2}\).

1.3.11 https://chem.libretexts.org/@go/page/55812
Structure of possible compounds with the molecular formula C H O : (a) octanoic acid (caprylic acid), (b) hexyl acetate,
8 16 2

(c) pentyl proponate, (d) 2-ethyl hexanoic acid, (e) valproic acid (VPA), (f) cyclohexanedimethanol (CHDM), and (g)
2,2,4,4-tetramethyl-1,3-cyclobutandiol (CBDO).

Binary compounds
By using combustion analysis, the chemical formula of a binary compound containing oxygen can also be determined. This is
particularly helpful in the case of combustion of a metal which can result in potential oxides of multiple oxidation states.

Example 1.3.4:

A sample of iron weighing 1.7480 g is combusted in the presence of excess oxygen. A metal oxide ( \ce{Fe_{x}O_{y})} is
formed with a mass of 2.4982 g. Determine the chemical formula of the oxide product and the oxidation state of Fe.
Step 1: Subtract the mass of Fe from the mass of the oxide to determine the mass of oxygen in the product.

2.4982 g Fex Oy − 1.7480 g Fe = 0.7502 g O

Step 2: Using the molar masses of Fe and O, calculate the molar amounts of each element.
1 mol Fe 
1.7480g Fe × = 0.031301 mol Fe 
55.845 g Fe 

1 mol O 
0.7502  g  × = 0.04689  mol O 
16.00  g O

1.3.12 https://chem.libretexts.org/@go/page/55812
Step 3: Divide the larger molar amount by the smaller molar amount. In some cases, the ratio is not made up of two integers.
Convert the numerator of the ratio to an improper fraction and rewrite the ratio in whole numbers as shown.
2
0.031301  mol Fe  0.6675 mol Fe mol Fe 2 mol Fe
3
= = =
0.04689 mol O 1 mol O 1 mol O 3 mol O

Therefore, the chemical formula of the oxide is Fe 2


O
3
, and Fe has a 3+ oxidation state.

Exercise 1.3.4

A sample of copper weighing 7.295 g is combusted in the presence of excess oxygen. A metal oxide (Cu x Oy ) is formed with a
mass of 8.2131 g. Determine the chemical formula of the oxide product and the oxidation state of Cu.

Answer
The chemical formula is Cu 2
O , and Cu has a 1+ oxidation state..

Bibliography
J. A. Dumas, Ann. Chem. Pharm., 1841, 38, 141.
H. Goldwhite, J. Chem. Edu., 1978, 55, 366.
A. Lavoisier, Traité Élémentaire de Chimie, 1789, 2, 493.
J. Von Liebig, Annalen der Physik und Chemie, 1831, 21, 1.
A. Linan and F. A. Williams, Fundamental Aspects of Combustion, Oxford University Press, New York (1993).
J. M. McBride, "Combustion Analysis," Chemistry 125, Yale University.
W. Prout, Philos. T. R. Soc. Lond., 1827, 117, 355.
D. Shriver and P. Atkins, Inorganic Chemistry, 5th Ed., W. H. Freeman and Co., New York (2009).
W. Vining et. al., General Chemistry, 1st Ed., Cengage, Brooks/Cole Cengage Learning, University of Massachusetts Amherst
(2014).
J. Warnatz, U. Maas, and R. W. Dibble, Combustion: Physical and Chemical Fundamentals, Modeling and Simulation,
Experiments, Pollutant Formation, 3rd Ed., Springer, Berlin (2001)

1.3: Introduction to Combustion Analysis is shared under a CC BY 4.0 license and was authored, remixed, and/or curated by Pavan M. V. Raja &
Andrew R. Barron via source content that was edited to conform to the style and standards of the LibreTexts platform; a detailed edit history is
available upon request.

1.3.13 https://chem.libretexts.org/@go/page/55812
1.4: Introduction to Atomic Absorption Spectroscopy
Brief overview of atomic absorption spectroscopy
History of atomic absorption spectroscopy
The earliest spectroscopy was first described by Marcus Marci von Kronland in 1648 by analyzing sunlight as is passed through
water droplets and thus creating a rainbow. Further analysis of sunlight by William Hyde Wollaston (Figure 1.4.1) led to the
discovery of black lines in the spectrum, which in 1820 Sir David Brewster (Figure 1.4.2) explained as absorption of light in the
sun’s atmosphere.

Figure 1.4.1 : English chemist and physicist William Hyde Wollaston (1659 - 1724).Figure 1.4.2 : Scottish physicist,
mathematician, astronomer, inventor, writer and university principal Sir David Brewster (1781 - 1868).
Robert Bunsen (Figure 1.4.3) and Gustav Kirchhoff (Figure 1.4.4) studied the sodium spectrum and came to the conclusion that
every element has its own unique spectrum that can be used to identify elements in the vapor phase. Kirchoff further explained the
phenomenon by stating that if a material can emit radiation of a certain wavelength, that it may also absorb radiation of that
wavelength. Although Bunsen and Kirchoff took a large step in defining the technique of atomic absorption spectroscopy (AAS), it
was not widely utilized as an analytical technique except in the field of astronomy due to many practical difficulties.

1.4.1 https://chem.libretexts.org/@go/page/55813
Figure 1.4.3 : German chemist Robert Bunsen (1811 - 1899). Figure 1.4.4 : German physicist Gustav Robert Kirchhoff (1824 -
1887).Figure 1.4.5 : British physicist Sir Alan Walsh (1916 - 1988).
In 1953, Alan Walsh (Figure 1.4.5) drastically improved the AAS methods. He advocated AAS to many instrument manufacturers,
but to no avail. Although he had improved the methods, he hadn’t shown how it could be useful in any applications. In 1957, he
discovered uses for AAS that convinced manufactures market the first commercial AAS spectrometers. Since that time, AAS's
popularity has fluctuated as other analytical techniques and improvements to the methods are made.

Theory of atomic absorption spectroscopy


In order to understand how atomic absorption spectroscopy works, some background information is necessary. Atomic theory
began with John Dalton (Figure 1.4.6) in the 18th century when he proposed the concept of atoms, that all atoms of an element are
identical, and that atoms of different elements can combine to form molecules. In 1913, Niels Bohr (Figure 1.4.7) revolutionized
atomic theory by proposing quantum numbers, a positively charged nucleus, and electrons orbiting around the nucleus in the what
became known as the Bohr model of the atom. Soon afterward, Louis deBroglie (Figure 1.4.8) proposed quantized energy of
electrons, which is an extremely important concept in AAS. Wolfgang Pauli (Figure 1.4.9) then elaborated on deBroglie’s theory
by stating that no two electrons can share the same four quantum numbers. These landmark discoveries in atomic theory are
necessary in understanding the mechanism of AAS.

1.4.2 https://chem.libretexts.org/@go/page/55813
Figure 1.4.6 : English chemist, physicist, and meteorologist John Dalton FRS (1766 - 1844).Figure 1.4.7 : Danish physicist Niels
Henrik David Bohr (1885 - 1962).Figure 1.4.8 : French physicist and a Nobel laureate Louis de Broglie (1892 - 1987). (public
domain).Figure 1.4.9 : Austrian physicist Wolfgang Pauli (1900 - 1958).
Atoms have valence electrons, which are the outermost electrons of the atom. Atoms can be excited when irradiated, which creates
an absorption spectrum. When an atom is excited, the valence electron moves up an energy level. The energies of the various
stationary states, or restricted orbits, can then be determined by these emission lines. The resonance line is then defined as the
specific radiation absorbed to reach the excited state.

1.4.3 https://chem.libretexts.org/@go/page/55813
The Maxwell-Boltzmann equation gives the number of electrons in any given orbital. It relates the distribution to the thermal
temperature of the system (as opposed to electronic temperature, vibrational temperature, or rotational temperature). Plank
proposed radiation emitted energy in discrete packets (quanta),
E = hν (1.4.1)

which can be related to Einstein’s equation


2
E = mc (1.4.2)

Both atomic emission and atomic absorption spectroscopy can be used to analyze samples. Atomic emission spectroscopy
measures the intensity of light emitted by the excited atoms, while atomic absorption spectroscopy measures the light absorbed by
atomic absorption. This light is typically in the visible or ultraviolet region of the electromagnetic spectrum. The percentage is then
compared to a calibration curve to determine the amount of material in the sample. The energy of the system can be used to find the
frequency of the radiation, and thus the wavelength through the combination of equations 1.4.2 and 1.4.3.
ν = c/λ (1.4.3)

Because the energy levels are quantized, only certain wavelengths are allowed and each atom has a unique spectrum. There are
many variables that can affect the system. For example, if the sample is changed in a way that increases the population of atoms,
there will be an increase in both emission and absorption and vice versa. There are also variables that affect the ratio of excited to
unexcited atoms such as an increase in temperature of the vapor.

Applications of Atomic Absorption Spectroscopy


There are many applications of atomic absorption spectroscopy (AAS) due to its specificity. These can be divided into the broad
categories of biological analysis, environmental and marine analysis, and geological analysis.

Biological analysis
Biological samples can include both human tissue samples and food samples. In human tissue samples, AAS can be used to
determine the amount of various levels of metals and other electrolytes, within tissue samples. These tissue samples can be many
things including but not limited to blood, bone marrow, urine, hair, and nails. Sample preparation is dependent upon the sample.
This is extremely important in that many elements are toxic in certain concentrations in the body, and AAS can analyze what
concentrations they are present in. Some examples of trace elements that samples are analyzed for are arsenic, mercury, and lead.
An example of an application of AAS to human tissue is the measurement of the electrolytes sodium and potassium in plasma. This
measurement is important because the values can be indicative of various diseases when outside of the normal range. The typical
method used for this analysis is atomization of a 1:50 dilution in strontium chloride (SrCl ) using an air-hydrogen flame. The
2

sodium is detected at its secondary line (330.2 nm) because detection at the first line would require further dilution of the sample
due to signal intensity. The reason that strontium chloride is used is because it reduces ionization of the potassium and sodium ions,
while eliminating phosphate’s and calcium’s interference.
In the food industry, AAS provides analysis of vegetables, animal products, and animal feeds. These kinds of analyses are some of
the oldest application of AAS. An important consideration that needs to be taken into account in food analysis is sampling. The
sample should be an accurate representation of what is being analyzed. Because of this, it must be homogenous, and many it is
often needed that several samples are run. Food samples are most often run in order to determine mineral and trace element
amounts so that consumers know if they are consuming an adequate amount. Samples are also analyzed to determine heavy metals
which can be detrimental to consumers.

Environmental and marine analysis


Environmental and marine analysis typically refers to water analysis of various types. Water analysis includes many things ranging
from drinking water to waste water to sea water. Unlike biological samples, the preparation of water samples is governed more by
laws than by the sample itself. The analytes that can be measured also vary greatly and can often include lead, copper, nickel, and
mercury.
An example of water analysis is an analysis of leaching of lead and zinc from tin-lead solder into water. The solder is what binds
the joints of copper pipes. In this particular experiment, soft water, acidic water, and chlorinated water were all analyzed. The
sample preparation consisted of exposing the various water samples to copper plates with solder for various intervals of time. The

1.4.4 https://chem.libretexts.org/@go/page/55813
samples were then analyzed for copper and zinc with air-acetylene flame AAS. A deuterium lamp was used. For the samples that
had copper levels below 100 µg/L, the method was changed to graphite furnace electrothermal AAS due to its higher sensitivity.

Geological analysis
Geological analysis encompasses both mineral reserves and environmental research. When prospecting mineral reserves, the
method of AAS used needs to be cheap, fast, and versatile because the majority of prospects end up being of no economic use.
When studying rocks, preparation can include acid digestions or leaching. If the sample needs to have silicon content analyzed,
acid digestion is not a suitable preparation method.
An example is the analysis of lake and river sediment for lead and cadmium. Because this experiment involves a solid sample,
more preparation is needed than for the other examples. The sediment was first dried, then grounded into a powder, and then was
decomposed in a bomb with nitric acid (HNO ) and perchloric acid (HClO ). Standards of lead and cadmium were prepared.
3 4

Ammonium sulfate ([NH ][SO ]]) and ammonium phosphate ([NH ][3 PO ]]) were added to the samples to correct for the
4 4 4 4

interferences caused by sodium and potassium that are present in the sample. The standards and samples were then analyzed with
electrothermal AAS.

Instrumentation
Atomizer
In order for the sample to be analyzed, it must first be atomized. This is an extremely important step in AAS because it determines
the sensitivity of the reading. The most effective atomizers create a large number of homogenous free atoms. There are many types
of atomizers, but only two are commonly used: flame and electrothermal atomizers.
Flame atomizer
Flame atomizers (Figure 1.4.10) are widely used for a multitude of reasons including their simplicity, low cost, and long length of
time that they have been utilized. Flame atomizers accept an aerosol from a nebulizer into a flame that has enough energy to both
volatilize and atomize the sample. When this happens, the sample is dried, vaporized, atomized, and ionized. Within this category
of atomizers, there are many subcategories determined by the chemical composition of the flame. The composition of the flame is
often determined based on the sample being analyzed. The flame itself should meet several requirements including sufficient
energy, a long length, non-turbulent, and safe.

Figure 1.4.10 : A schematic diagram of a flame atomizer showing the oxidizer inlet (1) and fuel inlet (2).
Electrothermal atomizer
Although electrothermal atomizers were developed before flame atomizers, they did not become popular until more recently due to
improvements made to the detection level. They employ graphite tubes that increase temperature in a stepwise manner.
Electrothermal atomization first dries the sample and evaporates much of the solvent and impurities, then atomizes the sample, and
then rises it to an extremely high temperature to clean the graphite tube. Some requirements for this form of atomization are the
ability to maintain a constant temperature during atomization, have rapid atomization, hold a large volume of solution, and emit
minimal radiation. Electrothermal atomization is much less harsh than the method of flame atomization.

1.4.5 https://chem.libretexts.org/@go/page/55813
Figure 1.4.11 :Schematic diagram of an electrothermal atomizer showing the external gas flow inlet (1), the external gas flow outlet
(2), the internal gas flow outlet (3), the internal gas flow inlet (4), and the light beam (5).

Radiation source
The radiation source then irradiates the atomized sample. The sample absorbs some of the radiation, and the rest passes through the
spectrometer to a detector. Radiation sources can be separated into two broad categories: line sources and continuum sources. Line
sources excite the analyte and thus emit its own line spectrum. Hollow cathode lamps and electrodeless discharge lamps are the
most commonly used examples of line sources. On the other hand, continuum sources have radiation that spreads out over a wider
range of wavelengths. These sources are typically only used for background correction. Deuterium lamps and halogen lamps are
often used for this purpose.

Spectrometer
Spectrometers are used to separate the different wavelengths of light before they pass to the detector. The spectrometer used in
AAS can be either single-beam or double-beam. Single-beam spectrometers only require radiation that passes directly through the
atomized sample, while double-beam spectrometers (Figure 1.4.12), as implied by the name, require two beams of light; one that
passes directly through the sample, and one that does not pass through the sample at all. (Insert diagrams) The single-beam
spectrometers have less optical components and therefore suffer less radiation loss. Double-beam monochromators have more
optical components, but they are also more stable over time because they can compensate for changes more readily.

Figure 1.4.12 : A schematic of a double-beam spectrometer showing the 50/50 beam splitters (1) and the mirrors (2).

Obtaining Measurements
Sample preparation
Sample preparation is extremely varied because of the range of samples that can be analyzed. Regardless of the type of sample,
certain considerations should be made. These include the laboratory environment, the vessel holding the sample, storage of the
sample, and pretreatment of the sample.
Sample preparation begins with having a clean environment to work in. AAS is often used to measure trace elements, in which case
contamination can lead to severe error. Possible equipment includes laminar flow hoods, clean rooms, and closed, clean vessels for
transportation of the sample. Not only must the sample be kept clean, it also needs to be conserved in terms of pH, constituents, and
any other properties that could alter the contents.
When trace elements are stored, the material of the vessel walls can adsorb some of the analyte leading to poor results. To correct
for this, perfluoroalkoxy polymers (PFA), silica, glassy carbon, and other materials with inert surfaces are often used as the storage

1.4.6 https://chem.libretexts.org/@go/page/55813
material. Acidifying the solution with hydrochloric or nitric acid can also help prevent ions from adhering to the walls of the vessel
by competing for the space. The vessels should also contain a minimal surface area in order to minimize possible adsorption sites.
Pretreatment of the sample is dependent upon the nature of the sample. See Table 1.4.1 for sample pretreatment methods.
Table 1.4.1 Sample pretreatment methods for AAS.
Sample Examples Pretreatment method

Digestion if interference causing substituents


Aqueous solutions Water, beverages, urine, blood
are present
Solid matter must either be removed by
filtration, centrifugation or digestion, and then
Suspensions Water, beverages, urine, blood
the methods for aqueous solutions can be
followed
Either direct measurement with AAS or diltion
with organic material followed by
Organic liquids Fuels, oils measurement with AAS, standards must
contain the analyte in the same form as the
sample

Solids Foodstuffs, rocks Digestion followed by electrothermal AAS

Calibration curve
In order to determine the concentration of the analyte in the solution, calibration curves can be employed. Using standards, a plot of
concentration versus absorbance can be created. Three common methods used to make calibration curves are the standard
calibration technique, the bracketing technique, and the analyte addition technique.
Standard calibration technique

This technique is the both the simplest and the most commonly used. The concentration of the sample is found by comparing its
absorbance or integrated absorbance to a curve of the concentration of the standards versus the absorbances or integrated
absorbances of the standards. In order for this method to be applied the following conditions must be met:
Both the standards and the sample must have the same behavior when atomized. If they do not, the matrix of the standards
should be altered to match that of the sample.
The error in measuring the absorbance must be smaller than that of the preparation of the standards.
The samples must be homogeneous.
The curve is typically linear and involves at least five points from five standards that are at equidistant concentrations from each
other (Figure 1.4.13). This ensures that the fit is acceptable. A least means squares calculation is used to linearly fit the line. In
most cases, the curve is linear only up to absorbance values of 0.5 to 0.8. The absorbance values of the standards should have the
absorbance value of a blank subtracted.

1.4.7 https://chem.libretexts.org/@go/page/55813
Figure 1.4.13 : An example of a calibration curve made for the standard calibration technique.
Bracketing Technique
The bracketing technique is a variation of the standard calibration technique. In this method, only two standards are necessary with
concentrations c and c . They bracket the approximate value of the sample concentration very closely. Applying Equation 1.4.4 to
1 2

determines the value for the sample, where c and A are the concentration and adsorbance of the unknown, and A and A are
x x 1 2

the adsorbance for c and c , respectively.


1 2

(Ax − A1 ) (c1 − c2 )
cx = + c1 (1.4.4)
A2 − A1

This method is very useful when the concentration of the analyte in the sample is outside of the linear portion of the calibration
curve because the bracket is so small that the portion of the curve being used can be portrayed as linear. Although this method can
be used accurately for nonlinear curves, the further the curve is from linear the greater the error will be. To help reduce this error,
the standards should bracket the sample very closely.
Analyte Addition Technique
The analyte addition technique is often used when the concomitants in the sample are expected to create many interferences and the
composition of the sample is unknown. The previous two techniques both require that the standards have a similar matrix to that of
the sample, but that is not possible when the matrix is unknown. To compensate for this, the analyte addition technique uses an
aliquot of the sample itself as the matrix. The aliquots are then spiked with various amounts of the analyte. This technique must be
used only within the linear range of the absorbances.

Measurement Interference
Interference is caused by contaminants within the sample that absorb at the same wavelength as the analyte, and thus can cause
inaccurate measurements. Corrections can be made through a variety of methods such as background correction, addition of
chemical additives, or addition of analyte.
Table 1.4.2 : Examples of interference in AAS.
Interference type Cause of interference Result Example Correction measures

Very rare, with the only


Typically doesn't occur in
Spectral profile of two Higher experimental plausable problem being
practical situations, so there
Atomic line overlap elements are within 0.01 absorption value than the that of copper (324.754
is no established correction
nm of each other real value nm) and europium
method
(324.753 nm)

Spectral profile of an Higher experimental Calcium hydroxide and


Molecular band and line
element overlaps with absorption value than the barium at 553.6 nm in a Background correction
overlap
molecular band real value air-acetylene flame

1.4.8 https://chem.libretexts.org/@go/page/55813
Interference type Cause of interference Result Example Correction measures

atoms are ionized at the


Add an ionization
temperature of the Lower experimental Problems commonly occur
Ionization (vapor-phase or suppressor (or buffer) to
flame/furnace, which absorption value than real with cesium, potassium,
cation enhancement) both the sample and the
decreases the amount of value and sodium
standards
free atoms

Solid particles scatter the


High in samples with
beam of light lowering the Higher experimental
many refractory elements, Matrix modifaction and/or
Light scattering intensity of the beam absorption value than the
highest at UV wavelengths background correction
entering the real value
(add specific example)
monochromater

Calcium and phosphate Increase the temperature of


The chemical being
ions form calcium the flame if flame AAS is
analyzed is contained Lower experimental
phosphate which is then being used, use a releasing
Chemical withing a compound in the absorption value than real
converted to calcium chemical, or standard
analyte that is not value
pyrophosphate which is addition for electrothermal
atomized
stable in high heat AAS

If physical properties of
the sample and the
standards are different, Can vary in either Viscosity differences, Alter the standards to have
Physical atomization can be direction depending upon surface tension differences, similar physical properties
affected thus affecting the the conditions etc to the samples
number of free atom
population

In electrothermal
Chlorides are very volatile,
atomization, interference
so they need to be
will occur if the rate of Change the matrix by
Can vary in either converted to a less volatile
volatilization is not the standard addition, or
Volitalization direction depending upon form. Often this is done by
same for the sample as for selectively volatileze
the conditions the addition of nitrate or
the standard, which is components of the matrix
slufate. Zinc and lead are
often caused by a heavy
also highly problamatic
matrix

Bibliography
L. Ebon, A. Fisher and S. J. Hill, An Introduction to Analytical Atomic Spectrometry, Ed. E. H. Evans, Wiley, New York (1998).
B. Welz and M. Sperling, Atomic Absorption Spectrometry, 3rd Ed, Wiley-VCH, New York (1999).
J. W. Robinson, Atomic Spectroscopy, 2nd Ed. Marcel Dekker, Inc., New York (1996).
K. S. Subramanian, Water Res., 1995, 29, 1827.
M. Sakata and O. Shimoda, Water Res., 1982, 16, 231.
J. C. Van Loon, Analytical Atomic Absorption Spectroscopy Selected Methods, Academic Press, New York (1980).

1.4: Introduction to Atomic Absorption Spectroscopy is shared under a CC BY 4.0 license and was authored, remixed, and/or curated by Pavan
M. V. Raja & Andrew R. Barron via source content that was edited to conform to the style and standards of the LibreTexts platform; a detailed
edit history is available upon request.

1.4.9 https://chem.libretexts.org/@go/page/55813
1.5: ICP-AES Analysis of Nanoparticles
What is ICP-AES?
Inductively coupled plasma atomic emission spectroscopy (ICP-AES) is a spectral method used to determine very precisely the
elemental composition of samples; it can also be used to quantify the elemental concentration with the sample. ICP-AES uses high-
energy plasma from an inert gas like argon to burn analytes very rapidly. The color that is emitted from the analyte is indicative of
the elements present, and the intensity of the spectral signal is indicative of the concentration of the elements that is present. A
schematic view of a typical experimental set-up is shown here.

Figure 1.5.1 : Schematic representation of an ICP-AES set-up.

How does ICP-AES work?


ICP-AES works by the emission of photons from analytes that are brought to an excited state by the use of high-energy plasma.
The plasma source is induced when passing argon gas through an alternating electric field that is created by an inductively couple
coil. When the analyte is excited the electrons try to dissipate the induced energy moving to a ground state of lower energy, in
doing this they emit the excess energy in the form of light. The wavelength of light emitted depends on the energy gap between the
excited energy level and the ground state. This is specific to the element based on the number of electrons the element has and
electron orbital’s are filled. In this way the wavelength of light can be used to determine what elements are present by detection of
the light at specific wavelengths.
As a simple example consider the situation when placing a piece of copper wire into the flame of a candle. The flame turns green
due to the emission of excited electrons within the copper metal, as the electrons try to dissipate the energy incurred from the
flame, they move to a more stable state emitting energy in the form of light. The energy gap between the excited state to the ground
state (ΔE dictates the color of the light or wavelength of the light, Equation 1.5.1, where h is Plank's constant (6.626×10-34
m2kg/s), and ν is the frequency of the emitted light.

ΔE = hν (1.5.1)

The wavelength of light is indicative of the element present. If another metal is placed in the flame such as iron a different color
flame will be emitted because the electronic structure of iron is different from that of copper. This is a very simple analogy for what
is happening in ICP-AES and how it is used to determine what elements are present. By detecting the wavelength of light that is
emitted from the analyte one can deduce what elements are be present.
Naturally if there is a lot of the material present then there will be an accumulative effect making the intensity of the signal large.
However, if there were very little materials present the signal would be low. By this rationale one can create a calibration curve
from analyte solutions of known concentrations, whereby the intensity of the signal changes as a function of the concentration of
the material that is present. When measuring the intensity from a sample of unknown concentration the intensity from this sample
can be compared to that from the calibration curve, so this can be used to determine the concentration of the analytes within the
sample.

1.5.1 https://chem.libretexts.org/@go/page/55818
ICP-AES of Nanoparticles to Determine Elemental Composition
As with any sample being studied by ICP-AES nanoparticles need to be digested so that all the atoms can be vaporized in the
plasma equally. If a metal containing nanoparticle were not digested using a strong acid to bring the metals atoms into solution, the
form of the particle could hinder some of the material being vaporized. The analyte would not be detected even though it is present
in the sample and this would give an erroneous result. Nanoparticles are often covered with a protective layer of organic ligands
and this must be removed also. Further to this the solvent used for the nanoparticles may also be an organic solution and this should
be removed as it too will not be miscible in the aqueous medium.
Several organic solvents have low vapor pressures so it is relatively easy to remove the solvent by heating the samples, removing
the solvent by evaporation. To remove the organic ligands that are present on the nanoparticle, choric acid can be used. This is a
very strong acid and can break down the organic ligands readily. To digest the particles and get the metal into solution concentrated
nitric acid is often used.
A typical protocol may use 0.5 mL of concentrated nanoparticle solution and digest this with 9.5 mL of concentrated nitric acid
over the period of a few days. After which 0.5 mL of the digested solution is placed in 9.5 mL of nanopure water. The reason why
nanopure water is used is because DI water or regular water will have some amount of metals ions present and these will be
detected by the ICP-AES measurement and will lead to figures that are not truly representative of the analyte concentration alone.
This is especially pertinent when there is a very a low concentration of metal analyte to be detected, and is even more a problem
when the metal to be detected is commonly found in water such as iron. Once the nanopure water and digested solution are
prepared then the sample is ready for analysis.
Another point to consider when doing ICP-AES on nanoparticles to determine chemical compositions, includes the potential for
wavelength overlap. The energy that is released in the form of light is unique to each element, but elements that are very similar in
atomic structure will have emission wavelengths that are very similar to one another. Consider the example of iron and cobalt, these
are both transition metals and sit right beside each other on the periodic table. Iron has an emission wavelength at 238.204 nm and
cobalt has an emission wavelength at 238.892 nm. So if you were to try determine the amount of each element in an alloy of the
two you would have to select another wavelength that would be unique to that element, and not have any wavelength overlap to
other analytes in solution. For this case of iron and cobalt it would be wiser to use a wavelength for iron detection of 259.940 nm
and a wavelength detection of 228.616 nm. Bearing this in mind a good rule of thumb is to try use the wavelength of the analyte
that affords the best detection primarily. But if this value leads to a possible wavelength overlap of within 15 nm wavelength with
another analyte in the solution then another choice should be made of the detection wavelength to prevent wavelength overlap from
occurring.
Some people have also used the ICP-AES technique to determine the size of nanoparticles. The signal that is detected is determined
by the amount of the material that is present in solution. If very dilute solutions of nanoparticles are being analyzed, particles are
being analyzed one at a time, i.e., there will be one nanoparticle per droplet in the nebulizer. The signal intensity would then differ
according to the size of the particle. In this way the ICP-AES technique could be used to determine the concentration of the
particles in the solution as well as the size of the particles.

Calculations for ICP Concentrations


In order to performe ICP-AES stock solutions must be prepared in dilute nitric acid solutions. To do this a concentrated solution
should be diluted with nanopure water to prepare 7 wt% nitric acid solutions. If the concentrated solution is 69.8 wt% (check the
assay amount that is written on the side of the bottle) then the amount to dilute the solution will be as such:
The density (d ) of HNO is 1.42 g/mL
3

Molecular weight (M ) of HNO is 63.01


W 3

Concentrated percentage 69.8 wt% from assay. First you must determine the molarity of the concentrated solution:
 Molarity  = [(%)(d)/ (MW )] × 10 (1.5.2)

For the present assay amount, the figure will be calculated as follows

M = [(69.8)(1.42)/(63.01)] × 10

∴ M = 15.73

1.5.2 https://chem.libretexts.org/@go/page/55818
This is the initial concentration CI . To determine the molarity of the 7% solution we again use Equation 1.5.2 to find the final
concentration C .
F

M = [(7)(1.42)/(63.01)] × 10

∴ M = 1.58

We use these figures to determine the amount of dilution required to dilute the concentrated nitric acid to make it a 7% solution.
 mass  1 ×  concentration  1 =  mass  F ×  concentration  F (1.5.3)

Now as we are talking about solutions the amount of mass will be measured in mL, and the concentration will be measured as a
molarity. MI and MF have been calculated above.
mL1 ∗ C1 = mLF ∗ CF (1.5.4)

∴ mL1 = [ mLF ∗ CF ] / C1 (1.5.5)

In addition, the amount of dilute solution will be dependent on the user and how much is required by the user to complete the ICP
analysis, for the sake of argument let’s say that we need 10 mL of dilute solution, this is mLF:
mL1 = [10 ∗ 1.58]/15.73 (1.5.6)

∴ mL1 = 10.03mL (1.5.7)

This means that 10.03 mL of the concentrated nitric acid (69.8%) should be diluted up to a total of 100 mL with nanopure water.
Now that you have your stock solution with the correct percentage then you can use this solution to prepare your solutions of
varying concentration. Let’s take the example that the stock solution that you purchase from a supplier has a concentration of 100
ppm of analyte, which is equivalent to 1 μg/mL.
In order to make your calibration curve more accurate it is important to be aware of two issues. Firstly, as with all straight-line
graphs, the more points that are used then the better the statistics is that the line is correct. But, secondly, the more measurements
that are used means that more room for error is introduced to the system, to avoid these errors from occurring one should be very
vigilant and skilled in the use of pipetting and diluting of solutions. Especially when working with very low concentration solutions
a small drop of material making the dilution above or below the exactly required amount can alter the concentration and hence
affect the calibration deleteriously. The premise upon which the calculation is done is based on Equation 1.5.4, whereby C refers to
concentration in ppm, and mL refers to mass in mL.
The choice of concentrations to make will depend on the samples and the concentration of analyte within the samples that are being
analyzed. For first time users it is wise to make a calibration curve with a large range to encompass all the possible outcomes.
When the user is more aware of the kind of concentrations that they are producing in their synthesis then they can narrow down the
range to fit the kind of concentrations that they are anticipating.
In this example we will make concentrations ranging from 10 ppm to 0.1 ppm, with a total of five samples. In a typical ICP-AES
analysis about 3 mL of solution is used, however if you have situations with substantial wavelength overlap then you may have
chosen to do two separate runs and so you will need approximately 6 mL solution. In general it is wise to have at least 10 mL of
solution to prepare for any eventuality that may occur. There will also be some extra amount needed for samples that are being used
for the quality control check. For this reason 10 mL should be a sufficient amount to prepare of each concentration.
We can define the unknowns in the equation as follows:
CI = concentration of concentrated solution (ppm)
CF = desired concentration (ppm)
MI = initial mass of material (mL)
MF = mass of material required for dilution (mL)
The methodology adopted works as follows. Make the high concentration solution then take from that solution and dilute further to
the desired concentrations that are required.
Let's say the concentration of the stock solution from the supplier is 100 ppm of analyte. First we should dilute to a concentration
of 10 ppm. To make 10 mL of 10 ppm solution we should take 1 mL of the 100 ppm solution and dilute it up to 10 mL with
nanopure water, now the concentration of this solution is 10 ppm. Then we can take from the 10 ppm solution and dilute this down

1.5.3 https://chem.libretexts.org/@go/page/55818
to get a solution with 5 ppm. To do this take 5 mL of the 10 ppm solution and dilute it to 10 mL with nanopure water, then you will
have a solution of 10 mL that is 5 ppm concentration. And so you can do this successively taking aliquots from each solution
working your way down at incremental steps until you have a series of solutions that have concentrations ranging from 10 ppm all
the way down to 0.1 ppm or lower, as required.

ICP-AES at work
While ICP-AES is a useful method for quantifying the presence of a single metal in a given nanoparticle, another very important
application comes from the ability to determine the ratio of metals within a sample of nanoparticles.
In the following examples we can consider the bi-metallic nanoparticles of iron with copper. In a typical synthesis 0.75 mmol of
+

Fe(acac) is used to prepare iron-oxide nanoparticle of the form Fe O . It is possible to replace a quantity of the Fe ions with
n
3 3 4

another metal of similar charge. In this manner bi-metallic particles were made with a precursor containing a suitable metal. In this
example the additional metal precursor will be Cu(acac) . 2

Keep the total metal concentration in this example is 0.75 mmol. So if we want to see the effect of having 10% of the metal in the
reaction as copper, then we will use 10% of 0.75 mmol, that is 0.075 mmol Cu(acac) , and the corresponding amount of iron is
2

0.675 mmol Fe(acac) . We can do this for successive increments of the metals until you make 100% copper oxide particles.
3

Subsequent Fe and Cu ICP-AES of the samples will allow the determination of Fe : Curatio that is present in the nanoparticle.
This can be compared to the ratio of Fe and Cuthat was applied as reactants. The graph shows how the percentage of Fe in the
nanoparticle changes as a function of how much Fe is used as a reagent.

Figure 1.5.2 : Change in iron percentage in the Fe-Cu-O nanoparticles as a function of how much iron precursor is used in the
synthesis of the nanoparticles.

Determining Analyte Concentration


Once the nanoparticles are digested and the ICP-AES analysis has been completed you must turn the figures from the ICP-AES
analysis into working numbers to determine the concentration of metals in the solution that was synthesized initially.
Let's first consider the nanoparticles that are of one metal alone. The figure given by the analysis in this case is given in units of
mg/L, this is the value in ppm's. This figure was recorded for the solution that was analyzed, and this is of a dilute concentration

1.5.4 https://chem.libretexts.org/@go/page/55818
compared to the initial synthesized solution because the particles had to be digested in acid first, then diluted further into nanopure
water.
As mentioned above in the experimental 0.5 mL of the synthesized nanoparticles were first digested in 9.5 mL of concentrated
nitric acid. Then when the digestion was complete 0.5 mL of this solution was dissolved in 9.5 mL of nanopure water. This was the
final solution that was analyzed using ICP, and the concentration of metal in this solution will be far lower than that of the original
solution. In this case the amount of analyte in the final solution being analyzed is 1/20th that of the total amount of material in the
solution that was originally synthesized.

Calculating Concentration in ppm


Let us take an example that upon analysis by ICP-AES the amount of Fe detected is 6.38 mg/L. First convert the figure to mg/mL,
−3
6.38 mg/L × 1/1000 L/mL = 6.38‘x10  mg/mL

The amount of material was diluted to a total volume of 10 mL. Therefore we should multiply this value by 10 mL to see how
much mass was in the whole container.
−3 −2
6.38 × 10  mg/mL × 10 mL = 6.38 × 10  mg

This is the total mass of iron that was present in the solution that was analyzed using the ICP device. To convert this amount to ppm
we should take into consideration the fact that 0.5 mL was initially diluted to 10 mL, to do this we should divide the total mass of
iron by this amount that it was diluted to.
−2
6.38 × 10  mg/0.5 mL = 0.1276 mg/mL

This was the total amount of analyte in the 10 mL solution that was analyzed by the ICP device, to attain the value in ppm it should
be mulitplied by a thousand, that is then 127.6 ppm of Fe.

Determining Concentration of Original Solution


We now need to factor in the fact that there were several dilutions of the original solution first to digest the metals and then to
dissolve them in nanopure water, in all there were two dilutions and each dilution was equivalent in mass. By diluting 0.5 mL to 10
mL , we are effectively diluting the solution by a factor of 20, and this was carried out twice.

0.1276 mg/mL × 20 = 2.552 mg/mL

This is the amount of analyte in the solution of digested particles, to covert this to ppm we should multiply it by 1/1000 mL/L, in
the following way:
L
2.552 mg/mL ∗ ×1/1000mL/L = 2552 mg/L

This is essentially your answer now as 2552 ppm. This is the amount of Fe in the solution of digested particles. This was made by
diluting 0.5 mL of the original solution into 9.5 mL concentrated nitric acid, which is the same as diluting by a factor of 20. To
calculate how much analyte was in the original batch that was synthesized we multiply the previous value by 20 again. This is the
final amount of Fe concentration of the original batch when it was synthesized and made soluble in hexanes.

2552 ppm × 20 = 51040 ppm

Calculating Stoichiometric Ratio


Moving from calculating the concentration of individual elements now we can concentrate on the calculation of stoichiometric
ratios in the bi-metallic nanoparticles.
Consider the case when we have the iron and the copper elements in the nanoparticle. The amounts determined by ICP are:
Iron = 1.429 mg/L.
Copper = 1.837 mg/L.
We must account for the molecular weights of each element by dividing the ICP obtained value, by the molecular weight for that
particular element. For iron this is calculated by
1.429 mg/L
= 0.0211
55.85

1.5.5 https://chem.libretexts.org/@go/page/55818
,
and thus this is molar ratio of iron. On the other hand the ICP returns a value for copper that is given by:
1.837mg/L
= 0.0289
63.55

To determine the percentage iron we use this equation, which gives a percentage value of 42.15% Fe.
 molar ratio of iron 
% Fe  = [ ] × 100
 sum of molar ratios 

We work out the copper percentage similarly, which leads to an answer of 57.85% Cu.
 molar ratio of copper
% Cu = [ ] × 100
 sum of molar ratios 

In this way the percentage iron in the nanoparticle can be determined as function of the reagent concentration prior to the synthesis
(Figure 1.5.2 ).

Determining Concentration of Nanoparticles in Solution


The previous examples have shown how to calculate both the concentration of one analyte and the effective shared concentration of
metals in the solution. These figures pertain to the concentration of elemental atoms present in solution. To use this to determine
the concentration of nanoparticles we must first consider how many atoms that are being detected are in a nanoparticle. Let us
consider that the Fe O nanoparticles are of 7 nm diameter. In a 7 nm particle we expect to find 20,000 atoms. However in this
3 4

analysis we have only detected Fe atoms, so we must still account for the number of oxygen atoms that form the crystal lattice also.
For every 3 Fe atoms, there are 4 O atoms. But as iron is slightly larger than oxygen, it will make up for the fact there is one less Fe
atom. This is an over simplification but at this time it serves the purpose to make the reader aware of the steps that are required to
take when judging nanoparticles concentration. Let us consider that half of the nanoparticle size is attributed to iron atoms, and the
other half of the size is attributed to oxygen atoms.
As there are 20,000 atoms total in a 7 nm particle, and then when considering the effect of the oxide state we will say that for every
10,000 atoms of Fe you will have a 7 nm particle. So now we must find out how many Fe atoms are present in the sample so we
can divide by 10,000 to determine how many nanoparticles are present.
In the case from above, we found the solution when synthesized had a concentration 51,040 ppm Fe atoms in solution. To
determine how how many atoms this equates to we will use the fact that 1 mole of material has the Avogadro number of atoms
present.

51040 ppm = 51040 mg/L = 51.040 g/L

1 mole of iron weighs 55.847 g. To determine how many moles we now have, we divide the values like this:
51.040 g/L
= 0.9139  mol/L 
55.847 g

The number of atoms is found by multiplying this by Avogadro’s number (6.022x1023):


23 23
(0.9139  mol/L) × (6.022 × 10  atoms ) = 5.5 × 10   atoms/L 

For every 10,000 atoms we have a nanoparticle (NP) of 7 nm diameter, assuming all the particles are equivalent in size we can then
divide the values. This is the concentration of nanoparticles per liter of solution as synthesized.
23 19
(5.5 × 10  atoms/ L ) /(10, 000 atoms/NP) = 5.5 × 10  NP/L

Combined Surface Area


One very interesting thing about nanotechnology that nanoparticles can be used for is their incredible ratio between the surface
areas compared with the volume. As the particles get smaller and smaller the surface area becomes more prominent. And as much
of the chemistry is done on surfaces, nanoparticles are good contenders for future use where high aspect ratios are required.
In the example above we considered the particles to be of 7 nm diameters. The surface area of such a particle is 1.539 x10-16 m2. So
the combined surface area of all the particles is found by multiplying each particle by their individual surface areas.

1.5.6 https://chem.libretexts.org/@go/page/55818
−16 2 19 2
(1.539 × 10 m ) × (5.5 × 10  mathrmN P /L) = 8465 m /L

To put this into context, an American football field is approximately 5321 m2. So a liter of this nanoparticle solution would have
the same surface area of approximately 1.5 football fields. That is allot of area in one liter of solution when you consider how much
material it would take to line the football field with thin layer of metallic iron. Remember there is only about 51 g/L of iron in this
solution!

Bibliography
http://www.ivstandards.com/extras/pertable/
A. Scheffer, C. Engelhard, M. Sperling, and W. Buscher, W. Anal. Bioanal. Chem., 2008, 390, 249.
H. Nakamuru, T. Shimizu, M. Uehara, Y. Yamaguchi, and H. Maeda, Mater. Res. Soc., Symp. Proc., 2007, 1056, 11.
S. Sun and H. Zeng, J. Am. Chem. Soc., 2002, 124, 8204.
C. A. Crouse and A. R. Barron, J. Mater. Chem., 2008, 18, 4146.

1.5: ICP-AES Analysis of Nanoparticles is shared under a CC BY 4.0 license and was authored, remixed, and/or curated by Pavan M. V. Raja &
Andrew R. Barron via source content that was edited to conform to the style and standards of the LibreTexts platform; a detailed edit history is
available upon request.

1.5.7 https://chem.libretexts.org/@go/page/55818
1.6: ICP-MS for Trace Metal Analysis
Inductively coupled plasma mass spectroscopy (ICP-MS) is an analytical technique for determining trace multi-elemental and
isotopic concentrations in liquid, solid, or gaseous samples. It combines an ion-generating argon plasma source with the sensitive
detection limit of mass spectrometry detection. Although ICP-MS is used for many different types of elemental analysis, including
pharmaceutical testing and reagent manufacturing, this module will focus on its applications in mineral and water studies. Although
akin to ICP-AES (inductively coupled plasma atomic emission spectroscopy), ICP-MS has significant differences, which will be
mentioned as well.

Basic Instrumentation and Operation


As shown in Figure 1.6.1 there are several basic components of an ICP-MS instrument, which consist of a sampling interface, a
peristaltic pump leading to a nebulizer, a spray chamber, a plasma torch, a detector, an interface, and ion-focusing system, a mass-
separation device, and a vacuum chamber, maintained by turbo molecular pumps. The basic operation works as follows: a liquid
sample is pumped into the nebulizer to convert the sample into a spray. An internal standard, such as germanium, is pumped into a
mixer along with the sample prior to nebulization to compensate for matrix effects. Large droplets are filtered out, and small
droplets continue into the plasma torch, turning to ions. The mass separation device separates these ions based on their mass-to-
charge ratio. An ion detector then converts these ions into an electrical signal, which is multiplied and read by computer software.

Figure 1.6.1 : Scheme depicting the basic components of an ICP-MS system. Adapted from R. Thomas, Practical Guide to ICP-
MS: A Tutorial for Beginners, CRC Press, Boca Raton, 2nd edn. (2008).
The main difference between ICP-MS and ICP-AES is the way in which the ions are generated and detected. In ICP-AES, the ions
are excited by vertical plasma, emitting photons that are separated on the basis of their emission wavelengths. As implied by the
name, ICP-MS separates the ions, generated by horizontal plasma, on the basis of their mass-to-charge ratios (m/z). In fact, caution
is taken to prevent photons from reaching the detector and creating background noise. The difference in ion formation and detection
methods has a significant impact on the relative sensitivities of the two techniques. While both methods are capable of very fast,
high throughput multi-elemental analysis (~10 - 40 elements per minute per sample), ICP-MS has a detection limit of a few ppt to a
few hundred ppm, compared to the ppb-ppm range (~1 ppb - 100 ppm) of ICP-AES. ICP-MS also works over eight orders of
magnitude detection level compared to ICP-AES’ six. As a result of its lower sensitivity, ICP-MS is a more expensive system. One
other important difference is that only ICP-MS can distinguish between different isotopes of an element, as it segregates ions based
on mass. A comparison of the two techniques is summarized in this table.
Table 1.6.1 : Comparison of ICP-MS and ICP-AES.
ICP-MS ICP-AES

Plasma Horizontal: generates cations Vertical: excites atoms, which emit photons

Ion detection Mass-to-charge ratio Wavelength of emitted light

Detection limit 1-10 ppt 1-10 ppb

1.6.1 https://chem.libretexts.org/@go/page/55820
ICP-MS ICP-AES

Working range 8 orders of magnitude 6 orders of magnitude

Throughput 20-30 elements per minute 10-40 elements per minute

Isotope detection Yes No

Cost ~$150,000 ~$50,000

Multi-element detection Yes Yes

Much greater in number and more complicated


Spectral interferences Predictable, less than 300
to correct

Electrothermal vaporization, laser ablation,


Routine accessories Rare
high-performance liquid chromatography, etc.

Sample Preparation
With such small sample sizes, care must be taken to ensure that collected samples are representative of the bulk material. This is
especially relevant in rocks and minerals, which can vary widely in elemental content from region to region. Random, composite,
and integrated sampling are each different approaches for obtaining representative samples.
Because ICP-MS can detect elements in concentrations as minute as a few nanograms per liter (parts per trillion), contamination is
a very serious issue associated with collecting and storing samples prior to measurements. In general, use of glassware should be
minimized, due to leaching impurities from the glass or absorption of analyte by the glass. If glass is used, it should be washed
periodically with a strong oxidizing agent, such as chromic acid (H Cr O ), or a commercial glass detergent. In terms of sample
2 2 7

containers, plastic is usually better than glass, polytetrafluoroethylene (PTFE) and Teflon® being regarded as the cleanest plastics.
However, even these materials can contain leachable contaminants, such as phosphorus or barium compounds. All containers,
pipettes, pipette tips, and the like should be soaked in 1 - 2% HNO . Nitric acid is preferred over HCl HCl, which can ionize in
3

the plasma to form Cl O and Ar Cl , which have the same mass-to-charge ratios as V and As , respectively. If
35 16 + 40 35 + 51 + 75 +

possible, samples should be prepared as close as possible to the ICP-MS instrument without being in the same room.
With the exception of solid samples analyzed by laser ablation ICP-MS, samples must be in liquid or solution form. Solids are
ground into a fine powder with a mortar and pestle and passed through a mesh sieve. Often the first sample is discarded to prevent
contamination from the mortar or sieve. Powders are then digested with ultrapure concentrated acids or oxidizing agents, like
chloric acid (HClO ), and diluted to the correct order of magnitude with 1 - 2% trace metal grade nitric acid.
3

Once in liquid or solution form, the samples must be diluted with 1 - 2% ultrapure HClO to a low concentration to produce a
3

signal intensity lower than about 106 counts. Not all elements have the same concentration to intensity correlation; therefore, it is
safer to test unfamiliar samples on ICP-AES first. Once properly diluted, the sample should be filtered through a 0.25 - 0.45 μm
membrane to remove particulates.
Gaseous samples can also be analyzed by direct injection into the instrument. Alternatively, gas chromatography equipment can be
coupled to an ICP-MS machine for separation of multiple gases prior to sample introduction.

Standards
Multi- and single-element standards can be purchased commercially, and must be diluted further with 1 - 2% nitric acid to prepare
different concentrations for the instrument to create a calibration curve, which will be read by the computer software to determine
the unknown concentration of the sample. There should be several standards, encompassing the expected concentration of the
sample. Completely unknown samples should be tested on less sensitive instruments, such as ICP-AES or EDXRF (energy
dispersive X-ray fluorescence), before ICP-MS.

Limitations of ICP-MS
While ICP-MS is a powerful technique, users should be aware of its limitations. Firstly, the intensity of the signal varies with each
isotope, and there is a large group of elements that cannot be detected by ICP-MS. This consists of H, He and most gaseous
elements, C, and elements without naturally occurring isotopes, including most actinides.
There are many different kinds of interferences that can occur with ICP-MS, when plasma-formed species have the same mass as
the ionized analyte species. These interferences are predictable and can be corrected with element correction equations or by

1.6.2 https://chem.libretexts.org/@go/page/55820
evaluating isotopes with lower natural abundances. Using a mixed gas with the argon source can also alleviate the interference.
The accuracy of ICP-MS is highly dependent on the user’s skill and technique. Standard and sample preparations require utmost
care to prevent incorrect calibration curves and contamination. As exemplified below, a thorough understanding of chemistry is
necessary to predict conflicting species that can be formed in the plasma and produce false positives. While an inexperienced user
may be able to obtain results fairly easily, those results may not be trustworthy. Spectral interference and matrix effects are
problems that the user must work diligently to correct.

Applications: Analysis of Mineral and Water Samples


In order to illustrate the capabilities of ICP-MS, various geochemical applications as described. The chosen examples are
representative of the types of studies that rely heavily on ICP-MS, highlighting its unique capabilities.

Trace Elemental Analysis of Minerals


With its high throughput, ICP-MS has made sensitive analysis of multi-element detection in rock and mineral samples feasible.
Studies of trace components in rock can reveal information about the chemical evolution of the mantle and crust. For example,
spinel peridotite xenoliths (Figure 1.6.2 ), which are igneous rock fragments derived from the mantle, were analyzed for 27
elements, including lithium, scandium and titanium at the parts per million level and yttrium, lutetium, tantalum, and hafnium in
parts per billion. X-ray fluorescence was used to complement ICP-MS, detecting metals in bulk concentrations. Both liquid and
solid samples were analyzed, the latter being performed using laser-ablation ICP-MS, which points out the flexibility of the
technique for being used in tandem with others. In order to prepare the solution samples, optically pure minerals were sonicated in
3 M HCl, then 5% HF, then 3 M HCl again and dissolved in distilled water. The solid samples were converted into plasma by laser
ablation prior to injection into the nebulizer of the LA-ICP-MS instrument. The results showed good agreement between the laser
ablation and solution methods. Furthermore, this comprehensive study shed light on the partitioning behavior of incompatible
elements, which, due to their size and charge, have difficulty entering cation sites in minerals. In the upper mantle, incompatible
trace elements, especially barium, niobium and tantalum, were found to reside in glass pockets within the peridotite samples.

1.6.3 https://chem.libretexts.org/@go/page/55820
Figure 1.6.2 : Crystal structure of a typical spinel, general formula A 2 + 3 +
B
2
2 −
O
4
.

Trace Elemental Analysis of Water


Another important area of geology that requires knowledge of trace elemental compositions is water analysis. In order to
demonstrate the full capability of ICP-MS as an analytical technique in this field, researchers aim to use the identification of trace
metals present in groundwater to determine a fingerprint for a particular water source. In one study the analysis of four different
Nevada springs determined trace metal analysis in parts per billion and even parts per trillion (ng/L). Because they were present is
such low concentrations, samples containing rare earth elements lutetium, thulium, and terbium were preconcentrated by a cation
exchange column to enable detection at 0.05 ppt. For some isotopes, special corrections necessary to account for false positives,
which are produced by plasma-formed molecules with the same mass-to-charge ratio as the isotopic ions. For instance, false
positives for Sc (m/z = 45) or Ti (m/z = 47) could result from CO H (m/z = 45) or PO (m/z = 47); and BaO (m/z = 151, 153)
2
+ + +

conflicts with Eu-151 and Eu-153. In the latter case, barium has many isotopes (134, 135, 136, 137, 138) in various abundances,
Ba-138 comprising 71.7% barium abundance. ICP-MS detects peaks corresponding to BaO for all isotopes. Thus researchers
+

were able to approximate a more accurate europium concentration by monitoring a non-interfering barium peak and extrapolating
back to the concentration of barium in the system. This concentration was subtracted out to give a more realistic europium
concentration. By employing such strategies, false positives could be taken into account and corrected. Additionally, 10 ppb
internal standard was added to all samples to correct for changes in sample matrix, viscosity and salt buildup throughout collection.
In total, 54 elements were detected at levels spanning seven orders of magnitude. This study demonstrates the incredible sensitivity
and working range of ICP-MS.

1.6.4 https://chem.libretexts.org/@go/page/55820
Determination of Arsenic Content
Elemental analysis in water is also important for the health of aquatic species, which can ultimately affect the entire food chain,
including people. With this in mind, arsenic content was determined in fresh water and aquatic organisms in Hayakawa River in
Kanagawa, Japan, which has very high arsenic concentrations due to its hot spring source in Owakudani Valley. While water
samples were simply filtered and prior to analysis, organisms required special preparation, in order to be compatible with the
sampler. Organisms collected for this studied included water bug, green macroalga, fish, and crustaceans. For total As content
determination, the samples were freeze-dried to remove all water from the sample in order to know the exact final volume upon
resuspension. Next, the samples were ground into a powder, followed by soaking in nitric acid, heating at 110 °C. The sample then
underwent heating with hydrogen peroxide, dilution, and filtering through a 0.45 μm membrane. This protocol served to oxidize the
entire sample and remove large particles prior to introduction into the ICP-MS instrument. Samples that are not properly digested
can build up on the plasma torch and cause expensive damage to the instrument. Since the plasma converts the sample into various
ion constituents, it is unnecessary to know the exact oxidized products prior to sample introduction. In addition to total As content,
the As concentration of different organic arsenic-containing compounds (arsenicals) produced in the organisms was measured by
high performance liquid chromatography coupled to ICP-MS (HPLC/ICP-MS). The arsenicals were separated by HPLC before
travelling into the ICP-MS instrument for As concentration determination. For this experiment, the organic compounds were
extracted from biological samples by dissolving freeze-dried samples in methanol/water solutions, sonicating, and centrifuging.
The extracts were dried under vacuum, redissolved in water, and filtered prior to loading. This did not account for all compounds,
however, because over 50% arsenicals were nonsoluble in aqueous solution. One important plasma side product to account for was
ArCl , which has the same mass-to-charge ratio (m/z = 75) as As. This was corrected by oxidizing the arsenic ions within the mass
+

separation device in the ICP-MS vacuum chamber to generate AsO , with m/z 91. The total arsenic concentration of the samples
+

ranged from 17 - 18 ppm.

Bibliography
R. Thomas, Practical Guide to ICP-MS: A Tutorial for Beginners, CRC Press, Boca Raton, 2nd edn. (2008).
K. J. Stetzenbach, M. Amano, D. K. Kreamer, and V. F. Hodge. Ground Water, 1994, 32, 976.
S. M. Eggins, R. L. Rudnick, and W. F. McDonough, Earth Planet. Sci. Lett., 1998, 154, 53.
S. Miyashita, M. Shimoya, Y. Kamidate, T. Kuroiwa, O. Shikino, S. Fujiwara, K. A. Francesconi, and T. Kaise. Chemosphere,
2009, 75, 1065.

1.6: ICP-MS for Trace Metal Analysis is shared under a CC BY 4.0 license and was authored, remixed, and/or curated by Pavan M. V. Raja &
Andrew R. Barron via source content that was edited to conform to the style and standards of the LibreTexts platform; a detailed edit history is
available upon request.

1.6.5 https://chem.libretexts.org/@go/page/55820
1.7: Ion Selective Electrode Analysis
Introduction
Ion selective electrode (ISE) is an analytical technique used to determine the activity of ions in aqueous solution by measuring the
electrical potential. ISE has many advantages compared to other techniques, including:
1. It is relatively inexpensive and easy to operate.
2. It has wide concentration measurement range.
3. As it measure the activity, instead of concentration, it is particularly useful in biological/medical application.
4. It is a real-time measurement, which means it can monitor the change of activity of ion with time.
5. It can determine both positively and negatively charged ions.
Based on these advantages, ISE has wide variety of applications, which is reasonable considering the importance of measuring ion
activity. For example, ISE finds its use in pollution monitoring in natural waters (CN-, F-, S-, Cl-, etc.), food processing (NO3-,
NO2- in meat preservatives), Ca2+ in dairy products, and K+ in fruit juices, etc.

Measurement setup
Before focusing on how ISE works, it would be good to get an idea what ISE setup looks like and the component of the ISE
instrument. Figure 1.7.1 shows the basic components of ISE setup. It has an ion selective electrode, which allows measured ions to
pass, but excludes the passage of the other ions. Within this ion selective electrode, there is an internal reference electrode, which is
made of silver wire coated with solid silver chloride, embedded in concentrated potassium chloride solution (filling solution)
saturated with silver chloride. This solution also contains the same ions as that to be measured. There is also a reference electrode
similar to ion selective electrode, but there is no to-be-measured ion in the internal electrolyte and the selective membrane is
replaced by porous frit, which allows the slow passage of the internal filling solution and forms the liquid junction with the external
text solution. The ion selective electrode and reference electrode are connected by a milli-voltmeter. Measurment is accomplished
simply by immersing the two electrodes in the same test solution.

1.7.1 https://chem.libretexts.org/@go/page/55822
Figure 1.7.1 : Measurement setup of ISE.

Theory of How ISE Works


There are commonly more than one types of ions in solution. So how ISE manage to measure the concentration of certain ion in
solution without being affected by other ions? This is done by applying a selective membrane at the ion selective electrode, which
only allows the desired ion to go in and out. At equilibrium, there is potential difference existing between two sides of the
membrane, and it is governed by the concentration of the tested solution described by Nernst equation EQ, where E is potential, E0
is a constant characteristic of a particular ISE, R is the gas constant (8.314 J/K.mol), T is the temperature (in K), n is the charge of
the ion and F is Faraday constant (96,500 coulombs/mol). To make it relevant, the measured potential difference is proportional to
the logarithm of ion concentration. Thus, the relationship between potential difference and ion concentration can be determined by
measuring the potential of two solutions of already-known ion concentration and a plot based on the measured potential and
logarithm of the ion concentration. Based on this plot, the ion concentration of an unknown solution can be known by measuring
the potential and corresponding it to the plot.
0
E =E + (2.030 RT /nF ) log C (1.7.1)

Example Application: Determination of Fluoride Ion


Fluoride is added into drinking water and toothpaste to prevent dental caries and thus the determination of its concentration is of
great importance to human health. Here, we will give some data and calculations to show how the concentration of fluoride ion is
determined and have a glance at how relevant ISE is to our daily life. According to Nernst equation, (Equation 1.7.1), in this case n
= 1, T = 25 °C and E0, R, F are constants and thus this equation can be simplied as
E = K + S log C (1.7.2)

The first step is to obtain a calibration curve for fluoride ion and this can be done by preparing several fluoride standard solution
with known concentration and making a plot of E versus log C.
Table 1.7.1 : Measurement results. Data from http://zimmer.csufresno.edu/~davidz/...uorideISE.html.

1.7.2 https://chem.libretexts.org/@go/page/55822
Concentration (mg/L) log C E (mV)

200.0 2.301 -35.6

100.0 2.000 -17.8

50.00 1.699 0.4

25.00 1.398 16.8

12.50 1.097 34.9

6.250 0.796 52.8

3.125 0.495 70.4

1.563 0.194 89.3

0.781 0.107 107.1

0.391 0.408 125.5

0.195 0.709 142.9

Figure 1.7.2 : Plot of E versus log C. Based on data from http://zimmer.csufresno.edu/~davidz/...uorideISE.html.


From the plot we can clearly identify the linear relationship between E versus log C with slope measured at -59.4 mV, which is
very closed to the theoretical value -59.2 mV at 25 °C. This plot can give the concentration of any solution containing fluoride ion
within the range of 0.195 mg/L and 200 mg/L by measuring the potential of the unknown solution.

Limit of ISE
Though ISE is a cost-effective and useful technique, it has some drawbacks that cannot be avoided. The selective ion membrane
only allows the measured ions to pass and thus the potential is only determined by this particular ion. However, the truth is there is
no such membrane that only permits the passage of one ion, and so there are cases when there are more than one ions that can pass
the membrane. As a result, the measured potential are affected by the passage of the “unwanted” ions. Also, because of its
dependence on ion selective membrane, one ISE is only suitable for one ion and this may be inconvenient sometimes. Another
problem worth noticing is that ion selective measures the concentration of ions in equilibrium at the surface of the membrane
surface. This does matter much if the solution is dilute but at higher concentrations, the inter-ionic interactions between the ions in
the solution tend to decrease the mobility of ions and thus the concentration near the membrane would be lower than that in the
bulk. This is one source of inaccuracy of ISE. To better analyze the results of ISE, we have to be aware of these inherent limitations
of it.

1.7.3 https://chem.libretexts.org/@go/page/55822
Bibliography
D. S. Papastathopoulos and M. I. Karayannis, J. Chem. Edu., 1980, 57, 904.
J. E. O'Reilly, J. Chem. Edu., 1979, 56, 279.
F. Scholz, Electroanalytical Methods: Guide to Experiments and Application, 2nd edition, Springer, Berlin (2010).
R. Greef, R. Peat, L. M. Peter, D. Pletcher, and J. Robinson, Instrumental Methods in Electrochemistry, Ellis Horwood,
Chichester (1985).

1.7: Ion Selective Electrode Analysis is shared under a CC BY 4.0 license and was authored, remixed, and/or curated by Pavan M. V. Raja &
Andrew R. Barron via source content that was edited to conform to the style and standards of the LibreTexts platform; a detailed edit history is
available upon request.

1.7.4 https://chem.libretexts.org/@go/page/55822
1.8: A Practical Introduction to X-ray Absorption Spectroscopy
X-ray absorption spectroscopy (XAS) is a technique that uses synchrotron radiation to provide information about the electronic,
structural, and magnetic properties of certain elements in materials. This information is obtained when X-rays are absorbed by an
atom at energies near and above the core level binding energies of that atom. Therefore, a brief description about X-rays,
synchrotron radiation and X-ray absorption is provided prior to a description of sample preparation for powdered materials.

X-rays and Synchrotron Radiation


X-rays were discovered by the Wilhelm Röntgen in 1895 (figure 1.8.1). They are a form of electromagnetic radiation, in the same
manner as visible light but with a very short wavelength, around 0.25 - 25 Å. As electromagnetic radiation, X-rays have a specific
energy. The characteristic range is defined by soft versus hard X-rays. Soft X-rays cover the range from hundreds of eV to a few
KeV, and the hard X-rays have an energy range from a few KeV up to around 100 KeV.

Figure 1.8.1 : German physicist Wilhelm Conrad Röntgen (1845 –1923) who received the first Nobel Prize in Physics in 1901 for
the production and use of X-rays.
X-rays are commonly produced by X-ray tubes, when high-speed electrons strike a metal target. The electrons are accelerated by a
high voltage towards the metal target; X-rays are produced when the electrons collide with the nuclei of the metal target.
Synchrotron radiation is generated when particles are moving at really high velocities and are deflected along a curved trajectory
by a magnetic field. The charged particles are first accelerated by a linear accelerator (LINAC) (figure 1.8.2); then, they are
accelerated in a booster ring that injects the particles moving almost at the speed of light into the storage ring. There, the particles
are accelerated toward the center of the ring each time their trajectory is changed so that they travel in a closed loop. X-rays with a
broad spectrum of energies are generated and emitted tangential to the storage ring. Beamlines are placed tangential to the storage
ring to use the intense X-ray beams at a wavelength that can be selected varying the set up of the beamlines. Those are well suited
for XAS measurements because the X-ray energies produced span 1000 eV or more as needed for an XAS spectrum.

1.8.1 https://chem.libretexts.org/@go/page/55824
Figure 1.8.2 : Scheme of a synchrotron and the particle trajectory inside it. Adapted from S. D. Kelly, D. Hesterberg, and B. Ravel
in Methods of Soil Analysis: Part 5, Mineralogical Methods, Ed. A. L. Urely and R. Drees, Soil Science Society of America Book
Series, Madison (2008).

X-ray Absorption
Light is absorbed by matter through the photoelectric effect. It is observed when an X-ray photon is absorbed by an electron in a
strongly bound core level (such as the 1s or 2p level) of an atom (figure 1.8.3). In order for a particular electronic core level to
participate in the absorption, the binding energy of this core level must be less than the energy of the incident X-ray. If the binding
energy is greater than the energy of the X-ray, the bound electron will not be perturbed and will not absorb the X-ray. If the binding
energy of the electron is less than that of the X-ray, the electron may be removed from its quantum level. In this case, the X-ray is
absorbed and any energy in excess of the electronic binding energy is given as kinetic energy to a photo-electron that is ejected
from the atom.

Figure 1.8.3 : A schematic representation of the photoelectric effect when a photon with the right energy hits an electron, it is
expelled.
When X-ray absorption is discussed, the primary concern is about the absorption coefficient, µ, which gives the probability that X-
rays will be absorbed according to Beer’s Law where I0 is the X-ray intensity incident on a sample, t is the sample thickness, and I
is the intensity transmitted through the sample.
−μt
I = I0 e (1.8.1)

1.8.2 https://chem.libretexts.org/@go/page/55824
The absorption coefficient, µE, is a smooth function of energy, with a value that depends on the sample density ρ, the atomic
number Z, atomic mass A, and the X-ray energy E roughly as
4
ρZ
μE ≈ (1.8.2)
3
AE

When the incident X-ray has energy equal to that of the binding energy of a core-level electron, there is a sharp rise in absorption:
an absorption edge corresponding to the promotion of this core level to the continuum. For XAS, the main concern is the intensity
of µ, as a function of energy, near and at energies just above these absorption edges. An XAS measurement is simply a measure of
the energy dependence of µ at and above the binding energy of a known core level of a known atomic species. Since every atom
has core-level electrons with well-defined binding energies, the element to probe can be selected by tuning the X-ray energy to an
appropriate absorption edge. These absorption edge energies are well-known. Because the element of interest is chosen in the
experiment, XAS is element-specific.

X-ray Absorption Fine Structure


X-ray absorption fine structure (XAFS) spectroscopy, also named X-ray absorption spectroscopy, is a technique that can be applied
for a wide variety of disciplines because the measurements can be performed on solids, gasses, or liquids, including moist or dry
soils, glasses, films, membranes, suspensions or pastes, and aqueous solutions. Despites its broad adaptability with the kind of
material used, there are samples which limits the quality of an XAFS spectrum. Because of that, the sample requirements and
sample preparation is reviewed in this section as well the experiment design which are vital factors in the collection of good data
for further analysis.

Experiment Design
The main information can be obtained using XAFS spectra consist in small changes in the absorption coefficient (E), which can be
measured directly in a transmission mode or indirectly using a fluorescence mode. Therefore, a good signal to noise ratio is
required (better than 103). In order to obtain this signal to noise ratio, an intense beam is required (on the order 1010 photons/second
or better), with the energy bandwidth of 1 eV or less, and the capability of scanning the energy of the incident beam over a range of
about 1 KeV above the edge in a time range of seconds or few minutes. As a result, synchrotron radiation is preferred further than
other kind of X-ray sources previously mentioned.
Beamline Setup
Despite the setup of a synchrotron beamline is mostly done by the assistance of specialist beamline scientists, nevertheless, it is
useful to understand the system behind the measurement. The main components of a XAFS beamline, as shown in figure below, are
as follows:
A harmonic rejection mirror to reduce the harmonic content of the X-ray beam.
A monochromator to choose the X-ray energy.
A series of slits which defines the X-ray profile.
A sample positioning stage.
The detectors, which can be a single ionization detector or a group of detectors to measure the X-ray intensity.

Figure 1.8.4 : Schematic of the basic components of a XAFS beamline.


Slits are used to define the X-ray beam profile and to block unwanted X-rays. Slits can be used to increase the energy resolution of
the X-ray incident on the sample at the expense of some loss in X-ray intensity. They are either fixed or adjustable slits. Fixed slits

1.8.3 https://chem.libretexts.org/@go/page/55824
have a pre-cut opening of heights between 0.2 and 1.0 mm and a width of some centimeters. Adjustable slits use metal plates that
move independently to define each edge of the X-ray beam.
Monochromator
The monochromator is used to select the X-ray energy incident on the sample. There are two main kinds of X-ray monochromators:
1. The double-crystal monochromator, which consists of two parallel crystals.
2. The channel-cut monochromator, which is a single crystal with a slot cut nearly through it.
Most monochromator crystals are made of silicon or germanium and are cut and polished such that a particular atomic plane of the
crystal is parallel to the surface of the crystal as Si(111), Si(311), or Ge(111). The energy of X-rays diffracted by the crystal is
controlled by rotating the crystals in the white beam.
Harmonic rejection mirrors
The harmonic X-ray intensity needs to be reduced, as these X-rays will adversely affect the XAS measurement. A common method
for removing harmonic X-rays is using a harmonic rejection mirror. This mirror is usually made of Si for low energies, Rh for X-
ray energies below the Rh absorption edge at 23 keV, or Pt for higher X-ray energies. The mirror is placed at a grazing angle in the
beam such that the X-rays with fundamental energy are reflected toward the sample, while the harmonic X-rays are not.
Detectors
Most X-ray absorption measurements use ionization detectors. These contain two parallel plates separated by a gas-filled space that
the X-rays travel through. Some of the X-rays ionize the gas particles. A voltage bias applied to the parallel plates separates the gas
ions, creating a current. The applied voltage should give a linear detector response for a given change in the incident X-ray
intensity. There are also other kinds as fluorescence and electron yield detectors.
Transmission and Fluorescence Modes
X-ray Absorption measurements can be performed in several modes: transmission, fluorescence and electron yield; where the two
first are the most common. The choice of the most appropriate mode to use in one experiment is a crucial decision.
The transmission mode is the most used because it only implies the measure of the X-ray flux before and after the beam passes the
sample. Therefore, the adsorption coefficient is defined as
I0
μE = ln( ) (1.8.3)
I

Transmission experiments are standard for hard X-rays, because the use of soft X-rays implies the use the samples thinner than 1
μm. Also, this mode should be used for concentrated samples. The sample should have the right thickness and be uniform and free
of pinholes.
The fluorescence mode measures the incident flux I0 and the fluorescence X-rays If that are emitted following the X-ray absorption
event. Usually the fluorescent detector is placed at 90° to the incident beam in the horizontal plane, with the sample at an angles,
commonly 45°, with respect to the beam, because in that position there is not interference generated because of the initial X-ray
flux (I0). The use of fluorescence mode is preferred for thicker samples or lower concentrations, even ppm concentrations or lower.
For a highly concentrated sample, the fluorescence X-rays are reabsorbed by the absorber atoms in the sample, causing an
attenuation of the fluorescence signal, it effect is named as self-absorption and is one of the most important concerns in the use of
this mode.

Sample Preparation for XAS


Sample Requirements
Uniformity
The samples should have a uniform distribution of the absorber atom, and have the correct absorption for the measurement. The X-
ray beam typically probes a millimeter-size portion of the sample. This volume should be representative of the entire sample.
Thickness
For transmission mode samples, the thickness of the sample is really important. It supposes to be a sample with a given thickness, t,
where the total adsorption of the atoms is less than 2.5 adsorption lengths, µEt ≈ 2.5; and the partial absorption due to the absorber
atoms is around one absorption length ∆ µEt ≈ 1, which corresponds to the step edge.

1.8.4 https://chem.libretexts.org/@go/page/55824
The thickness to give ∆ µEt = 1 is as

1 1.66 ∑i ni Mi
t = = (1.8.4)
Δμ ρ ∑i ni [ σi (E+ ) − σi (E− )]

where ρ is the compound density, n is the elemental stoichiometry, M is the atomic mass, σE is the adsorption cross-section in
barns/atom (1 barn = 10-24 cm2) tabulated in McMaster tables, and E+and E- are the just above and below the energy edge. This
calculation can be accomplished using the free download software HEPHAESTUS.
Total X-ray Adsorption
For non-concentrate samples, the total X-ray adsorption of the sample is the most important. It should be related to the area
concentration of the sample (ρt, in g/cm2). The area concentration of the sample multiplied by the difference of the mass
adsorption coefficient (ΔµE/ρ) give the edge step, where a desired value to obtain a good measure is a edge step equal to one,
(ΔµE/ρ)ρt ≈ 1 .

The difference of the mass adsorption coefficient is given by

ΔμE ΔμE ΔμE


( ) = ∑ fi [ ( ) −( ) ] (1.8.5)
ρ ρ ρ
i,( E+ ) i,( E− )

where (µE/ρ) is the mass adsorption coefficient just above (E ) and below (E ) of the edge energy and f is the mass fraction of
i + − i

the element i. Multiplying the area concentration, \(ρt\), for the cross-sectional area of the sample holder, amount of sample needed
is known.

Sample Preparation
As was described in last section, there are diluted solid samples, which can be prepared onto big substrates or concentrate solid
samples which have to be prepared in thin films. Both methods are following described.
Liquid and gases samples can also be measured, but the preparation of those kind of sample is not discussed in this paper because it
depends in the specific requirements of each sample. Several designs can be used as long they avoid the escape of the sample and
the material used as container does not absorb radiation at the energies used for the measure.
Method 1
1. The materials needed are showed in this figure: Kapton tape and film, a thin spatula, tweezers, scissors, weigh paper, mortar and
pestle, and a sample holder. The sample holder can be made from several materials, as polypropylene, polycarbonate or Teflon.

Figure 1.8.5 : Several tools are needed for the sample preparation using Method 1.
2. Two small squares of Kapton film are cut. One of them is placed onto the hole of the sample holder as shown figure 1.8.6a. A
piece of Kapton tape is placed onto the sample holder trying to minimize any air burble onto the surface and keeping the film as
was previously placed figure 1.8.6b. A side of the sample holder is now sealed in order to fill the hole (figure 1.8.7).

1.8.5 https://chem.libretexts.org/@go/page/55824
Figure 1.8.6 : Preparing one face of the sample holder by (a) positioning a small piece of Kapton film onto the hole, which is
held in place by Kapton tape (b).

Figure 1.8.7 : The side of the sample holder is closed.


3. Before fill the sample holder, make sure your sample is a fine powder. Use the mortar to grind the sample.

Figure 1.8.8 : The sample is ground to be sure the grain size of the sample is homogeneous and small enough.
4. Fill the hole with the powder. Make sure you have extra powder onto the hole (figure 1.8.9a). With the spatula press the
powder. The sample has to be as compact as possible (figure 1.8.9b).

Figure 1.8.9 : The sample holder is filled by (a) adding extra powder onto the hole then (b) compacting the sample with the
spatula.
5. Clean the surface of the slide. Repeat the step 2. Your sample loaded in the sample holder should look as picture below:

1.8.6 https://chem.libretexts.org/@go/page/55824
Figure 1.8.10 : Sample loaded and sealed into the sample holder.
Method 2
1. The materials needed are showed in photo: Kapton tape, tweezers, scissors, weigh paper, mortar and pestle, tape and aluminum
foil.

Figure 1.8.11 : Several utensils are needed for the sample preparation using Method 2.
2. Aluminum foil is placed as the work-area base. Kapton tape is place from one corner to the opposite one as shown figure
1.8.12. Tape is put onto the extremes to fix it. In this case yellow tape was used in order to show where the tape should be

placed but is better use Scotch invisible tape for the following steps.

Figure 1.8.12 : Preparation of the work-area.


3. The weigh paper is placed under the Kapton tape in one of the extremes. Sample is added onto that Kapton tape extreme. The
function of the weigh paper is further recuperation of extra sample.

1.8.7 https://chem.libretexts.org/@go/page/55824
Figure 1.8.13 : Add the sample onto an extreme of the Kapton tape.
4. With one finger, the sample is dispersed along the Kapton tape, always in the same direction and taking care that the weigh
paper is under the tape area is being used (figure 1.8.14a). The finger should be slid several times making pressure in order to
have a homogeneous and complete cover film (figure 1.8.14b).

Figure 1.8.14 : Making a thin film with a solid sample by (a) dispersing the solid along the Kapton tape and (b) repeated sliding
several times to obtain a homogeneous film.
5. The final sample covered Kapton tape should look like figure 1.8.15. Cut the extremes in order to a further manipulation of the
film.

Figure 1.8.15 : A complete thin film.


6. Using the tweezers, fold the film taking care that is well aligned and there fold is complete plane. figure 1.8.16a shows the first
folding, generating a 2 layers film. figure 1.8.16b and figure 1.8.16c shows the second and third folding, obtaining a 4 and 8
layers film. Sometimes a 4 layers film is good enough. You always can fold again to obtain bigger signal intensity.

Figure 1.8.16 : Folding of the thin film simple once results in a two layer film (a) and after a second and third folding four and
eight layers films are obtained (b and c, respectively).

Bibliography
B. D. Cullity and S. R. Stock. Elements of X-ray Diffraction, Prentice Hall, Upper Saddle River (2001).
F. Hippert, E. Geissler, J. L. Hodeau, E. Lelièvre-Berna, and J. R. Regnard. Neutron and X-ray Spectroscopy, Springer,
Dordrecht (2006).
G. Bunker. Introduction to XAFS: A practical guide to X-ray Absorption Fine Structure Spectroscopy, Cambridge University
Press, Cambridge (2010).

1.8.8 https://chem.libretexts.org/@go/page/55824
S. D. Kelly, D. Hesterberg, and B. Ravel in Methods of Soil Analysis: Part 5, Mineralogical Methods, Ed. A. L. Urely and R.
Drees, Soil Science Society of America Book Series, Madison (2008).

1.8: A Practical Introduction to X-ray Absorption Spectroscopy is shared under a CC BY 4.0 license and was authored, remixed, and/or curated
by Pavan M. V. Raja & Andrew R. Barron via source content that was edited to conform to the style and standards of the LibreTexts platform; a
detailed edit history is available upon request.

1.8.9 https://chem.libretexts.org/@go/page/55824
1.9: Neutron Activation Analysis (NAA)
Introduction
Neutron activation analysis (NAA) is a non-destructive analytical method commonly used to determine the identities and
concentrations of elements within a variety of materials. Unlike many other analytical techniques, NAA is based on nuclear rather
than electronic transitions. In NAA, samples are subjected to neutron radiation (i.e., bombarded with neutrons), which causes the
elements in the sample to capture free neutrons and form radioactive isotopes, such as
59 1 60
Co + n → Co (1.9.1)
27 0 27

.
The excited isotope undergoes nuclear decay and loses energy by emitting a series of particles that can include neutrons, protons,
alpha particles, beta particles, and high-energy gamma ray photons. Each element on the periodic table has a unique emission and
decay path that allows the identity and concentration of the element to be determined.

History
Almost eighty years ago in 1936, George de Hevesy and Hilde Levi published the first paper on the process of neutron activation
analysis. They had discovered that rare earth elements such as dysprosium became radioactive after being activated by thermal
neutrons from a radon-beryllium (266Ra + Be) source. Using a Geiger counter to count the beta particles emitted, Hevesy and Levi
were able to identify the rare earth elements by half-life. This discovery led to the increasingly popular process of inducing
radioactivity and observing the resulting nuclear decay in order to identify an element, a process we now know as NAA. In the
years immediately following Hevesy and Levi’s discovery, however, the advancement of this technique was restricted by the lack
of stable neutron sources and adequate spectrometry equipment. Even with the development of charged-particle accelerators in the
1930s, analyzing multi-element samples remained time-consuming and tedious. The method was improved in the mid-1940s with
the availability of the X-10 reactor at the Oak Ridge National Laboratory, the first research-type nuclear reactor. As compared with
the earlier neutron sources used, this reactor increased the sensitivity of NAA by a factor of a million. Yet the detection step of
NAA still revolved around Geiger or proportional counters; thus, many technological advancements were still to come. As
technology has progressed in the recent decades, the NAA method has grown tremendously, and scientists now have a plethora of
neutron sources and detectors to choose from when analyzing a sample with NAA.

Sample preparation
In order to analyze a material with NAA, a small sample of at least 50 milligrams must be obtained from the material, usually by
drilling. It is suggested that two different samples are obtained from the material using two drill bits of different compositions. This
will show any contamination from the drill bits and, thus, minimize error. Prior to irradiation, the small samples are encapsulated in
vials of either quartz or high purity linear polyethylene.

Instrument
How it Works
Neutron activation analysis works through the processes of neutron activation and radioactive decay. In neutron activation,
radioactivity is induced by bombarding a sample with free neutrons from a neuron source. The target atomic nucleus captures a free
neutron and, in turn, enters an excited state. This excited and therefore unstable isotope undergoes nuclear decay, a process in
which the unstable nucleus emits a series of particles that can include neutrons, protons, alpha, and beta particles in an effort to
return to a low-energy, stable state. As suggested by the several different particles of ionizing radiation listed above, there are many
different types of nuclear decay possible. These are summarized in the figure below.

1.9.1 https://chem.libretexts.org/@go/page/55826
Figure 1.9.1 : Transition diagram illustrating the changes in neutron number N and atomic number Z for different nuclear decay
modes – alpha decay (α), normal beta decay (β-), positron emission (β+), electron capture (EC), proton emission (p), and neutron
emission (n). Permission to copy granted via the GNU Free Documentation License.
An additional type of nuclear decay is that of gamma radiation (denoted as γ), a process in which the excited nucleus emits high-
energy gamma ray photons. There is no change in either neutron number N or atomic number Z, yet the nucleus undergoes a
nuclear transformation involving the loss of energy. In order to distinguish the higher energy parent nucleus (prior to gamma decay)
from the lower energy daughter nucleus (after gamma decay), the mass number of the parent nucleus is labeled with the letter m,
which means “metastable.” An example of gamma radiation with the element technetium is shown here.
99m 99 0
Tc → Tc + γ (1.9.2)
43 43 0

In NAA, the radioactive nuclei in the sample undergo both gamma and particle nuclear decay. The figure below presents a
schematic example of nuclear decay. After capturing a free neutron, the excited 60mCo nucleus undergoes an internal transformation
by emitting gamma rays. The lower-energy daughter nucleus 60Co, which is still radioactive, then emits a beta particle. This results
in a high-energy 60Ni nucleus, which once again undergoes an internal transformation by emitting gamma rays. The nucleus then
reaches the stable 60Ni state.

Figure 1.9.2 : Scheme of neutron activation analysis with 59Co as the target nucleus.
Although alpha and beta particle detectors do exist, most detectors used in NAA are designed to detect the gamma rays that are
emitted from the excited nuclei following neutron capture. Each element has a unique radioactive emission and decay path that is
scientifically known. Thus, based on the path and the spectrum produced by the instrument, NAA can determine the identity and
concentration of the element.

Neutron Sources
As mentioned above, there are many different neutron sources that can be used in modern-day NAA. A chart comparing three
common sources is shown in the table below.
Table 1.9.1 : Different neutron sources.

1.9.2 https://chem.libretexts.org/@go/page/55826
Source type Description Example(s) Typical output

Certain isotopes undergo 226Ra(Be), 124Sb(Be), 241Am(Be), 105-107 s-1 GBq-1 or 2.2 1012 s-1 g-
Isotopic neutron sources spontaneous fission and release 252Cf 1 for 252Cf
neutrons as they decay.

Particle accelerators produce Acceleration of deuterium ions


108-1010 s-1 for the first deuterium
neutrons by colliding hydrogen, toward a target containing
Particle accelerators or neutron on deuterium reactions and 109-
deuterium, and tritium with target deuterium or tritium, resulting in
generators 1011 s-1 for deuterium on tritium
nuclei such as deuterium, tritium, the reactions 2H(2H,n)3He and
3H(2H,n)4He reactions
lithium, and beryllium.

Within nuclear reactors, large


atomic nuclei absorbs neutrons
and undergo nuclear fission. The 235U
Nuclear research reactors and 239Pu 1015-1018 m-2 s-1
nuclei split into lighter nuclei,
which releases energy, radiation,
and free neutrons.

Gamma and Particle Detectors


As mentioned earlier, most detectors used in NAA are designed to detect the gamma rays emitted from the decaying nucleus. Two
widely used gamma detectors are the scintillation type and the semiconductor type. The former uses a sensitive crystal, often
sodium iodide that is doped with thallium (NaI(Tl)), that emits light when gamma rays strike it. Semiconductor detectors, on the
other hand, use germanium to form a diode that produces a signal in response to gamma radiation. The signal produced is
proportional to the energy of the emitted gamma radiation. Both types of gamma detectors have excellent sensitivity with detection
limits ranging from 0.1 to 106 nanogram element per gram sample, but semiconductor type detectors usually have superior
resolution.
Furthermore, particles detectors designed to detect the alpha and beta particles that are emitted in nuclear decay are also available;
however, gamma detectors are favorable. Particle detectors require a high vacuum since atmospheric gases in the air can absorb and
affect the emission of these particles. Gamma rays are not affected in this way.

Variations/Parameters
INAA versus RNAA
Instrumental neutron activation analysis (INAA) is the simplest and most widely used form of NAA. It involves the direct
irradiation of the sample, meaning that the sample does not undergo any chemical separation or treatment prior to detection. INAA
can only be used if the activity of the other radioactive isotopes in the sample does not interfere with the measurement of the
element(s) of interest. Interference often occurs when the element(s) of interest are present in trace or ultratrace amounts. If
interference does occur, the activity of the other radioactive isotopes must be removed or eliminated. Radiochemical separation is
one way to do this. NAA that involves sample decomposition and elemental separation is known as radiochemical neutron
activation analysis (RNAA). In RNAA, the interfering elements are separated from the element(s) of interest through an
appropriate separation method. Such methods include extractions, precipitations, distillations, and ion exchanges. Inactive elements
and matrices are often added to ensure appropriate conditions and typical behavior for the element(s) of interest. A schematic
comparison of INAA and RNAA is shown below.

1.9.3 https://chem.libretexts.org/@go/page/55826
Figure 1.9.3 : Schematic Comparison of INAA and RNAA. Adapted from Neutron Activation Analysis Online, www.naa-
online.net/theory/types-of-naa/, (accessed February 2014).

ENAA versus FNAA


Another experimental parameter that must be considered is the kinetic energy of the neutrons used for irradiation. In epithermal
neutron activation analysis (ENAA), the neutrons – known as epithermal neutrons – are partially moderated in the reactor and have
kinetic energies between 0.5 eV to 0.5 MeV. These are lower-energy neutrons as compared to fast neutrons, which are used in fast
neutron activation analysis (FNAA). Fast neutrons are high-energy, unmoderated neutrons with kinetic energies above 0.5 MeV.

PGNAA versus DGNAA


The final parameter to be discussed is the time of measurement. The nuclear decay products can be measured either during or after
neutron irradiation. If the gamma rays are measured during irradiation, the procedure is known as prompt gamma neutron
activation analysis (PGNAA). This is a special type of NAA that requires additional equipment including an adjacent gamma
detector and a neutron beam guide. PGNAA is often used for elements with rapid decay rates, elements with weak gamma emission
intensities, and elements that cannot easily be determined by delayed gamma neutron activation analysis (DGNAA) such as
hydrogen, boron, and carbon. In DGNAA, the emitted gamma rays are measured after irradiation. DGNAA procedures include
much longer irradiation and decay periods than PGNAA, often extending into days or weeks. This means that DGNAA is ideal for
long-lasting radioactive isotopes. A schematic comparison of PGNAA and DGNAA is shown below.

Figure 1.9.4 : Schematic Comparison of PGNAA and DGNAA. Adapted from Neutron Activation Analysis Online, www.naa-
online.net/theory/types-of-naa/, (accessed February 2014).

Examples
Characterizing archaeological materials
Throughout recent decades, NAA has often been used to characterize many different types of samples including archaeological
materials. In 1961, the Demokritos nuclear reactor, a water moderated and cooled reactor, went critical at low power at the National
Center for Scientific Research “Demokritos” (NCSR “Demokritos”) in Athens, Greece. Since then, NCSR “Demokritos” has been
a leading center for the analysis of archaeological materials.

1.9.4 https://chem.libretexts.org/@go/page/55826
Ceramics, carbonates, silicates, and steatite are routinely analyzed at NCSR “Demokritos” with NAA. A routine analysis begins by
weighing and placing 130 milligrams of the powdered sample into a polyethylene vial. Two batches of ten vials, eight samples and
two standards, are then irradiated in the Demokritos nuclear reactor for 45 minutes at a thermal neutron flux of 6 x 1013 neutrons
cm-2 s-1. The first measurement occurs seven days after irradiation. The gamma ray emissions of both the samples and standards
are counted with a germanium gamma detector (semiconductor type) for one hour. This measurement determines the
concentrations of the following elements: As, Ca, K, La, Lu, Na, Sb, Sm, U, and Yb. A second measurement is performed three
weeks after irradiation in which the samples and standards are counted for two hours. In this measurement, the concentrations of
the following elements are determined: Ba, Ce, Co, Cr, Cs, Eu, Fe, Hf, Nd, Ni, Rb, Sc, Ta, Tb, Th, Zn, and Zr.
Using the method described above, NCSR “Demokritos” analyzed 195 samples of black-on-red painted pottery from the late
Neolithic age in what is now known as the Black-On-Red Pottery Project. An example of black-on-red painted pottery is shown
here.

Figure 1.9.5 : Example of black-on-red painted pottery from the late Neolithic age. Reproduced from V. Kilikoglou, A. P. Grimanis,
A. Tsolakidou, A. Hein, D. Malalmidou, and Z. Tsirtsoni, Archaeometry, 2007, 49, 301. Copyright: John Wiley and Sons, Inc.,
(2007).
This project aimed to identify production patterns in this ceramic group and explore the degree of standardization, localization, and
scale of production from 14 sites throughout the Strymonas Valley in northern Greece. A map of the area of interest is provided
below in figure 1.9.6. NCSR “Demokritos” also sought to analyze the variations in pottery traditions by differentiating so-called
ceramic recipes. By using NAA, NCSR “Demokritos” was able to determine the unique chemical make-ups of the many pottery
fragments. The chemical patterning revealed through the analyses suggested that the 195 samples of black-on-red Neolithic pottery
came from four distinct productions areas with the primary production area located in the valley of the Strymon and Angitis rivers.
Although distinct, the pottery from the four different geographical areas all had common technological and stylistic characteristics,
which suggests that a level of standardization did exist throughout the area of interest during the late Neolithic age.

1.9.5 https://chem.libretexts.org/@go/page/55826
Figure 1.9.6 : Map of Strymonas Valley in northern Greece. Reproduced from V. Kilikoglou, A. P. Grimanis, A. Tsolakidou, A.
Hein, D. Malalmidou, and Z. Tsirtsoni, Archaeometry, 2007, 49, 301. Copyright: John Wiley and Sons, Inc., (2007).

Determining elemental concentrations in blood


Additionally, NAA has been used in hematology laboratories to determine specific elemental concentrations in blood and provide
information to aid in the diagnosis and treatment of patients. Identifying abnormalities and unusual concentrations of certain
elements in the bloodstream can also aid in the prediction of damage to the organ systems of the human body.
In one study, NAA was used to determine the concentrations of sodium and chlorine in blood serum. In order to investigate the
accuracy of the technique in this setting, 26 blood samples of healthy male and female donors – aged between 25 and 60 years and
weighing between 50 and 85 kilograms – were selected from the Paulista Blood Bank in São Paulo. The samples were initially
irradiated for 2 minutes at a neutron flux ranging from approximately 1 x 1011 to 6 x 1011 neutrons cm-2 s-1 and counted for 10
minutes using a gold activation detector. The procedure was later repeated using a longer irradiation time of 10 minutes. The
determined concentrations of sodium and chlorine were then compared to standard values. The NAA analyses resulted in
concentrations that strongly agreed with the adopted reference value. For example, the chlorine concentration was found to be 3.41
- 3.68 µg/µL of blood, which correlates closely to the reference value of 3.44 - 3.76 µg/µL of blood. This illustrates that NAA can
accurately measure elemental concentrations in a variety of materials including blood samples.

Limitations
Although NAA is an accurate (~5%) and precise (<0.1%) multi-element analytical technique, it has several limitations that should
be addressed. Firstly, samples irradiated in NAA will remain radioactive for a period of time (often years) following the analysis
procedures. These radioactive samples require special handling and disposal protocols. Secondly, the number of the available
nuclear reactors has declined in recent years. In the United States, only 31 nuclear research and test reactors are currently licensed
and operating. A map of these reactors shown here.

1.9.6 https://chem.libretexts.org/@go/page/55826
Figure 1.9.7 : Map of US nuclear research and test reactors. Taken from Map of Research and Test Reactor Sites,
www.nrc.gov/reactors/operatin...-reactors.html, (accessed February 2014).
As a result of the declining number of reactors and irradiation facilities in the nation, the cost of neutron activation analysis has
increased. The popularity of NAA has declined in recent decades due to both the increasing cost and the development of other
successful multi-element analytical methods such as inductively coupled plasma atomic emission spectroscopy (ICP-AES).

Bibliography
Z. B. Alfassi, Activation Analysis, CRC Press, Boca Raton (1990).
P. Bode, A. Byrne, Z. Chai, A. Chatt, V. Dimic, T. Z. Hossain, J. Kučera, G. C. Lalor, and R. Parthasarathy, Report of an
Advisory Group Meeting Held in Vienna, 22-26 June 1998, IAEA, Vienna, 2001, 1.
V. P. Guinn, Bio. Trace Elem. Res., 1990, 26-27, 1.
L. Hamidatou, H. Slamene, T. Akhal, and B. Zouranen, Imaging and Radioanalytical Techniques in Interdisciplinary Research
– Fundamentals and Cutting Edge Applications, ed. F. Kharfi, InTech, Rijeka (2013).
V. Kilikoglou, A. P. Grimanis, A. Tsolakidou, A. Hein, D. Malalmidou, and Z. Tsirtsoni, Archaeometry, 2007, 49, 301.
S. S. Nargolwalla and E. P. Przybylowicz, Activation Analysis with Neutron Generators, Wiley, New York, 39th edn. (1973).
M. Pollard and C. Heron, Archaeological Chemistry, Royal Society of Chemistry, Cambridge (1996).
B. Zamboi, L. C. Oliveira, and L. Dalaqua Jr., Americas Nuclear Energy Symposium, Miami, 2004.
Neutron Activation Analysis Online, www.naa-online.net/theory/types-of-naa/, (accessed February 2014).
Map of Research and Test Reactor Sites, www.nrc.gov/reactors/operating/map-nonpower-reactors.html, (accessed February
2014).

1.9: Neutron Activation Analysis (NAA) is shared under a CC BY 4.0 license and was authored, remixed, and/or curated by Pavan M. V. Raja &
Andrew R. Barron via source content that was edited to conform to the style and standards of the LibreTexts platform; a detailed edit history is
available upon request.

1.9.7 https://chem.libretexts.org/@go/page/55826
1.10: Total Carbon Analysis
Introduction
Carbon is one of the more abundant elements on the planet; all living things and many non-living things have some form of carbon
in them. Having the ability to measure and characterize the carbon content of a sample is of extreme value in a variety of different
industries and research environments.
Total carbon (TC) content is just one important piece of information that is needed by analysts concerned with the carbon content
of a sample. Having the knowledge of the origin of carbon in the sample, whether it be derived from organic or inorganic material,
is also of extreme importance. For example, oil companies are interested in finding petroleum, a carbon containing material derived
from organic matter, knowing the carbon content and the type of carbon in a sample of interest can mean the difference between
investing millions of dollars and not doing so. Regulatory agencies like the U.S. Environmental Protection Agency (EPA) is
another such example, where regulation of the carbon content and character of that carbon is essential for environmental and
human health.
Considering the importance of identifying and quantifying the carbon content of an analyte, it may be surprising to learn that there
is no one method to measure the carbon content of a sample. Unlike other techniques, no fancy instrument is required (although
some exists that can be useful). In fact, methods to measure the different forms of carbon (organic or inorganic) are different
themselves because they take advantage of the different properties characteristics to the carbon content you are measuring, in fact
you will most likely use multiple techniques to fully characterize the carbon content of a sample, not just one.
Measurements of carbon content are related, and therefore measurement of either total carbon content (TC), total inorganic carbon
content (TIC) and total organic carbon content (TOC) is related to the other two by
TC = TIC + TOC (1.10.1)

.
This means that measurement of two variables can indirectly give you the third, as there are only two classes of carbon: organic
carbon and inorganic carbon.
Herein several of the methods used in measuring the TOC, TIC and TC for samples will be outlined. Not all samples require the
same kind of instruments and methods. The goal of this module is to get the reader to see the simplicity of some of these methods
and understand the need for such quantification and analysis.

Measurement of Total Organic Carbon (TOC)


Sample and Sample Preparation
The total organic carbon content for a variety of different samples can be determined; there are very few samples that cannot be
measured for total carbon content. Before treatment, a sample must be homogenized, whereby a sample is mixed or broken such
that a measurement done on the sample can be representative of the entire sample. For example, if our sample were a rock, we
would want to make sure that the inner core of the rock, which could have a different composition than the outer surface, were
being measured as well. Not homogenizing the sample would lead to inconsistent and perhaps irreproducable results. Techniques
for homogenization vary wildly, depending on the sample, different techniques exist.

Dissolution of Total Inorganic Carbon


In order to measure the organic carbon content in a sample, the inorganic sources of carbon, which exist in the form of carbonate
and bicarbonate salts and minerals, must be removed from the sample. This is typically done by treating the sample with non-
oxidative acids such as H2SO4 and HCl, releasing CO2 and H2O, as shown
2 HCl + CaCO ⟶ CaCl + CO +H O (1.10.2)
3 2 2 2

HCl + NaHCO ⟶ NaCl + H O + CO (1.10.3)


3 2 2

Non oxidative acids are chosen such that minimal amounts of organic carbon are affected. Although the selection of acid chosen to
remove the inorganic sources of carbon is important; depending on your measurement technique, acids may interfere with the
measurement. For example, in the wet measurement technique that will be discussed later, the counter ion Cl- will add systematic
error to the measurement.

1.10.1 https://chem.libretexts.org/@go/page/55828
Treatment of a sample with acid is intended to dissolve all inorganic forms of carbon in the sample. In selectively digesting and
dissolving inorganic forms of carbon, be it aqueous carbonates or bicarbonates or trapped CO2, one can selectively remove
inorganic sources of carbon from organic ones; thereby leaving behind, in theory, only organic carbon in the sample.
It becomes apparent, in this treatment, the importance of sample homogenization. Using the rock example again. If a rock is treated
with acid without homogenizing, the inorganic carbon at the surface of the sample may be dissolved. Only with homogenization
can the acid dissolve in inorganic carbon on the inside of the rock. Otherwise this inorganic carbon may be interpreted as organic
carbon, leading to gross errors in total organic carbon determination.
Shortcomings in the Dissolution of Inorganic Carbon
A large problem and a potential source of error in technique measurement are the assumptions that have to be made, particularly in
the case of TOC measurement, that all of the inorganic carbon has been washed away and separated from the sample. There is no
way to distinguish TOC or TIC spectroscopically, the experimenter is forced to assume that they are looking at is all organic carbon
or all inorganic carbon, when in reality there may be some of both still on the sample.

Quantitative Measurement of TOC


Most TOC quantification methods are destructive in nature. The destructive nature of the methods means that none of the sample
may be recovered. Of the methods, there are two destructive techniques that will be discussed in this module. The first is the wet
method to measure TOC of solid sediment samples, and the second is a the dry combustion.

Wet Methods
Sample Preparation
Following sample pre-treatment with inorganic acids to dissolve away any inorganic material from the sample, a known amount of
potassium dichromate (K2Cr2O7) in concentrated sulfuric acid are added to the sample as per the Walkey-Black procedure, a well
known wet technique. The amount of dichromate and H2SO4 added can vary depending on the expected organic carbon content of
the sample, typically enough H2SO4 is added such that the solid potassium dichromate dissolves in solution.The mixture of
potassium dichromate with H2SO4 is an exothermic one, meaning that heat is evolved from the solution. As the dichromate reacts
according to
2− 0 + 3+
2 Cr O +3 C + 16 H ⟶ 4 Cr + 3 CO +8 H O (1.10.4)
2 7 2 2

The solution will bubble away CO2. Because the only source of carbon in the sample is in theory the organic forms of carbon
(assuming adequate pre-treatment of the sample to remove the inorganic forms of carbon), the evolved CO2 comes from organic
sources of carbon.
Elemental forms of carbon in this method present problems for oxidation of elemental carbon to CO2, meaning that not all of the
carbon will be converted to CO2, which will lead to an underestimation of total organic carbon content in the quantification steps.
In order to facilitate the oxidation of elemental carbon, the digestive solution of dichromate and H2SO4 is heated at 150°C for some
time (~30 min, depending on total carbon content in the sample and the amount of dichromate added). It is important that the
solution not be heated above 150 oC, as decomposition of the dichromate solution.
Other shortcomings, in addition to incomplete digestion, exist with this method. Fe2+ and Cl- in the sample can interfere with the
chromate solution, Fe2+ can be oxidized to Fe3+ and Cl- can form CrO2Cl2 leading to systematic error towards higher organic
carbon content. Conversely MnO2, like dichromate, will oxidize organic carbon, thereby leading to a negative bias and an
underestimation of TOC content in samples.
In order to counteract these biases, several additives can be used in the pre-treatment process. Fe2+ can be oxidized with mild
oxidant phosphoric acid, which will not oxidize organic carbon. Treatment of the digestive solution with AgSO2 can precipitate
silver chloride. MnO2 interferences can be dealt with using FeSO4, where the oxidation power of the manganese is dealt with by
taking the iron(II) sulfate to the +3 oxidation state. Any excess iron(II) can be dealt with using phosphoric acid.
Quantification of TOC
What follows sample treatment, where all of the organic carbon has been digested, is a titration to oxidize the excess dichromate in
the sample. Comparing the excess that is titrated to the amount that was originally added to the original solution, one can do
stoichiometric calculations according to Equation 1.10.4 and calculate the amount of dichromate that oxidized the organic carbon
in the sample, thereby allowing the determination of TOC in the sample.

1.10.2 https://chem.libretexts.org/@go/page/55828
How this titration is run is up to the user. Manual, potentiometric, titrations are all available to the investigator doing the TOC
measurement, as well as some others.
Manual titrations are similar to any other type of manual titration method. An indicator must be used in manual titrations, and in
the case of this wet method, commercially available “ferroin” is used. Titrant is typically ferrous ammonium sulfate. Titrant is
added until equivalence is reached. Indicative of reaching equivalence is color change catalyzed by the indicator. Depending on
the sample measured color change may be difficult to notice.
Insertion of platinum electrodes to the sample can be used to measure conductance of sample using potentiometric tirtration.
When sample reached endpoint, conductance will essentially be 0 or whatever the endpoint of the solution was set to. This
method presents several advantages over manual titration methods because titration can be automated to respond to feedback
from platinum electrodes so equivalence point determination is not color dependent.
Alternative to titration methods, capture of evolved CO2 presents another pheasable quantification method, as oxidized organic
carbon will be evolved as CO2. CO2 can be captured on absorbent material such as ascarite or other tared absorbent, whose
mass change as a result of absorbed CO2 can be measured, or the absorbed CO2 could be desorbed and quantified via IR non-
dispersive cell.
Disadvantages of Wet Technique
Measurement of TOC via the described wet techniques is a rather crude method to measure organic carbon content in a sample.
The technique relies on several assumptions that in reality are not wholly accurate, leading to TOC values that are in reality an
approximate.
The treatment with acid to remove the inorganic forms of carbon assumes that all of the inorganic carbon is removed and
washed away in the acid treatment, but in reality this is probably not true, as some inorganic carbon will cling to the sample and
be quantified incorrectly.
In the digestion process, which assumes that all of the carbon in the sample— which is already presumed to be entirely organic
carbon—is completely converted carbon dioxide, taking no account for the possible solubility of the carbon dioxide in the wet
sample or incomplete oxidation of carbon in the sample.
The wet method to measure TOC relies on the use of dichromate, while a very good oxidant, is a very toxic reagent with which
to analysis.
TOC Measurement of Water
As mentioned previously, measurement of TOC levels in water is extremely valuable to regulatory agencies concerned with water
quality. The presence of organic carbon in a substance that should have no carbon is of concern. Measurement of TOC in water
uses a variant of the wet method in order to avoid highly toxic oxidants: typically a persulfate salt is used as an oxidant instead of
dichromate.

Figure 1.10.1 : Structural representation of persulfate salt, in this case potassium salt. Breaking of oxygen-oxygen bond responsible
for radical-induced oxidation.
The procedure for measuring TOC levels in water is essentially the same as in the typical wet oxidation technique. The water is
first acidified to remove inorganic sources of carbon. Now because water is being measured, one cannot simply wash away the
inorganic carbon. The inorganic carbon escapes from the water solution as CO2. The remaining carbon in the solution is thought to
be organic. Treatment of the solution with persulfate will do nothing. Irradiation of the solution treated with persulfate with UV
radiation or heating will activate a radical species. This radical species will mediate oxidation of the organic carbon to CO2, which
can then be quantified by similar methods as the traditional wet oxidation technique.

Dry Methods
As an alternative to technique for TOC measurement, dry techniques present several advantages over wet techniques. Dry
techniques frequently involve the measurement of evolved carbon from the combustion of a sample. In this section of the module,
TOC measurements using dry techniques will be discussed.

1.10.3 https://chem.libretexts.org/@go/page/55828
Sample Pre-treatment
Like in the wet-oxidation case, measurement of TOC by dry techniques requires the removal of inorganic forms of carbon, and
therefore samples are treated with inorganic acids to do so. The inorganic acids are washed away and theoretically only organic
forms of carbon remain. Before combustion of the sample, the treated sample must be completely dried so as to remove any
moisture from the sample. In the case where non-volatile organics are present, or where little concern about the escape of organic
material exists (e.g., rock samples or Kerogen), sample can be placed in a 100 °C oven overnight. In the case where evolution of
organic matter at slightly elevated temperatures is a problem, drying can be done under vacuum and under the presence of deterite.
Volatile organics are difficult to measure using dry techniques because the sample needs to be without moisture, and removal of
moisture by any technique will most likely remove volatile organics.
Sample Quantification
As mentioned before, quantification of TOC in the dry quantification method will proceed via complete combustion of the sample
in a carbon free atmosphere (typically a pure oxygen atmosphere). Quantification of sample is performed via non-dispersive
infrared detection cell. A characteristic asymmetric stretching at 2350 cm-1 can be seen for CO2. The intensity of this infrared
signal CO2 is proportional to the quantity of CO2 in the sample. Therefore, in order to translate signal intensity to amount, a
calibration curve is constructed from known amounts of pure calcium carbonate, looking specifically at the intensity of the CO2
peak. One may point out that calcium carbonate is an inorganic source of carbon, but it is important to note that the source of
carbon has no effect on its quantification. Preparation of a calibration curve follows similar preparation as to an analyte, while no
pre-treatment with acid is needed, the standards must be thoroughly dried in an oven. When a sample is ready to be analyzed, it is
first weighed on some form of analytical balance, and then placed in the combustion analyzer, such as a LECO analyzer, where the
oven and the non-dispersive IR cell are one machine.

Figure 1.10.2 : Example of a LECO analyzer. Samples placed in small cream colored trays and combusted in oven under oxygen
atmosphere (big box lower right) computer output to show carbon content. Material used with permission by LECO Corporation
Combustion proceeds at temperatures in the excess of 1350 oC in a stream of pure oxygen. Comparing the intensity of your
characteristic IR peak to the intensities of the characteristic IR peaks of your known standards, the TOC of the sample can be
determined. By comparing the mass of the sample to the mass of carbon obtained from the analyzer, the % organic carbon in the
sample can be determined according to

% TOC  =  mass carbon/mass sample  (1.10.5)

Use of this dry technique is most common for rock and other solid samples. In the oil and gas industry, it is extremely important to
know the organic carbon content of rock samples in order to ascertain production viability of a well. The sample can be loaded in
the LECO combustion analyzer and pyrolyzed in order to quantify TOC.

1.10.4 https://chem.libretexts.org/@go/page/55828
Measurement of Total Carbon (TC)
As shown in Equation \ref {eq:TC} the total carbon in a sample (TC) is the sum of the inorganic forms of carbon and organic forms
of carbon in a sample.
It is known that no other sources of carbon contribute to the TC determination because no other sources of carbon exist. So in
theory, if one could quantify the TOC by a method described in the previous section, and follow that with a measurement of the
TIC in the pre-treatment acid waste, one could find the TC of a sample by summing the value obtained for TIC and the value
obtained for TOC. However, in TC quantification this is hardly done: partly in order to avoid propagation of error associated with
the other two methods, also cost restraints.
In measuring TC of a sample, the same dry technique of combustion of the sample is used, just like in the quantification of TOC.
The same analyzer used to measure TOC can handle a TC measurement. No sample pre-treatment with acid is needed, so it is
important to remember that the characteristic peak of CO2 now seen is representative of the carbon of the entire sample. Now using
Equation \ref {eq:TC}, the TIC carbon of the sample can be found as well. Subtraction of the TOC from the measured TC in the
analyzer gives the value for TIC.

Measurement of total inorganic carbon (TIC)


Direct methods to measure the TIC of a sample, in addition to indirect measurement by taking advantage of Equation, are possible.
Typical TIC measurements are done on water samples, where the alkalinity and hardness of water is a result of inorganic
carbonates, be it bicarbonate or carbonate. Treatment of these types of samples follows similar procedures to treatment of samples
for organic carbon. A sample of water is acidified, such that the equilibrium, Equation obeys Le Chatelier’s principle and favors the
release of CO2. The CO2 released can be measured in a variety of different ways

As with the combustion technique for measuring TC and TOC, measurement of the intensity of the characteristic IR stretch for CO2
compared to standards can be used to quantity of TIC in a sample. However, in this case, it is emission of IR radiation that is
measured, not absorption. An instrument that can do such a measurement is a FIRE-TIC, meaning Flame IR emission. This
instrument consists of a purge like devices connected to a FIRE detector.

Figure 1.10.3 : FIRE-TIC instrument. Sample is placed in a degassing purge box, in which typically helium or IR another IR
inactive gas. As the gas passes through the sample CO2 is released from sample.

Summary
Measurement of Carbon content is crucial for a lot of industries. In this module you have seen a variety of ways to measure Total
Carbon TC, as well as the source of that carbon, whether it be organic in nature (TOC), or inorganic (TIC). This information is
extremely important for several industries: from oil exploration, where information on carbon content is needed to evaluate a
formation’s production viability, to regulatory agencies, where carbon content and its origin are needed to ensure quality control
and public safety.
TOC, TC, TIC measurements do have significant limitations. Mostly all techniques are destructive in nature, meaning that sample
cannot be recovered. Further limitations include assumptions that have to be made in the measurement. In TOC measurement for
example, assumptions that all TIC has been removed in pretreatments with acid have to be made, as well as that all organic carbon

1.10.5 https://chem.libretexts.org/@go/page/55828
is completely oxidized to CO2. In TIC measurements, it is assumed that all carbon sources are removed from the sample and
detected. Several things can be done to promote these conditions so as to make such assumptions valid.
All measurements cost money, because TOC, TIC, and TC are all related by Equation, more frequently than not only two
measurements are done, and the third value is found by using their relation to one another.

Bibliography
Z. A, Wang, S. N. Chu, and K. A. Hoering, Environ. Sci. Technol., 2013, 47, 7840.
B. A. Schumacher, Methods for the determination of Total Organic Carbon (TOC) in Soils and Sediments. U.S. Environmental
Protection Agency, Washington, DC, EPA/600/R-02/069 (NTIS PB2003-100822), 2002
B.B. Bernard, H. Bernard, and J.M. Brooks: Determination of Total Carbon, Total Organic Carbon and Inorganic Carbon in
Sediments, College Station, Texas, USA, DI-Brooks International and B&B Laboratories, Inc., www.tdi-
bi.com/analytical_ser...environmental/ NOAA_methods/TOC.pdf (accessed October 21, 2011).
Julie, The Blogsicle. www.theblogsicle.com/?p=345
Schlumberger Ltd., Oilfield Review Autumn 2011, Schlumberger Ltd (2011), 43.
S. W. Kubala, D. C. Tilotta, M. A. Busch, and K. W. Busch, Anal. Chem., 1989, 61, 1841.
University of Georgia School CAES CAES Publications, University of Georgia Cooperative Extension Circular 922,
http://www.caes.uga.edu/publications...cfm?pk_id=7895.

1.10: Total Carbon Analysis is shared under a CC BY 4.0 license and was authored, remixed, and/or curated by Pavan M. V. Raja & Andrew R.
Barron via source content that was edited to conform to the style and standards of the LibreTexts platform; a detailed edit history is available upon
request.

1.10.6 https://chem.libretexts.org/@go/page/55828
1.11: Fluorescence Spectroscopy
Introduction
Atomic fluorescence spectroscopy (AFS) is a method that was invented by Winefordner and Vickers in 1964 as a means to analyze
the chemical concentration of a sample. The idea is to excite a sample vapor with the appropriate UV radiation, and by measuring
the emitting radiation, the amount of the specific element being measured could be quantified. In its most basic form, AFS consists
of a UV light source to excite the sample, a monochromator, a detector and a readout device (figure 1.11.1). Cold vapor atomic
fluorescence spectroscopy (CVAFS) uses the same technique as AFS, but the preparation of the sample is adapted specifically to
quantify the presence of heavy metals that are volatile, such as mercury, and allows for these elements to be measured at room
temperature.

Figure 1.11.1 The basic setup for CVAFS. *The monochromator can be in either position in the scheme.

Theory
The theory behind CVAFS is that as the sample absorbs photons from the radiation source, it will enter an excited state. As the
atom falls back into the ground state from its excited vibrational state(s), it will emit a photon, which can then be measured to
determine the concentration. In its most basic sense, this process is represented by 1.11.1, where PF is the power given off as
photons from the sample, Pabs is the power of the radiation absorbed by the sample, and φ is the proportionality factor of the energy
lost due to collisions and interactions between the atoms present, and not due to photon emission.

PF   =  ψ Pabs (1.11.1)

Sample Preparation

For CVAFS, the sample must be digested, usually with an acid to break down the compound being tested so that all metal atoms in
the sample are accessible to be vaporized. The sample is put into a bubbler, usually with an agent that will convert the element to
its gaseous species. An inert gas carrier such as argon is then passed through the bubbler to carry the metal vapors to the
fluorescence cell. It is important that the gas carrier is inert, so that the signal will only be absorbed and emitted by the sample in
question and not the carrier gas.
Atomic Fluorescence Spectroscopy
Once the sample is loaded into the cell, a collimated (almost parallel) UV light source passes through the sample so that it will
fluoresce. A monochromator is often used, either between the light source and the sample, or between the sample and the detector.
These two different setups are referred to as excitation or emission spectrum, respectively. In an excitation spectrum, the light
source is kept at a constant wavelength via the monochromator, and multiple wavelengths of emitted light are gathered, whereas in
the emission spectrum, only the specified wavelength of light emitted from the sample is measured, but the sample is exposed to
multiple wavelengths of light from the excitatory source. The fluorescence will be detected by a photomultiplier tube, which is
extremely light sensitive, and a photodiode is used to convert the light into voltage or current, which can then in turn be interpreted
into the amount of the chemical present.

Detecting Mercury Using Gold Amalgamation and Cold Vapor Atomic Fluorescence Spectroscopy
Introduction
Mercury poisoning can damage the nervous system, kidneys, and also fetal development in pregnant women, so it is important to
evaluate the levels of mercury present in our environment. Some of the more common sources of mercury are in the air (from
industrial manufacturing, mining, and burning coal), the soil (deposits, waste), water (byproduct of bacteria, waste), and in food
(especially seafood). Although regulation for food, water and air mercury content differs, EPA regulation for mercury content in
water is the lowest, and it cannot exceed 2 ppb (27 µg/L).

1.11.1 https://chem.libretexts.org/@go/page/55829
In 1972, J. F. Kopp et al. first published a method to detect minute concentrations of mercury in soil, water, and air using gold
amalgamation and cold vapor atomic fluorescence spectroscopy. While atomic absorption can also measure mercury
concentrations, it is not as sensitive or selective as cold vapour atomic fluorescence spectroscopy (CVAFS).
Sample Preparation
As is common with all forms of atomic fluorescence spectroscopy (AFS) and atomic absorption spectrometry (AES), the sample
must be digested, usually with an acid, to break down the compounds so that all the mercury present can be measured. The sample
is put in the bubbler with a reducing agent such as stannous chloride (SnCl2) so that Hg0 is the only state present in the sample.
Gold Amalgam and CVAFS
Once the mercury is in its elemental form, the argon enters the bubbler through a gold trap, and carries the mercury vapors out of
the bubbler to the first gold trap, after first passing through a soda lime (mixture of Ch(OH)2, NaOH, and KOH) trap where any
remaining acid or water vapors are caught. After all the mercury from the sample is absorbed by the first gold trap, it is heated to
450 °C, which causes the mercury absorbed onto the gold trap to be carried by the argon gas to the second gold trap. Once the
mercury from the sample has been absorbed by the second trap, it is heated to 450 °C, releasing the mercury to be carried by the
argon gas into the fluorescence cell, where light at a wavelength of 253.7 nm will be used for mercury samples. The detection limit
for mercury using gold amalgamation and CVAFS is around 0.05 ng/L, but the detection limit will vary due to the equipment being
used, as well as human error.
Calculating CVAFS concentrations
A standard solution of mercury should be made, and from this dilutions will be used to make at least five different standard
solutions. Depending on the detection limit and what is being analyzed, the concentrations in the standard solutions will vary. Note
that what other chemicals the standard solutions contain will depend upon how the sample is digested.

Example 1
A 1.00 g/mL Hg (1 ppm) working solution is made, and by dilution, five standards are made from the working solution, at 5.0,
10.0, 25.0, 50.0, and 100.0 ng/L (ppt). If these five standards give peak heights of 10 units, 23 units, 52 units, 110 units, and 207
units, respectively, then 1.11.2 is used to calculate the calibration factor, where CFx is the calibration factor, Ax is the area of the
peak or peak height, and Cx is the concentration in ng/L of the standard, 1.11.3.
CFx   =  AX / CX (1.11.2)

10/5.0 ng/L  =  2.00 units L/ng (1.11.3)

The calibration factors for the other four standards are calculated in the same fashion: 2.30, 2.08, 2.20, and 2.07, respectively. The
average of the five calibration factors is then taken, 1.11.4.

CFm   =  (2.00  +  2.30  +  2.08  +  2.20  +  2.07)/5  =  2.13 units L/ng (1.11.4)

Now to calculate the concentration of mercury in the sample, 1.11.5 is used, where As is the area of the peak sample, CFm is the
mean calibration factor, Vstd is the volume of the standard solution minus the reagents added, and Vsmp is the volume of the initial
sample (total volume minus volume of reagents added). If As is measured at 49 units, Vstd = 0.47 L, and Vsmp = 0.26 L, then the
concentration can be calculated, 1.11.6.

[Hg] (ng/L)  =  (As / CFm ) ⋅ (Vstd / Vsmp ) (1.11.5)

49 units/2.13 units L/ng) ⋅ (0.47 L/0.26 L)  =  43.2 ng/L of Hg present (1.11.6)

Sources of Error
Contamination from the sample collection is one of the biggest sources of error: if the sample is not properly collected or
hands/gloves are not clean, this can tamper with the concentration. Also, making sure the glassware and equipment is clean from
any sources of contamination.
Furthermore, sample vials that are used to store mercury-containing samples should be made out of borosilicate glass or
fluoropolymer, because mercury can leach or absorb other materials, which could cause an inaccurate concentration reading.

1.11.2 https://chem.libretexts.org/@go/page/55829
The Application of Fluorescence Spectroscopy in the Mercury Ion Detection
Mercury in the Environment
Mercury pollution has become a global problem and seriously endangers human health. Inorganic mercury can be easily released
into the environment through a variety of anthropogenic sources, such as the coal mining, solid waste incineration, fossil fuel
combustion, and chemical manufacturing. It can also be released through the nonanthropogenic sources in the form of forest fires,
volcanic emissions, and oceanic emission.
Mercury can be easily transported into the atmosphere as the form of the mercury vapor. The atmospheric deposition of mercury
ions leads to the accumulation on plants, in topsoil, in water, and in underwater sediments. Some prokaryotes living in the
sediments can convert the inorganic mercury into methylmercury, which can enter food chain and finally is ingested by human.
Mercury seriously endangers people’s health. One example is that many people died due to exposure to methylmercury through
seafood consumption in Minamata, Japan. Exposure in the organic mercury causes a serious of neurological problems, such as
prenatal brain damage, cognitive and motion disorders, vision and hearing loss, and even death. Moreover, inorganic mercury also
targets the renal epithelial cells of the kidney, which results in tubular necrosis and proteinuria.
The crisis of mercury in the environment and biological system compel people to carry out related work to confront the challenge.
To design and implement new mercury detection tools will ultimately aid these endeavors. Therefore, in this paper, we will mainly
introduce fluorescence molecular sensor, which is becoming more and more important in mercury detection due to its easy use, low
cost and high efficiency.
Introduction of Fluorescence Molecular Sensors
Fluorescence molecular sensor, one type of fluorescence molecular probe, can be fast, reversible response in the recognition
process. There are four factors, selectivity, sensitivity, in-situ detection, and real time, that are generally used to evaluate the
performance of the sensor. In this paper, four fundamental principles for design fluorescence molecular sensors are introduced.

Photoinduced Electron Transfer (PET)


Photoinduced electron transfer is the most popular principle in the design of fluorescence molecular sensors. The characteristic
structure of PET sensors includes three parts as shown in Figure 1.11.2:
The fluorophore absorbs the light and emits fluorescence signal.
The receptor selectively interacts with the guest.
A spacer connects the fluorophore and receptor together to form an integral system and successfully, effectively transfers the
recognition information from receptor to fluorophore.

Figure 1.11.2 The general view of the principle of PET fluorescence molecular sensor.
In the PET sensors, photoinduced electron transfer makes the transfer of recognition information to fluorescence signal between
receptor and fluorophore come true. Figure 1.11.2 shows the detailed process of how PET works in the fluorescence molecular
sensor. The receptor could provide the electron to the vacated electoral orbital of the excited fluorophore. The excited electron in
the fluorophore could not come back the original orbital, resulting in the quenching of fluorescence emission. The coordination of
receptor and guest decreased the electron donor ability of receptor reduced or even disrupted the PET process, then leading to the
enhancement of intensity of fluorescence emission. Therefore, the sensors had weak or no fluorescence emission before the
coordination. However, the intensity of fluorescence emission would increase rapidly after the coordination of receptor and gust.

Intramolecular Charge Transfer (ICT)


Intramolecular charge transfer (ICT) is also named photoinduced charge transfer. The characteristic structure of ICT sensors
includes only the fluorophore and recognition group, but no spacer. The recognition group directly binds to the fluorophore. The

1.11.3 https://chem.libretexts.org/@go/page/55829
electron withdrawing or electron donating substituents on the recognition group plays an important role in the recognition. When
the recognition happens, the coordination between the recognition group and guest affects the electron density in the fluorophore,
resulting in the change of fluorescence emission in the form of blue shift or red shift.

Excimer
When the two fluorophores are in the proper distance, an intermolecular excimer can be formed between the excited state and
ground state. The fluorescence emission of the excimer is different with the monomer and mainly in the form of new, broad, strong,
and long wavelength emission without fine structures. The proper distance determines the formation of excimer, therefore
modulation of the distance between the two fluorophores becomes crucial in the design of the sensors based on this mechanism.
The fluorophores have long lifetime in the singlet state to be easily forming the excimers. They are often used in such sensors.

Fluorescence Resonance Energy Transfer (FRET)


FRET is a popular principle in the design of the fluorescence molecular sensor. In one system, there are two different fluorophores,
in which one acts as a donor of excited state energy to the receptor of the other. As shown in Figure 1.11.2, the receptor accepts the
energy from the excited state of the donor and gives the fluorescence emission, while the donor will return back to the electronic
ground state. There are three factors affecting the performance of FRET. They are the distance between the donor and the acceptor,
the proper orientation between the donor emission dipole moment and acceptor absorption moment, and the extent of spectral
overlap between the donor emission and acceptor absorption spectrum (Figure 1.11.3).

Figure 1.11.2 A schematic fluorescence resonance energy transfer system.

Figure 1.11.3 Diagram showing the spectral overlap for fluorescence resonance energy transfer system.
Introduction of Fluorescence Spectroscopy

Fluorescence
Fluorescence is a process involving the emission of light from any substance in the excited states. Generally speaking, fluorescence
is the emission of electromagnetic radiation (light) by the substance absorbed the different wavelength radiation. Its absorption and
emission is illustrated in the Jablonski diagram (Figure 1.11.4), a fluorophore is excited to higher electronic and vibrational state
from ground state after excitation. The excited molecules can relax to lower vibrational state due to the vibrational relaxation and,
then further retune to the ground state in the form of fluorescence emission.

1.11.4 https://chem.libretexts.org/@go/page/55829
Figure 1.11.4 Jablonski diagram of fluorescence.

Instrumentation
Most spectrofluorometers can record both excitation and emission spectra. They mainly consists of four parts: light sources,
monochromators, optical filters and detector (Figure 1.11.5).

Figure 1.11.5 Schematic representation of a fluorescence spectrometer.

Light Sources
Light sources that can emit wavelength of light over the ultraviolet and the visible range can provide the excitation energy. There
are different light sources, including arc and incandescent xenon lamps, high-pressure mercury (Hg) lamps, Xe-Hg arc lamps, low
pressure Hg and Hg-Ar lamps, pulsed xenon lamps, quartz-tungsten halogen (QTH) lamps, LED light sources, etc. The proper light
source is chosen based on the application.

Monochromators
Prisms and diffraction gratings are two mainly used types of monocharomators, which help to get the experimentally needed
chromatic light with a wavelength range of 10 nm. Typically, the monocharomators are evaluated based on dispersion, efficiency,
stray light level and resolution.

Optical Filters
Optical filters are used in addition to monochromators in order to further purifying the light. There are two kinds of optical filters.
The first one is the colored filter, which is the most traditional filter and is also divided into two catagories: monochromatic filter
and long-pass filter. The other one is thin film filter that is the supplement for the former one in the application and being gradually
instead of colored filter.

Detector
An InGaAs array is the standard detector used in many spectrofluorometers. It can provide rapid and robust spectral
characterization in the near-IR.

1.11.5 https://chem.libretexts.org/@go/page/55829
Applications

PET Fluorescence Sensor


As a PET sensor 2-{5-[(2-{[bis-(2-ethylsulfanyl-ethyl)-amino]-methyl}-phenylamino)-methyl]-2-chloro-6-hydroxy-3-oxo-3H-
xanthen-9-yl}-benzoic acid (MS1) (Figure 1.11.6) shows good selectivity for mercury ions in buffer solution (pH = 7, 50 mM
PIPES, 100 mM KCl). From Figure 1.11.7, it is clear that, upon the increase of the concentration of Hg2+ ions, the coordination
between the sensor and Hg2+ ions disrupted the PET process, leading to the increase of the intensity of fluorescence emission with
slight red shift to 528 nm. Sensor MS1 also showed good selectivity for Hg2+ ions over other cations of interest as shown in the
right bars in Figure 1.11.8; moreover, it had good resistance to the interference from other cations when detected Hg2+ ions in the
mixture solution excluding Cu2+ ions as shown in the dark bars in the Figure 1.11.8.

Figure 1.11.6 Structure of the PET fluorescence sensor 2-{5-[(2-{[bis-(2-ethylsulfanyl-ethyl)-amino]-methyl}-phenylamino)-


methyl]-2-chloro-6-hydroxy-3-oxo-3H-xanthen-9-yl}-benzoic acid.

Figure 1.11.7 Fluorescence spectra of sensor MS1 (1 µM) upon addition of Hg2+ (0 - 3 µM) in buffer solution (pH = 7, 50 mM
PIPES, 100 mM KCl) with an excitation of 500 nm.

Figure 1.11.8 The selectivity of MS1 for Hg2+ ions in the presence of other cations of interest. The light bars represent the
emission of MS1 in the presence of 67 equiv of the interested cations. The dark bars represent the change in integrated emission
that occurs upon subsequent addition of 67 equiv of Hg2+ to the mixed solution.
ICT Fluorescence Sensor

2,2',2'',2'''-(3-(benzo[d]thiazol-2-yl)-2-oxo-2-H-chromene-6,7-diyl) bis(azanetriyl)tetrakis(N-(2-hydroxyethyl)acetamide) (RMS)


(Figure 1.11.9) has been shown to be an ICT fluorescence sensor. From Figure 1.11.10, it is clear that, with the gradual increase of
the concentration of Hg2+ ions, fluorescence emission spectra revealed a significant blue shift, which was about 100-nm emission
band shift from 567 to 475 nm in the presence of 40 equiv of Hg2+ ions. The fluorescence change came from the coexistence of two
electron-rich aniline nitrogen atoms in the electron-donating receptor moiety, which prevented Hg2+ ions ejection from them

1.11.6 https://chem.libretexts.org/@go/page/55829
simultaneously in the excited ICT fluorophore. Sensor RMS also showed good selectivity over other cations of interest. As shown
in Figure 1.11.11, it is easy to find that only Hg2+ ions can modulate the fluorescence of RMS in a neutral buffered water solution.

Figure 1.11.9 Structure of the ICT fluorescence sensor 2,2',2'',2'''-(3-(benzo[d]thiazol-2-yl)-2-oxo-2-H-chromene-6,7-


diyl)bis(azanetriyl) tetrakis(N-(2-hydroxyethyl)acetamide) (RMS).

Figure 1.11.10 Fluorescence spectra of RMS (5 µM) upon addition of Hg2+ (0 µM to 200 µM) in 0.05 M phosphate-buffered water
solution (pH 7.5) with an excitation of 390 nm.

Figure 1.11.11 Fluorescence response of 10 µM RMS in the presence of 20 equiv of different cations of interest at the same
condition: control (0), Cd2+ (1), Hg2+ (2), Fe3+ (3), Zn2+ (4), Ag+ (5), Co2+ (6), Cu2+ (7), Ni2+ (8), and Pb2+ (9).
Excimer Fluorescence Sensor
The (NE,N'E)-2,2'-(ethane-1,2-diyl-bis(oxy))bis(N-(pyren-4-ylmethylene)aniline) (BA) (Figure 1.11.12 is the excimer fluorescence
sensor. As shown in Figure 1.11.13, when BA existed without mercury ions in the mixture of HEPES-CH3CN (80:20, v/v, pH 7.2),
it only had the weak monomer fluorescence emission. Upon the increase of the concentration of mercury ions in the solution of
BA, a strong excimer fluorescence emission at 462 nm appeared and increased with the change of the concentration of mercury
ions. From Figure 1.11.14, it is clear that BA showed good selectivity for mercury ions. Moreover, it had good resistance to the
interference when detecting mercury ions in the mixture solution.

Figure 1.11.12 Structure of the excimer fluorescence sensor (NE,N'E)-2,2'-(ethane-1,2-diyl-bis(oxy))bis(N-(pyren-4-ylmethylene)


aniline) (BA).

1.11.7 https://chem.libretexts.org/@go/page/55829
Figure 1.11.13 Fluorescence spectra of BA (1 µM) upon addition of Hg2+ (0 µM to 10 µM) in the mixture of HEPES-CH3CN
(80:20, v/v, pH 7.2) with an excitation of 365 nm.

Figure 1.11.14 Fluorescence response of BA (1 µM) with 10 equiv of other cations of interest in the same condition. Bars represent
the final (F) over the initial (F0) integrated emission. The red bars represent the addition of the competing metal ion to a 1 µM
solution of BA. The blue bars represent the change of the emission that occurs upon the subsequent addition of 10 µM Hg2+ to the
above solution.
FRET Fluorescence Sensor
The calix[4]arene derivative bearing two pyrene and rhodamine fluorophores (CPR) (Figure 1.11.15) is a characteristic FRET
fluorescence sensor. Fluorescence titration experiment of CPR (10.0 μM) with Hg2+ ions was carried out in CHCl3/CH3CN (50:50,
v/v) with an excitation of 343 nm. As shown in Figure 1.11.16, upon gradual increase the concentration of Hg2+ ions in the
solution of CPR, the increased fluorescence emission of the ring-opened rhodamine at 576 nm was observed with a concomitantly
declining excimer emission of pyrene at 470 nm. Moreover, an isosbestic point centered at 550 nm appeared. This change in the
fluorescence emission demonstrated that an energy from the pyrene excimer transferred to rhodamine, resulting from the trigger of
Hg2+ ions. Figure 1.11.17 showed that CPR had good resistance to other cations of interest when detected Hg2+ ions, though Pb2+
ions had little interference in this process.

Figure 1.11.15 Structure of the FRET fluorescence sensor calix[4]arene derivative (CPR) bearing two pyrene and rhodamine
fluorophores.

1.11.8 https://chem.libretexts.org/@go/page/55829
Figure 1.11.16 Fluorescence spectra of CPR (10.0 μM) in CHCl3/CH3CN (50:50, v/v) upon addition of different concentrations of
Hg(ClO4)2 (0 μM to 30 μM).

Figure 1.11.17 Competition experiment of 10.0 μM CPR at 576 nm with 10 equiv of other cations of interest in the presence of
Hg2+ (3 equiv) in the same condition. F0 and F denote the fluorescence intensity of CPR and Hg2+ ions and the interested metal
ions in the presence of CPR and Hg2+ ions.

1.11: Fluorescence Spectroscopy is shared under a CC BY 4.0 license and was authored, remixed, and/or curated by Pavan M. V. Raja & Andrew
R. Barron via source content that was edited to conform to the style and standards of the LibreTexts platform; a detailed edit history is available
upon request.

1.11.9 https://chem.libretexts.org/@go/page/55829
1.12: An Introduction to Energy Dispersive X-ray Spectroscopy
Introduction
Energy-dispersive X-ray spectroscopy (EDX or EDS) is an analytical technique used to probe the composition of a solid materials.
Several variants exist, but the all rely on exciting electrons near the nucleus, causing more distant electrons to drop energy levels to
fill the resulting “holes.” Each element emits a different set of X-ray frequencies as their vacated lower energy states are refilled, so
measuring these emissions can provide both qualitative and quantitative information about the near-surface makeup of the sample.
However, accurate interpretation of this data is dependent on the presence of high-quality standards, and technical limitations can
compromise the resolution.

Physical Underpinnings
In the quantum mechanical model of the atom, an electron’s energy state is defined by a set of quantum numbers. The primary
quantum number, n, provides the coarsest description of the electron’s energy level, and all the sublevels that share the same
primary quantum number are sometimes said to comprise an energy “shell.” Instead of describing the lowest-energy shell as the “n
= 1 shell,” it is more common in spectroscopy to use alphabetical labels: The K shell has n = 1, the L shell has n = 2, the M shell
has n = 3, and so on. Subsequent quantum numbers divide the shells into subshells: one for K, three for L, and five for M.
Increasing primary quantum numbers correspond with increasing average distance from the nucleus and increasing energy (Figure
1.12.1). An atom’s core shells are those with lower primary quantum numbers than the highest occupied shell, or valence shell.

Figure 1.12.1 diagram of the core electronic energy levels of an atom, with the lowest energy shell, K, nearest the nucleus. Circles
are used here for convenience – they are not meant to represent the shapes of the electron’s orbitals. Adapted from Introduction to
Energy Dispersive X-ray Spectroscopy (EDS), micron.ucr.edu/public/manuals/EDS-intro.pdf.
Transitions between energy levels follow the law of conservation of energy. Excitation of an electron to a higher energy state
requires an input of energy from the surroundings, and relaxation to a lower energy state releases energy to the surroundings. One
of the most common and useful ways energy can be transferred into and out of an atom is by electromagnetic radiation. Core shell
transitions correspond to radiation in the X-ray portion of the spectrum; however, because the core shells are normally full by
definition, these transitions are not usually observed.
X-ray spectroscopy uses a beam of electrons or high-energy radiation (see instrument variations, below) to excite core electrons to
high energy states, creating a low-energy vacancy in the atoms’ electronic structures. This leads to a cascade of electrons from
higher energy levels until the atom regains a minimum-energy state. Due to conservation of energy, the electrons emit X-rays as
they transition to lower energy states. It is these X-rays that are being measured in X-ray spectroscopy. The energy transitions are
named using the letter of the shell where ionization first occurred, a Greek letter denoting the group of lines that transition belongs
to, in order of decreasing importance, and a numeric subscript ranking the peak's the intensity within that group. Thus, the most
intense peak resulting from ionization in the K shell would be Kα1 (Figure 1.12.2). Since each element has a different nuclear
charge, the energies of the core shells and, more importantly, the spacing between them vary from one element to the next. While
not every peak in an element’s spectrum is exclusive to that element, there are enough characteristic peaks to be able to determine
composition of the sample, given sufficient resolving power.

1.12.1 https://chem.libretexts.org/@go/page/55830
Figure 1.12.2 A diagram of the energy transitions after the excitation of a gold atom. The arrows show the direction the vacancy
moves when the higher energy electrons move down to refill the core. Adapted from Introduction to Energy Dispersive X-ray
Spectroscopy (EDS), micron.ucr.edu/public/manuals/EDS-intro.pdf.

Instrumentation and Sample Preparation


Instrument variations
There are two common methods for exciting the core electrons off the surface atoms. The first is to use a high-energy electron
beam like the one in a scanning electron microscope (SEM). The beam is produced by an electron gun, in which electrons emitted
thermionically from a hot cathode are guided down the column by an electric field and focused by a series of negatively charged
“lenses.” X-rays emitted by the sample strike a lithium-drifted silicon p-i-n junction plate. This promotes electrons in the plate into
the conduction band, inducing a voltage proportional to the energy of the impacting X-ray which generally falls between about 1
and 10 keV. The detector is cooled to liquid nitrogen temperatures to reduce electronic noise from thermal excitations.
It is also possible to use X-rays to excite the core electrons to the point of ionization. In this variation, known as energy-dispersive
X-ray fluorescence analysis (EDXRFA or XRF), the electron column is replaced by an X-ray tube and the X-rays emitted by the
sample in response to the bombardment are called secondary X-rays, but these variants are otherwise identical.
Regardless of the excitation method, subsequent interactions between the emitted X-rays and the sample can lead to poor resolution
in the X-ray spectrum, producing a Gaussian-like curve instead of a sharp peak. Indeed, this spreading of energy within the sample
combined with the penetration of the electron or X-ray beam leads to the analysis of a roughly 1 µm3 volume instead of only the
surface features. Peak broadening can lead to overlapping peaks and a generally misleading spectrum. In cases where a normal
EDS spectrum is inadequately resolved, a technique called wavelength-dispersive X-ray spectroscopy (WDS) can be used. The
required instrument is very similar to the ones discussed above, and can use either excitation method. The major difference is that
instead of having the X-rays emitted by the sample hit the detector directly, they first encounter an analytical crystal of know lattice
dimensions. Bragg’s law predicts that the strongest reflections off the crystal will occur for wavelengths such that the path
difference between a rays reflecting from consecutive layers in the lattice is equal to an integral number of wavelengths. This is
represented mathematically as 1.12.1, where n is an integer, λ is the wavelength of impinging light, d is the distance between layers
in the lattice, and θ is the angle of incidence. The relevant variables for the equation are labeled in Figure 1.12.3.
nλ  =  2d sin θ (1.12.1)

Figure 1.12.3 A diagram of a light beam impinging on a crystal lattice. If the light meets the criterion nλ = 2d sin(θ), Bragg’s law
predicts that the waves reflecting off each layer of the lattice interfere constructively, leading to a strong signal. Adapted from D.
Henry, N. Eby, J. Goodge, and D. Mogk, X-ray Reflection in Accordance with Bragg’s Law,
http://serc.carleton.edu/research_education/geochemsheets/BraggsLaw.html

1.12.2 https://chem.libretexts.org/@go/page/55830
By moving the crystal and the detector around the Rowland circle, the spectrometer can be tuned to examine specific wavelengths (
1.12.1). Generally, an initial scan across all wavelengths is taken first, and then the instrument is programmed to more closely

examine the wavelengths that produced strong peaks. The resolution available with WDS is about an order of magnitude better than
with EDS because the analytical crystal helps filter out the noise of subsequent, non-characteristic interactions. For clarity, “X-ray
spectroscopy” will be used to refer to all of the technical variants just discussed, and points made about EDS will hold true for XRF
unless otherwise noted.

Figure 1.12.4 A schematic of a typical WDS instrument. The analytical crystal and the detector can be moved around an arc known
as the Rowland Circle. This grants the operator the ability to change the angle between the sample, the crystal, and the detector,
thereby changing the X-ray wavelength that would satisfy Bragg’s law. The sample holder is typically stationary. Adapted from D.
Henry and J. Goodge, Wavelength-Dispersive X-ray Spectroscopy (WDS),
http://serc.carleton.edu/research_education/geochemsheets/wds.html.
Sample Preparation
Compared with some analytical techniques, the sample preparation required for X-ray spectroscopy or any of the related methods
just discussed is trivial. The sample must be stable under vacuum, since the sample chamber is evacuated to prevent the atmosphere
from interfering with the electron beam or X-rays. It is also advisable to have the surface as clean as possible; X-ray spectroscopy
is a near-surface technique, so it should analyze the desired material for the most part regardless, but any grime on the surface will
throw off the composition calculations. Simple qualitative readings can be obtained from a solid of any thickness, as long as it fits
in the machine, but for reliable quantitative measurements, the sample should be shaved as thin as possible.

Data Interpretation
Qualitative analysis, the determination of which elements are present in the sample but not necessarily the stoichiometry, relies on
empirical standards. The energies of the commonly used core shell transitions have been tabulated for all the natural elements.
Since combinations of elements can act differently than a single element alone, standards with compositions as similar as possible
to the suspected makeup of the sample are also employed. To determine the sample’s composition, the peaks in the spectrum are
matched with peaks from the literature or standards.
Quantitative analysis, the determination of the sample’s stoichiometry, needs high resolution to be good enough that the ratio of the
number of counts at each characteristic frequency gives the ratio of those elements in the sample. It takes about 40,000 counts for
the spectrum to attain a 2σ precision of ±1%. It is important to note, however, that this is not necessarily the same as the empirical
formula, since not all elements are visible. Spectrometers with a beryllium window between the sample and the detector typically
cannot detect anything lighter than sodium. Spectrometers equipped with polymer based windows can quantify elements heavier
than beryllium. Either way, hydrogen cannot be observed by X-ray spectroscopy.
X-ray spectra are presented with energy in keV on the x-axis and the number of counts on the y-axis. The EDX spectra of biotite
and NIST glass K309 are shown as examples (Figure 1.12.5 and Figure 1.12.6 respectively). Biotite is a mineral similar to mica
which has the approximate chemical formula K(Mg,Fe)3AlSi3O10(F,OH)2. Strong peaks for manganese, aluminum, silicon,
potassium, and iron can be seen in the spectrum. The lack of visible hydrogen is expected, and the absence of oxygen and fluorine
peaks suggests the instrument had a beryllium window. The titanium peak is small and unexpected, so it may only be present in
trace amounts. K309 is a mix of glass developed by the National Institute for Standards and Technology. The spectrum shows that
it contains significant amounts of silicon, aluminum, calcium, oxygen, iron, and barium. The large peak at the far left is the carbon
signal from the carbon substrate the glass was placed on.

1.12.3 https://chem.libretexts.org/@go/page/55830
Figure 1.12.5 EDS spectrum of biotite. Silicon, aluminum, manganese, potassium, magnesium, iron, and titanium are all
identifiable, though titanium appears to be only a trace component. Adapted from J. Goodge, Energy-Dispersive X-ray
Spectroscopy (EDS), http://serc.carleton.edu/research_education/geochemsheets/eds.html.

Figure 1.12.6 EDS spectrum of NIST K309 glass. Silicon, aluminum, barium, calcium, iron, and oxygen are identifiable in the
spectrum. Adapted from J. Goldstein, D. Newbury, D. Joy, C. Lyman, P. Echlin, E. Lifshin, L.Sawyer, and J. Michael, Scanning
Electron Microscopy and X-ray Microanalysis, 3rd, Springer, New York (2003).

Limitations
As has just been discussed, X-ray spectroscopy is incapable of seeing elements lighter than boron. This is a problem given the
abundance of hydrogen in natural and man-made materials. The related techniques X-ray photoelectron spectroscopy (XPS) and
Auger spectroscopy are able to detect Li and Be, but are likewise unable to measure hydrogen.
X-ray spectroscopy relies heavily on standards for peak identification. Because a combination of elements can have noticeably
different properties from the individual constituent elements in terms of X-ray fluorescence or absorption, it is important to use a
standard as compositionally similar to the sample as possible. Naturally, this is more difficult to accomplish when examining new
materials, and there is always a risk of the structure of the sample being appreciably different than expected.
The energy-dispersive variants of X-ray spectroscopy sometimes have a hard time distinguishing between emissions that are very
near each other in energy or distinguishing peaks from trace elements from background noise. Fortunately, the wavelength-
dispersive variants are much better at both of these. The rough, stepwise curve in Figure 1.12.7 represents the EDS spectrum of
molybdenite, a mineral with the chemical formula MoS2. Broadened peaks make it difficult to distinguish the molybdenum signals
from the sulfur ones. Because WDS can select specific wavelengths, it has much better resolution and can pinpoint the separate
peaks more accurately. Similarly, the trace silicon signal in the EDS spectrum of the nickel-aluminum-manganese alloy in Figure
1.12.8a is barely distinguishable as a bump in the baseline, but the WDS spectrum in Figure 1.12.8b clearly picks it up.

Figure 1.12.7 A comparison of the EDS (yellow) and WDS spectra (light blue) of a sample of molybdenite. The sulfur and
molybdenum peaks are unresolved in the EDS spectrum, but are sharp and distinct in the WDS spectrum. Adapted from Oxford
Instruments, The power of WDS sensitivity and resolution, www.x-raymicroanalysis.com/x-ray-microanalysis-
explained/pages/detectors/wave1.htm.

1.12.4 https://chem.libretexts.org/@go/page/55830
Figure 1.12.8 (A) The EDS spectrum of an alloy comprised primarily of sodium, aluminum, and manganese. Silicon is a trace
element in the alloy, but is not discernible in the spectrum. (B) The WDS spectrum of the same alloy in the region around the
characteristic silicon peak. In this measurement, the silicon emission stands out quite clearly. Adapted from Oxford Instruments,
The power of WDS sensitivity and resolution, www.x-raymicroanalysis.com/x-ray-microanalysis-
explained/pages/detectors/wave1.htm.

1.12: An Introduction to Energy Dispersive X-ray Spectroscopy is shared under a CC BY 4.0 license and was authored, remixed, and/or curated
by Pavan M. V. Raja & Andrew R. Barron via source content that was edited to conform to the style and standards of the LibreTexts platform; a
detailed edit history is available upon request.

1.12.5 https://chem.libretexts.org/@go/page/55830
1.13: X-ray Photoelectron Spectroscopy
XPS Analysis of Modified Substances
Introduction
X-Ray photoelectron spectroscopy (XPS), also known as electron spectroscopy for chemical analysis (ESCA), is one of the most
widely used surface techniques in materials science and chemistry. It allows the determination of atomic composition of the sample
in a non-destructive manner, as well as other chemical information, such as binding constants, oxidation states and speciation. The
sample under study is subjected to irradiation by a high energy X-ray source. The X-rays penetrate only 5 – 20 Å into the sample,
allowing for surface specific, rather than bulk chemical, analysis. As an atom absorbs the X-rays, the energy of the X-ray will cause
a K-shell electron to be ejected, as illustrated by Figure 1.13.1. The K-shell is the lowest energy shell of the atom. The ejected
electron has a kinetic energy (KE) that is related to the energy of the incident beam (hν), the electron binding energy (BE), and the
work function of the spectrometer (φ) (1.13.1). Thus, the binding energy of the electron can be calculated.

Figure 1.13.1 Excitation of an electron from an atom's K-shell.


BE  =  hν   −  KE  −  ψs (1.13.1)

Table 1.13.1 shows the binding energy of the ejected electron, and the orbital from which the electron is ejected, which is
characteristic of each element. The number of electrons detected with a specific binding energy is proportional to the number of
corresponding atoms in the sample. This then provides the percent of each atom in the sample.
Table 1.13.1 Binding energies for select elements in their elemental forms.
Element Binding Energy (eV)

Carbon (C) (1s) 284.5 - 285.1

Nitrogen (N) (1s) 396.1 - 400.5

Oxygen (O) (1s) 526.2 - 533.5

Silicon (Si) (2p) 98.8 - 99.5

Sulfur (S) (2p3/2) 164.0 - 164.3

Iron (Fe) (2p3/2) 706.8 - 707.2

Gold (Au) (4f7/2) 83.8 - 84.2

The chemical environment and oxidation state of the atom can be determined through the shifts of the peaks within the range
expected (Table 1.13.2). If the electrons are shielded then it is easier, or requires less energy, to remove them from the atom, i.e.,
the binding energy is low. The corresponding peaks will shift to a lower energy in the expected range. If the core electrons are not
shielded as much, such as the atom being in a high oxidation state, then just the opposite occurs. Similar effects occur with
electronegative or electropositive elements in the chemical environment of the atom in question. By synthesizing compounds with
known structures, patterns can be formed by using XPS and structures of unknown compounds can be determined.
Table 1.13.2 Binding energies of electrons in various compounds.
Compound Binding Energy (eV)

COH (C 1s) 286.01 - 286.8

CHF (C 1s) 287.5 - 290.2

Nitride (N 1s) 396.2 - 398.3

1.13.1 https://chem.libretexts.org/@go/page/55831
Compound Binding Energy (eV)

Fe2O3 (from O, 1s) 529.5 - 530.2

Fe2O3 (from Fe, 2p3/2) 710.7 - 710.9

FeO (from Fe 2p3/2) 709.1 - 709.5

SiO2 (from O, 2s) 532.5 - 533.3

SiO2 (from Si, 2p) 103.2 - 103.9

Sample preparation is important for XPS. Although the technique was originally developed for use with thin, flat films, XPS can be
used with powders. In order to use XPS with powders, a different method of sample preparation is required. One of the more
common methods is to press the powder into a high purity indium foil. A different approach is to dissolve the powder in a quickly
evaporating solvent, if possible, which can then be drop-casted onto a substrate. Using sticky carbon tape to adhere the powder to a
disc or pressing the sample into a tablet are an option as well. Each of these sample preparations are designed to make the powder
compact, as powder not attached to the substrate will contaminate the vacuum chamber. The sample also needs to be completely
dry. If it is not, solvent present in the sample can destroy the necessary high vacuum and contaminate the machine, affecting the
data of the current and future samples.
Analyzing Functionalized Surfaces

Depth Pro ling


When analyzing a sample (Figure 1.13.2 a) by XPS, questions often arise that deal with layers of the sample. For example, is the
sample homogenous, with a consistent composition throughout, or layered, with certain elements or components residing in
specific places in the sample? (Figure 1.13.2 b,c). A simple way to determine the answer to this question is to perform a depth
analysis. By sputtering away the sample, data can be collected at different depths within the sample. It should be noted that
sputtering is a destructive process. Within the XPS instrument, the sample is subjected to an Ar+ ion beam that etches the surface.
This creates a hole in the surface, allowing the X-rays to hit layers that would not have otherwise been analyzed. However, it
should be realized that different surfaces and layers may be etched at different rates, meaning the same amount of etching does not
occur during the same amount of time, depending on the element or compound currently being sputtered.

Figure 1.13.2 Schematic representation of analysis of (a) an homogeneous sample, as compared to (b) an homogeneous layers in a
sample, and (c) an inhomogeneous layers in a sample.
It is important to note that hydrocarbons sputter very easily and can contaminate the high vacuum of the XPS instrument and thus
later samples. They can also migrate to a recently sputtered (and hence unfunctionalized) surface after a short amount of time, so it
is imperative to sputter and take a measurement quickly, otherwise the sputtering may appear to have had no effect.

Functionalized Films
When running XPS, it is important that the sample is prepared correctly. If it is not, there is a high chance of ruining not only data
acquisition, but the instrument as well. With organic functionalization, it is very important to ensure the surface functional group
(or as is the case with many functionalized nanoparticles, the surfactant) is immobile on the surface of the substrate. If it is removed
easily in the vacuum chamber, it not only will give erroneous data, but it will contaminate the machine, which may then

1.13.2 https://chem.libretexts.org/@go/page/55831
contaminate future samples. This is particularly important when studying thiol functionalization of gold samples, as thiol groups
bond strongly with the gold. If there is any loose thiol group contaminating the machine, the thiol will attach itself to any gold
sample subsequently placed in the instrument, providing erroneous data. Fortunately, with the above exception, preparing samples
that have been functionalized is not much different than standard preparation procedures. However, methods for analysis may have
to be modified in order to obtain good, consistent data.
A common method for the analysis of surface modified material is angle resolved X-ray photoelectron spectroscopy (ARXPS).
ARXPS is a non-destructive alternative to sputtering, as it relies upon using a series of small angles to analyze the top layer of the
sample, giving a better picture of the surface than standard XPS. ARXPS allows for the analysis of the topmost layer of atoms to be
analyzed, as opposed to standard XPS, which will analyze a few layers of atoms into the sample, as illustrated in Figure 1.13.3.
ARXPS is often used to analyze surface contaminations, such as oxidation, and surface modification or passivation. Though the
methodology and limitations are beyond the scope of this module, it is important to remember that, like normal XPS, ARXPS
assumes homogeneous layers are present in samples, which can give erroneous data, should the layers be heterogeneous.

Figure 1.13.3 Schematic representation of (a) a standard XPS analysis and (b) ARXPS on a multilayer sample.
Limitations of XPS
There are many limitations to XPS that are not based on the samples or preparation, but on the machine itself. One such limitation
is that XPS cannot detect hydrogen or helium. This, of course, leads to a ratio of elements in the sample that is not entirely
accurate, as there is always some amount of hydrogen. It is a common fallacy to assume the percent of atoms obtained from XPS
data are completely accurate due to this presence of undetected hydrogen (Table 1.13.1).
It is possible to indirectly measure the amount of hydrogen in a sample using XPS, but it is not very accurate and has to be done in
a roundabout, often time consuming manner. If the sample contains hydrogen with a partial positive charge (i.e. OH), the sample
can be washed in sodium naphthalenide (C10H8Na). This replaces this hydrogen with sodium, which can then be measured. The
sodium to oxygen ratio that is obtained infers the hydrogen to oxygen ratio, assuming that all the hydrogen atoms have reacted.
XPS can only give an average measurement, as the electrons lower down in the sample will lose more energy as they pass other
atoms while the electrons on the surface retain their original kinetic energy. The electrons from lower layers can also undergo
inelastic or elastic scattering, seen in Figure 1.13.4. This scattering may have a significant impact on data at higher angles of
emission. The beam itself is also relatively wide, with the smallest width ranging from 10 – 200 μm, lending to the observed
average composition inside the beam area. Due to this, XPS cannot differentiate sections of elements if the sections are smaller
than the size of the beam.

Figure 1.13.4 : Schematic representation of (a) no scattering, (b) inelastic scattering, and (c) elastic scattering.
Sample reaction or degredation are important considerations. Caution should be exercised when analyzing polymers, as they are
often chemically active and X-rays will provide energy to start degrading the polymer, altering the properties of the sample. One
method found to help overcome this particular limitation is to use angle-resolved X-ray photoelectron spectroscopy (ARXPS). XPS
can often reduce certain metal salts, such as Cu2+. This reduction will give peaks that indicate a certain set of properties or
chemical environments when it could be completely different. It needs to be understood that charges can build up on the surface of
the sample due to a number of reasons, specifically due to the loss of electrons during the XPS experiment. The charge on the
surface will interact with the electrons escaping from the sample, affecting the data obtained. If the charge collecting is positive, the
electrons that have been knocked off will be attracted to the charge, slowing the electrons. The detector will pick up a lower kinetic

1.13.3 https://chem.libretexts.org/@go/page/55831
energy of the electrons, and thus calculate a different binding energy than the one expected, giving peaks which could be labeled
with an incorrect oxidation state or chemical environment. To overcome this, the spectra must be charge referenced by one of the
following methods: using the naturally occurring graphite peak as a reference, sputtering with gold and using the gold peak as a
reference or flooding the sample with the ion gun and waiting until the desired peak stops shifting.

Limitations with Surfactants and Sputtering


While it is known that sputtering is destructive, there are a few other limitations that are not often considered. As mentioned above,
the beam of X-rays is relatively large, giving an average composition in the analysis. Sputtering has the same limitation. If the
surfactant or layers are not homogeneous, then when the sputtering is finished and detection begins, the analysis will show a
homogeneous section, due to the size of both the beam and sputtered area, while it is actually separate sections of elements.
The chemistry of the compounds can be changed with sputtering, as it removes atoms that were bonded, changing the oxidation
state of a metal or the hybridization of a non-metal. It can also introduce charges if the sample is non-conducting or supported on a
non-conducting surface.

Using XPS to Analyze Metal Nanoparticles


Introduction
X-ray photoelectron spectroscopy (XPS) is a surface technique developed for use with thin films. More recently, however, it has
been used to analyze the chemical and elemental composition of nanoparticles. The complication of nanoparticles is that they are
neither flat nor larger than the diameter of the beam, creating issues when using the data obtained at face value. Samples of
nanoparticles will often be large aggregates of particles. This creates problems with the analysis acquisition, as there can be a
variety of cross-sections, as seen in Figure 1.13.5. This acquisition problem is also compounded by the fact that the surfactant may
not be completely covering the particle, as the curvature of the particle creates defects and divots. Even if it is possible to create a
monolayer of particles on a support, other issues are still present. The background support will be analyzed with the particle, due to
their small size and the size of the beam and the depth at which it can penetrate.

Figure 1.13.5 Different cross-sections of analysis possible on a nanoparticle.


Many other factors can introduce changes in nanoparticles and their properties. There can be probe, environmental, proximity, and
sample preparation effects. The dynamics of particles can wildly vary depending on the reactivity of the particle itself. Sputtering
can also be a problem. The beam used to sputter will be roughly the same size or larger than the particles. This means that what
appears in the data is not a section of particle, but an average composition of several particles.
Each of these issues needs to be taken into account and preventative measures need to be used so the data is the best representation
possible.
Sample Preparation

Sample preparation of nanoparticles is very important when using XPS. Certain particles, such as iron oxides without surfactants,
will interact readily with oxygen in the air. This causes the particles to gain a layer of oxygen contamination. When the particles are
then analyzed, oxygen appears where it should not and the oxidation state of the metal may be changed. As shown by these
particles, which call for handling, mounting and analysis without exposure to air, knowing the reactivity of the nanoparticles in the
sample is very important even before starting analysis. If the reactivity of the nanoparticle is known, such as the reactivity of
oxygen and iron, then preventative steps can be taken in sample preparation in order to obtain the best analysis possible.
When preparing a sample for XPS, a powder form is often used. This preparation, however, will lead to aggregation of
nanoparticles. If analysis is performed on such a sample, the data obtained will be an average of composition of each nanoparticle.

1.13.4 https://chem.libretexts.org/@go/page/55831
If composition of a single particle is what is desired, then this average composition will not be sufficient. Fortunately, there are
other methods of sample preparation. Samples can be supported on a substrate, which will allow for analysis of single particles. A
pictorial representation in Figure 1.13.6 shows the different types of samples that can occur with nanoparticles.

Figure 1.13.6 Representation of (a) a theoretical isolated nanoparticles, (b) nanoparticles suspended on a substrate, (c) an aggregate
of nanoparticles, and (d) a powdered form of nanoparticles.
Analysis Limitations
Nanoparticles are dynamic; their properties can change when exposed to new chemical environments, leading to a new set of
applications. It is the dynamics of nanoparticles that makes them so useful and is one of the reasons why scientists strive to
understand their properties. However, it is this dynamic ability that makes analysis difficult to do properly. Nanoparticles are easily
damaged and can change properties over time or with exposure to air, light or any other environment, chemical or otherwise.
Surface analysis is often difficult because of the high rate of contamination. Once the particles are inserted into XPS, even more
limitations appear.

Probe Effects
There are often artifacts introduced from the simple mechanism of conducting the analysis. When XPS is used to analyze the
relatively large surface of thin films, there is small change in temperature as energy is transferred. The thin films, however, are
large enough that this small change in energy has to significant change to its properties. A nanoparticle is much smaller. Even a
small amount of energy can drastically change the shape of particles, in turn changing the properties, giving a much different set of
data than expected.
The electron beam itself can affect how the particles are supported on a substrate. Theoretically, nanoparticles would be considered
separate from each other and any other chemical environments, such as solvents or substrates. This, however, is not possible, as the
particles must be suspended in a solution or placed on a substrate when attempting analysis. The chemical environment around the
particle will have some amount of interaction with the particle. This interaction will change characteristics of the nanoparticles,
such as oxidation states or partial charges, which will then shift the peaks observed. If particles can be separated and suspended on
a substrate, the supporting material will also be analyzed due to the fact that the X-ray beam is larger than the size of each
individual particle. If the substrate is made of porous materials, it can adsorb gases and those will be detected along with the
substrate and the particle, giving erroneous data.

Environmental Effects
Nanoparticles will often react, or at least interact, with their environments. If the particles are highly reactive, there will often be
induced charges in the near environment of the particle. Gold nanoparticles have a well-documented ability to undergo plasmon
interactions with each other. When XPS is performed on these particles, the charges will change the kinetic energy of the electrons,
shifting the apparent binding energy. When working with nanoparticles that are well known for creating charges, it is often best to
use an ion gun or a coating of gold. The purpose of the ion gun or gold coating is to try to move peaks back to their appropriate
energies. If the peaks do not move, then the chance of there being no induced charge is high and thus the obtained data is fairly
reliable.

1.13.5 https://chem.libretexts.org/@go/page/55831
Proximity Effects
The proximity of the particles to each other will cause interactions between the particles. If there is a charge accumulation near one
particle, and that particle is in close proximity with other particles, the charge will become enhanced as it spreads, affecting the
signal strength and the binding energies of the electrons. While the knowledge of charge enhancement could be useful to potential
applications, it is not beneficial if knowledge of the various properties of individual particles is sought.
Less isolated (i.e., less crowded) particles will have different properties as compared to more isolated particles. A good example of
this is the plasmon effect in gold nanoparticles. The closer gold nanoparticles are to each other, the more likely they will induce the
plasmon effect. This can change the properties of the particles, such as oxidation states and partial charges. These changes will then
shift peaks seen in XPS spectra. These proximity effects are often introduced in the sample preparation. This, of course, shows why
it is important to prepare samples correctly to get desired results.

Conclusions
Unfortunately there is no good general procedure for all nanoparticles samples. There are too many variables within each sample to
create a basic procedure. A scientist wanting to use XPS to analyze nanoparticles must first understand the drawbacks and
limitations of using their sample as well as how to counteract the artifacts that will be introduced in order to properly use XPS.
One must never make the assumption that nanoparticles are flat. This assumption will only lead to a misrepresentation of the
particles. Once the curvature and stacking of the particles, as well as their interactions with each other are taken into account, XPS
can be run.

1.13: X-ray Photoelectron Spectroscopy is shared under a CC BY 4.0 license and was authored, remixed, and/or curated by Pavan M. V. Raja &
Andrew R. Barron via source content that was edited to conform to the style and standards of the LibreTexts platform; a detailed edit history is
available upon request.

1.13.6 https://chem.libretexts.org/@go/page/55831
1.14: Auger Electron Spectroscopy
Basic Principles
Auger electron spectroscopy (AES) is one of the most commonly employed surface analysis techniques. It uses the energy of
emitted electrons to identify the elements present in a sample, similar to X-ray photoelectron spectroscopy (XPS). The main
difference is that XPS uses an X-ray beam to eject an electron while AES uses an electron beam to eject an electron. In AES, the
sample depth is dependent on the escape energy of the electrons. It is not a function of the excitation source as in XPS. In AES, the
collection depth is limited to 1-5 nm due to the small escape depth of electrons, which permits analysis of the first 2 - 10 atomic
layers. In addition, a typical analysis spot size is roughly 10 nm. A representative AES spectrum illustrating the number of emitted
electrons, N, as a function of kinetic energy, E, in direct form (red) and in differentiated form (black) is shown in Figure 1.14.1.

Figure 1.14.1 AES survey spectrum (red) and differentiated spectrum (black) of an oxidized Fe-Cr-Nb alloy. Adapted from H. J.
Mathieu in Surface Analysis: The Principal Techniques, 2nd Edition, Ed. J. C. Vickerman, Wiley-VCH, Weinheim (2011).
Like XPS, AES measures the kinetic energy (Ek) of an electron to determine its binding energy (Eb). The binding energy is
inversely proportional to the kinetic energy and can be found from 1.14.1, where hν is the energy of the incident photon and ΔΦ is
the difference in work function between the sample and the detector material.

Eb   =  hν   −  Ek   +  ΔΦ (1.14.1)

Since the Eb is dependent on the element and the electronic environment of the nucleus, AES can be used to distinguish elements
and their oxidation states. For instance, the energy required to remove an electron from Fe3+ is more than in Fe0. Therefore, the
Fe3+ peak will have a lower Ek than the Fe0 peak, effectively distinguishing the oxidation states.

Auger Process
An Auger electron comes from a cascade of events. First, an electron beam comes in with sufficient energy to eject a core electron
creating a vacancy (see Figure 1.14.2a). Typical energies of the primary electrons range from 3 - 30 keV. A secondary electron
(imaging electron) of higher energy drops down to fill the vacancy (see Figure 1.14.2 b) and emits sufficient energy to eject a
tertiary electron (Auger electron) from a higher shell (see Figure 1.14.2 c).

Figure 1.14.2 Schematic diagram of the Auger process.


The shells from which the electrons move from lowest to highest energy are described as the K shell, L shell, and M shell. This
nomenclature is related to quantum numbers. Explicitly, the K shell represents the 1s orbital, the L shell represents the 2s and 2p
orbitals, and the M shell represents the 3s, 3p, and 3d orbitals. The cascade of events typically begins with the ionization of a K
shell electron, followed by the movement of an L shell electron into the K shell vacancy. Then, either an L shell electron or M shell
electron is ejected. It depends on the element, which peak is prevalent but often both peaks will be present. The peak seen in the
spectrum is labeled according to the shells involved in the movement of the electrons. For example, an electron ejected from a gold
atom could be labeled as Au KLL or Au KLM.
The intensity of the peak depends on the amount of material present, while the peak position is element dependent. Auger
transitions characteristic of each elements can be found in the literature. Auger transitions of the first forty detectable elements are

1.14.1 https://chem.libretexts.org/@go/page/55832
listed in Table 1.14.1.
Table 1.14.1 Selected AES transitions and their corresponding kinetic energy. Adapted from H. J. Mathieu in Surface Analysis: The Principal
Techniques, Second Edition, Ed. J. C. Vickerman, Wiley-VCH, Weinheim (2011).
Kinetic Energy of Transition
Atomic Number Element AES transition
(eV)

3 Li KLL 43

4 Be KLL 104

5 B KLL 179

6 C KLL 272

7 N KLL 379

8 O KLL 508

9 F KLL 647

11 Na KLL 990

12 Mg KLL 1186

13 Al LMM 68

14 Si LMM 92

15 P LMM 120

16 S LMM 152

17 Cl LMM 181

19 K KLL 252

20 Ca LMM 291

21 Sc LMM 340

22 Ti LMM 418

23 V LMM 473

24 Cr LMM 529

25 Mn LMM 589

26 Fe LMM 703

27 Co LMM 775

28 Ni LMM 848

29 Cu LMM 920

30 Zn LMM 994

31 Ga LMM 1070

32 Ge LMM 1147

33 As LMM 1228

34 Se LMM 1315

35 Br LMM 1376

39 Y MNN 127

40 Zr MNN 147

41 Nb MNN 167

42 Mo MNN 186

1.14.2 https://chem.libretexts.org/@go/page/55832
Instrumentation
Important elements of an Auger spectrometer include a vacuum system, an electron source, and a detector. AES must be performed
at pressures less than 10-3 pascal (Pa) to keep residual gases from adsorbing to the sample surface. This can be achieved using an
ultra-high-vacuum system with pressures from 10-8 to 10-9 Pa. Typical electron sources include tungsten filaments with an electron
beam diameter of 3 - 5 μm, LaB6 electron sources with a beam diameter of less than 40 nm, and Schottky barrier filaments with a
20 nm beam diameter and high beam current density. Two common detectors are the cylindrical mirror analyzer and the concentric
hemispherical analyzer discussed below. Notably, concentric hemispherical analyzers typically have better energy resolution.
Cylindrical Mirror Analyzer (CMA)
A CMA is composed of an electron gun, two cylinders, and an electron detector (Figure 1.14.2). The operation of a CMA involves
an electron gun being directed at the sample. An ejected electron then enters the space between the inner and outer cylinders (IC
and OC). The inner cylinder is at ground potential, while the outer cylinder’s potential is proportional to the kinetic energy of the
electron. Due to its negative potential, the outer cylinder deflects the electron towards the electron detector. Only electrons within
the solid angle cone are detected. The resulting signal is proportional to the number of electrons detected as a function of kinetic
energy.

Figure 1.14.3 Schematic of a cylindrical mirror analyzer.


Concentric Hemispherical Analyzer (CHA)
A CHA contains three parts (Figure 1.14.4):
1. A retarding and focusing input lens assembly
2. An inner and outer hemisphere (IH and OH)
3. An electron detector

Figure 1.14.4 Schematic of a concentric hemispherical analyzer.


Electrons ejected from the surface enter the input lens, which focuses the electrons and retards their energy for better resolution.
Electrons then enter the hemispheres through an entrance slit. A potential difference is applied on the hemispheres so that only
electrons with a small range of energy differences reach the exit. Finally, an electron detector analyzes the electrons.

Applications
AES has widespread use owing to its ability to analyze small spot sizes with diameters from 5 μm down to 10 nm depending on the
electron gun. For instance, AES is commonly employed to study film growth and surface-chemical composition, as well as grain
boundaries in metals and ceramics. It is also used for quality control surface analyses in integrated circuit production lines due to
short acquisition times. Moreover, AES is used for areas that require high spatial resolution, which XPS cannot achieve. AES can
also be used in conjunction with transmission electron microscopy (TEM) and scanning electron microscopy (SEM) to obtain a
comprehensive understanding of microscale materials, both chemically and structurally. As an example of combining techniques to

1.14.3 https://chem.libretexts.org/@go/page/55832
investigate microscale materials, Figure 1.14.5 shows the characterization of a single wire from a Sn-Nb multi-wire alloy. Figure
1.14.5 a is a SEM image of the singular wire and Figure 1.14.5 b is a schematic depicting the distribution of Nb and Sn within the

wire. Point analysis was performed along the length of the wire to determine the percent concentrations of Nb and Sn.

Figure 1.14.5 Analysis of a Sn-Nb wire. (a) SEM image of the wire, (b) schematic of the elemental distribution, and (c) graphical
representation of point analysis giving the percent concentration of Nb and Sn. Adapted from H. J. Mathieu in Surface Analysis:
The Principal Techniques, Second Edition, Ed. J. C. Vickerman, Wiley-VCH, Weinheim (2011).
AES is widely used for depth profiling. Depth profiling allows the elemental distributions of layered samples 0.2 – 1 μm thick to be
characterized beyond the escape depth limit of an electron. Varying the incident and collection angles, and the primary beam energy
controls the analysis depth. In general, the depth resolution decreases with the square root of the sample thickness. Notably, in
AES, it is possible to simultaneously sputter and collect Auger data for depth profiling. The sputtering time indicates the depth and
the intensity indicates elemental concentrations. Since, the sputtering process does not affect the ejection of the Auger electron,
helium or argon ions can be used to sputter the surface and create the trench, while collecting Auger data at the same time. The
depth profile does not have the problem of diffusion of hydrocarbons into the trenches. Thus, AES is better for depth profiles of
reactive metals (e.g., gold or any metal or semiconductor). Yet, care should be taken because sputtering can mix up different
elements, changing the sample composition.

Limitations
While AES is a very valuable surface analysis technique, there are limitations. Because AES is a three-electron process, elements
with less than three electrons cannot be analyzed. Therefore, hydrogen and helium cannot be detected. Nonetheless, detection is
better for lighter elements with fewer transitions. The numerous transition peaks in heavier elements can cause peak overlap, as can
the increased peak width of higher energy transitions. Detection limits of AES include 0.1 – 1% of a monolayer, 10-16 – 10-15 g of
material, and 1012 – 1013 atoms/cm2.
Another limitation is sample destruction. Although focusing of the electron beam can improve resolution; the high-energy electrons
can destroy the sample. To limit destruction, beam current densities of greater than 1 mA/cm2 should be used. Furthermore,
charging of the electron beam on insulating samples can deteriorate the sample and result in high-energy peak shifts or the
appearance of large peaks.

1.14: Auger Electron Spectroscopy is shared under a CC BY 4.0 license and was authored, remixed, and/or curated by Pavan M. V. Raja &
Andrew R. Barron via source content that was edited to conform to the style and standards of the LibreTexts platform; a detailed edit history is
available upon request.

1.14.4 https://chem.libretexts.org/@go/page/55832
1.15: Rutherford Backscattering of Thin Films
Introduction
One of the main research interests of the semiconductor industry is to improve the performance of semiconducting devices and to
construct new materials with reduced size or thickness that have potential application in transistors and microelectronic devices.
However, the most significant challenge regarding thin film semiconductor materials is measurement. Properties such as the
thickness, composition at the surface, and contamination, all are critical parameters of the thin films. To address these issues, we
need an analytical technique which can measure accurately through the depth of the of the semiconductor surface without
destruction of the material. Rutherford backscattering spectroscopy is a unique analysis method for this purpose. It can give us
information regarding in-depth profiling in a non-destructive manner. However X-ray photo electron spectroscopy (XPS), energy
dispersive X-ray analysis (EDX) and Auger electron spectroscopy are also able to study the depth-profile of semiconductor films.
Table 1.15.1 demonstrates the comparison between those techniques with RBS.
Table 1.15.1 Comparison between different thin film analysis techniques.
Method Destructive Incident Particle Outgoing Particle Detection Limit Depth Resolution

RBS No Ion Ion ~1 10 nm

XPS Yes X-ray photon Electron ~0.1-1 ~1 µm

EDX Yes Electron X-ray photon ~0.1 1.5 µm

Auger Yes Electron Electron ~0.1-1 1.5 nm

Basic Concept of Rutherford Backscattering Spectroscopy


At a basic level, RBS demonstrates the electrostatic repulsion between high energy incident ions and target nuclei. The specimen
under study is bombarded with monoenergetic beam of 4He+ particles and the backscattered particles are detected by the detector-
analysis system which measures the energies of the particles. During the collision, energy is transferred from the incident particle to
the target specimen atoms; the change in energy of the scattered particle depends on the masses of incoming and target atoms. For
an incident particle of mass M1, the energy is E0 while the mass of the target atom is M2. After the collision, the residual energy E
of the particle scattered at angle Ø can be expressed as:
2
E  =  k E0 (1.15.1)

−−−−−−−−−−−−−
2 2
(M1  cos(θ)  +  √ M   −  M si n2 θ )
2 1

 k  =   (1.15.2)
M1   +  M2

where k is the kinematic scattering factor, which is actually the energy ratio of the particle before and after the collision. Since k
depends on the masses of the incident particle and target atom and the scattering angle, the energy of the scattered particle is also
determined by these three parameters. A simplified layout of backscattering experiment is shown in Figure 1.15.1.

Figure 1.15.1 Schematic representation of the experimental setup for Rutherford backscattering analysis.
The probability of a scattering event can be described by the differential scattering cross section of a target atom for scattering an
incoming particle through the angle Ø into differential solid angle as follows,

1.15.1 https://chem.libretexts.org/@go/page/55833
−−−−−−−−−−−−
M1 2
2
[cosθ  +  √ 1  −  ( sinθ) ]
dσR zZe2 M2
  =( )  =   (1.15.3)
−−−−−−−−−−−−
dϕ 2 E0 sin(2θ) M1
√ 1  −  ( sinθ)2
M2

where dσR is the effective differential cross section for the scattering of a particle. The above equation may looks complicated but it
conveys the message that the probability of scattering event can be expressed as a function of scattering cross section which is
proportional to the zZ when a particle with charge ze approaches the target atom with charge Ze.
Helium ions not scattered at the surface lose energy as they traverse the solid. They lose energy due to interaction with electrons in
the target. After collision the He particles lose further energy on their way out to the detector. We need to know two quantities to
measure the energy loss, the distance Δt that the particles penetrate into the target and the energy loss ΔE in this distance Figure
1.15.2. The rate of energy loss or stopping power is a critical component in backscattering experiments as it determines the depth

profile in a given experiment.

Figure 1.15.2 Components of energy loss for a ion beam that scatters from depth t. First, incident beam loses energy through
interaction with electrons ΔEin. Then energy lost occurs due to scattering Ec. Finally outgoing beam loses energy for interaction
with electrons ΔEout. Adapted from L. C. Feldman and J. W. Mayer, Fundamentals of Surface and Thin Film Analysis , North
Holland-Elsevier, New York (1986).
In thin film analysis, it is convenient to assume that total energy loss ΔE into depth t is only proportional to t for a given target. This
assumption allows a simple derivation of energy loss in backscattering as more complete analysis requires many numerical
techniques. In constant dE/dx approximation, total energy loss becomes linearly related to depth t, Figure 1.15.3.

Figure 1.15.3 Variation of energy loss with the depth of the target in constant dE/dx approximation.

Experimental Set-up
The apparatus for Rutherford backscattering analysis of thin solid surface typically consist of three components:
1. A source of helium ions.
2. An accelerator to energize the helium ions.
3. A detector to measure the energy of scattered ions.

1.15.2 https://chem.libretexts.org/@go/page/55833
There are two types of accelerator/ion source available. In single stage accelerator, the He+ source is placed within an insulating
gas-filled tank (Figure 1.15.4). It is difficult to install new ion source when it is exhausted in this type of accelerator. Moreover, it
is also difficult to achieve particles with energy much more than 1 MeV since it is difficult to apply high voltages in this type of
system.

Figure 1.15.4 Schematic representation of a single stage accelerator.


Another variation is “tandem accelerator.” Here the ion source is at ground and produces negative ion. The positive terminal is
located is at the center of the acceleration tube (Figure 1.15.5). Initially the negative ion is accelerated from ground to terminal. At
terminal two-electron stripping process converts the He- to He++. The positive ions are further accelerated toward ground due to
columbic repulsion from positive terminal. This arrangement can achieve highly accelerated He++ ions (~ 2.25 MeV) with
moderate voltage of 750 kV.

Figure 1.15.5 Schematic representation of a tandem accelerator.


Particles that are backscattered by surface atoms of the bombarded specimen are detected by a surface barrier detector. The surface
barrier detector is a thin layer of p-type silicon on the n-type substrate resulting p-n junction. When the scattered ions exchange
energy with the electrons on the surface of the detector upon reaching the detector, electrons get promoted from the valence band to
the conduction band. Thus, each exchange of energy creates electron-hole pairs. The energy of scattered ions is detected by simply
counting the number of electron-hole pairs. The energy resolution of the surface barrier detector in a standard RBS experiment is
12 - 20 keV. The surface barrier detector is generally set between 90° and 170° to the incident beam. Films are usually set normal to
the incident beam. A simple layout is shown in Figure 1.15.6.

Figure 1.15.6 Schematic representation general setup where the surface barrier detector is placed at angle of 165° to the
extrapolated incident beam.

Depth Profile Analysis


As stated earlier, it is a good approximation in thin film analysis that the total energy loss ΔE is proportional to depth t. With this
approximation, we can derive the relation between energy width ΔE of the signal from a film of thickness Δt as follows,

1.15.3 https://chem.libretexts.org/@go/page/55833
dE 1 dE
ΔE  =  Δt(k   +    ) (1.15.4)
dxin cos θ dxout

where Ø = lab scattering angle.


It is worth noting that k is the kinematic factor defined in equation above and the subscripts “in” and “out” indicate the energies at
which the rate of loss of energy or dE/dx is evaluated. As an example, we consider the backscattering spectrum, at scattering angle
170°, for 2 MeV He++ incidents on silicon layer deposited onto 2 mm thick niobium substrate Figure 1.15.7.

Figure 1.15.7 The backscattering spectrum for 2.0 MeV He ions incident on a silicon thin film deposited onto a niobium substrate.
Adapted from P. D. Stupik, M. M. Donovan, A. R. Barron, T. R. Jervis and M. Nastasi, Thin Solid Films, 1992, 207, 138.
The energy loss rate of incoming He++ or dE/dx along inward path in elemental Si is ≈24.6 eV/Å at 2 MeV and is ≈26 eV/Å for the
outgoing particle at 1.12 MeV (Since K of Si is 0.56 when the scattering angle is 170°, energy of the outgoing particle would be
equal to 2 x 0.56 or 1.12 MeV) . Again the value of ΔESi is ≈133.3 keV. Putting the values into above equation we get
133.6 keV
Δt ≈ (1.15.5)
eV 1 eV
(0.56 ∗ 24.6  )  +  ( ∘
  ∗  26  )
Å cos 170 Å

133.3 keV
=  (1.15.6)
13.77 eV / Å  +  29.985 eV / Å

133.3 keV
=  (1.15.7)
40.17eV / Å

=  3318  Å (1.15.8)

Hence a Si layer of ca. 3300 Å thickness has been deposited on the niobium substrate. However we need to remember that the
value of dE/dx is approximated in this calculation.

Quantitative Analysis
In addition to depth profile analysis, we can study the composition of an element quantitatively by backscattering spectroscopy.
The basic equation for quantitative analysis is

Y   =  σΩQN Δt (1.15.9)

Where Y is the yield of scattered ions from a thin layer of thickness Δt, Q is the number of incident ions and Ω is the detector solid
angle, and NΔt is the number of specimen atoms (atom/cm2). Figure 1.15.8 shows the RBS spectrum for a sample of silicon
deposited on a niobium substrate and subjected to laser mixing. The Nb has reacted with the silicon to form a NbSi2 interphase
layer. The Nb signal has broadened after the reaction as show in Figure 1.15.8.
We can use ratio of the heights HSi/HNb of the backscattering spectrum after formation of NbSi2 to determine the composition of
the silicide layer. The stoichiometric ratio of Nb and Si can be approximated as,

1.15.4 https://chem.libretexts.org/@go/page/55833
NSi [ HSi   ∗  σSi ]
  ≈ (1.15.10)
NN b [ HN b   ∗  σN b ]

Hence the concentration of Si and Nb can be determined if we can know the appropriate cross sections σSiand σNb. However the
yield in the backscattering spectra is better represented as the product of signal height and the energy width ΔE. Thus
stoichiometric ratio can be better approximated as

NSi [ HSi   ∗  ΔESi   ∗  σSi ]


  ≈ (1.15.11)
NN b [ HN b   ∗  ΔEN b   ∗   σN b ]

Limitations
It is of interest to understand the limitations of the backscattering technique in terms of the comparison with other thin film analysis
technique such as AES, XPS and SIMS (Table 1.15.1). AES has better mass resolution, lateral resolution and depth resolution than
RBS. But AES suffers from sputtering artifacts. Compared to RBS, SIMS has better sensitivity. RBS does not provide any
chemical bonding information which we can get from XPS. Again, sputtering artifact problems are also associated in XPS. The
strength of RBS lies in quantitative analysis. However, conventional RBS systems cannot analyze ultrathin films since the depth
resolution is only about 10 nm using surface barrier detector.

Summary
Rutherford Backscattering analysis is a straightforward technique to determine the thickness and composition of thin films (< 4000
Å). Areas that have been lately explored are the use of backscattering technique in composition determination of new
superconductor oxides; analysis of lattice mismatched epitaxial layers, and as a probe of thin film morphology and surface
clustering.

1.15: Rutherford Backscattering of Thin Films is shared under a CC BY 4.0 license and was authored, remixed, and/or curated by Pavan M. V.
Raja & Andrew R. Barron via source content that was edited to conform to the style and standards of the LibreTexts platform; a detailed edit
history is available upon request.

1.15.5 https://chem.libretexts.org/@go/page/55833
1.16: An Accuracy Assessment of the Refinement of Crystallographic Positional
Metal Disorder in Molecular Solid Solutions
Introduction
Crystallographic positional disorder is evident when a position in the lattice is occupied by two or more atoms; the average of
which constitutes the bulk composition of the crystal. If a particular atom occupies a certain position in one unit cell and another
atom occupies the same position in other unit cells, the resulting electron density will be a weight average of the situation in all the
unit cells throughout the crystal. Since the diffraction experiment involves the average of a very large number of unit cells (ca.
1018 in a crystal used for single crystal X-ray diffraction analysis), minor static displacements of atoms closely simulate the effects
of vibrations on the scattering power of the “average” atom. Unfortunately, the determination of the “average” atom in a crystal
may be complicated if positional disorder is encountered.
Crystal disorder involving groups such as CO, CN and Cl have been documented to create problems in assigning the correct
structure through refinement procedures. While attempts have been made to correlate crystallographic lattice parameters with bulk
chemical composition of the solution from which single crystal was grown, there has been little effort to correlate crystallographic
site occupancy with chemical composition of the crystal from which single crystal diffraction data was obtained. These are two
very different issues that must be considered when solving a crystal structure with site occupancy disorder.
What is the relationship of a single crystal to the bulk material?
Is the refinement of a site-occupancy-factor actually gives a realistic value for % occupancy when compared to the "actual"
% composition for that particular single crystal?
The following represents a description of a series of methods for the refinement of a site occupancy disorder between two atoms
(e.g., two metal atoms within a mixture of isostructural compounds).

Methods for X-ray Diffraction Determination of Positional Disorder in Molecular Solid Solutions
An atom in a structure is defined by several parameters: the type of atom, the positional coordinates (x, y, z), the occupancy factor
(how many “atoms” are at that position) and atomic displacement parameters (often called temperature or thermal parameters). The
latter can be thought of as being a “picture” of the volume occupied by the atom over all the unit cells, and can be isotropic (1
parameter defining a spherical volume) or anisotropic (6 parameters defining an ellipsoidal volume). For a “normal” atom, the
occupancy factor is fixed as being equal to one, and the positions and displacement parameters are “refined” using least-squares
methods to values in which the best agreement with the observed data is obtained. In crystals with site-disorder, one position is
occupied by different atoms in different unit cells. This refinement requires a more complicated approach. Two broad methods may
be used: either a new atom type that is the appropriate combination of the different atoms is defined, or the same positional
parameters are used for different atoms in the model, each of which has occupancy values less than one, and for which the sum is
constrained to total one. In both approaches, the relative occupancies of the two atoms are required. For the first approach, these
occupancies have to be defined. For the second, the value can be refined. However, there is a relationship between the thermal
parameter and the occupancy value so care must be taken when doing this. These issues can be addressed in several ways.
Method 1
The simplest assumption is that the crystal from which the X-ray structure is determined represents the bulk sample was
crystallized. With this value, either a new atom type can be generated that is the appropriate combination of the measured atom
type 1 (M) and atom type 2 (M’) percent composition or two different atoms can be input with the occupancy factor set to reflect
the percent composition of the bulk material. In either case the thermal parameters can be allowed to refine as usual.
Method 2
The occupancy values for two atoms (M and M’) are refined (such that their sum was equal to 1), while the two atoms are
constrained to have the same displacement parameters.
Method 3
The occupancy values (such that their sum was equal to 1) and the displacement parameters are refined independently for the two
atoms.

1.16.1 https://chem.libretexts.org/@go/page/55834
Method 4
Once the best values for occupancy is obtained using either Methods 2 or 3, these values were fixed and the displacement
parameters are allowed to refine freely.

A Model System
Metal β-diketonate complexes (Figure 1.16.1) for metals in the same oxidation state are isostructural and often isomorphous. Thus,
crystals obtained from co-crystallization of two or more metal β-diketonate complexes [e.g., Al(acac)3 and Cr(acac)3] may be
thought of as a hybrid of the precursors; that is, the metal position in the crystal lattice may be defined as having the average metal
composition.

Figure 1.16.1 Molecular structure of M(acac)3, a typical metal β-diketonate complex.


A series of solid solutions of Al(acac)3 and Cr(acac)3 can be prepared for study by X-ray diffraction, by the crystallization from
acetone solutions of specific mixtures of Al(acac)3 and Cr(acac)3 (Table 1.16.1, Column 1). The pure derivatives and the solid
solution, Al1-xCrx(acac)3, crystallize in the monoclinic space group P21/c with Z = 4.
Table 1.16.1 Variance in chromium concentrations (%) for samples of Al1-xCrx(acac)3 crystallized from solutions of Al(acac)3 and Cr(acac)3.
aConcentration too low to successfully refine the Cr occupancy.

Composition as Refined from X-ray


Solution Composition (% Cr) WDS Composition of Single Crystal (% Cr)
Diffraction (% Cr)

13 1.9 ± 0.2 0a

2 2.1 ± 0.3 0a

20 17.8 ± 1.6 17.3 ± 1.8

26 26.7 ± 1.7 28.3 ± 1.9

18 48.5 ± 4.9 46.7 ± 2.1

60 75.1 ± 4.1 72.9 ± 2.4

80 91.3 ± 1.2 82.3 ± 3.1

Substitution of Cr for Al in the M(acac)3 structure could possibly occur in a random manner, i.e., a metal site has an equal
probability of containing an aluminum or a chromium atom. Alternatively, if the chromium had preference for specific sites a super
lattice structure of lower symmetry would be present. Such an ordering is not observed since all the samples show no additional
reflections other than those that may be indexed to the monoclinic cell. Therefore, it may be concluded that the Al(acac)3 and
Cr(acac)3 do indeed form solid solutions: Al1-xCrx(acac)3.
Electron microprobe analysis, using wavelength-dispersive spectrometry (WDS), on the individual crystal from which X-ray
crystallographic data was collected provides the “actual” composition of each crystal. Analysis was performed on at least 6 sites on
each crystal using a 10 μm sized analysis spot providing a measure of the homogeneity within the individual crystal for which X-
ray crystallographic data was collected. An example of a SEM image of one of the crystals and the point analyses is given in Figure
1.16.2. The data in Table 1.16.1 and Figure 1.16.2 demonstrate that while a batch of crystals may contain individual crystals with

different compositions, each individual crystal is actually reasonably homogenous. There is, for most samples, a significant
variance between the molar Al:Cr ratio in the bulk material and an individual crystal chosen for X-ray diffraction. The variation in
Al:Cr ratio within each individual crystal (±10%) is much less than that between crystals.

1.16.2 https://chem.libretexts.org/@go/page/55834
Figure 1.16.2 SEM image of a representative crystal used for WDS and X-ray diffraction analysis showing the location and results
for the WDS analysis. The 10 μm sized analysis spots are represented by the white dots. Adapted from B. D. Fahlman, Ph.D.
Thesis, Rice University, 2000.

Comparison of the Methods


Method 1
Since Method 1 does not refine the %Cr and relies on an input for the Al and Cr percent composition of the "bulk" material, i.e.,
the %Cr in the total mass of the material (Table 1.16.1, Column 1), as opposed to the analysis of the single crystal on which X-ray
diffraction was performed, (Table 1.16.1, Column 2), the closer these values were to the "actual" value determined by WDS for the
crystal on which X-ray diffraction was performed (Table 1.16.1, Column 1 vs 2) then the closer the overall refinement of the
structure to those of Methods 2 - 4.
While this assumption is obviously invalid for many of the samples, it is one often used when bulk data (for example, from NMR)
is available. However, as there is no reason to assume that one crystal is completely representative of the bulk sample, it is unwise
to rely only on such data.
Method 2
This method always produced final, refined, occupancy values that were close to those obtained from WDS (Table 1.16.1). This
approach assumes that the motion of the central metal atoms is identical. While this is obviously not strictly true as they are of
different size, the results obtained herein imply that this is a reasonable approximation where simple connectivity data is required.
For samples where the amount of one of the elements (i.e., Cr) is very low so low a good refinement can not often be obtained. In
theses cases, when refining the occupancy values, that for Al would exceed 1 while that of Cr would be less than 1!
Method 3
In some cases, despite the interrelationship between the occupancy and the displacement parameters, convergence was obtained
successfully. In these cases the refined occupancies were both slightly closer to those observed from WDS than the occupancy
values obtained using Method 2. However, for some samples with higher Cr content the refinement was unstable and would not
converge. Whether this observation was due to the increased percentage of Cr or simply lower data quality is not certain.
While this method does allow refinement of any differences in atomic motion between the two metals, it requires extremely high
quality data for this difference to be determined reliably.
Method 4
This approach adds little to the final results.

Correlation between Analyzed Composition and Refined Composition


Figure 1.16.3 shows the relationship between the chromium concentration (%Cr) determined from WDS and the refinement of X-
ray diffraction data using Methods 2 or 3 (labeled in Figure 1.16.3. Clearly there exists a good correlation, with only a slight
divergence at high Cr concentration. This is undoubtedly a consequence of trying to refine a low fraction of a light atom (Al) in the
presence of a large fraction of a heavier atom (Cr). X-ray diffraction is, therefore, an accurate method of determining the M:M'
ratios in crystalline solid solution.

1.16.3 https://chem.libretexts.org/@go/page/55834
Figure 1.16.3 Comparison of the chromium concentration determined from WDS (with error) and refinement of X-ray diffraction
data (with error) using Methods 2 or 3. Adapted from S. G. Bott, B. D. Fahlman, M. L. Pierson, and A. R. Barron, J. Chem. Soc.,
Dalton Trans., 2001, 2148.

1.16: An Accuracy Assessment of the Refinement of Crystallographic Positional Metal Disorder in Molecular Solid Solutions is shared under a
CC BY 4.0 license and was authored, remixed, and/or curated by Pavan M. V. Raja & Andrew R. Barron via source content that was edited to
conform to the style and standards of the LibreTexts platform; a detailed edit history is available upon request.

1.16.4 https://chem.libretexts.org/@go/page/55834
1.17: Principles of Gamma-ray Spectroscopy and Applications in Nuclear Forensics
Introduction
Gamma-ray (γ-ray) spectroscopy is a quick and nondestructive analytical technique that can be used to identify various radioactive
isotopes in a sample. In gamma-ray spectroscopy, the energy of incident gamma-rays is measured by a detector. By comparing the
measured energy to the known energy of gamma-rays produced by radioisotopes, the identity of the emitter can be determined.
This technique has many applications, particularly in situations where rapid nondestructive analysis is required.

Background Principles
Radioactive Decay
The field of chemistry typically concerns itself with the behavior and interactions of stable isotopes of the elements. However,
elements can exist in numerous states which are not stable. For example, a nucleus can have too many neutrons for the number of
protons it has or contrarily, it can have too few neutrons for the number of protons it has. Alternatively, the nuclei can exist in an
excited state, wherein a nucleon is present in an energy state that is higher than the ground state. In all of these cases, the unstable
state is at a higher energy state and the nucleus must undergo some kind of decay process to reduce that energy.
There are many types of radioactive decay, but type most relevant to gamma-ray spectroscopy is gamma decay. When a nucleus
undergoes radioactive decay by α or β decay, the resultant nucleus produced by this process, often called the daughter nucleus, is
frequently in an excited state. Similar to how electrons are found in discrete energy levels around a nucleus, nucleons are found in
discrete energy levels within the nucleus. In γ decay, the excited nucleon decays to a lower energy state and the energy difference is
emitted as a quantized photon. Because nuclear energy levels are discrete, the transitions between energy levels are fixed for a
given transition. The photon emitted from a nuclear transition is known as a γ-ray.
Radioactive Decay Kinetics and Equilibria
Radioactive decay, with few exceptions, is independent of the physical conditions surrounding the radioisotope. As a result, the
probability of decay at any given instant is constant for any given nucleus of that particular radioisotope. We can use calculus to see
how the number of parent nuclei present varies with time. The time constant, λ, is a representation of the rate of decay for a given
nuclei, 1.17.1.
dN
  =   − λdt (1.17.1)
N

If the symbol N0 is used to represent the number of radioactive nuclei present at t = 0, then 1.17.2 describes the number of nuclei
present at some given time.
−λt
N   =  N0 e (1.17.2)

The same equation can be applied to the measurement of radiation with some sort of detector. The count rate will decrease from
some initial count rate in the same manner that the number of nuclei will decrease from some initial number of nuclei.
The decay rate can also be represented in a way that is more easily understood. The equation describing half-life (t1/2) is shown in
1.17.3.

ln 2
t1/2   =   (1.17.3)
λ

The half-life has units of time and is a measure of how long it takes for the number of radioactive nuclei in a given sample to
decrease to half of the initial quantity. It provides a conceptually easy way to compare the decay rates of two radioisotopes. If one
has a the same number of starting nuclei for two radioisotopes, one with a short half-life and one with a long half-life, then the
count rate will be higher for the radioisotope with the short half-life, as many more decay events must happen per unit time in order
for the half-life to be shorter.
When a radioisotope decays, the daughter product can also be radioactive. Depending upon the relative half-lives of the parent and
daughter, several situations can arise: no equilibrium, a transient equilibrium, or a secular equilibrium. This module will not discuss
the former two possibilities, as they are off less relevance to this particular discussion.
Secular equilibrium takes place when the half-life of the parent is much longer than the half-life of the daughter. In any arbitrary
equilibrium, the ratio of atoms of each can be described as in 1.17.4.

1.17.1 https://chem.libretexts.org/@go/page/55835
NP λD   −  λP
  =  (1.17.4)
ND λP

Because the half-life of the parent is much, much greater than the daughter, as the parent decays, the observed amount of activity
changes very little.
NP λD
  =  (1.17.5)
ND λP

This can be rearranged to show that the activity of the daughter should equal the activity of the parent.

AP   =  AD (1.17.6)

Once this point is reached, the parent and the daughter are now in secular equilibrium with one another and the ratio of their
activities should be fixed. One particularly useful application of this concept, to be discussed in more detail later, is in the analysis
of the refinement level of long-lived radioisotopes that are relevant to trafficking.

Detectors
Scintillation Detector
A scintillation detector is one of several possible methods for detecting ionizing radiation. Scintillation is the process by which
some material, be it a solid, liquid, or gas, emits light in response to incident ionizing radiation. In practice, this is used in the form
of a single crystal of sodium iodide that is doped with a small amount of thallium, referred to as NaI(Tl). This crystal is coupled to
a photomultiplier tube which converts the small flash of light into an electrical signal through the photoelectric effect. This
electrical signal can then be detected by a computer.
Semiconductor Detector
A semiconductor accomplishes the same effect as a scintillation detector, conversion of gamma radiation into electrical pulses,
except through a different route. In a semiconductor, there is a small energy gap between the valence band of electrons and the
conduction band. When a semiconductor is hit with gamma-rays, the energy imparted by the gamma-ray is enough to promote
electrons to the conduction band. This change in conductivity can be detected and a signal can be generated correspondingly.
Germanium crystals doped with lithium, Ge(Li), and high-purity germanium (HPGe) detectors are among the most common types.
Advantages and Disadvantages
Each detector type has its own advantages and disadvantages. The NaI(Tl) detectors are generally inferior to Ge(Li) or HPGe
detectors in many respects, but are superior to Ge(Li) or HPGe detectors in cost, ease of use, and durability. Germanium-based
detectors generally have much higher resolution than NaI(Tl) detectors. Many small photopeaks are completely undetectable on
NaI(Tl) detectors that are plainly visible on germanium detectors. However, Ge(Li) detectors must be kept at cryogenic
temperatures for the entirety of their lifetime or else they rapidly because incapable of functioning as a gamma-ray detector.
Sodium iodide detectors are much more portable and can even potentially be used in the field because they do not require
cryogenic temperatures so long as the photopeak that is being investigated can be resolved from the surrounding peaks.
Gamma Spectrum Features
There are several dominant features that can be observed in a gamma spectrum. The dominant feature that will be seen is the
photopeak. The photopeak is the peak that is generated when a gamma-ray is totally absorbed by the detector. Higher density
detectors and larger detector sizes increase the probability of the gamma-ray being absorbed.
The second major feature that will be observed is that of the Compton edge and distribution. The Compton edge arises due to
Compton Effect, wherein a portion of the energy of the gamma-ray is transferred to the semiconductor detector or the scintillator.
This occurs when the relatively high energy gamma ray strikes a relatively low energy electron. There is a relatively sharp edge to
the Compton edge that corresponds to the maximum amount of energy that can be transferred to the electron via this type of
scattering. The broad peak lower in energy than the Compton edge is the Compton distribution and corresponds to the energies that
result from a variety of scattering angles. A feature in Compton distribution is the backscatter peak. This peak is a result of the
same effect but corresponds to the minimum energy amount of energy transferred. The sum of the energies of the Compton edge
and the backscatter peak should yield the energy of the photopeak.
Another group of features in a gamma spectrum are the peaks that are associated with pair production. Pair production is the
process by which a gamma ray of sufficiently high energy (>1.022 MeV) can produce an electron-positron pair. The electron and

1.17.2 https://chem.libretexts.org/@go/page/55835
positron can annihilate and produce two 0.511 MeV gamma photons. If all three gamma rays, the original with its energy reduced
by 1.022 MeV and the two annihilation gamma rays, are detected simultaneously, then a full energy peak is observed. If one of the
annihilation gamma rays is not absorbed by the detector, then a peak that is equal to the full energy less 0.511 MeV is observed.
This is known as an escape peak. If both annihilation gamma rays escape, then a full energy peak less 1.022 MeV is observed. This
is known as a double escape peak.

Example of Experiments
Determination of Depleted Uranium
Natural uranium is composed mostly of 238U with low levels of 235U and 234U. In the process of making enriched uranium,
uranium with a higher level of 235U, depleted uranium is produced. Depleted uranium is used in many applications particularly for
its high density. Unfortunately, uranium is toxic and is a potential health hazard and is sometimes found in trafficked radioactive
materials, so it is important to have a methodology for detection and analysis of it.
One easy method for this determination is achieved by examining the spectrum of the sample and comparing it qualitatively to the
spectrum of a sample that is known to be natural uranium. This type of qualitative approach is not suitable for issues that are of
concern to national security. Fortunately, the same approach can be used in a quantitative fashion by examining the ratios of various
gamma-ray photopeaks.
The concept of a radioactive decay chain is important in this determination. In the case of 238U, it decays over many steps to 206Pb.
In the process, it goes through 234mPa, 234Pa, and 234Th. These three isotopes have detectable gamma emissions that are capable of
being used quantitatively. As can be seen in Table 1.17.1, the half-life of these three emitters is much less than the half-life of 238U.
As a result, these should exist in secular equilibrium with 238U. Given this, the ratio of activity of 238U to each daughter products
should be 1:1. They can thus be used as a surrogate for measuring 238U decay directly via gamma spectroscopy. The total activity
of the 238U can be determined by 1.17.8, where A is the total activity of 238U, R is the count rate of the given daughter isotope, and
B is the probability of decay via that mode. The count rate may need to be corrected for self-absorption of the sample is particularly
thick. It may also need to be corrected for detector efficiency if the instrument does not have some sort of internal calibration.
A = R/B (1.17.7)

Table 1.17.1 Half-lives of pertinent radioisotopes in the 238U decay chain


Isotope Half-life
238U 4.5 x 10^{9} years
234Th 24.1 days
234mPa 1.17 minutes

Example 1
Question
A gamma spectrum of a sample is obtained. The 63.29 keV photopeak associated with 234Th was found to have a count rate
of 5.980 kBq. What is the total activity of 238U present in the sample?
Answer
234
Th exists in secular equilibrium with 238U. The total activity of 234Th must be equal to the activity of the 238U. First, the
observed activity must be converted to the total activity using Equation A=R/B. It is known that the emission probability for
the 63.29 kEv gamma-ray for 234Th is 4.84%. Therefore, the total activity of 238U in the sample is 123.6 kBq.
The count rate of 235U can be observed directly with gamma spectroscopy. This can be converted, as was done in the case of 238U
above, to the total activity of 235U present in the sample. Given that the natural abundances of 238U and 235U are known, the ratio of
the expected activity of 238U to 235U can be calculated to be 21.72 : 1. If the calculated ratio of disintegration rates varies
significantly from this expected value, then the sample can be determined to be depleted or enriched.

Example 2

1.17.3 https://chem.libretexts.org/@go/page/55835
Question
As shown above, the activity of 238U in a sample was calculated to be 123.6 kBq. If the gamma spectrum of this sample
shows a count rate 23.73 kBq at the 185.72 keV photopeak for 235U, can this sample be considered enriched uranium? The
emission probability for this photopeak is 57.2%.
Answer
As shown in the example above, the count rate can be converted to a total activity for 235U. This yields a total activity of
41.49 kBq for 235U. The ratio of activities of 238U and 235U can be calculated to be 2.979. This is lower than the expected
ratio of 21.72, indicating that the 235U content of the sample greater than the natural abundance of 235U.
This type of calculation is not unique to 238U. It can be used in any circumstance where the ratio of two isotopes needs to be
compared so long as the isotope itself or a daughter product it is in secular equilibrium with has a usable gamma-ray photopeak.
Determination of the Age of Highly-enriched Uranium
Particularly in the investigation of trafficked radioactive materials, particularly fissile materials, it is of interest to determine how
long it has been since the sample was enriched. This can help provide an idea of the source of the fissile material—if it was
enriched for the purpose of trade or if it was from cold war era enrichment, etc.
When uranium is enriched, 235U is concentrated in the enriched sample by removing it from natural uranium. This process will
separate the uranium from its daughter products that it was in secular equilibrium with. In addition, when 235U is concentrated in
the sample, 234U is also concentrated due to the particulars of the enrichment process. The 234U that ends up in the enriched sample
will decay through several intermediates to 214Bi. By comparing the activities of 234U and 214Bi or 226Ra, the age of the sample can
be determined.
AU
2
ABi   =  ARa   =   λT h λRa T (1.17.8)
2

In 1.17.8, ABi is the activity of 214Bi, ARais the activity of 226Ra, AU is the activity of 234U, λTh is the decay constant for 230Th, λRa
is the decay constant for 226Ra, and T is the age of the sample. This is a simplified form of a more complicated equation that holds
true over all practical sample ages (on the order of years) due to the very long half-lives of the isotopes in question. The results of
this can be graphically plotted as they are in Figure 1.17.1.

Figure 1.17.1 Ratio of 226Ra/234U (= 214Bi/234U) plotted versus age based on 1.17.8 . This can be used to determine how long ago a
sample was enriched based on the activities of 234U and 226Ra or 214Bi in the sample.

Example 3
Question
The gamma spectrum for a sample is obtained. The count rate of the 121 keV 234U photopeak is 4500 counts per second and the
associated emission probability is 0.0342%. The count rate of the 609.3 keV 214Bi photopeak is 5.83 counts per second and the
emission probability is 46.1%. How old is the sample?
Answer

1.17.4 https://chem.libretexts.org/@go/page/55835
The observed count rates can be converted to the total activities for each radionuclide. Doing so yields a total activity for 234U of
4386 kBq and a total activity for 214Bi of 12.65 Bq. This gives a ratio of 9.614 x 10-7. Using Figure 1.17.1, as graphed this
indicates that the sample must have been enriched 22.0 years prior to analysis.

1.17: Principles of Gamma-ray Spectroscopy and Applications in Nuclear Forensics is shared under a CC BY 4.0 license and was authored,
remixed, and/or curated by Pavan M. V. Raja & Andrew R. Barron via source content that was edited to conform to the style and standards of the
LibreTexts platform; a detailed edit history is available upon request.

1.17.5 https://chem.libretexts.org/@go/page/55835
CHAPTER OVERVIEW
2: Physical and Thermal Analysis
2.1: Melting Point Analysis
2.2: Molecular Weight Determination
2.3: BET Surface Area Analysis of Nanoparticles
2.4: Dynamic Light Scattering
2.5: Zeta Potential Analysis
2.6: Viscosity
2.7: Electrochemistry
2.8: Thermal Analysis
2.9: Electrical Permittivity Characterization of Aqueous Solutions
2.10: Dynamic Mechanical Analysis
2.11: Finding a Representative Lithology

2: Physical and Thermal Analysis is shared under a CC BY 4.0 license and was authored, remixed, and/or curated by Pavan M. V. Raja & Andrew
R. Barron via source content that was edited to conform to the style and standards of the LibreTexts platform; a detailed edit history is available
upon request.

1
2.1: Melting Point Analysis
Melting point (Mp) is a quick and easy analysis that may be used to qualitatively identify relatively pure samples (approximately
<10% impurities). It is also possible to use this analysis to quantitatively determine purity. Melting point analysis, as the name
suggests, characterizes the melting point, a stable physical property, of a sample in a straightforward manner, which can then be
used to identify the sample.

Equipment
Although different designs of apparatus exist, they all have some sort of heating or heat transfer medium with a control, a
thermometer, and often a backlight and magnifying lens to assist in observing melting (Figure 2.1.1 ). Most models today utilize
capillary tubes containing the sample submerged in a heated oil bath. The sample is viewed with a simple magnifying lens. Some
new models have digital thermometers and controls and even allow for programming. Programming allows more precise control
over the starting temperature, ending temperature and the rate of change of the temperature.

Figure 2.1.1 A Thomas Hoover melting point apparatus. The tower (A) contains a thermometer with a reflective view (B), so that
the sample and temperature may be monitored simultaneously. The magnifying lens (C) allows better viewing of samples and lies
above the heat controller (D).

Sample Preparation
For melting point analysis, preparation is straight forward. The sample must be thoroughly dried and relatively pure ( <10%
impurities). The dry sample should then be packed into a melting point analysis capillary tube, which is simply a glass capillary
tube with only one open end. Only 1 to 3 mm of sample is needed for sufficient analysis. The sample needs to be packed down into
the closed end of the tube. This may be done by gently tapping the tube or dropping it upright onto a hard surface (Figure 2.1.2 ).
Some apparatuses have a vibrator to assist in packing the sample. Finally the tube should be placed into the machine. Some models
can accommodate multiple samples.

2.1.1 https://chem.libretexts.org/@go/page/55838
Figure 2.1.2 Schematic showing how to pack dried sample into a melting point analysis capillary tube: (a) using a spatula, push a
sufficient amount of sample into the tube opening, (b) using a tapping motion or dropping the tube, pack the sample into the closed
end, (c) the sample is ready to be loaded into the apparatus.

Recording Data
Performing analysis is different from machine to machine, but the overall process is the same (Figure 2.1.3 ). If possible, choose a
starting temperature, ending temperature, and rate of change of temperature. If the identity of the sample is known, base the starting
and ending temperatures from the known melting point of the chemical, providing margins on both sides of the range. If using a
model without programming, simply turn on the machine and monitor the rate of temperature change manually.

Exp 3 Melting Point Determination

Figure 2.1.3 A video discussing sample preparation, recording data and melting point analysis in general. Made by Indiana
University-Purdue University Indianapolis chemistry department.
Visually inspect the sample as it heats. Once melting begins, note the temperature. When the sample is completely melted, note the
temperature again. That is the melting point range for the sample. Pure samples typically have a 1 - 2 °C melting point range,
however, this may be broadened due to colligative properties.

Interpreting Data
There are two primary uses of melting point analysis data. The first is for qualitative identification of the sample, and the second is
for quantitative purity characterization of the sample.
For identification, compare the experimental melting point range of the unknown to literature values. There are several vast
databases of these values. Obtain a pure sample of the suspected chemical and mix a small amount of the unknown with it and
conduct melting point analysis again. If a sharp melting point range is observed at similar temperatures to the literature values, then
the unknown has likely been identified correctly. Conversely, if the melting point range is depressed or broadened, which would be
due to colligative properties, then the unknown was not successfully identified.
To characterize purity, first the identity of the solvent (the main constituent of the sample) and the identity of the primary solute
need to be known. This may be done using other forms of analysis, such as gas chromatography-mass spectroscopy coupled with a
database. Because melting point depression is unique between chemicals, a mixed melting curve comparing molar fractions of the
two constituents with melting point needs to either be obtained or prepared (Figure 2.1.4 ). Simply prepare standards with known
molar fraction ratios, then perform melting point analysis on each standard and plot the results. Compare the melting point range of

2.1.2 https://chem.libretexts.org/@go/page/55838
the experimental sample to the curve to identify the approximate molar fractions of the constituents. This sort of purity
characterization cannot be performed if there are more than two primary components to the sample.

Figure 2.1.4 A mixed melting curve for naphthalene and biphenyl. Non-pure samples exhibit melting point depression due to
colligative properties. Adapted from “Melting Point Analysis”, Chem 211L, Clark College protocol.

Specificity and Accuracy


Melting point analysis is fairly specific and accurate given its simplicity. Because melting point is a unique physical characteristic
of a substance, melting point analysis does have high specificity. Although, many substances have similar melting points, so having
an idea of possible chemicals in mind can greatly narrow down the choices. The thermometers used are also accurate. However,
melting point is dependent on pressure as well, so experimental results can vary from literature values, especially at extreme
locations, i.e., places of high altitude. The biggest source of error stems from the visual detection of melting by the experimenter.
Controlling the change rate and running multiple trials can lessen the degree of error introduced at this step.

Advantages of Melting Point Analysis


Melting point analysis is a quick, relatively easy, and inexpensive preliminary analysis if the sample is already mostly pure and has
a suspected identity. Additionally, analysis requires small samples only.

Limitations of Melting Point Analysis


As with any analysis, there are certain drawbacks to melting point analysis. If the sample is not solid, melting point analysis cannot
be done. Also, analysis is destructive of the sample. For qualitative identification analysis, there are now more specific and accurate
analyses that exist, although they are typically much more expensive. Also, samples with more than one solute cannot be analyzed
quantitatively for purity.

2.1: Melting Point Analysis is shared under a CC BY 4.0 license and was authored, remixed, and/or curated by Pavan M. V. Raja & Andrew R.
Barron via source content that was edited to conform to the style and standards of the LibreTexts platform; a detailed edit history is available upon
request.

2.1.3 https://chem.libretexts.org/@go/page/55838
2.2: Molecular Weight Determination
Solution Molecular Weight of Small Molecules
The cryoscopic method was formally introduced in the 1880’s when François-Marie Raoult published how solutes depressed the
freezing points of various solvents such as benzene, water, and formic acid. He concluded from his experimentation “if one
molecule of a substance can be dissolved in one-hundred molecules of any given solvent then the solvent temperature is lowered by
a specific temperature increment”. Based on Raoult’s research, Ernst Otto Beckmann invented the Beckmann thermometer and the
associated freezing - point apparatus, which was a significant improvement in measuring freezing - point depression values for a
pure solvent. The simplicity, ease, and accuracy of this apparatus has allowed it to remain as a current standard with few
modifications for molecular weight determination of unknown compounds.

Figure 2.2.1 French chemist François-Marie Raoult (1830 - 1901).

Figure 2.2.2 German chemist Ernst Otto Beckmann (1853 - 1923).

Figure 2.2.3 Beckmann differential thermometer and freezing point depression apparatus
The historical significance of Raoult and Beckmann’s research, among many other investigators, has revolutionized a physical
chemistry technique that is currently applied to a vast range of disciplines from food science to petroleum fluids. For example,
measured cryoscopic molecular weights of crude oil are used to predict the viscosity and surface tension for necessary fluid flow
calculations in pipeline.
Freezing Point Depression
Freezing point depression is a colligative property in which the freezing temperature of a pure solvent decreases in proportion to
the number of solute molecules dissolved in the solvent. The known mass of the added solute and the freezing point of the pure
solvent information permit an accurate calculation of the molecular weight of the solute.
In Equation 2.2.1 the freezing point depression of a non-ionic solution is described. Where ∆Tf is the change in the initial and final
temperature of the pure solvent, Kf is the freezing point depression constant for the pure solvent, and m (moles solute/kg solvent) is
the molality of the solution.

ΔTf = Kf m (2.2.1)

For an ionic solution shown in Figure 2.2.2, the dissociation particles must be accounted for with the number of solute particles per
formula unit, i (the van’t Hoff factor).

ΔTf = Kf mi  (2.2.2)

2.2.1 https://chem.libretexts.org/@go/page/55839
Cryoscopic Procedures

Cryoscopic Apparatus
For cryoscopy, the apparatus to measure freezing point depression of a pure solvent may be representative of the Beckmann
apparatus previously shown in Figure 2.2.3. The apparatus consists of a test tube containing the solute dissolved in a pure solvent,
stir bar or magnetic wire and closed with a rubber stopper encasing a mercury thermometer. The test tube component is immersed
in an ice-water bath in a beaker. An example of the apparatus is shown in Figure 2.2.4. The rubber stopper and stir bar/wire stirrer
are not shown in the figure.

Figure 2.2.4 An example of a cryoscopic apparatus. Adapted from www.lahc.cc.ca.us/classes/che...ng%20Point.pdf

Sample and Solvent Selection


The cryoscopic method may be used for a wide range of samples with various degrees of polarity. The solute and solvent selection
should follow the premise of like dissolved like or in terms of Raoult’s principle of the dissolution of one molecule of solute in one-
hundred molecules of a solvent. The most common solvents such as benzene are generally selected because it is unreactive,
volatile, and miscible with many compounds.Table 2.2.1 shows the cryoscopic constants (Kf) for the common solvents used for
cryoscopy. A complete list of Kf values are available in Knovel Critical Tables.
Table 2.2.1 : Cryoscopic constants (Kf) for common solvents used for cryoscopy.
Compound Kf

Acetic Acid 3.90

Benzene 5.12

Camphor 39.7

Carbon disulfide 3.8

Carbon tetrachloride 30

Chloroform 4.68

Cyclohexane 20.2

Ethanol 1.99

Naphthalene 6.8

Phenol 7.27

2.2.2 https://chem.libretexts.org/@go/page/55839
Compound Kf

Water 1.86

Cryoscopic Method
The detailed information about the procedure used for cryoscopy is shown below:
Allow the solution to stir continuously to avoid supercooling
1. Weigh (15 to 20 grams) of the pure solvent in a test tube and record the measured weight value of the pure solvent.
2. Place a stir bar or wire stirrer in the test tube and close with a rubber stopper that has a hole to encase a mercury thermometer.
3. Place a mercury thermometer in the rubber stopper hole.
4. Immerse the test tube apparatus in an ice-water bath.
5. Allow the solvent to stir continuously and equilibrate to a few degrees below the freezing point of the solvent.
6. Record the temperature at which the solvent reaches the freezing point, which remains at a constant temperature reading.
7. Repeat the freezing point data collection for at least two more measurements without a difference less than 0.5 °C between the
measurements.
8. Weigh a quantity of the solute for investigation and record the measured value.
9. Add the weighed solute to the test tube containing the pure solvent.
10. Re - close the test tube with a rubber stopper encasing a mercury thermometer.
11. Re-immerse the test tube in an ice water bath and allow the mixture to stir to fully dissolve the solute in the pure solvent.
12. Measure the freezing point and record the temperature value.
The observed freezing point of the solution is when the temperature reading remains constant.

Sample calculation to determine molecular weight


Sample Data Set
Table 2.2.2 represents an example of a data set collection for cryoscopy.
Table 2.2.2 Example data set collection for cryoscopy
Parameter Trial 1 Trial 2 Trial 3 Avg.

Mass of cyclohexane (g) 9.05 9.00 9.04 9.03

Mass of unknown solute


0.4000 0.41010 0.4050 0.4050
(g)

Freezing point of
6.5°C 6.5°C 6.5°C 6.5°C
cyclohexane (°C)

Freezing point of
cyclohexane mixed with 4.2°C 4.3°C 4.2°C 4.2°C
unknown solute (°C)

Calculation of molecular weight using the freezing point depression equation


Calculate the freezing point (Fpt) depression of the solution (TΔf) from Equation 2.2.3
T Δf =  (F pt of  pure solvent)  −  (F pt of  solution)  (2.2.3)

∘ ∘
T Δf =  6.5 C − 4.2 C (2.2.4)


T Δf =  2.3 (2.2.5)

Calculate the molal concentration, m, of the solution using the freezing point depression and Kf (see \label{4})
T Δf = Kf m  (2.2.6)

∘ ∘
m = (2.3 C )/(20.2 C /molal) (2.2.7)

m = 0.113molal (2.2.8)

2.2.3 https://chem.libretexts.org/@go/page/55839
m = g(solute)/kg(solvent) (2.2.9)

Calculate the MW of the unknown sample.


i = 1 for covalent compounds in 2.2.2
Kf (g solute)
MW = (2.2.10)
ΔTf (kgsolvent)


20.2 C ∗ kg/moles × 0.405 g
MW = (2.2.11)

2.3 C × 0.00903 kg

MW =  393 g/mol (2.2.12)

Problems
1. Nicotine (Figure 2.2.5 is an extracted pale yellow oil from tobacco leaves that dissolves in water at temperatures less than 60°C.
What is the molality of nicotine in an aqueous solution that begins to freeze at -0.445°C? See Table 2.2.1 for Kf values.

Figure 2.2.5 The chemical structure of nicotine.


2. If the solution used in Problem 1 is obtained by dissolving 1.200 g of nicotine in 30.56 g of water, what is the molar mass of
nicotine?
3. What would be the freezing point depression when 0.500 molal of Ca(NO3)2 is dissolved in 60 g of water?
4. Calculate the number of weighted grams of Ca(NO3)2 added to the 60 g of water to achieve the freezing point depression from
Problem 3.
Answers
1.

2.

3.

4.

Molecular Weight of Polymers


Knowledge of the molecular weight of polymers is very important because the physical properties of macromolecules are affected
by their molecular weight. For example, shown in Figure 2.2.6 the interrelation between molecular weight and strength for a
typical polymer. Dependence of mechanical strength on polymer molecular weight. Adapted from G. Odian, Principles of
Polymerization, 4th edition, Willey-Interscience, New York (2004).

2.2.4 https://chem.libretexts.org/@go/page/55839
Figure 2.2.6 A diagram of the typical curve associating mechanical strength and molecular weight

The melting point of polymers are also slightly depend on their molecular weight. Figure 2.2.7 shows relationship between
molecular weight and melting temperatures of polyethylene (Figure 2.2.8 ) Most linear polyethylenes have melting temperatures
near 140 °C. The approach to the theoretical asymptote, that is a line whose distance to a given curve tends to zero, indicative that a
theoretical polyethylene of infinite molecular weight (i.e., M = ∞) would have a melting point of 145 °C.
The molecular weight-melting temperature relationship for the alkane series. Adapted from L. H. Sperling, Introduction to physical
polymer science, 4th edition, Wiley-Interscience, New York (2005).

Figure 2.2.7 A diagram of the asymptotic approach of the melting point of a polymer to a specific value.

Figure 2.2.8 Structure of Polyethylene


There are several ways to calculate molecular weight of polymers like number average of molecular weight, weight average of
molecular weight, Z-average molecular weight, viscosity average molecular weight, and distribution of molecular weight.
Molecular Weight Calculations
Number average of molecular weight (Mn)
Number average of molecular weight is measured to determine number of particles. Number average of molecular weight is the
total weight of polymer, 2.2.13, divided by the number of polymer molecules, 2.2.14 . The number average molecular weight (Mn)
is given by 2.2.15 , where Mi is molecular weight of a molecule of oligomer n, and Ni is number of molecules of that molecular
weight.

T otal weight =  Σ Mi Ni (2.2.13)
i=1


T otal number =  Σ Ni (2.2.14)
i=1


Σ Mi Ni
i=1
Mn = (2.2.15)

Σ Ni
i=1

Example 2.2.8

Consider a polymer sample comprising of 5 moles of polymer molecules having molecular weight of 40.000 g/mol and 15
moles of polymer molecules having molecular weight of 30.000 g/mol.

Weight average of molecular weight (MW)

2.2.5 https://chem.libretexts.org/@go/page/55839
Weight average of molecular weight (MW) is measured to determine the mass of particles. MW defined as 2.2.16 , where Mi is
molecular weight of oligomer n, and Ni is number of molecules of that molecular weight.
∞ 2
Σ Ni (Mi )
i=1
MW = (2.2.16)

Σ Ni Mi
i=1

Example:
Consider the polymer described in the previous problem.

Calculate the MW for a polymer sample comprising of 9 moles of polymer molecules having molecular weight of 30.000 g/mol and
5 moles of polymer molecules having molecular weight of 50.000 g/mol.

Answer:

Z-average molecular weight (MZ)


The Z-average molecular weight (Mz) is measured in some sedimentation equilibrium experiments. Mz isn’t common technique for
molecular weight of polymers. The molar mass depends on size and mass of the molecules. The ultra centrifugation techniques
employ to determine Mz. Mz emphasizes large particles and it defines the EQ, where Mi is molecular weight and Ni is number of
molecules.
3
ΣNi M
i
MW = (2.2.17)
2
Σ Ni M i

Consider the polymer described in the previous problem.

Viscosity average molecular weight (MV)


One of the ways to measure the average molecular weight of polymers is viscosity of solution. Viscosity of a polymer depend on
concentration and molecular weight of polymers. Viscosity techniques is common since it is experimentally simple. Viscosity
average molecular weight defines as 2.2.18 , where Mi is molecular weight and Ni is number of molecules, a is a constant which
depend on the polymer-solvent in the viscosity experiments. When a is equal 1, Mv is equal to the weight average molecular
weight, if it isn’t equal 1 it is between weight average molecular weight and the number average molecular weight.
1+a
ΣNi M 1
i
( ) 2
(2.2.18)
ΣNi Mi

Distribution of molecular weight


Molecular weight distribution is one of the important characteristic of polymer because it affects polymer properties. A typical
molecular distribution of polymers show in 2.2.6. There are various molecular weights in the range of curve. The distribution of
sizes in a polymer sample isn't totally defined by its central tendency. The width and shape of distribution must be known. It is

2.2.6 https://chem.libretexts.org/@go/page/55839
always true that the various range molecular weight is 2.2.19 . The equality is occurring when all polymer in the sample have the
same molecular weight.
MN ≥ MV ≥ MW ≥ MZ ≥ MZ+1 (2.2.19)

Figure 2.2.6 A schematic plot of a distribution of molecular weights along with the rankings of the various average molecular
weights. Adapted from J. A. Nairn, Oregon State University (2003).
Molecular weight analysis of polymers
Gel permeation chromatography (GPC)
Gel permeation chromatography is also called size exclusion chromatography. It is widely used method to determine high
molecular weight distribution. In this technique, substances separate according to their molecule size. Firstly, large molecules begin
to elute then smaller molecules are eluted Figure 2.2.7. The sample is injected into the mobile phase then the mobile phase enters
into the columns. Retention time is the length of time that a particular fraction remains in the column. As shown in Figure 2.2.7,
while the mobile phase passes through the porous particles, the area between large molecules and small molecules is getting
increase. GPC gives a full molecular distribution, but its cost is high.

Figure 2.2.7 Solvent flow through column. Adapted from A. M. Striegel, W. W. Yau, J. J. Kirkland, and D. D. Bly. Modern Size-
Exclusion Liquid Chromatography- Practice of Gel Permeation and Gel Filtration Chromatography, 2nd Edition. Hoboken. N.J.
(2009).
According to basic theory of GPC, the basic quantity measured in chromatography is the retention volume, 2.2.20, where V0 is
mobile phase volume and Vp is the volume of a stationary phase. K is a distribution coefficient related to the size and types of the
molecules.
Ve = V0 + Vp K (2.2.20)

The essential features of gel permeation chromatography are shown in Figure 2.2.8. Solvent leaves the solvent supply, then solvent
is pumped through a filter. The desired amount of flow through the sample column is adjusted by sample control valves and the
reference flow is adjusted that the flow through the reference and flow through the sample column reach the detector in common

2.2.7 https://chem.libretexts.org/@go/page/55839
front. The reference column is used to remove any slight impurities in the solvent. In order to determine the amount of sample, a
detector is located at the end of the column. Also, detectors may be used to continuously verify the molecular weight of species
eluting from the column. The flow of solvent volume is as well monitored to provide a means of characterizing the molecular size
of the eluting species.

Figure 2.2.8 Schematic of gel permeation chromatography system.


As an example, consider the block copolymer of ethylene glycol (PEG, Figure 2.2.9 ) and poly(lactide) (PLA, Figure 2.2.10 ), i.e.,
Figure 2.2.11. In the first step starting with a sample of PEG with a Mn of 5,700 g/mol. After polymerization, the molecular weight
increased because of the progress of lactide polymerization initiated from end of PEG chain. Varying composition of PEG-PLA
shown in Table 2.2.3 can be detected by GPC (Figure 2.2.12 ).

Figure 2.2.9 The structure of polyethyleneglycol (PEG).

Figure 2.2.10 The ring-opening polymerization of lactide to polylactide.

Figure 2.2.11 The structure of PEG-PLA block copolymer.

Figure 2.2.12 Gel permeation chromatogram of (a) PEG (MW = 5,700 g/mol) and (b) PEG-PLA block copolymer (MW = 11,000
g/mol). Adapted from K. Yasugi, Y. Nagasaki, M. Kato, K. Kataoka, J. Control. Release, 1999, 62, 89.
Table 2.2.3 Characteristics of PEG-PLA block copolymer with varying composition. Adapted from K. Yasugi, Y. Nagasaki, M. Kato, and K.
Kataoka, J. Control Release , 1999, 62, 89
Mw/Mn of block Weight ratio of PLA
Polymer Mn of PEG Mw/Mn of PEG Mn of PLA
copolymer to PEG

PEG-PLA (41-12) 4100 1.05 1200 1.05 0.29

2.2.8 https://chem.libretexts.org/@go/page/55839
PEG-PLA (60-30) 6000 1.03 3000 1.08 0.50

PEG-PLA (57-54) 5700 1.03 5400 1.08 0.95

PEG-PLA (61-78) 6100 1.03 7800 1.11 1.28

Light-scattering
One of the most used methods to characterize the molecular weight is light scattering method. When polarizable particles are
placed in the oscillating electric field of a beam of light, the light scattering occurs. Light scattering method depends on the light,
when the light is passing through polymer solution, it is measure by loses energy because of absorption, conversion to heat and
scattering. The intensity of scattered light relies on the concentration, size and polarizability that is proportionality constant which
depends on the molecular weight. Figure 2.2.13 shows light scattering off a particle in solution.

Figure 2.2.13 Modes of scattering of light in solution.


A schematic laser light-scattering is shown in Figure 2.2.14. A major problem of light scattering is to prepare perfectly clear
solutions. This problem is usually accomplished by ultra-centrifugation. A solution should be as possible as clear and dust free to
determine absolute molecular weight of polymer. The advantages of this method, it doesn’t need calibration to obtain absolute
molecular weight and it can give information about shape and Mw information. Also, it can be performed rapidly with less amount
of sample and absolute determinations of the molecular weight can be measured. The weaknesses of the method is high price and
most times it requires difficult clarification of the solutions.

Figure 2.2.14 Schematic representation of light scattering. Adapted from J. A. Nairn, polymer characterization, Material science
and engineering 5473, spring 2003.
The weight average molecular weight value of scattering polymers in solution related to their light scattering properties that define
by 2.2.21 , where K is the wave vector, that defined by 2.2.22 . C is solution concentration, R(θ) is the reduced Rayleigh ratio, P(θ)
the particle scattering function, θ is the scattering angle, A is the osmotic virial coefficients, where n0 solvent refractive index, λ the
light wavelength and Na Avagadro’s number. The particle scattering function is given by 2.2.23 , where Rz is the radius of gyration.

KC /R(θ)  =  1/ MW (P (θ)  +  2 A2 C   +  3 A3 C2   +  . . . ) (2.2.21)

2 2 2 2
K  =  2 π n (dn/dC ) / Na λ (2.2.22)
0

2 2 2 2 2
1/(P (θ))  =  1 + 16 π n (Rz )si n (θ/2)3 λ (2.2.23)
0

2.2.9 https://chem.libretexts.org/@go/page/55839
Weight average molecular weight of a polymer is found from extrapolation of data in the form of a Zimm plot ( Figure 2.2.15 ).
Experiments are performed at several angles and at least at 4 different concentrations. The straight line extrapolations provides Mw.

Figure 2.2.15 A typical Zimm plot of light scattering data. Adapted from M. P. Stevens, Polymer Chemistry an Introduction, 3rd
edition, Oxford University Press, Oxford (1999).
X-ray Scattering
X-rays are a form of electromagnetic wave with wavelengths between 0.001 nm and 0.2 nm. X-ray scattering is particularly used
for semicrystalline polymers which includes thermoplastics, thermoplastic elastomers, and liquid crystalline polymers. Two types
of X-ray scattering are used for polymer studies:
1. Wide-angle X-ray scattering (WAXS) which is used to study orientation of the crystals and the packing of the chains.
2. Small-angle X-ray scattering (SAXS) which is used to study the electron density fluctuations that occur over larger distances as
a result of structural inhomogeneities.
Schematic representation of X-ray scattering shows in Figure 2.2.16.

Figure 2.2.16 Schematic diagram of X-ray scattering. Adapted from B. Chu, and B. S. Hsiao, Chem. Rev. 2001,101, 1727.
At least two SAXS curves are required to determine the molecular weight of a polymer. The SAXS procedure to determine the
molecular weight of polymer sample in monomeric or multimeric state solution requires the following conditions.
a. The system should be monodispersed.
b. The solution should be dilute enough to avoid spatial correlation effects.
c. The solution should be isotropic.
d. The polymer should be homogenous.
Osometer
Osmometry is applied to determine number average of molecular weight (Mn). There are two types of osmometer:
1. Vapor pressure osmometry (VPO).
2. Membrane osmometry.
Vapor pressure osmometry measures vapor pressure indirectly by measuring the change in temperature of a polymer solution on
dilution by solvent vapor and is generally useful for polymers with Mn below 10,000–40,000 g/mol. When molecular weight is
more than that limit, the quantity being measured becomes very small to detect. A typical vapor osmometry shows in the Figure
2.2.17. Vapor pressure is very sensitive because of this reason it is measured indirectly by using thermistors to measure voltage

changes caused by changes in temperature.

2.2.10 https://chem.libretexts.org/@go/page/55839
Figure 2.2.17 Schematic vapor pressure osmometry. Adapted from http://www.gallay.com.au/node/186
Membrane osmometry is absolute technique to determine Mn(Figure 2.2.18 ). The solvent is separated from the polymer solution
with semipermeable membrane that is strongly held between the two chambers. One chamber is sealed by a valve with a transducer
attached to a thin stainless steel diaphragm which permits the measurement of pressure in the chamber continuously. Membrane
osmometry is useful to determine Mn about 20,000-30,000 g/mol and less than 500,000 g/mol. When Mn of polymer sample more
than 500,000 g/mol, the osmotic pressure of polymer solution becomes very small to measure absolute number average of
molecular weight. In this technique, there are problems with membrane leakage and symmetry. The advantages of this technique is
that it doesn’t require calibration and it gives an absolute value of Mn for polymer samples.

Figure 2.2.18 Schematic representative of membrane osmometry. Adapted from


http://www.flickr.com/photos/mitopencourseware/3327963527/
Summary
Properties of polymers depend on their molecular weight. There are different kind of molecular weight and each can be measured
by different techniques. The summary of these techniques and molecular weight is shown in the Table 2.2.4.
Table 2.2.4 Summary of techniques of molecular weight of polymers.
Method Type of Molecular Weight Range of Application

Light Scattering MW ∞

Membrane Osmometry Mn 104 -106

Vapor Phase Osmometry Mn 40,000

X-ray scattering Mw, n, z 102

Size Exclusion Chromatography and its Application in Polymer Science


Size exclusion chromatography (SEC) is a useful technique that is specifically applicable to high-molecular-weight species, such as
polymer. It is a method to sort molecules according to their sizes in solution. The sample solution is injected into the column, which
is filled with rigid, porous, materials, and is carried by the solvent through the packed column. The sizes of molecules are
determined by the pore size of the packing particle in the column within which the separation occurs.
For polymeric materials, the molecular weight (Mw) or molecular size plays a key role in determining the mechanical, bulk, and
solution properties of materials. It is known that the sizes of polymeric molecules depend on their molecular weights, side chain

2.2.11 https://chem.libretexts.org/@go/page/55839
configurations, molecular interaction, and so on. For example, the exclusion volume of polymers with rigid side group is larger
than those with soft long side chains. Therefore, in order to determine the molecular weight and molecular weight distribution of a
polymer, one of the most widely applied methods is gel-permeation chromatography.
Gel permeation chromatography (GPC) is a term used for when the separation technique size exclusion chromatography (SEC) is
applied to polymers.
The primary purpose and use of the SEC technique is to provide molecular weight distribution information about a particular
polymeric material. Typically, in about 30 minutes using standard SEC, the complete molecular weight distribution of a polymer as
well as all the statistical information of the distribution can be determined. Thus, SEC has been considered as a technique
essentially supplanting classical molecular weight techniques. To apply this powerful technique, there is some basic work that
needs to be done before its use. The selection of an appropriate solvent and the column, as well as the experimental conditions, are
important for proper separation of a sample. Also, it is necessary to have calibration curves in order to determine the relative
molecular weight from a given retention volume/time.
It is well known that both the majority of natural and synthetic polymers are polydispersed with respect to molar mass. For
synthetic polymers, the more mono-dispersed a polymer can be made, the better the understanding of its inherent properties will be
obtained.
Polymer Properties
A polymer is a large molecule (macromolecule) composed of repeating structural units typically connected by covalent chemical
bonds. Polymers are common materials that are widely used in our lives. One of the most important features which distinguishes
most synthetic polymers from simple molecular compounds is the inability to assign an exact molar mass to a polymer. This is a
consequence of the fact that during the polymerization reaction the length of the chain formed is determined by several different
events, each of which have different reaction rates. Hence, the product is a mixture of chains of different length due to the random
nature of growth. In addition, some polymers are also branched (rather than linear) as a consequence of alternative reaction steps.
The molecular weight (Mw) and molecular weight distribution influences many of the properties of polymers:
Processability - the suitability of the polymer to physical processing.
Glass-transition temperature - refers to the transformation of a glass-forming liquid into a glass.
Solution viscosity - measure of the resistance of a fluid which is being deformed by either shear stress or tensile stress.
Hardness - a measure of how resistant a polymer is to various kinds of permanent shape change when a force is applied.
Melt viscosity - the rate of extrusion of thermoplastics through an orifice at a prescribed temperature and load.
Tear strength - a measure of the polymers resistance to tearing.
Tensile strength - as indicated by the maxima of a stress-strain curve and, in general, is the point when necking occurs upon
stretching a sample.
Stress-crack resistance - the formation of cracks in a polymer caused by relatively low tensile stress and environmental
conditions.
Brittleness - the liability of a polymer to fracture when subjected to stress.
Impact resistance - the relative susceptibility of polymers to fracture under stresses applied at high speeds.
Flex life - the number of cycles required to produce a specified failure in a specimen flexed in a prescribed manner.
Stress relaxation - describes how polymers relieve stress under constant strain.
Toughness - the resistance to fracture of a polymer when stressed.
Creep strain - the tendency of a polymer to slowly move or deform permanently under the influence of stresses.
Drawability - The ability of fiber-forming polymers to undergo several hundred percent permanent deformation, under load, at
ambient or elevated temperatures.
Compression - the result of the subjection of a polymer to compressive stress.
Fatigue - the failure by repeated stress.
Tackiness - the property of a polymer being adhesive or gummy to the touch.
Wear - the erosion of material from the polymer by the action of another surface.
Gas permeability - the permeability of gas through the polymer.
Consequently, it is important to understand how to determine the molecular weight and molecular weight distribution.
Molecular Weight
Simpler pure compounds contain the same molecular composition for the same species. For example, the molecular weight of any
sample of styrene will be the same (104.16 g/mol). In contrast, most polymers are not composed of identical molecules. The

2.2.12 https://chem.libretexts.org/@go/page/55839
molecular weight of a polymer is determined by the chemical structure of the monomer units, the lengths of the chains and the
extent to which the chains are interconnected to form branched molecules. Because virtually all polymers are mixtures of many
large molecules, we have to resort to averages to describe polymer molecular weight.
The polymers produced in polymerization reactions have lengths which are distributed according to a probability function which is
governed by the polymerization reaction. To define a particular polymer weight average, the average molecular weight Mavg is
defined by 2.2.24 Where Ni is the number of molecules with molecular weight Mi.
a
ΣNi M
i
Mavg   =   (2.2.24)
a−1
ΣNi M
i

There are several possible ways of reporting polymer molecular weight. Three commonly used molecular weight descriptions are:
the number average (Mn), weight average (Mw), and z-average molecular weight (Mz). All of three are applicable to different
constant a in 2.2.25 and are shown in Figure 2.2.19.

Figure 2.2.19 Distribution of molar masses for a polymer sample.


When a = 1, the number average molecular weight, 2.2.25 .
ΣNi Mi w
Mn, avg  =     =  (2.2.25)
ΣNi N

When a = 2, the number average molecular weight, 2.2.26 .


2
ΣNi M ΣNi Mi
i
Mn, avg  =     =  (2.2.26)
ΣNi Mi w

When a = 2, the number average molecular weight, 2.2.27 .


3 2
ΣNi M ΣNi M
i i
Mn, avg  =     =  (2.2.27)
2
ΣNi M ΣNi Mi
i

Bulk properties weight average molecular weight, Mw is the most useful one, because it fairly accounts for the contributions of
different sized chains to the overall behavior of the polymer, and correlates best with most of the physical properties of interest.
There are various methods published to detect these three different primary average molecular weights respectively. For instance, a
colligative method, such as osmotic pressure, effectively calculates the number of molecules present and provides a number
average molecular weight regardless of their shape or size of polymers. The classical van’t Hoff equation for the osmotic pressure
of an ideal, dilute solution is shown in 2.2.28 .
π RT
  =  (2.2.28)
c Mn

The weight average molecular weight of a polymer in solution can be determined by either measuring the intensity of light
scattered by the solution or studying the sedimentation of the solute in an ultracentrifuge. From light scattering method which is
depending on the size rather than the number of molecules, weight average molecular weight is obtained. This work requires
concentration fluctuations which are the main source of the light scattered by a polymer solution. The intensity of the light
scattering of polymer solution is often expressed by its turbidity τ which is given in Rayleigh’s law in 2.2.29 . Where iθ is scattered
intensity at only one angle θ, r is the distance from the scattering particle to the detection point, and I0 is the incident intensity.

2.2.13 https://chem.libretexts.org/@go/page/55839
2
16πiΘ r
τ  =   (2.2.29)
2
3 I0 (1 + cos Θ)

The intensity scattered by molecules (Ni) of molecular weight (Mi) is proportional to NiMi2. Thus, the total light scattered by all
molecules is described in 2.2.30 , where c is the total weight of the sample ∑NiMi.
2
π ΣNi M
i
    =  MW , avg (2.2.30)
c ΣNi Mi

Poly-disperse index (PDI)


The polydispersity index (PDI), is a measure of the distribution of molecular mass in a given polymer sample. As shown in Figure
2.2.19, it is the result of the definitions that Mw ≥ Mn. The equality of Mw and Mn would correspond with a perfectly uniform

(monodisperse) sample. The ratio of these average molecular weights is often used as a guide to the dispersity of the chain lengths
in a polymer sample. The greater Mw / Mn is, the greater the dispersity is.
The properties of a polymer sample are strongly dependent on the way in which the weights of the individual molecules are
distributed about the average. The ratio Mw/Mn gives sufficient information to characterize the distribution when the mathematical
form of the distribution curve is known.
Generally, the narrow molecular weight distribution materials are the models for much of work aimed at understanding the
materials’ behaviors. For example, polystyrene and its block copolymer polystyrene-b-polyisoprene have quite narrow distribution.
As a result, narrow molecular weight distribution materials are a necessary requirement when people study their behavior, such as
self-assembly behavior for block copolymer. Nonetheless, there are still lots of questions for scientists to explore the influence of
polydispersity. For example, research on self-assembly which is one of the interesting fields in polymer science shows that we
cannot throw polydispersity away.
Setup of SEC Equipment
In SEC, sample components migrate through the column at different velocities and elute separately from the column at different
times. In liquid chromatography and gas chromatography, as a solute moves along with the carrier fluid, it is at times held back
either by surface of the column packing, by stationary phase, or by both. Unlike gas chromatography (GC) and liquid
chromatography (LC), molecular size, or more precisely, molecular hydrodynamic volume governs the separation process of SEC,
not varied by the type of mobile phase. The smallest molecules are able to penetrate deeply into pores whereas the largest
molecules are excluded by the smaller pore sizes. Figure 2.2.20 shows the regular instrumental setup of SEC.

Figure 2.2.20 Regular instrumentation for size exclusion chromatography (SEC).


The properties of mobile phase are still important in that it is supposed to be strong affinity to stationary phase and be a good
solubility to samples. The purpose of well soluble of sample is to make the polymer be perfect coil suspending in solution. Thus, as
a mixture of solutes of different size passes through a column packed with porous particles. As shown in Figure 2.2.21, it clearly
depicts the general idea for size separation by SEC. the main setup of SEC emphasizes three concepts: stationary phase (column),
mobile phase (solvent) and sample preparation.

2.2.14 https://chem.libretexts.org/@go/page/55839
Figure 2.2.21 Development and detection of size separation by SEC. Adapted from A. M. Striegel, W. W. Yau, J. J. Kirkland, and
D. D. Bly. Modern Size-Exclusion Liquid Chromatography- Practice of Gel Permeation and Gel Filtration Chromatography, 2nd
Edition. Hoboken. N.J. (2009).
Solvent Selection
Solvent selection for SEC involves a number if considerations, such as convenience, sample type, column packing, operating
variables, safety, and purity.
For samples concern, the solvents used for mobile phase of SEC are limited to those follows following criteria:
The solvent must dissolve the sample completely.
The solvent has different properties with solute in the eluent: typically with solvent refractive index (RI) different from the
sample RI by ± 0.05 unit of more, or more than 10% of incident energy for UV detector.
The solvent must not degrade the sample during use. Otherwise, the viscosity of eluent will gradually increase over times.
The solvent is not corrosive to any components of the equipment.
Therefore, several solvents are qualified to be solvents such as THF, chlorinated hydrocarbons (chloroform, methylene chloride,
dichloroethane, etc), aromatic hydrocarbons (benzene, toluene , trichlorobenzene, etc).
Normally, high purity of solvent (HPLC-grade) is recommended. The reasons are to avoid suspended particulates that may abrade
the solvent pumping system or cause plugging of small-particle columns, to avoid impurities that may generate baseline noise, and
to avoid impurities that are concentrated due to evaporation of solvent.
Column Selection
Column selection of SEC depends mainly on the desired molecular weight range of separation and the nature of the solvents and
samples. Solute molecules should be separated solely by the size of gels without interaction of packing materials. Better column
efficiencies and separations can be obtained with small particle packing in columns and high diffusion rates for sample solutes.
Furthermore, optimal performance of an SEC packing materials involves high resolution and low column backpressure.
Compatible solvent and column must be chosen because, for example, organic solvent is used to swell the organic column packing
and used to dissolve and separate the samples.
It is convenient that columns are now usually available from manufacturers regarding the various types of samples. They provide
the information such as maximum tolerant flow rates, backpressure tolerances, recommended sample concentration, and injection
volumes, etc. Nonetheless, users have to notice a few things upon using columns:
Vibration and extreme temperatures should be avoided because these will post irreversible damage on columns.
For aqueous mobile phase, it is unwise to allow the extreme pH solutions staying in the columns for a long period of time.
The columns should be stored with some neat organic mobile phase, or aqueous mobile phase with pH range 2 - 8 to prevent
degradation of packing when not in use.
Sample Preparation

2.2.15 https://chem.libretexts.org/@go/page/55839
The sample solutions are supposed to be prepared in dilute concentration (less than 2 mg/mL) for several concerns. For polymer
samples, samples must be dissolved in the solvent same as used for mobile phase except some special cases. A good solvent can
dissolve a sample in any proportion in a range of temperatures. It is a slow process for dissolution because the rate determining step
is solvent diffusion into polymers to produce swollen gels. Then, gradual disintegration of gels makes sample-solvent mixture truly
become solution. Agitation and warming the mixtures are useful methods to speed up sample preparation.
It is recommended to filter the sample solutions before injecting into columns or storing in sample vials in order to get rid of
clogging and excessively high pressure problems. If unfortunately the excessively high pressure or clogging occur due to higher
concentration of sample solution, raising the column temperature will reduce the viscosity of the mobile phase, and may be helpful
to redissolve the precipitated or adsorbed solutes in the column. Back flushing of the columns should only be used as the last resort.
Analysis of SEC Data

The size exclusion separation mechanism is based on the effective hydrodynamic volume of the molecule, not the molecular
weight, and therefore the system must be calibrated using standards of known molecular weight and homogeneous chemical
composition. Then, the curve of sample is used to compare with calibration curve and obtain information relative to standards. The
further step is required to covert relative molecular weight into absolute molecular weight of a polymer.
Calibration
The purpose of calibration in SEC is to define the relationship between molecular weight and retention volume/time in the chosen
permeation range of column set and to calculate the relative molecular weight to standard molecules. There are several calibration
methods are commonly employed in modern SEC: direct standard calibration, poly-disperse standard calibration, universal
calibration.
The most commonly used calibration method is direct standard calibration. In the direct standard calibration method, narrowly
distributed standards of the same polymer being analyzed are used. Normally, narrow-molecular weight standards commercially
available are polystyrene (PS). The molecular weight of standards are measured originally by membrane osmometry for number-
average molecular weight, and by light scattering for weight-average molecular weight as described above. The retention volume at
the peak maximum of each standard is equated with its stated molecular weight.

Figure 2.2.22 Calibration curve for a size-exclusion.


Relative Mw versus absolute Mw
The molecular weight and molecular weight distribution can be determined from the calibration curves as described above. But as
the relationship between molecular weight and size depends on the type of polymer, the calibration curve depends on the polymer
used, with the result that true molecular weight can only be obtained when the sample is the same type as calibration standards. As
Figure 2.2.23 depicted, large deviations from the true molecular weight occur in the instance of branched samples because the
molecular density of these is higher than in the linear chains.

2.2.16 https://chem.libretexts.org/@go/page/55839
Figure 2.2.23 SEC elution of linear and branched samples of similar hydrodynamic volumes, but different molecular weights. S.
Mori, and H. G. Barth. Size Exclusion Chromatography, Springer, New York. (1999).
Light-scattering detector is now often used to overcome the limitations of conventional SEC. These signals depend only on
concentration, not on molecular weight or polymer size. For instance, for LS detector, 2.2.31 applies:
2
LS Signal  =  KLS ⋅ (dn/dc ) ⋅ MW ⋅ c (2.2.31)

Where KLS is an apparatus-specific sensitivity constant, dn/dc is the refractive index increment and c is concentration. Therefore,
accurate molecular weight can be determined while the concentration of the sample is known without calibration curve.
A Practical Example
The syntheses of poly(3-hexylthiophene) are well developed during last decade. It is an attractive polymer due to its potential as
electronic materials. Due to its excellent charge transport performances and high solubility, several studies discuss its further
improvement such as making block copolymer even triblock copolymer. The details are not discussed here. However, the
importance of molecular weight and molecular weight distribution is still critical.
As shown in Figure 2.2.24, they studied the mechanism of chain-growth polymerization and successfully produced low
polydispersity P3HT. The figure also demonstrates that the molecule with larger molecular size/ or weight elutes out of the column
earlier than those which has smaller molecular weight.
The real molecular weight of P3HT is smaller than the molecular weight relative to polystyrene. In this case, the backbone of P3HT
is harder compared with polystyrenes’ backbone because of the position of aromatic groups. It results in less flexibility. We can
briefly judge the authentic molecular weight of the synthetic polymer according to its molecular structure.

Figure 2.2.24 Synthesis of a well-defined poly(3-hexylthiphene) (HT-P3HT).

Figure 2.2.25 GPC profiles of HT-P3HT obtained by the polymerization. Adapted from R. Miyakoshi, A. Yokoyama, and T.
Yokozawa, Macromol. Rapid Commun., 2004, 25, 1663.

2.2: Molecular Weight Determination is shared under a CC BY 4.0 license and was authored, remixed, and/or curated by Pavan M. V. Raja &
Andrew R. Barron via source content that was edited to conform to the style and standards of the LibreTexts platform; a detailed edit history is
available upon request.

2.2.17 https://chem.libretexts.org/@go/page/55839
2.3: BET Surface Area Analysis of Nanoparticles
Introduction
In the past few years, nanotechnology research has expanded out of the chemistry department and into the fields of medicine,
energy, aerospace and even computing and information technology. With bulk materials, the surface area to volume is insignificant
in relation to the number of atoms in the bulk, however when the particles are only 1 to 100 nm across, different properties begin to
arise. For example, commercial grade zinc oxide has a surface area range of 2.5 to 12 m2/g while nanoparticle zinc oxide can have
surface areas as high as 54 m2/g . The nanoparticles have superior UV blocking properties when compared to the bulk material,
making them useful in applications such as sunscreen. Many useful properties of nanoparticles rise from their small size, making it
very important to be able to determine their surface area.

Overview of BET Theory


The BET theory was developed by Stephen Brunauer (Figure 2.3.1 ), Paul Emmett (Figure 2.3.2 ), and Edward Teller (Figure
2.3.3 ) in 1938. The first letter of each publisher’s surname was taken to name this theory. The BET theory was an extension of the

Langmuir theory, developed by Irving Langmuir (Figure 2.3.4 ) in 1916.

Figure 2.3.1 Hungarian chemist Stephen Brunauer (1903-1986). Adapted from K. S. Sing, Langmuir, 1987, 3, 2 (Copyright:
American Chemical Society)

Figure 2.3.2 American chemical engineer Paul H. Emmett (1900 - 1985). Adapted from B.H. Davis, J. Phys. Chem., 1986, 90,
4702 (Copyright: American Chemical Society).

2.3.1 https://chem.libretexts.org/@go/page/55840
Figure 2.3.3 Hungarian born theoretical physicist Edward Teller (1908 – 2003) shown in 1958 as the director of Lawrence
Livermore National Laboratory was known as "the father of the hydrogen bomb".

Figure 2.3.4 American chemist and physicist Irving Langmuir (1881 - 1957). Adapted from J. Chem. Educ., 1933, 10, 65
(Copyright: American Chemical Society).
The Langmuir theory relates the monolayer adsorption of gas molecules (Figure 2.3.5 ), also called adsorbates, onto a solid surface
to the gas pressure of a medium above the solid surface at a fixed temperature to 2.3.1 , where θ is the fractional cover of the
surface, P is the gas pressure and α is a constant.
α⋅P
Θ  =   (2.3.1)
1  +  (α ⋅ P )

Figure 2.3.5 Schematic of the adsorption of gas molecules onto the surface of a sample showing (a) the monolayer adsorption
model assumed by the Langmuir theory and (b) s the multilayer adsorption model assumed by the BET theory.
The Langmuir theory is based on the following assumptions:
All surface sites have the same adsorption energy for the adsorbate, which is usually argon, krypton or nitrogen gas. The surface
site is defined as the area on the sample where one molecule can adsorb onto.
Adsorption of the solvent at one site occurs independently of adsorption at neighboring sites.
Activity of adsorbate is directly proportional to its concentration.
Adsorbates form a monolayer.

2.3.2 https://chem.libretexts.org/@go/page/55840
Each active site can be occupied only by one particle.
The Langmuir theory has a few flaws that are addressed by the BET theory. The BET theory extends the Langmuir theory to
multilayer adsorption (Figure 2.3.1 ) with three additional assumptions:
Gas molecules will physically adsorb on a solid in layers infinitely.
The different adsorption layers do not interact.
The theory can be applied to each layer.
How does BET Work?
Adsorption is defined as the adhesion of atoms or molecules of gas to a surface. It should be noted that adsorption is not confused
with absorption, in which a fluid permeates a liquid or solid. The amount of gas adsorbed depends on the exposed surface are but
also on the temperature, gas pressure and strength of interaction between the gas and solid. In BET surface area analysis, nitrogen
is usually used because of its availability in high purity and its strong interaction with most solids. Because the interaction between
gaseous and solid phases is usually weak, the surface is cooled using liquid N2 to obtain detectable amounts of adsorption. Known
amounts of nitrogen gas are then released stepwise into the sample cell. Relative pressures less than atmospheric pressure is
achieved by creating conditions of partial vacuum. After the saturation pressure, no more adsorption occurs regardless of any
further increase in pressure. Highly precise and accurate pressure transducers monitor the pressure changes due to the adsorption
process. After the adsorption layers are formed, the sample is removed from the nitrogen atmosphere and heated to cause the
adsorbed nitrogen to be released from the material and quantified. The data collected is displayed in the form of a BET isotherm,
which plots the amount of gas adsorbed as a function of the relative pressure. There are five types of adsorption isotherms possible.
Type I Isotherm
Type I is a pseudo-Langmuir isotherm because it depicts monolayer adsorption (Figure 2.3.6 ). A type I isotherm is obtained when
P/Po < 1 and c > 1 in the BET equation, where P/Po is the partial pressure value and c is the BET constant, which is related to the
adsorption energy of the first monolayer and varies from solid to solid. The characterization of microporous materials, those with
pore diameters less than 2 nm, gives this type of isotherm.

Figure 2.3.6 The isotherm plots the volume of gas adsorbed onto the surface of the sample as pressure increases. Adapted from S.
Brunauer L. S. Deming, W. E. Deming, and E. Teller, J. Am. Chem. Soc., 1940, 62, 1723.
Type II Isotherm
A type II isotherm (Figure 2.3.7 ) is very different than the Langmuir model. The flatter region in the middle represents the
formation of a monolayer. A type II isotherm is obtained when c > 1 in the BET equation. This is the most common isotherm
obtained when using the BET technique. At very low pressures, the micropores fill with nitrogen gas. At the knee, monolayer
formation is beginning and multilayer formation occurs at medium pressure. At the higher pressures, capillary condensation occurs.

Figure 2.3.7 The isotherm plots the volume of gas adsorbed onto the surface of the sample as pressure increases. Adapted from S.
Brunauer, L. S. Deming, W. E. Deming, and E. Teller, J. Am. Chem. Soc., 1940, 62, 1723.
Type III Isotherm
A type III isotherm (Figure 2.3.8 ) is obtained when the c < 1 and shows the formation of a multilayer. Because there is no
asymptote in the curve, no monolayer is formed and BET is not applicable.

2.3.3 https://chem.libretexts.org/@go/page/55840
Figure 2.3.8 Brunauer, L. S. Deming, W. E. Deming, and E. Teller, J. Am. Chem. Soc., 1940, 62, 1723.

Type IV Isotherm
Type IV isotherms (Figure 2.3.9 ) occur when capillary condensation occurs. Gases condense in the tiny capillary pores of the solid
at pressures below the saturation pressure of the gas. At the lower pressure regions, it shows the formation of a monolayer followed
by a formation of multilayers. BET surface area characterization of mesoporous materials, which are materials with pore diameters
between 2 - 50 nm, gives this type of isotherm.

Figure 2.3.9 Brunauer, L. S. Deming, W. E. Deming, and E. Teller, J. Am. Chem. Soc., 1940, 62, 1723.
Type V Isotherm
Type V isotherms (Figure 2.3.10 ) are very similar to type IV isotherms and are not applicable to BET.

Figure 2.3.10 Brunauer L. S. Deming, W. E. Deming, and E. Teller, J. Am. Chem. Soc., 1940, 62, 1723.

Calculations
The BET Equation, 2.3.2 , uses the information from the isotherm to determine the surface area of the sample, where X is the
weight of nitrogen adsorbed at a given relative pressure (P/Po), Xm is monolayer capacity, which is the volume of gas adsorbed at
standard temperature and pressure (STP), and C is constant. STP is defined as 273 K and 1 atm.
1 1 C −1 P
= + ( ) (2.3.2)
X[(P0 /P ) − 1] Xm C Xm C P0

Multi-point BET

Ideally five data points, with a minimum of three data points, in the P/P0 range 0.025 to 0.30 should be used to successfully
determine the surface area using the BET equation. At relative pressures higher than 0.5, there is the onset of capillary
condensation, and at relative pressures that are too low, only monolayer formation is occurring. When the BET equation is plotted,
the graph should be of linear with a positive slope. If such a graph is not obtained, then the BET method was insufficient in
obtaining the surface area.
The slope and y-intercept can be obtained using least squares regression.
The monolayer capacity Xm can be calculated with 2.3.3 .
Once Xm is determined, the total surface area St can be calculated with the following equation, where Lav is Avogadro’s
number, Am is the cross sectional area of the adsorbate and equals 0.162 nm2 for an absorbed nitrogen molecule, and Mv is the

2.3.4 https://chem.libretexts.org/@go/page/55840
molar volume and equals 22414 mL, 2.3.4 .
1 C −1
Xm   = = (2.3.3)
s  +  i Cs

Xm Lav Am
S  = (2.3.4)
Mv

Single point BET can also be used by setting the intercept to 0 and ignoring the value of C. The data point at the relative pressure
of 0.3 will match up the best with a multipoint BET. Single point BET can be used over the more accurate multipoint BET to
determine the appropriate relative pressure range for multi-point BET.
Sample Preparation and Experimental Setup
Prior to any measurement the sample must be degassed to remove water and other contaminants before the surface area can be
accurately measured. Samples are degassed in a vacuum at high temperatures. The highest temperature possible that will not
damage the sample’s structure is usually chosen in order to shorten the degassing time. IUPAC recommends that samples be
degassed for at least 16 hours to ensure that unwanted vapors and gases are removed from the surface of the sample. Generally,
samples that can withstand higher temperatures without structural changes have smaller degassing times. A minimum of 0.5 g of
sample is required for the BET to successfully determine the surface area.
Samples are placed in glass cells to be degassed and analyzed by the BET machine. Glass rods are placed within the cell to
minimize the dead space in the cell. Sample cells typically come in sizes of 6, 9 and 12 mm and come in different shapes. 6 mm
cells are usually used for fine powders, 9 mm cells for larger particles and small pellets and 12 mm are used for large pieces that
cannot be further reduced. The cells are placed into heating mantles and connected to the outgas port of the machine.
After the sample is degassed, the cell is moved to the analysis port (Figure 2.3.11 ). Dewars of liquid nitrogen are used to cool the
sample and maintain it at a constant temperature. A low temperature must be maintained so that the interaction between the gas
molecules and the surface of the sample will be strong enough for measurable amounts of adsorption to occur. The adsorbate,
nitrogen gas in this case, is injected into the sample cell with a calibrated piston. The dead volume in the sample cell must be
calibrated before and after each measurement. To do that, helium gas is used for a blank run, because helium does not adsorb onto
the sample.

Figure 2.3.11 Schematic representation of the BET instrument. The degasser is not shown.
Shortcomings of BET

The BET technique has some disadvantages when compared to NMR, which can also be used to measure the surface area of
nanoparticles. BET measurements can only be used to determine the surface area of dry powders. This technique requires a lot of
time for the adsorption of gas molecules to occur. A lot of manual preparation is required.

The Surface Area Determination of Metal-Organic Frameworks


The BET technique was used to determine the surface areas of metal-organic frameworks (MOFs), which are crystalline
compounds of metal ions coordinated to organic molecules. Possible applications of MOFs, which are porous, include gas
purification and catalysis. An isoreticular MOF (IRMOF) with the chemical formula Zn4O(pyrene-1,2-dicarboxylate)3 (Figure
2.3.12 ) was used as an example to see if BET could accurately determine the surface area of microporous materials. The predicted

2.3.5 https://chem.libretexts.org/@go/page/55840
surface area was calculated directly from the geometry of the crystals and agreed with the data obtained from the BET isotherms.
Data was collected at a constant temperature of 77 K and a type II isotherm (Figure 2.3.13 ) was obtained.

Figure 2.3.12 The structure of catenated IRMOF-13. Orange and yellow represent non-catenated pore volumes. Green represents
catenated pore volume.

Figure 2.3.13 The BET isotherms of the zeolites and metal-organic frameworks. IRMOF-13 is symbolized by the black triangle
and red line. Adapted from Y.S. Bae, R.Q. Snurr, and O. Yazaydin, Langmuir, 2010, 26, 5478.
The isotherm data obtained from partial pressure range of 0.05 to 0.3 is plugged into the BET equation, 2.3.2 , to obtain the BET
plot (Figure 2.3.14 ).

Figure 2.3.14 BET plot of IRMOF-13 using points collected at the pressure range 0.05 to 0.3. The equation of the best-fit line and
R2 value are shown. Adapted from Y.S. Bae, R.Q. Snurr, and O. Yazaydin, Langmuir, 2010, 26, 5479.
Using 2.3.5 , the monolayer capactiy is determined to be 391.2 cm3/g.
1
Xm   = (2.3.5)
(2.66E  −  3)  +  (−5.212E  −  0.05)

Now that Xm is known, then 2.3.6 can be used to determine that the surface area is 1702.3 m2/g.
2 2 23
391.2c m ∗ 0.162nm ∗ 6.02 ∗ 10
S  = (2.3.6)
22.414 : L

2.3: BET Surface Area Analysis of Nanoparticles is shared under a CC BY 4.0 license and was authored, remixed, and/or curated by Pavan M. V.
Raja & Andrew R. Barron via source content that was edited to conform to the style and standards of the LibreTexts platform; a detailed edit
history is available upon request.

2.3.6 https://chem.libretexts.org/@go/page/55840
2.4: Dynamic Light Scattering
Dynamic light scattering (DLS), which is also known as photon correlation spectroscopy (PCS) or quasi-elastic light scattering
(QLS), is a spectroscopy method used in the fields of chemistry, biochemistry, and physics to determine the size distribution of
particles (polymers, proteins, colloids, etc.) in solution or suspension. In the DLS experiment, normally a laser provides the
monochromatic incident light, which impinges onto a solution with small particles in Brownian motion. And then through the
Rayleigh scattering process, particles whose sizes are sufficiently small compared to the wavelength of the incident light will
diffract the incident light in all direction with different wavelengths and intensities as a function of time. Since the scattering
pattern of the light is highly correlated to the size distribution of the analyzed particles, the size-related information of the sample
could be then acquired by mathematically processing the spectral characteristics of the scattered light.

Figure 2.4.1 : Scheme of Rayleigh scattering.


Herein a brief introduction of basic theories of DLS will be demonstrated, followed by descriptions and guidance on the instrument
itself and the sample preparation and measurement process. Finally, data analysis of the DLS measurement, and the applications of
DLS as well as the comparison against other size-determine techniques will be shown and summarized.

DLS Theory
The theory of DLS can be introduced utilizing a model system of spherical particles in solution. According to the Rayleigh
scattering (Figure 2.4.1), when a sample of particles with diameter smaller than the wavelength of the incident light, each particle
will diffract the incident light in all directions, while the intensity I is determined by 2.4.1 , where I and λ is the intensity and
0

wavelength of the unpolarized incident light, R is the distance to the particle, θ is the scattering angel, n is the refractive index of
the particle, and r is the radius of the particle.
2 4 2 2
1  + cos θ 2π n   −  1
6
I   =  I0 ( ) ( ) r (2.4.1)
2 2
2R λ n   +  2

If that diffracted light is projected as an image onto a screen, it will generate a “speckle" pattern (Figure 2.4.2 ); the dark areas
represent regions where the diffracted light from the particles arrives out of phase interfering destructively and the bright area
represent regions where the diffracted light arrives in phase interfering constructively.

Figure 2.4.2 Typical speckle pattern. A photograph of an objective speckle pattern. This is the light field formed when a laser beam
was scattered from a plastic surface onto a wall. (Public Domain; Epzcaw).
In practice, particle samples are normally not stationary but moving randomly due to collisions with solvent molecules as described
¯
¯¯¯¯¯¯¯¯¯¯¯
¯
by the Brownian motion, 2.4.2, where (Δx) is the mean squared displacement in time t, and D is the diffusion constant, which is
2

related to the hydrodynamic radius a of the particle according to the Stokes-Einstein equation, 2.4.3 , where kB is Boltzmann
constant, T is the temperature, and μ is viscosity of the solution. Importantly, for a system undergoing Brownian motion, small
particles should diffuse faster than large ones.
¯
¯¯¯¯¯¯¯¯¯¯¯
¯
2
(Δx)   =  2Δt (2.4.2)

2.4.1 https://chem.libretexts.org/@go/page/55841
kB T
D  = (2.4.3)
6πμa

As a result of the Brownian motion, the distance between particles is constantly changing and this results in a Doppler shift
between the frequency of the incident light and the frequency of the scattered light. Since the distance between particles also affects
the phase overlap/interfering of the diffracted light, the brightness and darkness of the spots in the “speckle” pattern will in turn
fluctuate in intensity as a function of time when the particles change position with respect to each other. Then, as the rate of these
intensity fluctuations depends on how fast the particles are moving (smaller particles diffuse faster), information about the size
distribution of particles in the solution could be acquired by processing the fluctuations of the intensity of scattered light. Figure
2.4.3 shows the hypothetical fluctuation of scattering intensity of larger particles and smaller particles.

Figure 2.4.3 Hypothetical fluctuation of scattering intensity of larger particles and smaller particles.
In order to mathematically process the fluctuation of intensity, there are several principles/terms to be understood. First, the
intensity correlation function is used to describe the rate of change in scattering intensity by comparing the intensity I(t) at time t to
the intensity I(t + τ) at a later time (t + τ), and is quantified and normalized by 2.4.4 and 2.4.5 , where braces indicate averaging
over t.

G2 (τ ) =  ⟨I (t)I (t  +  τ )⟩ (2.4.4)

⟨I (t)I (t  +  τ )⟩
g2 (τ ) = (2.4.5)
2
⟨I (t)⟩

Second, since it is not possible to know how each particle moves from the fluctuation, the electric field correlation function is
instead used to correlate the motion of the particles relative to each other, and is defined by 2.4.6 and 2.4.7 , where E(t) and E(t +
τ) are the scattered electric fields at times t and t+ τ.
G1 (τ ) =  ⟨E(t)E(t  +  τ )⟩ (2.4.6)

⟨E(t)E(t  +  τ )⟩
g1 (τ ) = (2.4.7)
⟨E(t)E(t)⟩

For a monodisperse system undergoing Brownian motion, g1(τ) will decay exponentially with a decay rate Γ which is related by
Brownian motion to the diffusivity by 2.4.8 , 2.4.9 , and 2.4.10 , where q is the magnitude of the scattering wave vector and q2
reflects the distance the particle travels, n is the refraction index of the solution and θ is angle at which the detector is located.
−Γτ
g1 (τ ) =  e (2.4.8)

2
Γ  =   − Dq (2.4.9)

4πn Θ
q = sin (2.4.10)
λ 2

For a polydisperse system however, g1(τ) can no longer be represented as a single exponential decay and must be represented as a
intensity-weighed integral over a distribution of decay rates G(Γ) by 2.4.11 where G(Γ) is normalized, 2.4.12 .

−Γτ
g1 (τ ) = ∫ G(Γ)e dΓ (2.4.11)
0

2.4.2 https://chem.libretexts.org/@go/page/55841

∫ G(Γ)dΓ  =  1 (2.4.12)
0

Third, the two correlation functions above can be equated using the Seigert relationship based on the principles of Gaussian random
processes (which the scattering light usually is), and can be expressed as 2.4.13 , where β is a factor that depends on the
experimental geometry, and B is the long-time value of g2(τ), which is referred to as the baseline and is normally equal to 1. Figure
2.4.4 shows the decay of g2(τ) for small size sample and large size sample.

2
g2 (τ ) =  B  +  β[ g1 (τ )] (2.4.13)

Figure 2.4.4 Decay of g2(τ) for small size sample and large size sample. Malvern Instruments Ltd., Zetasizer Nano Series User
Manual, 2004. Copyright: Malvern Instruments Ltd. (2004).
When determining the size of particles in solution using DLS, g2(τ) is calculated based on the time-dependent scattering intensity,
and is converted through the Seigert relationship to g1(τ) which usually is an exponential decay or a sum of exponential decays.
The decay rate Γ is then mathematically determined (will be discussed in section ) from the g1(τ) curve, and the value of diffusion
constant D and hydrodynamic radius a can be easily calculated afterwards.

Experimental
Instrument of DLS
In a typical DLS experiment, light from a laser passes through a polarizer to define the polarization of the incident beam and then
shines on the scattering medium. When the sizes of the analyzed particles are sufficiently small compared to the wavelength of the
incident light, the incident light will scatters in all directions known as the Rayleigh scattering. The scattered light then passes
through an analyzer, which selects a given polarization and finally enters a detector, where the position of the detector defines the
scattering angle θ. In addition, the intersection of the incident beam and the beam intercepted by the detector defines a scattering
region of volume V. As for the detector used in these experiments, a phototube is normally used whose dc output is proportional to
the intensity of the scattered light beam. Figure 2.4.5 shows a schematic representation of the light-scattering experiment.

Figure 2.4.5 A schematic representation of the light-scattering experiment. (CC BY-NC; Ümit Kaya via LibreTexts).
In modern DLS experiments, the scattered light spectral distribution is also measured. In these cases, a photomultiplier is the main
detector, but the pre- and postphotomultiplier systems differ depending on the frequency change of the scattered light. The three
different methods used are filter (f > 1 MHz), homodyne (f > 10 GHz), and heterodyne methods (f < 1 MHz), as schematically
illustrated in Figure 2.4.6. Note that that homodyne and heterodyne methods use no monochromator of “filter” between the

2.4.3 https://chem.libretexts.org/@go/page/55841
scattering cell and the photomultiplier, and optical mixing techniques are used for heterodyne method. shows the schematic
illustration of the various techniques used in light-scattering experiments.

Figure 2.4.6 Schematic illustration of the various techniques used in light-scattering experiments: (a) filter methods; (b) homodyne;
(c) heterodyne. B. J. Berne and R. Pecora, Dynamic Light Scattering: With Applications to Chemistry, Biology, and Physics, Dover,
Mineola, NY (2000). Copyright: Dover Publications (2000).
As for an actual DLS instrument, take the Zetasizer Nano (Malvern Instruments Ltd.) as an example (Figure 2.4.7), it actually
looks like nothing other than a big box, with components of power supply, optical unit (light source and detector), computer
connection, sample holder, and accessories. The detailed procedure of how to use the DLS instrument will be introduced
afterwards.

Figure 2.4.7 Photo of a DLS instrument at Rice University (Zetasizer Nano, Malvern Instruments Ltd.).

Sample Preparation
Although different DLS instruments may have different analysis ranges, we are usually looking at particles with a size range of nm
to μm in solution. For several kinds of samples, DLS can give results with rather high confidence, such as monodisperse
suspensions of unaggregated nanoparticles that have radius > 20 nm, or polydisperse nanoparticle solutions or stable solutions of
aggregated nanoparticles that have radius in the 100 - 300 nm range with a polydispersity index of 0.3 or below. For other more
challenging samples such as solutions containing large aggregates, bimodal solutions, very dilute samples, very small
nanoparticles, heterogeneous samples, or unknown samples, the results given by DLS could not be really reliable, and one must be
aware of the strengths and weaknesses of this analytical technique.
Then, for the sample preparation procedure, one important question is how much materials should be submit, or what is the optimal
concentration of the solution. Generally, when doing the DLS measurement, it is important to submit enough amount of material in
order to obtain sufficient signal, but if the sample is overly concentrated, then light scattered by one particle might be again
scattered by another (known as multiple scattering), and make the data processing less accurate. An ideal sample submission for
DLS analysis has a volume of 1 – 2 mL and is sufficiently concentrated as to have strong color hues, or opaqueness/turbidity in the
case of a white or black sample. Alternatively, 100 - 200 μL of highly concentrated sample can be diluted to 1 mL or analyzed in a
low-volume microcuvette.
In order to get high quality DLS data, there are also other issues to be concerned with. First is to minimize particulate
contaminants, as it is common for a single particle contaminant to scatter a million times more than a suspended nanoparticle, by
using ultra high purity water or solvents, extensively rinsing pipettes and containers, and sealing sample tightly. Second is to filter
the sample through a 0.2 or 0.45 μm filter to get away of the visible particulates within the sample solution. Third is to avoid probe
sonication to prevent the particulates ejected from the sonication tip, and use the bath sonication in stead.

2.4.4 https://chem.libretexts.org/@go/page/55841
Measurement
Now that the sample is readily prepared and put into the sample holder of the instrument, the next step is to actually do the DLS
measurement. Generally the DLS instrument will be provided with software that can help you to do the measurement rather easily,
but it is still worthwhile to understand the important parameters used during the measurement.
Firstly, the laser light source with an appropriate wavelength should be selected. As for the Zetasizer Nano series (Malvern
Instruments Ltd.), either a 633 nm “red” laser or a 532 nm “green” laser is available. One should keep in mind that the 633 nm laser
is least suitable for blue samples, while the 532 nm laser is least suitable for red samples, since otherwise the sample will just
absorb a large portion of the incident light.
Then, for the measurement itself, one has to select the appropriate stabilization time and the duration time. Normally, longer
striation/duration time can results in more stable signal with less noises, but the time cost should also be considered. Another
important parameter is the temperature of the sample, as many DLS instruments are equipped with the temperature-controllable
sample holders, one can actually measure the size distribution of the data at different temperature, and get extra information about
the thermal stability of the sample analyzed.
Next, as is used in the calculation of particle size from the light scattering data, the viscosity and refraction index of the solution are
also needed. Normally, for solutions with low concentration, the viscosity and refraction index of the solvent/water could be used
as an approximation.
Finally, to get data with better reliability, the DLS measurement on the same sample will normally be conducted multiple times,
which can help eliminate unexpected results and also provide additional error bar of the size distribution data.

Data Analysis
Although size distribution data could be readily acquired from the software of the DLS instrument, it is still worthwhile to know
about the details about the data analysis process.

Cumulant method
As is mentioned in the Theory portion above, the decay rate Γ is mathematically determined from the g1(τ) curve; if the sample
solution is monodispersed, g1(τ) could be regard as a single exponential decay function e-Γτ, and the decay rate Γ can be in turn
easily calculated. However, in most of the practical cases, the sample solution is always polydispersed, g1(τ) will be the sum of
many single exponential decay functions with different decay rates, and then it becomes significantly difficult to conduct the fitting
process.
There are however, a few methods developed to meet this mathematical challenge: linear fit and cumulant expansion for mono-
modal distribution, exponential sampling and CONTIN regularization for non-monomodal distribution. Among all these
approaches, cumulant expansion is most common method and will be illustrated in detail in this section.
Generally, the cumulant expansion method is based on two relations: one between g1(τ) and the moment-generating function of the
distribution, and one between the logarithm of g1(τ) and the cumulant-generating function of the distribution.
To start with, the form of g1(τ) is equivalent to the definition of the moment-generating function M(-τ, Γ) of the distribution G(Γ),
2.4.14 .


−Γτ
g1 (τ ) =   ∫ G(Γ)e dΓ  =  M (−τ , Γ) (2.4.14)
0

The mth moment of the distribution mm(Γ) is given by the mth derivative of M(-τ, Γ) with respect to τ, 2.4.15 .

m −Γτ
mm (Γ) =   ∫ G(Γ)Γ e dΓ∣−τ=0 (2.4.15)
0

Similarly, the logarithm of g1(τ) is equivalent to the definition of the cumulant-generating function K(-τ, Γ), EQ, and the mth
cumulant of the distribution km(Γ) is given by the mth derivative of K(-τ, Γ) with respect to τ, 2.4.16 and 2.4.17 .
ln g1 (τ ) = ln M (−τ , Γ)  =  K(−τ , Γ) (2.4.16)

m
d K(−τ , Γ)
km (Γ) = ∣−τ=0 (2.4.17)
m
d(−τ )

2.4.5 https://chem.libretexts.org/@go/page/55841
By making use of that the cumulants, except for the first, are invariant under a change of origin, the km(Γ) could be rewritten in
terms of the moments about the mean as 2.4.18 , 2.4.19 , 2.4.20 , and 2.4.21 where here μm are the moments about the mean,
defined as given in 2.4.22 .

k1 (τ ) =   ∫ G(Γ)ΓdΓ = Γ̄ (2.4.18)
0

k2 (τ ) =  μ2 (2.4.19)

k3 (τ ) =  μ3 (2.4.20)

2
k4 (τ ) =  μ4 − 3 μ ⋯ (2.4.21)
2


m
μm   =   ∫ G(Γ)(Γ  −  Γ̄) dΓ (2.4.22)
0

Based on the Taylor expansion of K(-τ, Γ) about τ = 0, the logarithm of g1(τ) is given as 2.4.23 .
k2 2
k3 3
k4 4
¯
ln g1 (τ ) =  K(−τ , Γ) =   − Γτ   + τ  − τ  + τ ⋯ (2.4.23)
2! 3! 4!

Importantly, if look back at the Seigert relationship in the logarithmic form, 2.4.24 .
ln(g2 (τ ) − B) = lnβ  +  2ln g1 (τ ) (2.4.24)

The measured data of g2(τ) could be fitted with the parameters of km using the relationship of 2.4.25 , where Γ ¯
(k1), k2, and k3
describes the average, variance, and skewness (or asymmetry) of the decay rates of the distribution, and polydispersity index
k2
γ  =  
¯
2
is used to indicate the width of the distribution. And parameters beyond k3 are seldom used to prevent overfitting the data.
Γ

Finally, the size distribution can be easily calculated from the decay rate distribution as described in theory section previously.
Figure 2.4.6 shows an example of data fitting using the cumulant method.
k2 k3
¯ 2 3
ln(g2 (τ ) − B) =]lnβ  +  2(−Γτ   + τ  − τ ⋯) (2.4.25)
2! 3!

Figure 2.4.8 : Sample data taken for POPC vesicles formed by extrusion through polycarbonate membranes. The curve through the
data is a fit of EQ to the data. The dashed curve shows the weighted residuals: the difference of the fit from the data divided by the
uncertainty in each point. B. J. Frisken, Appl. Optics, 2001, 40, 4087. Copyright: Optical Society of America (2001).
When using the cumulant expansion method however, one should keep in mind that it is only suitable for monomodal distributions
(Gaussian-like distribution centered about the mean), and for non-monomodal distributions, other methods like exponential
sampling and CONTIN regularization should be applied instead.

Three Index of Size Distribution


Now that the size distribution is able to be acquired from the fluctuation data of the scattered light using cumulant expansion or
other methods, it is worthwhile to understand the three kinds of distribution index usually used in size analysis: number weighted
distribution, volume weighted distribution, and intensity weighted distribution.
First of all, based on all the theories discussed above, it should be clear that the size distribution given by DLS experiments is the
intensity weighted distribution, as it is always the intensity of the scattering that is being analyzed. So for intensity weighted
distribution, the contribution of each particle is related to the intensity of light scattered by that particle. For example, using
Rayleigh approximation, the relative contribution for very small particles will be proportional to a6.

2.4.6 https://chem.libretexts.org/@go/page/55841
For number weighted distribution, given by image analysis as an example, each particle is given equal weighting irrespective of its
size, which means proportional to a0. This index is most useful where the absolute number of particles is important, or where high
resolution (particle by particle) is required.
For volume weighted distribution, given by laser diffraction as an example, the contribution of each particle is related to the
volume of that particle, which is proportional to a3. This is often extremely useful from a commercial perspective as the
distribution represents the composition of the sample in terms of its volume/mass, and therefore its potential money value.
When comparing particle size data for the same sample represented using different distribution index, it is important to know that
the results could be very different from number weighted distribution to intensity weighted distribution. This is clearly illustrated in
the example below (Figure 2.4.9 ), for a sample consisting of equal numbers of particles with diameters of 5 nm and 50 nm. The
number weighted distribution gives equal weighting to both types of particles, emphasizing the presence of the finer 5 nm particles,
whereas the intensity weighted distribution has a signal one million times higher for the coarser 50 nm particles. The volume
weighted distribution is intermediate between the two.

Figure 2.4.9 Example of number, volume and intensity weighted particle size distributions for the same sample. Malvern
Instruments Ltd., A Basic Guide to Particle Characterization, 2012. Copyright: Malvern Instrument Ltd. (2012).
Furthermore, based on the different orders of correlation between the particle contribution and the particle size a, it is possible to
convert particle size data from one type of distribution to another type of distribution, and that is also why the DLS software can
also give size distributions in three different forms (number, volume, and intensity), where the first two kinds are actually deducted
from the raw data of intensity weighted distribution.

An Example of an Application
As the DLS method could be used in many areas towards size distribution such as polymers, proteins, metal nanoparticles, or
carbon nanomaterials, here gives an example about the application of DLS in size-controlled synthesis of monodisperse gold
nanoparticles.
The size and size distribution of gold particles are controlled by subtle variation of the structure of the polymer, which is used to
stabilize the gold nanoparticles during the reaction. These variations include monomer type, polymer molecular weight, end-group
hydrophobicity, end-group denticity, and polymer concentration; a total number of 88 different trials have been conducted based on
these variations. By using the DLS method, the authors are able to determine the gold particle size distribution for all these trials
rather easily, and the correlation between polymer structure and particle size can also be plotted without further processing the data.
Although other sizing techniques such as UV-V spectroscopy and TEM are also used in this paper, it is the DLS measurement that
provides a much easier and reliable approach towards the size distribution analysis.

Comparison with TEM and AFM


Since DLS is not the only method available to determine the size distribution of particles, it is also necessary to compare DLS with
the other common-used general sizing techniques, especially TEM and AFM.
First of all, it has to be made clear that both TEM and AFM measure particles that are deposited on a substrate (Cu grid for TEM,
mica for AFM), while DLS measures particles that are dispersed in a solution. In this way, DLS will be measuring the bulk phase
properties and give a more comprehensive information about the size distribution of the sample. And for AFM or TEM, it is very
common that a relatively small sampling area is analyzed, and the size distribution on the sampling area may not be the same as the
size distribution of the original sample depending on how the particles are deposited.
On the other hand however, for DLS, the calculating process is highly dependent on the mathematical and physical assumptions
and models, which is, monomodal distribution (cumulant method) and spherical shape for the particles, the results could be
inaccurate when analyzing non-monomodal distributions or non-spherical particles. Yet, since the size determining process for

2.4.7 https://chem.libretexts.org/@go/page/55841
AFM or TEM is nothing more than measuring the size from the image and then using the statistic, these two methods can provide
much more reliable data when dealing with “irregular” samples.
Another important issue to consider is the time cost and complication of size measurement. Generally speaking, the DLS
measurement should be a much easier technique, which requires less operation time and also cheaper equipment. And it could be
really troublesome to analysis the size distribution data coming out from TEM or AFM images without specially programmed
software.
In addition, there are some special issues to consider when choosing size analysis techniques. For example, if the originally sample
is already on a substrate (synthesized by the CVD method), or the particles could not be stably dispersed within solution,
apparently the DLS method is not suitable. Also, when the particles tend to have a similar imaging contrast against the substrate
(carbon nanomaterials on TEM grid), or tend to self-assemble and aggregate on the surface of the substrate, the DLS approach
might be a better choice.
In general research work however, the best way to do size distribution analysis is to combine these analyzing methods, and get
complimentary information from different aspects. One thing to keep in mind, since the DLS actually measures the hydrodynamic
radius of the particles, the size from DLS measurement is always larger than the size from AFM or TEM measurement. As a
conclusion, the comparison between DLS and AFM/TEM is shown in Table 2.4.1.
Table 2.4.1 Comparison between DLS, AFM, and TEM.
DLS AFM/TEM

Sample Preparation Solution Substrate

Measurement Easy Difficult

Sampling Bulk Small area

Shape of Particles Sphere No Requirement

Polydispersity Low No Requirement

Size Range nm to um nm to um

Size Info. Hydrodynamic radius Physical size

Conclusion
In general, relying on the fluctuating Rayleigh scattering of small particles that randomly moves in solution, DLS is a very useful
and rapid technique used in the size distribution of particles in the fields of physics, chemistry, and bio-chemistry, especially for
monomodally dispersed spherical particles, and by combining with other techniques such as AFM and TEM, a comprehensive
understanding of the size distribution of the analyte can be readily acquired.

2.4: Dynamic Light Scattering is shared under a CC BY 4.0 license and was authored, remixed, and/or curated by Pavan M. V. Raja & Andrew R.
Barron via source content that was edited to conform to the style and standards of the LibreTexts platform; a detailed edit history is available upon
request.

2.4.8 https://chem.libretexts.org/@go/page/55841
2.5: Zeta Potential Analysis
Introduction
The physical properties of colloids (nanoparticles) and suspensions are strongly dependent on the nature and extent of the particle-
liquid interface. The behavior of aqueous dispersions between particles and liquid is especially sensitive to the ionic and electrical
structure of the interface.
Zeta potential is a parameter that measures the electrochemical equilibrium at the particle-liquid interface. It measures the
magnitude of electrostatic repulsion/attraction between particles and thus, it has become one of the fundamental parameters known
to affect stability of colloidal particles. It should be noted that that term stability, when applied to colloidal dispersions, generally
means the resistance to change of the dispersion with time. Figure 2.5.1 illustrates the basic concept of zeta potential.

Figure 2.5.1 Schematic representation of the ionic concentration and potential difference as a function of distance from the charged
surface of a particle suspended in a dispersion medium.
From the fundamental theory’s perspective, zeta potential is the electrical potential in the interfacial double layer (DL) at the
location of the slipping plane (shown in Figure 2.5.1 ). We can regard zeta potential as the potential difference between the
dispersion medium and the stationary layer of the fluid attached to the particle layer. Therefore, in experimental concerns, zeta
potential is key factor in processes such as the preparation of colloidal dispersions, utilization of colloidal phenomena and the
destruction of unwanted colloidal dispersions. Moreover, zeta potential analysis and measurements nowadays have a lot of real-
world applications. In the field of biomedical research, zeta potential measurement, in contrast to chemical methods of analysis
which can disrupt the organism, has the particular merit of providing information referring to the outermost regions of an organism.
It is also largely utilized in water purification and treatment. Zeta potential analysis has established optimum coagulation conditions
for removal of particulate matter and organic dyestuffs from aqueous waste products.

Brief History and Development of Zeta Potential


Zeta potential is a scientific term for electrokinetic potential in colloidal dispersions. In prior literature, it is usually denoted using
the Greek letter zeta, Ζ, hence it has obtained the name zeta potential as Ζ-potential. The earliest theory for calculating Zeta
potential from experimental data was developed by Marian Smoluchowski in 1903 (Figure 2.5.2 ). Even till today, this theory is
still the most well-known and widely used method for calculating zeta potential.

2.5.1 https://chem.libretexts.org/@go/page/55842
Figure 2.5.2 Portrait of Polish physicist Marian Smoluchowski (1872-1917) pioneer of statistical physics.
Interestingly, this theory was originally developed for electrophoresis. Later on, people started to apply his theory in calculation of
zeta potential. The main reason that this theory is powerful is because of its universality and validity for dispersed particles of any
shape and any concentration. However, there still some limitations to this early theory as it was mainly determined experimentally.
The main limitations are that Smoluchowski’s theory neglects the contribution of surface conductivity and only works for particles
which have sizes much larger than the interface layer, denoted as κa (1/κ is called Debye length and a is the particle radius).
Overbeek and Booth as early pioneers in this direction started to develop more theoretical and rigorous electrokinetic theories that
were able to incorporate surface conductivity for electrokinetic applications. Modern rigorous electrokinetic theories that are valid
almost any κa mostly are generated from Ukrainian (Dukhin) and Australian (O’Brien) scientists.
Principle of Zeta Potential Analysis
Electrokinetic Phenomena
Because an electric double-layer (EDL) exists between a surface and solution, then any relative motion between the rigid and
mobile parts of the EDL will result in the generation of an electrokinetic potential. As described above, zeta potential is essentially
a electrokinetic potential which rises from electrokinetic phenomena. So it is important to understand different situations where
electrokinetic potential can be produced. There are generally four fundamental ways which zeta potential can be produced, via
electrophoresis, electro-osmosis, streaming potential, and sedimentation potential as shown from Figure 2.5.3.

Figure 2.5.3 Relationship between the four types of electrokinetic phenomena (www.americanpharmaceuticalrev...2-
Measurement/)
Calculations of Zeta Potential
There are many different ways of calculating zeta potential . In this section, the methods of calculating zeta potential in
electrophoresis and electroosmosis will be introduced.

Zeta Potential in Electrophoresis


Electrophoresis is the movement of charged colloidal particles or polyelectrolytes, immersed in a liquid, under the influence of an
external electric field. In such case, the electrophoretic velocity, ve (ms-1) is the velocity during electrophoresis and the

2.5.2 https://chem.libretexts.org/@go/page/55842
electrophoretic mobility, ue (m 2 V -1 s -1 ) is the magnitude of the velocity divided by the magnitude of the electric field strength.
The mobility is counted positive if the particles move toward lower potential and negative in the opposite case. And therefore, we
have the relationship ve= ueE, where E is the externally applied field.
Thus, the formula accounted for zeta potential in electrophoresis case is given in EQ, where εrs is the relative permittivity of the
electrolyte solution, ε0 is the electric permittivity of vacuum and η is the viscosity.
εrs ε0 ζ
ue   = (2.5.1)
η

εrs ε0 ζ
ve   = E (2.5.2)
η

There are two cases regarding the size of κa:


1. κa < 1: the formula is similar, 2.5.3 .
2. κa > 1: the formula is rather complicated and we need to solve equation for zeta potential, 2.5.4 , where y eζ
=  eζ/kT , m is
about 0.15 for aqueous solution.
2 εrs ε0 ζ
ue = (2.5.3)
3 η

ek
y ln 2 ek
−ζy
6[ − {1 − e }]
3 ηe 3 2 ζ
ek
ue = y − (2.5.4)
ek
2 εrs ε0 kT 2 ka
−ζy

2+ e 2
2
1+3m/ζ

Zeta Potential in Electroosmosis


Electroosmosis is the motion of a liquid through an immobilized set of particles, a porous plug, a capillary, or a membrane, in
response to an applied electric field. Similar to electrophoresis, it has the electroosmotic velocity, veo (ms -1 ) as the uniform
velocity of the liquid far from the charged interface. Usually, the measured quantity is the volume flow rate of liquid divided by
electric field strength, Qeo,E (m 4 V -1 s -1 ) or diveided by the electric current, Qeo,I (m 3 C -1 ). Therefore, the relationship is given
by 2.5.5 .

Qeo =   ∫ ∫ veo dS (2.5.5)

Thus the formula accounted for Zeta potential in electroosmosis is given in EQ.
As with electrophoresis there are two cases regarding the size of κa:
κa >>1 and there is no surface conduction, where Ac is the cross-section area and KL is the bulk conductivity of particle.
σ

κa < 1, 2.5.8 , where Δu  = K


is the Dukhin number account for surface conductivity, K is the surface conductivity of the
KL
σ

particle.
−εrs ε0 ζ
Qeo,E = Ac (2.5.6)
η

−εrs ε0 ζ 1
Qeo,I = (2.5.7)
η KL

−εrs ε0 ζ 1
Qeo,I = (2.5.8)
η KL (1 + 2Δu)

Relationship Between Zeta Potential and Particle Stability in Electrophoresis


Using the above theoretical methods, we can calculate zeta potential for particles in electrophoresis. The following table
summarizes the stability behavior of the colloid particles with respect to zeta potential. Thus, we can use zeta potential to predict
the stability of colloidal particles in the electrokinetic phenomena of electrophoresis.
Table 2.5.1 Stability behavior of the colloid particles with respect to zeta potential.

2.5.3 https://chem.libretexts.org/@go/page/55842
Zeta Potential (mV) Stability behavior of the particles

0 to ±5 Rapid Coagulation or Flocculation

±10 to ±30 Incipient Instability

±30 to ±40 Moderate Stability

±40 to ±60 Good Stability

More than ±61 Excellent Stability

Instrumentation
In this section, a market-available zeta potential analyzer will be used as an example of how experimentally zeta potential is
analyzed. Figure 2.5.4 shows an example of a typical zeta potential analyzer for electrophoresis.

Figure 2.5.4 Typical zeta potential analyzer for electrophoresis.


The inside measuring principle is described in the following diagram, which shows the detailed mechanism of zeta potential
analyzer (Figure 2.5.5 ).

Figure 2.5.5 Mechanism of zeta potential analyzer for electrophoresis (zeta potential measurement, Microtec Co.,
Ltd.,http://nition.com/en/products/zeecom_s.htm )
When a voltage is applied to the solution in which particles are dispersed, particles are attracted to the electrode of the opposite
polarity, accompanied by the fixed layer and part of the diffuse double layer, or internal side of the "sliding surface". Using the
following formula below of this specific Analyzer and the computer program, we can obtain the zeta potential for electrophoresis
using this typical zeta potential analyzer (Figure 2.5.6.

Figure 2.5.6 Experimental formula of calculation of Zeta potential for electrophoresis (Zeta potential Measurement, Microtec Co.,
Ltd.,http://nition.com/en/products/zeecom_s.htm )

2.5: Zeta Potential Analysis is shared under a CC BY 4.0 license and was authored, remixed, and/or curated by Pavan M. V. Raja & Andrew R.
Barron via source content that was edited to conform to the style and standards of the LibreTexts platform; a detailed edit history is available upon
request.

2.5.4 https://chem.libretexts.org/@go/page/55842
2.6: Viscosity
Introduction
All liquids have a natural internal resistance to flow termed viscosity. Viscosity is the result of frictional interactions within a given
liquid and is commonly expressed in two different ways.
Dynamic Viscosity
The first is dynamic viscosity, also known as absolute viscosity, which measures a fluid’s resistance to flow. In precise terms,
dynamic viscosity is the tangential force per unit area necessary to move one plane past another at unit velocity at unit distance
apart. As one plane moves past another in a fluid, a velocity gradient is established between the two layers (Figure 2.6.1 ).
Viscosity can be thought of as a drag coefficient proportional to this gradient.

Figure 2.6.1 Fluid dynamics as one plane moves relative to a stationary plane through a liquid. The moving plane has area A and
requires force F to overcome the fluid’s internal resistance.
The force necessary to move a plane of area A past another in a fluid is given by Equation 2.6.1 where V is the velocity of the
liquid, Y is the separation between planes, and η is the dynamic viscosity.
V
F = ηA (2.6.1)
Y

V/Y also represents the velocity gradient (sometimes referred to as shear rate). Force over area is equal to τ, the shear stress, so the
equation simplifies to Equation 2.6.2 .
V
τ =η (2.6.2)
Y

For situations where V does not vary linearly with the separation between plates, the differential formula based on Newton’s
equations is given in Equation 2.6.3.
δV
τ =η (2.6.3)
δY

Kinematic Viscosity
Kinematic viscosity, the other type of viscosity, requires knowledge of the density, ρ, and is given by Equation 2.6.4 , where v is
the kinematic viscosity and the η is the dynamic viscosity.
η
ν = (2.6.4)
ρ

Units of Viscosity
Viscosity is commonly expressed in Stokes, Poise, Saybolt Universal Seconds, degree Engler, and SI units.

Dynamic Viscosity
The SI units for dynamic (absolute) viscosity is given in units of N·S/m2, Pa·S, or kg/(m·s), where N stands for Newton and Pa for
Pascal. Poise are metric units expressed as dyne·s/cm2 or g/(m·s). They are related to the SI unit by g/(m·s) = 1/10 Pa·S. 100
centipoise, the centipoise (cP) being the most used unit of viscosity, is equal to one Poise. Table 2.6.1 shows the interconversion
factors for dynamic viscosity.
Table 2.6.1 : The interconversion factors for dynamic viscosity.
Table 2.6.2 lists the dynamic viscosities of several liquids at various temperatures in centipoise. The effect of the temperature on viscosity is
clearly evidenced in the drastic drop in viscosity of water as the temperature is increased from near ambient to 60 degrees Celsius. Ketchup

2.6.1 https://chem.libretexts.org/@go/page/55849
has a viscosity of 1000 cP at 30 degrees Celsius or more than 1000 times that of water at the same temperature!
Unit Pa*S Dyne·s/cm2 or g/(m·s) (Poise) Centipoise (cP)

Pa*S 1 10 1000

Dyne·s/cm2 or g/(m·s) (Poise) 0.1 1 100

Centipoise (cP) 0.001 0.01 1

Table 2.6.2 : Viscosities of common liquids (*at 0% evaporation volume).


Liquid η (cP) Temperature (°C)

Water 0.89 25

Water 0.47 60

Milk 2.0 18

Olive Oil 107.5 20

Toothpaste 70,000 - 100,000 18

Ketchup 1000 30

Custard 1,500 85-90

Crude Oil (WTI)* 7 15

Kinematic Viscosity
The CGS unit for kinematic viscosity is the Stoke which is equal to 10-4 m2/s. Dividing by 100 yields the more commonly used
centistoke. The SI unit for viscosity is m2/s. The Saybolt Universal second is commonly used in the oilfield for petroleum products
represents the time required to efflux 60 milliliters from a Saybolt Universal viscometer at a fixed temperature according to ASTM
D-88. The Engler scale is often used in Britain and quantifies the viscosity of a given liquid in comparison to water in an Engler
viscometer for 200cm3 of each liquid at a set temperature.
Newtonian versus Non-Newtonian Fluids
One of the invaluable applications of the determination of viscosity is identifying a given liquid as Newtonian or non-Newtonian in
nature.
Newtonian liquids are those whose viscosities remain constant for all values of applied shear stress.
Non-Newtonian liquids are those liquids whose viscosities vary with applied shear stress and/or time.
Moreover, non-Newtonian liquids can be further subdivided into classes by their viscous behavior with shear stress:
Pseudoplastic fluids whose viscosity decreases with increasing shear rate
Dilatants in which the viscosity increases with shear rate.
Bingham plastic fluids, which require some force threshold be surpassed to begin to flow and which thereafter flow
proportionally to increasing shear stress.
Measuring Viscosity
Viscometers are used to measure viscosity. There are seven different classes of viscometer:
1. Capillary viscometers.
2. Orifice viscometers.
3. High temperature high shear rate viscometers.
4. Rotational viscometers.
5. Falling ball viscometers.
6. Vibrational viscometers.
7. Ultrasonic Viscometers.

Capillary Viscometers

2.6.2 https://chem.libretexts.org/@go/page/55849
Capillary viscometers are the most widely used viscometers when working with Newtonian fluids and measure the flow rate
through a narrow, usually glass tube. In some capillary viscometers, an external force is required to move the liquid through the
capillary; in this case, the pressure difference across the length of the capillary is used to obtain the viscosity coefficient.
Capillary viscometers require a liquid reservoir, a capillary of known dimensions, a pressure controller, a flow meter, and a
thermostat be present. These viscometers include, Modified Ostwald viscometers, Suspended-level viscometers, and Reverse-flow
viscometers and measure kinematic viscosity.
The equation governing this type of viscometry is the Pouisille law (Equation 2.6.5 ), where Q is the overall flowrate, ΔP, the
pressure difference, a, the internal radius of the tube, η, the dynamic viscosity, and l the path length of the fluid.
4
πΔP a
Q  = (2.6.5)
8ηl

Here, Q is equal to V/t; the volume of the liquid measured over the course of the experiment divided by the time required for it to
move through the capillary where V is volume and t is time.
For gravity-type capillary viscometers, those relying on gravity to move the liquid through the tube rather than an applied force,
Equation 2.6.6 is used to find viscosity, obtained by substituting the relation Equation 2.6.5 with the experimental values, where P
is pressure, ρ is density, g is the gravitational constant, and h is the height of the column.
4
πgha
η  = ρt (2.6.6)
8lV

An example of a capillary viscometer (Ostwald viscometer) is shown in Figure 2.6.2.

Figure 2.6.2 The capillary, submerged in an isothermal bath, is filled until the liquid lies at Mark 3. The liquid is then drawn up
through the opposite side of the tube. The time it takes for the liquid to travel from Mark 2 to Mark 1 is used to compute the
viscosity.

Ori ce Viscometers
Commonly found in the oil industry, orifice viscometers consist of a reservoir, an orifice, and a receiver. These viscometers report
viscosity in units of efflux time as the measurement consists of measuring the time it takes for a given liquid to travel from the
orifice to the receiver. These instruments are not accurate as the set-up does not ensure that the pressure on the liquid remains
constant and there is energy lost to friction at the orifice. The most common types of these viscometer include Redwood, Engler,
Saybolt, and Ford cup viscometers. A Saybolt viscometer is represented in Figure 2.6.3.

Figure 2.6.3 The time it takes for a 60 mL collection flask to fill is used to determine the viscosity in Saybolt units.

2.6.3 https://chem.libretexts.org/@go/page/55849
High Temperature, High Shear Rate Viscometers
These viscometers, also known as cylinder-piston type viscometers are employed when viscosities above 1000 poise, need to be
determined, especially of non-Newtonian fluids. In a typical set-up, fluid in a cylindrical reservoir is displaced by a piston. As the
pressure varies, this type of viscometry is well-suited for determining the viscosities over varying shear rates, ideal for
characterizing fluids whose primary environment is a high temperature, high shear rate environment, e.g., motor oil. A typical
cylinder-piston type viscometer is shown in Figure 2.6.4.

Figure 2.6.4 A typical cylinder-piston type viscometer.

Rotational Viscometers
Well-suited for non-Newtonian fluids, rotational viscometers measure the rate at which a solid rotates in a viscous medium. Since
the rate of rotation is controlled, the amount of force necessary to spin the solid can be used to calculate the viscosity. They are
advantageous in that a wide range of shear stresses and temperatures and be sampled across. Common rotational viscometers
include: the coaxial-cylinder viscometer, cone and plate viscometer, and coni-cylinder viscometer. A cone and plate viscometer is
shown in Figure 2.6.5.

Figure 2.6.5 A cone is spun by a rotor in a liquid paste along a plate. The response of the rotation of the cone is measured, thereby
determining viscosity.

Falling Ball Viscometer


This type of viscometer relies on the terminal velocity achieved by a balling falling through the viscous liquid whose viscosity is
being measured. A sphere is the simplest object to be used because its velocity can be determined by rearranging Stokes’ law
Equation 2.6.7 to Equation 2.6.8 , where r is the sphere’s radius, η the dynamic viscosity, v the terminal velocity of the sphere, σ
the density of the sphere, ρ the density of the liquid, and g the gravitational constant
4
3
6πrηv  =   π r (σ − ρ)g (2.6.7)
3

4 2
π r (σ − ρ)g
3
η  = (2.6.8)
6πv

A typical falling ball viscometric apparatus is shown in Figure 2.6.6.

2.6.4 https://chem.libretexts.org/@go/page/55849
Figure 2.6.6 The time taken for the falling ball to pass from mark 1 to mark 2 is used to obtain viscosity measurements.

Vibrational Viscometers
ften used in industry, these viscometers are attached to fluid production processes where a constant viscosity quality of the product
is desired. Viscosity is measured by the damping of an electrochemical resonator immersed in the liquid to be tested. The resonator
is either a cantilever, oscillating beam, or a tuning fork. The power needed to keep the oscillator oscillating at a given frequency,
the decay time after stopping the oscillation, or by observing the difference when waveforms are varied are respective ways in
which this type of viscometer works. A typical vibrational viscometer is shown in Figure 2.6.7.

Figure 2.6.7 A resonator produces vibrations in the liquid whose viscosity is to be tested. An external sensor detects the vibrations
with time, characterizing the material’s viscosity in realtime.

Ultrasonic Viscometers
This type of viscometer is most like vibrational viscometers in that it is obtaining viscosity information by exposing a liquid to an
oscillating system. These measurements are continuous and instantaneous. Both ultrasonic and vibrational viscometers are
commonly found on liquid production lines and constantly monitor the viscosity.

2.6: Viscosity is shared under a CC BY 4.0 license and was authored, remixed, and/or curated by Pavan M. V. Raja & Andrew R. Barron via
source content that was edited to conform to the style and standards of the LibreTexts platform; a detailed edit history is available upon request.

2.6.5 https://chem.libretexts.org/@go/page/55849
2.7: Electrochemistry
Cyclic Voltammetry Measurements
Introduction
Cyclic voltammetry (CV) is one type of potentiodynamic electrochemical measurements. Generally speaking, the operating process
is a potential-controlled reversible experiment, which scans the electric potential before turning to reverse direction after reaching
the final potential and then scans back to the initial potential, as shown in Figure 2.7.1 -a . When voltage is applied to the system
changes with time, the current will change with time accordingly as shown in Figure 2.7.1 -b. Thus the curve of current and
voltage, illustrated in Figure 2.7.1 -c, can be represented from the data, which can be obtained from Figure 2.7.1 -a and Figure
2.7.1 -b.

Figure 2.7.1 Potential wave changes with time (a); current response with time (b); current-potential representations (c). Adapted
from D. K. Gosser, Jr. Cyclic Voltammetry Simulation and Analysis of Reaction Mechanisms, Wiley-VCH, New York, (1993).
Cyclic voltammetry is a very important analytical characterization in the field of electrochemistry. Any process that includes
electron transfer can be investigated with this characterization. For example, the investigation of catalytical reactions, analyzing the
stoichiometry of complex compounds, and determining of the photovoltaic materials’ band gap. In this module, I will focus on the
application of CV measurement in the field of characterization of solar cell materials.
Although CV was first practiced using a hanging mercury drop electrode, based on the work of Nobel Prize winner Heyrovský
(Figure 2.7.2 ), it did not gain widespread until solid electrodes like Pt, Au and carbonaceous electrodes were used, particularly to
study anodic oxidations. A major advance was made when mechanistic diagnostics and accompanying quantitations became known
through the computer simulations. Now, the application of computers and related software packages make the analysis of data
much quicker and easier.

2.7.1 https://chem.libretexts.org/@go/page/55851
Figure 2.7.2 Czech chemist and inventor Jaroslav Heyrovský (1890 – 1967).
The Components of a CV System
As shown in Figure 2.7.3, the CV systems are as follows:
The epsilon includes potentiostat and current-voltage converter. The potentiostat is required for controlling the applied
potential, and a current-to-voltage converter is used for measuring the current, both of which are contained within the epsilon
(Figure 2.7.3.
The input system is a function generator (Figure 2.7.3. Operators can change parameters, including scan rate and scan range,
through this part. The output part is a computer screen, which can show data and curves directly to the operators.
All electrodes must work in electrolyte solution.
Sometimes, the oxygen and water in the atmosphere will dissolve in the solution, and will be deoxidized or oxidized when
voltage is applied. Therefore the data will be less accurate. To prevent this from happening, bubbling of an inert gas (nitrogen or
argon) is required.
The key component of the CV systems is the electrochemical cell which is connected to the epsilon part. Electrochemical cell
contains three electrodes, counter electrode (C in Figure 2.7.3 ) working electrode (W in Figure 2.7.3 ) and reference electrode
(R in Figure 2.7.3 ). All of them must be immersed in an electrolyte solution when working.

Figure 2.7.3 Components of cyclic voltammetry systems. Adapted from D. K. Gosser, Jr., Cyclic Voltammetry Simulation and
Analysis of Reaction Mechanisms, Wiley-VCH, NewYork, (1993).
In order to better understand the electrodes mentioned above, three kinds of electrodes will be discussed in more detail.
Counter electrodes (C in Figure 2.7.3 are non-reactive high surface area electrodes, for which the platinum gauze is the
common choice.
The working electrode in (W in Figure 2.7.3 ) is commonly an inlaid disc electrodes (Pt, Au, graphite, etc.) of well-defined area
are most commonly used. Other geometries may be available in appropriate circumstances, such as dropping or hanging
mercury hemisphere, cylinder, band, arrays, and grid electrodes.
For the reference electrode (R in Figure 2.7.3 ) aqueous Ag/AgCl or calomel half cells are commonly used, and can be obtained
commercially or easily prepared in the laboratory. Sometimes, a simple silver or platinum wire is used in conjunction with an

2.7.2 https://chem.libretexts.org/@go/page/55851
internal potential reference provided by ferrocene, when a suitable conventional reference electrode is not available. Ferrocene
undergoes a one-electron oxidation at a low potential, around 0.5 V versus a saturated calomel electrode (SCE). It is also been
used as standard in electrochemistry as Fc+/Fc = 0.64 V versus a normal hydrogen electrode (NHE).

Figure 2.7.4 The structure of (C5H5)2Fe (ferrocene).


Cyclic voltammetry systems employ different types of potential waveforms (Figure 2.7.4 ) that can be used to satisfy different
requirements. Potential waveforms reflect the way potential is applied to this system. These different types are referred to by
characteristic names, for example, cyclic voltammetry, and differential pulse voltammetry. The cyclic voltammetry analytical
method is the one whose potential waveform is generally an isosceles triangle (Figure 2.7.4 a).

Figure 2.7.5 Examples of different waveforms of CV systems, illustrating various possible cycles. Adapted from D. K. Gosser, Jr.,
Cyclic Voltammetry Simulation and Analysis of Reaction Mechanisms, Wiley-VCH, New York (1993).
Physical Principles of CV Systems
As mentioned above, there are two main parts of a CV system: the electrochemical cell and the epsilon. Figure 2.7.6 shows the
schematic drawing of circuit diagram in electrochemical cell.

Figure 2.7.6 Diagram of a typical cyclic voltammetry circuit layout. Adapted from R. G. Compton and C. E. Banks, Understanding
Voltammetry, World Scientific, Sigapore (2007).
In a voltammetric experiment, potential is applied to a system, using working electrode (W in Figure 2.7.7 ) and the reference
electrode (R = Figure 2.7.7 ) and the current response is measured using the working electrode and a third electrode, the counter
electrode (C in Figure 2.7.7 ). The typical current-voltage curve for ferricyanide/ferrocyanide, 2.7.1 , is shown in Figure 2.7.7.


Eeq   =  E   +  (0.059/n) log([reactant]/[product]) (2.7.1)

2.7.3 https://chem.libretexts.org/@go/page/55851
Figure 2.7.7 Typical curve of current-voltage curve for for ferricyanide/ferrocyanide, 2.7.1 .

What Useful Information Can We Get From The Data Collected


The information we are able to obtain from CV experimental data is the current-voltage curve. From the curve we can then
determine the redox potential, and gain insights into the kinetics of electron reactions, as well as determine the presence of reaction
intermediate.

Why CV For The Characterizations Of Solar Cell Materials


Despite some limitations, cyclic voltammetry is very well suited for a wide range of applications. Moreover, in some areas of
research, cyclic voltammetry is one of the standard techniques used for characterization. Due to its characteristic shapes of curves,
it has been considered as ‘electrochemical spectroscopy’. In addition, the system is quite easy to operate, and sample preparation is
relatively simple.
The band gap of a semiconductor is a very important value to be determined for photovoltaic materials. Figure 2.7.8 shows the
relative energy level involved in light harvesting of an organic solar cell. The energy difference (Eg) between the lowest
unoccupied molecular orbital (LUMO) and the highest occupied molecular orbital (HOMO), which determines the efficiency. The
oxidation and reduction of an organic molecule involve electron transfers (Figure 2.7.9 ), and CV measurements can be used to
determine the potential change during redox. Through the analysis of data obtained by the CV measurement the electronic band gap
is obtained.

Figure 2.7.8 Diagram showing energy level and light harvesting of and organic solar cell. Adapted from S. B. Darling, Energy
Environm. Sci., 2009, 2, 1266.

2.7.4 https://chem.libretexts.org/@go/page/55851
Figure 2.7.9 Diagram showing energy level and light harvesting of organic solar cell. Adapted from D. K. Gosser, Jr., Cyclic
Voltammetry Simulation and Analysis of Reaction Mechanisms, Wiley-VCH, New York (1993).

The Example Of The Analysis Of CV Data In Solar Cell Material Charecterization


Graphene nanoribbons (GNRs) are long, narrow sheets of graphene formed from the unzipping of carbon nanotubes (Figure 2.7.10
). GNRs can be both semiconducting and semi-metallic, depending on their width, and they represent a particularly versatile variety
of graphene. The high surface area, high aspect ratio, and interesting electronic properties of GNRs render them promising
candidates for applications of energy-storage materials.

Figure 2.7.10 Schematic for the “unzipping” of carbon nanotubes to produce graphene (Rice University).
Graphene nanoribbons can be oxidized to oxidized graphene nanoribbons (XGNRs), are readily soluble in water easily. Cyclic
voltammetry is an effective method to characterize the band gap of semiconductor materials. To test the band gap of oxidized
graphene nanoribbons (XGNRs), operating parameters can be set as follows:
0.1M KCl solution
Working electrode: evaporated gold on silicon.
Scan rate: 10 mV/s.
Scan range: 0 ~ 3000 mV for oxidization reaction; -3000 ~ 0 mV for reduction reaction.
Samples preparation: spin coat an aqueous solution of the oxidized graphene nanoribbons onto the working electrode, and dry at
100 °C.
To make sure that the results are accurate, two samples can be tested under the same condition to see whether the redox peaks are at
the same position. The amount of XGNRs will vary from sample to sample, thus the height of peaks will vary also. Typical curves
obtained from the oxidation reaction (Figure 2.7.9 a) and reduction reaction (Figure 2.7.9 b) are shown in Figure 2.7.10 and Figure
2.7.11, respectively.

2.7.5 https://chem.libretexts.org/@go/page/55851
Figure 2.7.11 Oxidation curves of two samples of XGNRs prepared under similar condition. The sample with lower concentration
is shown by the red curve, while the sample with higher concentration is shown as a black curve.

Figure 2.7.12 Reduction curves of two samples of XGNRs prepared under similar condition. The sample with lower concentration
is shown by the green curve, while the sample with higher concentration is shown as a black curve.
From the curves shown in Figure 2.7.11 and Figure 2.7.12 the following conclusions can be obtained:
Two reduction peak and onset is about -0.75 eV (i.e. Figure 2.7.9 b).
One oxidation peak with onset about 0.85 eV (i.e. Figure 2.7.9 a).
The calculated band gap = 1.60 eV
In conclusion, there are many applications for CV system, efficient method, and the application in the field of solar cell provides
the band gap information for research.

Applications of Cyclic Voltammetry in Proton Exchange Membrane Fuel Cells


Introduction
Proton exchange membrane fuel cells (PEMFCs) are one promising alternative to traditional combustion engines. This method
takes advantage of the exothermic hydrogen oxidation reaction in order to generate energy and water (Table 2.7.1 ).
Table 2.7.1 Summary of oxidation-reduction reactions in PEMFC in acidic and basic electrolytes.
Acidic Redox Potential at Basic Redox Potential at
Acidic Electrolyte Basic Electrolyte
STP (V) STP (V)

Anode half-reaction 2 H2 →  4 H
+
  +  4 e

2 H2   +  4O H

→  4 H2 O  +  4 e

Cathode half-reaction −
O2 + 4 e   +  4 H
+
→  2 H2 O 1.23 −
O2   +  4 e   +  2 H2 O →  4O H

0.401

The basic PEMFC consists of an anode and a cathode separated by a proton exchange membrane (Figure 2.7.13 ). This membrane
is a key component of the fuel cell because for the redox couple reactions to successfully occur, protons must be able to pass from
the anode to the cathode. The membrane in a PEMFC is usually composed of Nafion, which is a polyfluorinated sulfonic acid, and
exclusively allows protons to pass through. As a result, electrons and protons travel from the anode to the cathode through an
external circuit and through the proton exchange membrane, respectively, to complete the circuit and form water.

2.7.6 https://chem.libretexts.org/@go/page/55851
Figure 2.7.13 Schematic of a proton exchange membrane fuel cell (PEMFCs).
PEMFCs present many advantages compared to traditional combustion engines. They are more efficient and have a greater energy
density than traditional fossil fuels. Additionally, the fuel cell itself is very simple with few or no moving parts, which makes it
long-lasting, reliable, and very quiet. Most importantly, however, the operation of a PEMFC results in zero emissions as the only
byproduct is water (Table 2.7.2 ). However, the use of PEMFCs has been limited because of the slow reaction rate for the oxygen
reduction half-reaction (ORR). Reaction rates, k°, for reduction-oxidation reactions such as these tend to be on the order of 10-10 –
10-9 where 10-10 is the fastest reaction rate and 10-9 is the slowest reaction rate. Compared to the hydrogen oxidation half-reaction
(HOR), which has a reaction rate of k° = 1x10-10 cm/s, the reaction rate for the ORR is k° ~ 1x10-9 cm/s. Thus, the ORR is the
kinetic rate-limiting half-reaction and its reaction rate must be increased for PEMFCs to be a viable alternative to combustion
engines. Because cyclic voltammetry can be used to examine the kinetics of the ORR reaction, it is a critical technique in
evaluating potential solutions to this problem.
Table 2.7.2 Summary of advantages and disadvantages of PEFMCs as an alternative to combustion engines.
Advantages Disadvantages

More efficient than combustion ORR half-reaction too slow for commercial use

Greater energy density than fossil fuels Hydrogen fuel is not readily available

Water circulation must be managed to keep the proton exchange


Long-lasting
membrane hydrated

Reliable

Quiet

No harmful emissions

Cyclic Voltammetry
Overview
Cyclic voltammetry is a key electrochemical technique that, among its other uses, can be employed to examine the kinetics of
oxidation-reduction reactions in electrochemical systems. Specifically, data collected with cyclic voltammetry can be used to
determine the rate of reaction. In its simplest form, this technique requires a simple three electrode cell and a potentiostat Figure
2.7.14.

Figure 2.7.14 A simple three electrode cell.


A potential applied to the working electrode is varied linearly with time and the response in the current is measured Figure 2.7.14.
Typically the potential is cycled between two values once in the forward direction and once in the reverse direction. For example,
in Figure 2.7.15, the potential is cycled between 0.8V and -0.2V with the forward scan moving from positive to negative potential

2.7.7 https://chem.libretexts.org/@go/page/55851
and the reverse scan moving from negative to positive potential. Various parameters can be adjusted including the scan rate, the
number of scan cycles, and the direction of the potential scan i.e. whether the forward scan moves from positive to negative
voltages or vice versa. For publication, data is typically collected at a scan rate of 20 mV/s with at least 3 scan cycles.

Figure 2.7.15 Triangular waveform demonstrating the cycling of potential with time.
Reading a Voltammogram
From a cyclic voltammetry experiment, a graph called a voltammogram will be obtained. Because both the oxidation and reduction
half-reactions occur at the working electrode surface, steep changes in the current will be observed when either of these half-
reactions occur.A typical voltammogram will feature two peaks where one peak corresponds to the oxidation half-reaction and the
other to the reduction half-reaction. In an oxidation half-reaction in an electrochemical cell, electrons flow from the species in
solution to the electrode resulting in an anodic current, ia. Frequently, this oxidation peak appears when scanning from negative to
positive potentials (Figure 2.7.16 ). In a reduction half-reaction in an electrochemical cell, electrons flow from the electrode to the
species in solution, resulting in a cathodic current, ic. This type of current is most often observed when scanning from positive to
negative potentials. When the starting reactant is completely oxidized or completely reduced, peak anodic current, ipa, and peak
cathodic current, ipc, respectively, are reached. Then, the current decays as the oxidized or reduced species leaves the electrode
surface. The shape of these anodic and cathodic peaks can be modeled with the Nernst equation, 2.7.2 , where number of electrons
transferred and E˚’ (formal reduction potential) = (Epa + Epc)/2


Eeq   =  E   +  (0.059/n) log ([reactant]/[product]) (2.7.2)

Figure 2.7.16 Example of an idealized cyclic voltammogram. Reprinted with permission from P. T. Kissinger and W. R. Heineman,
J. Chem. Educ., 1981, 60, 702. Copyright 1983 American Chemical Society
Important Values from the Voltammogram

Several key pieces of information can be obtained through examination of the voltammogram including ipa, ipc, and the anodic and
cathodic peak potentials. ipa and ipcboth serve as important measures of catalytic activity: the larger the peak currents, the greater
the activity of the catalyst. Values for ipa and ipc can be obtained through one of two methods: physical examination of the graph or
the Randles-Sevick equation. To determine the peak potentials directly from the graph, a vertical tangent line from the peak current
is intersected with an extrapolated baseline. In contrast, the Randles-Sevick equation uses information about the electrode and the
experimental parameters to calculate the peak current, 2.7.3 ,where A = electrode area; D = diffusion coefficient; C =
concentration; v = scan rate.
5 3/2 1/2 12
ip   =  (2.69x 10 )n AD Cν (2.7.3)

2.7.8 https://chem.libretexts.org/@go/page/55851
Anodic peak potential, Epa, and cathodic peak potential, Epc, can also be obtained from the voltammogram by determining the
potential at which ipa and ipc respectively occur. These values are an indicator of the relative magnitude of the reaction rate. If the
exchange of electrons between the oxidizing and reducing agents is fast, they form an electrochemically reversible couple. These
redox couples fulfill the relationship: ΔEp = Epa – Epc ≡ 0.059/n. In contrast, a nonreversible couple will have a slow exchange of
electrons and ΔEp > 0.059/n. However, it is important to note that ΔEp is dependent on scan rate.
Analysis of Reaction Kinetics
The Tafel and Butler-Volmer equations allow for the calculation of the reaction rate from the current-potential data generated by the
voltammogram. In these analyses, the rate of the reaction can be expressed as two values: k° and io. k˚, the standard rate constant, is
a measure of how fast the system reaches equilibrium: the larger the value of k°, the faster the reaction. The exchange current
density, (io) is the current flow at the surface of the electrode at equilibrium: the larger the value of io, the faster the reaction. While
both io and k° can be used, io is more frequently used because it is directly related to the overpotential through the current-
overpotential and Butler-Volmer equations. When the reaction is at equilibrium, ko and io are related by 2.7.4 , where Co,eq and
CR,eq= equilibrium concentrations of the oxidized and reduced species respectively and a = symmetry factor.
∘ 1−a a
iO   =  nF k C C (2.7.4)
O,eq R,eq

Tafel equation
In its simplest form, the Tafel equation is expressed as 2.7.4 , where a and b can be a variety of constants. Any equation which has
the form of 2.7.5 is considered a Tafel equation.

E −E   =  a  +  b log(i) (2.7.5)

For example, the relationship between current, potential, the concentration of reactants and products, and k˚ can be expressed as
2.7.6 , where CO(0,t) and CR(0,t) = concentrations of the oxidized and reduced species respectively at a specific reaction time, F =

Faraday constant, R = gas constant, and T = temperature.


∘ ∘
[nf /RT ](E−E ) ∘ [anF /RT ](E−E )
CO (0, t)  −  CR (0, t)e   =  [i/nF k ][ e ] (2.7.6)

At very large overpotentials, this equation reduces to a Tafel equation, 2.7.7 , where a = -[RT/(1-a)nF]ln(io) and b = [RT/(1-a)nF].

E −E   =  [RT /(1 − a)nF ]ln(i)  −  [RT /(1 − a)nF ]ln(i0 ) (2.7.7)

The linear relationship between E-E˚ and log(i) can be exploited to determine io through the formation of a Tafel plot (Figure
2.7.17 ), E-E˚ versus log(i).The resulting anodic and cathodic branches of the graph have slopes of [(1-a)nF/2.3RT] and[-

anF/2.3RT], respectively. An extrapolation of these two branches results in a y-intercept = log(io). Thus, this plot directly relates
potential and current data collected by cyclic voltammetry to io.

Figure 2.7.17 Example of an idealized Tafel plot. Reprinted with the permission of Dr. Rob C.M. Jakobs under the GNU Free
Documentation License, Copyright 2010.
Butler-Volmer Equation
While the Butler-Volmer equation resembles the Tafel equation, and in some cases can even be reduced to the Tafel formulation, it
uniquely provides a direct relationship between io and Η. Without simplification, the Butler-Volmer equation is known as the
current-overpotential 2.7.8 .
∘ ∘
[anF /RT ](E−E ) [(1−a)nF /RT ](E−E )
i/ iO   =  CO (0, t)/ CO,eq ] e   −  [ CR (0, t)/ CR,eq ] e (2.7.8)

If the solution is well-stirred, the bulk and surface concentrations can be assumed to be equal and 2.7.8 can be reduced to Butler-
Volmer equation, 2.7.9 .

2.7.9 https://chem.libretexts.org/@go/page/55851
∘ ∘
{[anF /RT ](E−E )} [(1−a)nF /RT ](E−E )
I   =  iO [ e −e ] (2.7.9)

Cyclic Voltammetry in ORR Catalysis Research


Platinum Catalysis
While the issue of a slow ORR reaction rate has been addressed in many ways, it is most often overcome with the use of catalysts.
Traditionally, platinum catalysts have demonstrated the best performance at 30 °C, the ORR io on a Pt catalyst is 2.8 x 10-7 A/cm2
compared to the limiting case of ORR where io = 1 x 10-10A/cm2. Pt is particularly effective as a catalyst for the ORR in PEMFCs
because its binding energy for both O and OH is the closest to ideal of all the bulk metals, its activity is the highest of all the bulk
metals, its selectivity for O2 adsorption is close to 100%, and its extreme stability under a variety of acidic and basic conditions as
well as high operating voltages Figure 2.7.18.

Figure 2.7.18 Anodic sweeps of cyclic voltammograms of Pt, Pt3Sc, and Pt3Y in 0.1 M HClO4 at 20 mV/s. Reprinted by
permission from Macmillan Publishers Ltd: [Nature] J. Greeley, I. E. L. Stephens, A. S. Bondarenko, T. P. Johansson, H. A.
Hansen, T. F. Jaramillo, J. Rossmeisl, I. Chorkendorff, and J. K. Nørskov, Nat. Chem., 2009, 1, 552. Copyright 2009.
Metal-Nitrogren-Carbon Composite Catalysis
Nonprecious metal catalysts (NPMCs) show great potential to reduce the cost of the catalyst without sacrificing catalytic activity.
The best NPMCs currently in development have comparable or even better ORR activity and stability than platinum-based
catalysts in alkaline electrolytes; in acidic electrolytes, however, NPMCs perform significantly worse than platinum-based
catalysts.
In particular, transition metal-nitrogen-carbon composite catalysts (M-N-C) are the most promising type of NPMC. The highest-
performing members of this group catalyze the ORR at potentials within 60 mV of the highest-performing platinum catalysts
(Figure 2.7.19 ). Additionally, these catalysts have excellent stability: after 700 hours at 0.4 V, they do not show any performance
degradation. In a comparison of high-performing PANI-Co-C and PANI-Fe-C (PANI = polyaniline), Zelenay and coworkers used
cyclic voltammetry to compare the activity and performance of these two catalysts in H2SO4. The Co-PANI-C catalyst was found
to have no reduction-oxidation features on its voltammogram whereas Fe-PANI-C was found to have two redox peaks at ~0.64
(Figure 2.7.20 ). These Fe-PANI-C peaks have a full width at half maximum of ~100 mV, which is indicative of the reversible one-
electron Fe3+/Fe2+ reduction-oxidation (theoretical FWHM = 96 mV). Zelenay and coworkers also determined the exchange
current density using the Tafel analysis and found that Fe-PANI-C has a significantly greater io (io = 4 x 10-8 A/cm2) compared to
Co-PANI-C (io = 5 x 10-10 A/cm2). These differences not only demonstrate the higher ORR activity of Fe-PANI-C when compared
to Co-PANI-C, but also suggest that the ORR-active sites and reaction mechanisms are different for these two catalysts. While the
structure of Fe-PANI-C has been examined (Figure 2.7.21 ) the structure of Co-PANI-C is still being investigated.

Figure 2.7.19 Comparison of Fe-PANI-C and Pt/C catalysts in basic electrolyte. Reprinted by permission from Macmillan
Publishers Ltd: [Nature] H. T. Chung, J. H. Won, and P. Zelenay, Nat. Commun., 2013, 4, 1922, Copyright 2013.

2.7.10 https://chem.libretexts.org/@go/page/55851
Figure 2.7.20 Comparison of Co-PANI-C and Fe-PANI-C catalysts by cyclic voltammetry for PANI-Fe-C catalysts. Reproduced
from G. Wu, C.M. Johnston, N.H. Mack, K. Artyushkova, M. Ferrandon, M. Nelson, J.S. Lezama-Pacheco, S.D. Conradson, K.L
More, D.J. Myers, and P. Zelenay, J. Mater. Chem., 2011, 21, 11392-11405 with the permission of The Royal Society of Chemistry.

Figure 2.7.21 Synthetic scheme for Fe-PANI-C catalyst. Reprinted with the permission of the Royal Society of Chemistry under
the CC BY-NC 3.0 License: N. Daems, X. Sheng, Y. Alvarez-Gallego, I. F. J. Vankelecom, and P. P. Pescarmona, Green Chem.,
2016, 18, 1547. Copyright 2015.
While the majority of the M-N-C catalysts show some ORR activity, the magnitude of this activity is highly dependent upon a
variety of factors; cyclic voltammetry is critical in the examination of the relationships between each factor and catalytic activity.
For example, the activity of M-N-Cs is highly dependent upon the synthetic procedure. In their in-depth examination of Fe-PANI-C
catalysts, Zelenay and coworkers optimized the synthetic procedure for this catalyst by examining three synthetic steps: the first
heating treatment, the acid-leaching step, and the second heating treatment. Their synthetic procedure involved the formation of a
PANI-Fe-carbon black suspension that was vacuum-dried onto a carbon support. Then, the intact catalyst underwent a one-hour
heating treatment followed by acid leaching and a three-hour heating treatment. The heating treatments were performed at 900˚C,
which was previously determined to be the optimal temperature to achieve maximum ORR activity (Figure 2.7.21 ).
To determine the effects of the synthetic steps on the intact catalyst, the Fe-PANI-C catalysts were analyzed by cyclic voltammetry
after the first heat treatment (HT1), after the acid-leaching (AL), and after the second heat treatment (HT2). Compared to HT1,
both the AL and HT2 steps showed increases in the catalytic activity. Additionally, HT2 was found to increase the catalytic activity
even more than AL (Figure 2.7.22 ). Based on this data, Zelenay and coworkers concluded HT1 likely either creates active sites in
the catalytic surface while both the AL step removes impurities, which block the surface pores, to expose more active sites.
However, this step is also known to oxidize some of the catalytic area. Thus, the additional increase in activity after HT2 is likely a
result of “repairing” the catalytic surface oxidation.

Figure 2.7.22 Comparison of synthetic techniques by cyclic voltammetry for PANI-Fe-C catalysts. Reproduced from G. Wu, C. M.
Johnston, N. H. Mack, K. Artyushkova, M. Ferrandon, M. Nelson, J. S. Lezama-Pacheco, S. D. Conradson, K. L More, D. J.
Myers, and P. Zelenay, J. Mater. Chem., 2011, 21, 11392, with the permission of The Royal Society of Chemistry.

2.7.11 https://chem.libretexts.org/@go/page/55851
Conclusion
With further advancements in catalytic research, PEMFCs will become a viable and advantageous technology for the replacement
of combustion engines. The analysis of catalytic activity and reaction rate that cyclic voltammetry provides is critical in comparing
novel catalysts to the current highest-performing catalyst: Pt.

Chronocoulometry: A Technique for Electroplating


Fundamentals of Electrochemistry
A chemical reaction that involves a change in the charge of a chemical species is called an electrochemical reaction. As the name
suggests, these reactions involve electron transfer between chemicals. Many of these reactions occur spontaneously when the
various chemicals come in contact with one another. In order to force a nonspontaneous electrochemical reaction to occur, a driving
force needs to be provided. This is because every chemical species has a relative reduction potential. These values provide
information on the ability of the chemical to take extra electrons. Conversely, we can think if relative oxidation potentials, which
indicate the ability of a chemical to give away electrons. It is important to note that these values are relative and need to be defined
against a reference reaction. A list of standard reduction potentials (standard indicating measurement against the normal hydrogen
electrode as seen in (Figure 2.7.23 ) for common electrochemical half-reactions is given in Table 2.7.3. Nonspontaneous
electrochemical systems, often called electrolytic cells, as mentioned previously, require a driving force to occur. This driving force
is an applied voltage, which forces reduction of the chemical that is less likely to gain an electron.

Figure 2.7.23 A schematic diagram of a normal hydrogen electrode.


Table 2.7.3 List of standard reduction potentials of various half reactions.
Oxidant Reductant E° (V vs NHE)

2H2O +2e- H2 (g) +2OH- -0.8227

Cu2O (s) H2O + 2e- 2Cu (s) + 2OH- -0.360

Sn4+ + 2e- Sn2+ +0.15

Cu2+ + 2e- Cu (s) +0.337

O2 (g) + 2H+ + 2e- H2O2 (aq) +0.70

Design of an Electrochemical Cell


A schematic of an electrochemical cell is seen in Figure 2.7.24. Any electrochemical cell must have two electrodes – a cathode,
where the reduction half-reaction takes place, and an anode, where the oxidation half-reaction occurs. Examples of half reactions
can be seen in Table 2.7.3. The two electrodes are electrically connected in two ways – the electrolyte solution and the external
wire. The electrolyte solution typically includes a small amount of the electroactive analyte (the chemical species that will actually
participate in electron transfer) and a large amount of supporting electrolyte (the chemical species that assist in the movement of
charge, but are not actually involved in electron transfer. The external wire provides a path for the electrons to travel from the
oxidation half-reaction to the reduction half-reaction. As mentioned previously, when an electrolytic reaction (nonspontaneous) is
being forced to occur a voltage needs to be applied. This requires the wires to be connected to a potentiostat. As its name suggests,

2.7.12 https://chem.libretexts.org/@go/page/55851
a potentiostat controls voltage (i.e., “potentio” = potential measured in volts). The components of an electrochemical cell and their
functions are also given in Table 2.7.4.

Figure 2.7.24 Schematic of an electrochemical cell.


Table 2.7.4 Various components of an electrochemical cell and their respective functions.
Component Function

Electrode Interface between ions and electrons

Anode Electrode at which the oxidation half reaction takes place

Cathode Electrode at which the reduction half reaction takes place

Electrolyte solution Solution that contains supporting electrolyte and electroactive analyte

Supporting electrolyte Not a part of the faradaic process; only a part of the capacitive process

Electroactive analyte The chemical species responsible for all faradaic current

DC Voltage source; sets the potential difference between the cathode


Potentiostat
and anode

Wire Connects the electrodes to the potentiostat

Chronocoulometry: an Electroanalytical Technique


Theory
Chronocoulometry, as indicated by the name, is a technique in which the charge is measured (i.e. “coulometry”) as a function of
time (i.e., “chrono”). There are various types of coulometry. The one discussed here is potentiostatic coulometry in which the
potential (or voltage) is set and, as a result, charge flows through the cell. The input and output example graphs can be seen in
Figure 2.7.25. The input is a potential step that spans the reduction potential of the electroactive species. If this potential step is
performed in an electrochemical cell that does not contain and electroactive species, only capacitive current will flow (Figure
2.7.26 ), in which the ions migrate in such a way that charges are aligned (positive next to negative, but no charge is transferred.

Once an electroactive species is introduced into the system however, the faradaic current begins to flow. This current is a result of
the electron transfer between the electrode and the electroactive species.

2.7.13 https://chem.libretexts.org/@go/page/55851
Figure 2.7.25 Input potential step (a) and output charge transfer (b) as used in chronocoulometry.

Figure 2.7.26 Capacitive alignment (a) and faradaic charge transfer (b) – the two sources of current in an electrochemical cell.
Electroplating: an Application of Chronocoulometry
Electroplating is an electrochemical process that utilizes techniques such as chronocoulometry to electrodeposit a charged chemical
from a solution as a neutral chemical on the surface of another chemical. These chemicals are typically metals. The science of
electroplating dates back to the early 1800s when Luigi Valentino Brugnatelli (Figure 2.7.27 ) electroplated gold from solution
onto silver metals. By the mid 1800s, the process of electroplating was patented by cousins George and Henry Elkington (Figure
2.7.28 ). The Elkingtons brought electroplated goods to the masses by producing consumer products such as artificial jewelry and

other commemorative items (Figure 2.7.29 ).

Figure 2.7.27 Portrait of Luigi Valentino Brugnatelli (1761-1818).

2.7.14 https://chem.libretexts.org/@go/page/55851
Figure 2.7.28 Portrait of George Elkington (1801-1865)

Figure 2.7.29 A commemorative inkstand gilded using the process of electroplating.


Recent scientific studies have taken interest in studying electroplating. Trejo and coworkers have demonstrated that a quartz
microbalance can be used to measure the change in mass over time during electrodeposition via chronocoulometry. Figure 2.7.30 a
shows the charge transferred at various potential steps. Figure 2.7.30 b shows the change in mass as a function of potential step. It
is clear that the magnitude of the potential step is directly related to the amount of charge transferred and consequently the mass of
the electroactive species deposited.

Figure 2.7.30 Charge transferred over time at varied potentials (a) and mass transferred at varied potentials. Reproduced from A.
Mendez, L. E. Moron, L Ortiz-Frade, Y. Meas, R Ortega-Borges, G. Trejo, J. Electrochem. Soc., 2011, 158, F45. Copyright: The
Electrochemical Society, 2011.
The effect of electroplating via chronocoulometry on the localized surface plasmon resonance (LSPR) has been studied on metallic
nanoparticles. An LSPR is the collective oscillation of electrons as induced by an electric field (Figure 2.7.31 ). In various studies
by Mulvaney and coworkers, a clear effect on the LSPR frequency was seen as potentials were applied (Figure 2.7.32 ). In initial
studies, no evidence of electroplating was reported. In more recent studies by the same group, it was shown that nanoparticles
could be electroplated using chronocoulometry (Figure 2.7.33. Such developments can lead to an expansion of the applications of
both electroplating and plasmonics.

2.7.15 https://chem.libretexts.org/@go/page/55851
Figure 2.7.31 The localized surface plasmon resonance as induced by application of an electric field.

Figure 2.7.32 Shift in the localized surface plasmon resonance frequency as a result of applied potential step. Reproduced from T.
Ung, M. Giersig, D. Dunstan, and P. Mulvaney, Langmuir, 1997, 13, 1773. Copyright: American Chemical Society, 1997.

Figure 2.7.33 Use of chronocoulometry to electroplate nanoparticles. Reproduced from M. Chirea, S. Collins, X. Wei, and P.
Mulvaney, J. Phys. Chem. Lett., 2014, 5, 4331. Copyright: American Chemical Society, 2014

2.7: Electrochemistry is shared under a CC BY 4.0 license and was authored, remixed, and/or curated by Pavan M. V. Raja & Andrew R. Barron
via source content that was edited to conform to the style and standards of the LibreTexts platform; a detailed edit history is available upon
request.

2.7.16 https://chem.libretexts.org/@go/page/55851
2.8: Thermal Analysis
Thermogravimetric Analysis
TGA and SWNTS
Thermogravimetric analysis (TGA) and the associated differential thermal analysis (DTA) are widely used for the characterization
of both as-synthesized and side-wall functionalized single walled carbon nanotubes (SWNTs). Under oxygen, SWNTs will
pyrolyze leaving any inorganic residue behind. In contrast in an inert atmosphere since most functional groups are labile or
decompose upon heating and as SWNTs are stable up to 1200 °C, any weight loss before 800 °C is used to determine the
functionalization ratio of side-wall functionalized SWNTs. The following properties of SWNTs can be determined using this TGA;
1. The mass of metal catalyst impurity in as synthesized SWNTs.
2. The number of functional groups per SWNT carbon (CSWNT).
3. The mass of a reactive species absorbed by a functional group on a SWNT.
Quantitative determination of these properties are used to define the purity of SWNTs, and the extent of their functionalization.
An Overview of Thermogravimetric Analysis
The main function of TGA is the monitoring of the thermal stability of a material by recording the change in mass of the sample
with respect to temperature. Figure 2.8.1 shows a simple diagram of the inside of a typical TGA.

Figure 2.8.1 Schematic representation of a TGA apparatus.


Inside the TGA, there are two pans, a reference pan and a sample pan. The pan material can be either aluminium or platinum. The
type of pan used depends on the maximum temperature of a given run. As platinum melts at 1760 °C and alumium melts at 660 °C,
platinum pans are chosen when the maximum temperature exceeds 660 °C. Under each pan there is a thermocouple which reads the
temperature of the pan. Before the start of each run, each pan is balanced on a balance arm. The balance arms should be calibrated
to compensate for the differential thermal expansion between the arms. If the arms are not calibrated, the instrument will only
record the temperature at which an event occurred and not the change in mass at a certain time. To calibrate the system, the empty
pans are placed on the balance arms and the pans are weighed and zeroed.
As well as recording the change in mass, the heat flow into the sample pan (differential scanning calorimetry, DSC) can also be
measured and the difference in temperature between the sample and reference pan (differential thermal analysis, DTA). DSC is
quantitative and is a measure of the total energy of the system. This is used to monitor the energy released and absorbed during a
chemical reaction for a changing temperature. The DTA shows if and how the sample phase changed. If the DTA is constant, this
means that there was no phase change. Figure 2.8.2 shows a DTA with typical examples of an exotherm and an endotherm.

2.8.1 https://chem.libretexts.org/@go/page/55853
Figure 2.8.2 Simplified representation of the DTA for an exotherm and an endotherm.
When the sample melts, the DTA dips which signifies an endotherm. When the sample is melting it requires energy from the
system. Therefore the temperature of the sample pan decreases compared with the temperature of the reference pan. When the
sample has melted, the temperature of the sample pan increases as the sample is releasing energy. Finally the temperatures of the
reference and sample pans equilibrate resulting in a constant DTA. When the sample evaporates, there is a peak in the DTA. This
exotherm can be explained in the same way as the endotherm.
Typically the sample mass range should be between 0.1 to 10 mg and the heating rate should be 3 to 5 °C/min.
Determination of the Mass of Iron Catalyst Impurity in HiPCO SWNTs
SWNTs are typically synthesized using metal catalysts. Those prepared using the HiPco method, contain residual Fe catalyst. The
metal (i.e., Fe) is usually oxidized upon exposure to air to the appropriate oxide (i.e., Fe2O3). While it is sometimes unimportant
that traces of metal oxide are present during subsequent applications it is often necessary to quantify their presence. This is
particularly true if the SWNTs are to be used for cell studies since it has been shown that the catalyst residue is often responsible
for observed cellular toxicity.
In order to calculate the mass of catalyst residue the SWNTs are pyrolyzed under air or O2, and the residue is assumed to be the
oxide of the metal catalyst. Water can be added to the raw SWNTs, which enhances the low-temperature catalytic oxidation of
carbon. A typical TGA plot of a sample of raw HiPco SWNTs is shown in Figure 2.8.3.

Figure 2.8.3 The TGA of unpurified HiPco SWNTs under air showing the residual mass associated with the iron catalyst. Adapted
from I. W. Chiang, B. E. Brinson, A. Y. Huang, P. A. Willis, M. J. Bronikowski, J. L. Margrave, R. E. Smalley, and R. H. Hauge, J.
Phys. Chem. B, 2001, 105, 8297. Adapted from Chiang et al, 2001
The weight gain (of ca. 5%) at 300 °C is due to the formation of metal oxide from the incompletely oxidized catalyst. To determine
the mass of iron catalyst impurity in the SWNT, the residual mass must be calculated. The residual mass is the mass that is left in
the sample pan at the end of the experiment. From this TGA diagram, it is seen that 70% of the total mass is lost at 400 °C. This

2.8.2 https://chem.libretexts.org/@go/page/55853
mass loss is attributed to the removal of carbon. The residual mass is 30%. Given that this is due to both oxide and oxidized metal,
the original total mass of residual catalyst in raw HiPCO SWNTs is ca. 25%.
Determining the Number of Functional Groups on SWNTs
The limitation of using SWNTs in any practical applications is their solubility; for example SWNTs have little to no solubility in
most solvents due to aggregation of the tubes. Aggregation/roping of nanotubes occurs as a result of the high van der Waals binding
energy of ca. 500 eV per μm of tube contact. The van der Waals force between the tubes is so great, that it take tremendous energy
to pry them apart, making it very difficult to make combination of nanotubes with other materials such as in composite
applications. The functionalization of nanotubes, i.e., the attachment of “chemical functional groups”, provides the path to
overcome these barriers. Functionalization can improve solubility as well as processability, and has been used to align the
properties of nanotubes to those of other materials. In this regard, covalent functionalization provides a higher degree of fine-tuning
for the chemical and physical properties of SWNTs than non-covalent functionalization.
Functionalized nanotubes can be characterized by a variety of techniques, such as atomic force microscopy (AFM), transmission
electron microscopy (TEM), UV-vis spectroscopy, and Raman spectroscopy, however, the quantification of the extent of
functionalization is important and can be determined using TGA. Because any sample of functionalized-SWNTs will have
individual tubes of different lengths (and diameters) it is impossible to determine the number of substituents per SWNT. Instead the
extent of functionalization is expressed as number of substituents per SWNT carbon atom (CSWNT), or more often as
CSWNT/substituent, since this is then represented as a number greater than 1.
Figure 2.8.4 shows a typical TGA for a functionalized SWNT. In this case it is polyethyleneimine (PEI) functionalized SWNTs
prepared by the reaction of fluorinated SWNTs (F-SWNTs) with PEI in the presence of a base catalyst.

Figure 2.8.4 The TGA of SWNTs functionalized with polyethyleimine (PEI) under air showing the sequential loss of complexed
CO2 and decomposition of PEI.
In the present case the molecular weight of the PEI is 600 g/mol. When the sample is heated, the PEI thermally decomposes leaving
behind the unfunctionalized SWNTs. The initial mass loss below 100 °C is due to residual water and ethanol used to wash the
sample.
In the following example the total mass of the sample is 25 mg.
The initial mass, Mi = 25 mg = mass of the SWNTs, residues and the PEI.
After the initial moisture has evaporated there is 68% of the sample left. 68% of 25 mg is 17 mg. This is the mass of the
PEI and the SWNTs.
At 300 °C the PEI starts to decompose and all of the PEI has been removed from the SWNTs at 370 °C. The mass loss
during this time is 53% of the total mass of the sample. 53% of 25 mg is 13.25 mg.
The molecular weight of this PEI is 600 g/mol. Therefore there is 0.013 g / 600 g/mol = 0.022 mmole of PEI in the
sample.
15% of the sample is the residual mass, this is the mass of the decomposed SWNTs. 15% of 25 mg is 3.75 mg. The
molecular weight of carbon is 12 g/mol. So there is 0.3125 mmole of carbon in the sample.
There is 93.4 mol% of carbon and 6.5 mol% of PEI in the sample.

2.8.3 https://chem.libretexts.org/@go/page/55853
Determination of the Mass of a Chemical Absorbed by Functionalized SWNTs
Solid-state 13C NMR of PEI-SWNTs shows the presence of carboxylate substituents that can be attributed to carbamate formation
as a consequence of the reversable CO2 absorption to the primary amine substituents of the PEI. Desorption of CO2 is
accomplished by heating under argon at 75 °C.
The quantity of CO2 absorbed per PEI-SWNT unit may be determined by initially exposing the PEI-SWNT to a CO2 atmosphere to
maximize absorption. The gas flow is switched to either Ar or N2 and the sample heated to liberate the absorbed CO2 without
decomposing the PEI or the SWNTs. An example of the appropriate TGA plot is shown in Figure 2.8.5.

Figure 2.8.5 The TGA results of PEI(10000)-SWNT absorbing and desorbing CO2. The mass has been normalized to the lowest
mass recorded, which is equivalent to PEI(10000)-SWNT.
The sample was heated to 75 °C under Ar, and an initial mass loss due to moisture and/or atmospherically absorbed CO2 is seen. In
the temperature range of 25 °C to 75 °C the flow gas was switched from an inert gas to CO2. In this region an increase in m-
depenass is seen, the increase is due to CO2 absorption by the PEI (10000Da)-SWNT. Switching the carrier gas back to Ar resulted
in the desorption of the CO2.
The total normalized mass of CO2 absorbed by the PEI(10000)-SWNT can be calculated as follows;

Solution Outline
1. Minimum mass = mass of absorbant = Mabsorbant
2. Maximum mass = mass of absorbant and absorbed species = Mtotal
3. Absorbed mass = Mabsorbed = Mtotal - Mabsorbant
4. % of absorbed species= (Mabsorbed/Mabsorbant)*100
5. 1 mole of absorbed species = MW of absorbed species
6. Number of moles of absorbed species = (Mabsorbed/MW of absorbed species)
7. The number of moles of absorbed species absorbed per gram of absorbant= (1g/Mtotal)*(Number of moles of absorbed
species)

Solution
1. Mabsorbant = Mass of PEI-SWNT = 4.829 mg
2. Mtotal = Mass of PEI-SWNT and CO2 = 5.258 mg
3. Mabsorbed = Mtotal - Mabsorbant = 5.258 mg - 4.829 mg = 0.429 mg
4. % of absorbed species= % of CO2 absorbed = (Mabsorbed/Mabsorbant)*100 = (0.429/4.829)*100 = 8.8%
5. 1 mole of absorbed species = MW of absorbed species = MW of CO2 = 44 therefore 1 mole = 44g
6. Number of moles of absorbed species = (Mabsorbed/MW of absorbed species)= (0.429 mg / 44 g) = 9.75 μM
7. The number of moles of absorbed species absorbed per gram of absorbant =(1 g/Mtotal)*(Number of moles of absorbed
species) = (1 g/5.258 mg)*(9.75)= 1.85 mmol of CO2 absorbed per gram of absorbant

TGA/DSC-FTIR Charecterization of Oxide Nanoparticles

2.8.4 https://chem.libretexts.org/@go/page/55853
Metal Oxide Nanoparticles
The binary compound of one or more oxygen atoms with at least one metal atom that forms a structure ≤100 nm is classified as
metal oxide (MOx) nanoparticle. MOxnanoparticles have exceptional physical and chemical properties (especially if they are
smaller than 10 nm) that are strongly related to their dimensions and to their morphology. These enhanced features are due to the
increased surface to volume ratio which has a strong impact on the measured binding energies. Based on theoretical models,
binding or cohesive energy is inversely related to particle size with a linear relationship 2.8.1 .
−1
EN P = Ebulk /cdot[1 − c ⋅ r (2.8.1)

where ENP and Ebulk is the binding energy of the nanoparticle and the bulk binding energy respectively, c is a material constant and
r is the radius of the cluster. As seen from 2.8.1 , nanoparticles have lower binding energies than bulk material, which means lower
electron cloud density and therefore more mobile electrons. This is one of the features that have been identified to contribute to a
series of physical and chemical properties.
Synthesis of Metal Oxide Nanoparticles
Since today, numerous synthetic methods have been developed with the most common ones presented in Table 2.8.1. These
methods have been successfully applied for the synthesis of a variety of materials with 0-D to 3-D complex structures. Among
them, the solvothermal methods are by far the most popular ones due to their simplicity. Between the two classes of solvothermal
methods, slow decomposition methods, usually called thermal decomposition methods, are preferred over the hot injection methods
since they are less complicated, less dangerous and avoid the use of additional solvents.
Table 1 Methods for synthesizing MOx nanoparticles
Method Characteristics Advantages Disadvantages

a. Slow heating of M-precursor


Solvothermal a. Safe, easily carried out, variety a. Poor control of nucleation/
in the presence of
of M-precursors to use growth stages – Particle size
a. Slow decomposition ligand/surfactant precursor
b. Excellent control of particle b. Hazardous, Reproducibility
b. Hot injection b. Injection of M-precursor into
distribution depends on individual
solution at high Temp.

Use of organic molecules or


preexistent nanoparticles as High yield and high purity of Template removal in some cases
Template directed
templates for directing nanoparticles causes particle deformation or loss
nanoparticle formation

Ultrasound influence particle


Sonochemical Mild synthesis conditions Limited applicability
nucleation

Monodisperse particle formation,


Thermal evaporation of Metal Extremely high temperatures, and
Thermal evaporation excellent control in shape and
oxides vacuum system is required
structure

Use of catalyst that serves as a


Excellent control in shape and
Gas phase catalytic growth preferential site for absorbing Limited applicability
structure
Metal reactants

A general schematic diagram of the stages involving the nanoparticles formation is shown in Figure 2.8.6. As seen, first step is the
M-atom generation by dissociation of the metal-precursor. Next step is the M-complex formulation, which is carried out before the
actual particle assembly stage. Between this step and the final particle formulation, oxidation of the activated complex occurs upon
interaction with an oxidant substance. The x-axis is a function of temperature or time or both depending on the synthesis procedure.

2.8.5 https://chem.libretexts.org/@go/page/55853
Figure 2.8.6 Stages of nanoparticle synthesis.
In all cases, the particles synthesized consist of MOx nanoparticle structures stabilized by one or more types of ligand(s) as seen in
Figure 2.8.7. The ligands are usually long-chained organic molecules that have one more functional groups. These molecules
protect the nanoparticles from attracting each other under van der Waals forces and therefore prevent them from aggregating.

Figure 2.8.7 : Schematic representation of a surfactant/ligand stabilized nanoparticle.


Even though often not referred to specifically, all particles synthesized are stabilized by organic (hydrophilic, hydrophobic or
amphoteric) ligands. The detection and the understanding of the structure of these ligands can be of critical importance for
understanding the controlling the properties of the synthesized nanoparticles.
Metal Oxide Nanoparticles Synthesized via slow decomposition

In this work, we refer to MOx nanoparticles synthesized via slow decomposition of a metal complex. In Table 2.8.2, a number of
different MOxnanoparticles are presented, synthesized via metal complex dissociation. Metal–MOx and mixed MOx nanoparticles
are not discussed here.
Table 2.8.2 Examples of MOx nanoparticles synthesized via decomposition of metal complexes.
Metal Oxide Shape Size (approx.)

Cerium oxide dots 5-20 nm

Iron oxide dots, cubes 8.5-23.4 nm

Maganese oxide Multipods > 50 nm

Zinc oxide Hexagonal pyramid 15-25 nm

Cobalt oxide dots ~ 10 nm

Chromium oxide dots 12 nm

Vanadium oxide dots 9 - 15 nm

Molybdenum oxide dots 5 nm

Rhodium oxide dots, rods 16 nm

Palladium oxide dots 18 nm

Ruthenium oxide dots 9 - 14 nm

Zirconium oxide rods 7 x 30 nm

Barium oxide dots 20 nm

2.8.6 https://chem.libretexts.org/@go/page/55853
Metal Oxide Shape Size (approx.)

Magnesium oxide dots 4 - 8 nm

Calcium oxide dots, rods 7 - 12 nm

Nickel oxide dots 8 - 15 nm

Titanium oxide dots and rods 2.3 - 30 nm

Tin oxide dots 2 - 5 nm

Indium oxide dots ~ 5 nm

Samaria Square ~ 10 nm

A significant number of metal oxides synthesized using slow decomposition is reported in literature. If we use the periodic table to
map the different MOx nanoparticles (Figure 2.8.8 ), e notice that most of the alkali and transition metals generate MOx
nanoparticles, while only a few of the poor metals seem to do so, using this synthetic route. Moreover, two of the rare earth metals
(Ce and Sm) have been reported to successfully give metal oxide nanoparticles via slow decomposition.

Figure 2.8.8 “Periodic” table of MOx nanoparticles synthesized using the slow decomposition technique.
Among the different characterization techniques used for defining these structures, transition electron microscopy (TEM) holds the
lion’s share. Nevertheless, most of the modern characterization methods are more important when it comes to understanding the
properties of nanoparticles. X-ray photoelectron spectroscopy (XPS), X-ray diffraction (XRD), nuclear magnetic resonance
(NMR), IR spectroscopy, Raman spectroscopy, and thermogravimetric analysis (TGA) methods are systematically used for
characterization.
Synthesis and Characterization of WO3-x nanorods
The synthesis of WO3-x nanorods is based on the method published by Lee et al. A slurry mixture of Me3NO·2H2O, oleylamine and
W(CO)6 was heated up to 250 °C at a rate of 3 °C/min (Figure 2.8.9 ). The mixture was aged at this temperature for 3 hours before
cooling down to room temperature.

2.8.7 https://chem.libretexts.org/@go/page/55853
Figure 2.8.9 Experimental setup for synthesis of WO3-x nanorods.
Multiple color variations were observed between 100 - 250 °C with the final product having a dark blue color. Tungsten oxide
nanorods (W18O49 identified by XRD) with a diameter of 7±2 nm and 50±2 nm long were acquired after centrifugation of the
product solution. A TEM image of the W18O49 nanorods is shown in Figure 2.8.10.

Figure 2.8.10 TEM image of WO3-x nanorods.


Thermogravimetric Analysis (TGA)/Differential Scanning Calorimetry (DSC)
Thermogravimetric analysis (TGA) is a technique widely used for determining the organic and inorganic content of various
materials. Its basic rule of function is the high precision measurement of weight gain/loss with increasing temperature under inert
or reactive atmospheres. Each weight change corresponds to physical (crystallization, phase transformation) or chemical
(oxidation, reduction, reaction) processes that take place by increasing the temperature. The sample is placed into platinum or
alumina pan and along with an empty or standard pan are placed onto two high precision balances inside a high temperature oven.
A method for pretreating the samples is selected and the procedure is initiated. Differential scanning calorimetry (DSC) is a
technique usually accompanying TGA and is used for calculating enthalpy energy changes or heat capacity changes associated with
phase transitions and/or ligand-binding energy cleavage.
In Figure 2.8.11 the TGA/DSC plot acquired for the ligand decomposition of WO3-x nanorods is presented. The sample was heated
at constant rate under N2 atmosphere up to 195 °C for removing moisture and then up to 700 °C for removing the oleylamine
ligands. It is important to use an inert gas for performing such a study to avoid any premature oxidation and/or capping agent
combustion. 26.5% of the weight loss is due to oleylamine evaporations which means about 0.004 moles per gram of sample. After
isothermal heating at 700 °C for 25 min the flow was switched to air for oxidizing the ligand-free WO3-x to WO3. From the DSC
curve we noticed the following changes of the weight corrected heat flow:
1. From 0 – 10 min assigned to water evaporation.
2. From 65 – 75 min assigned to OA evaporation.
3. From 155 – 164 min assigned to WO3-x oxidation.

2.8.8 https://chem.libretexts.org/@go/page/55853
4. From 168 – 175 min is also due to further oxidation of W5+ atoms.
The heat flow increase during the WO3-x to WO3 oxidation is proportional to the crystal phase defects (or W atoms of oxidation
state +5) and can be used for performing qualitative studies between different WOx nanoparticles.

Figure 2.8.11 TGA/DSC plot for WO3-x nanorods.


The detailed information about the procedure used to acquire the TGA/DSC plot shown in Figure 2.8.11 is as follows.
Select gas (N2 with flow rate 50 mL/min.)
Ramp 20 °C/min to 200 °C.
Isothermal for 20 min.
Ramp 5 °C/min to 700 °C.
Isothermal for 25 min.
Select gas (air).
Isothermal for 20 min.
Ramp 10 °C/min to 850 °C.
Cool down
Fourier Transform Infrared Spectroscopy
Fourier transform infrared spectroscopy (FTIR) is the most popular spectroscopic method used for characterizing organic and
inorganic compounds. The basic modification of an FTIR from a regular IR instrument is a device called interferometer, which
generates a signal that allows very fast IR spectrum acquisition. For doing so, the generatated interferogram has to be “expanded”
using a Fourier transformation to generate a complete IR frequency spectrum. In the case of performing FTIR transmission studies
the intensity of the transmitted signal is measured and the IR fingerprint is generated 2.8.2 .
I cεl
T = =e (2.8.2)
L

Where I is the intensity of the samples, Ib is the intensity of the background, c is the concentration of the compound, ε is the molar
extinction coefficient and l is the distance that light travels through the material. A transformation of transmission to absorption
spectra is usually performed and the actual concentration of the component can be calculated by applying the Beer-Lambert law
2.8.3

A = −ln(T ) = cεl (2.8.3)

A qualitative IR-band map is presented in Figure 2.8.12.


The absorption bands between 4000 to 1600 cm-1 represent the group frequency region and are used to identify the stretching
vibrations of different bonds. At lower frequencies (from 1600 to 400 cm-1) vibrations due to intermolecular bond bending occurs
upon IR excitation and therefore are usually not taken into account.

2.8.9 https://chem.libretexts.org/@go/page/55853
Figure 2.8.12 Selected FTIR stretching and bending modes associated with the typical ligands used for nanoparticle stabilization.
TGA/DSC-FTIR Characterization
TGA/DSC is a powerful tool for identifying the different compounds evolved during the controlled pyrolysis and therefore provide
qualitative and quantitative information about the volatile components of the sample. In metal oxide nanoparticle synthesis
TGA/DSC-FTIR studies can provide qualitative and quantitative information about the volatile compounds of the nanoparticles.
TGA–FTIR results presented below were acquired using a Q600 Simultaneous TGA/DSC (SDT) instrument online with a Nicolet
5700 FTIR spectrometer. This system has a digital mass flow control and two gas inlets giving the capability to switch reacting gas
during each run. It allows simultaneous weight change and differential heat flow measurements up to 1500 °C, while at the same
time the outflow line is connected to the FTIR for performing gas phase compound identification. Grand-Schmidt thermographs
were usually constructed to present the species evolution with time in 3 dimensions.
Selected IR spectra are presented in Figure 2.8.13. Four regions with intense peaks are observed. Between 4000 – 3550 cm-1 due to
O-H bond stretching assigned to H2O that is always present and due to due to N-H group stretching that is assigned to the amine
group of oleylamine. Between 2400 – 2250 cm-1 due to O=C=O stretching, between 1900 – 1400 cm-1 which is mainly to C=O
stretching and between 800 – 400 cm-1 cannot be resolved as explained previously.

Figure 2.8.13 FTIR spectra of products from WO3-x pyrolysis.


The peak intensity evolution with time can be more easily observed in Figure 2.8.14 and Figure 2.8.15. As seen, CO2 evolution
increases significantly with time especially after switching our flow from N2 to air. H2O seems to be present in the outflow stream
up to 700 °C while the majority of the N-H amine peaks seem to disappear at about 75 min. C=N compounds are not expected to be
present in the stream which leaves bands between 1900 – 1400 cm-1 assigned to C=C and C=O stretching vibrations. Unsaturated
olefins resulting from the cracking of the oleylamine molecule are possible at elevated temperatures as well as the presence of CO
especially under N2atmosphere.

2.8.10 https://chem.libretexts.org/@go/page/55853
Figure 2.8.14 3D representation of FTIR Spectra of the volatile compounds of WO3-x.

Figure 2.8.15 Intensity profile of FTIR spectra of the volatile compounds formed from the pyrolysis of WO3-x.
From the above compound identification we can summarize and propose the following applications for TGA-FTIR. First, more
complex ligands, containing aromatic rings and maybe other functional groups may provide more insight in the ligand to MOx
interaction. Second, the presence of CO and CO2 even under N2 flow means that complete O2 removal from the TGA and the FTIR
cannot be achieved under these conditions. Even though the system was equilibrated for more than an hour, traces of O2 are
existent which create errors in our calculations.
Determination of Sublimation Enthalpy and Vapor Pressure for Inorganic and Metal-Organic Compounds by Thermogravimetric Analysis

Metal compounds and complexes are invaluable precursors for the chemical vapor deposition (CVD) of metal and non-metal thin
films. In general, the precursor compounds are chosen on the basis of their relative volatility and their ability to decompose to the
desired material under a suitable temperature regime. Unfortunately, many readily obtainable (commercially available) compounds
are not of sufficient volatility to make them suitable for CVD applications. Thus, a prediction of the volatility of a metal-organic
compounds as a function of its ligand identity and molecular structure would be desirable in order to determine the suitability of
such compounds as CVD precursors. Equally important would be a method to determine the vapor pressure of a potential CVD
precursor as well as its optimum temperature of sublimation.
It has been observed that for organic compounds it was determined that a rough proportionality exists between a compound’s
melting point and sublimation enthalpy; however, significant deviation is observed for inorganic compounds.
Enthalpies of sublimation for metal-organic compounds have been previously determined through a variety of methods, most
commonly from vapor pressure measurements using complex experimental systems such as Knudsen effusion, temperature drop

2.8.11 https://chem.libretexts.org/@go/page/55853
microcalorimetry and, more recently, differential scanning calorimetry (DSC). However, the measured values are highly dependent
on the experimental procedure utilized. For example, the reported sublimation enthalpy of Al(acac)3 (Figure 2.8.16 a where M =
Al, n = 3) varies from 47.3 to 126kJ/mol.

Figure 2.8.16 Structure of a typical metal β-diketonate complex. (a) acetylacetonate (acac); (b) trifluoro acetylacetonate (tfac), and
(c) hexafluoroacetylacetonate (hfac).
Thermogravimetric analysis offers a simple and reproducible method for the determination of the vapor pressure of a potential
CVD precursor as well as its enthalpy of sublimation.
Determination of Sublimation Enthalpy
The enthalpy of sublimation is a quantitative measure of the volatility of a particular solid. This information is useful when
considering the feasibility of a particular precursor for CVD applications. An ideal sublimation process involves no compound
decomposition and only results in a solid-gas phase change, i.e., 2.8.4
[M (L)n I ](solid) → [M (Ln )](vapor) (2.8.4)

Since phase changes are thermodynamic processes following zero-order kinetics, the evaporation rate or rate of mass loss by
sublimation (msub), at a constant temperature (T), is constant at a given temperature, 2.8.5 . Therefore, the msub values may be
directly determined from the linear mass loss of the TGA data in isothermal regions.
Δ[mass]
msub   =   (2.8.5)
Δt

The thermogravimetric and differential thermal analysis of the compound under study is performed to determine the temperature of
sublimation and thermal events such as melting. Figure 2.8.17 shows a typical TG/DTA plot for a gallium chalcogenide cubane
compound (Figure 2.8.18 ).

Figure. Adapted from E. G. Gillan, S. G. Bott, and A. R. Barron, Chem. Mater., 1997, 9, 3, 796.

Figure 2.8.18 Structure of gallium chalcogenide cubane compound, where E = S, Se, and R = CMe3, CMe2Et, CEt2Me, CEt3.

Data Collection

2.8.12 https://chem.libretexts.org/@go/page/55853
In a typical experiment 5 - 10 mg of sample is used with a heating rate of ca. 5 °C/min up to under either a 200-300 mL/min inert
(N2 or Ar) gas flow or a dynamic vacuum (ca. 0.2 Torr if using a typical vacuum pump). The argon flow rate was set to 90.0
mL/min and was carefully monitored to ensure a steady flow rate during runs and an identical flow rate from one set of data to the
next.
Once the temperature range is defined, the TGA is run with a preprogrammed temperature profile (Figure 2.8.19 ). It has been
found that sufficient data can be obtained if each isothermal mass loss is monitored over a period (between 7 and 10 minutes is
found to be sufficient) before moving to the next temperature plateau. In all cases it is important to confirm that the mass loss at a
given temperature is linear. If it is not, this can be due to either (a) temperature stabilization had not occurred and so longer times
should be spent at each isotherm, or (b) decomposition is occurring along with sublimation, and lower temperature ranges must be
used. The slope of each mass drop is measured and used to calculate sublimation enthalpies as discussed below.

Figure 2.8.19 A typical temperature profile for determination of isothermal mass loss rate.
As an illustrative example, Figure 2.8.20 displays the data for the mass loss of Cr(acac)3 (Figure 2.8.16 a, where M = Cr, n = 3 ) at
three isothermal regions under a constant argon flow. Each isothermal data set should exhibit a linear relation. As expected for an
endothermal phase change, the linear slope, equal to msub, increases with increasing temperature.

Figure 2.8.20 Plot of TGA results for Cr(acac)3 performed at different isothermal regions. Adapted from B. D. Fahlman and A. R.
Barron, Adv. Mater. Optics Electron., 2000, 10, 223.
Samples of iron acetylacetonate (Figure 2.8.16 a, where M = Fe, n = 3) may be used as a calibration standard through ΔHsub
determinations before each day of use. If the measured value of the sublimation enthalpy for Fe(acac)3 is found to differ from the
literature value by more than 5%, the sample is re-analyzed and the flow rates are optimized until an appropriate value is obtained.
Only after such a calibration is optimized should other complexes be analyzed. It is important to note that while small amounts (<
10%) of involatile impurities will not interfere with the ΔHsub analysis, competitively volatile impurities will produce higher
apparent sublimation rates.
It is important to discuss at this point the various factors that must be controlled in order to obtain meaningful (useful) msub data
from TGA data.
1. The sublimation rate is independent of the amount of material used but may exhibit some dependence on the flow rate of an
inert carrier gas, since this will affect the equilibrium concentration of the cubane in the vapor phase. While little variation was

2.8.13 https://chem.libretexts.org/@go/page/55853
observed we decided that for consistency msub values should be derived from vacuum experiments only.
2. The surface area of the solid in a given experiment should remain approximately constant; otherwise the sublimation rate (i.e.,
mass/time) at different temperatures cannot be compared, since as the relative surface area of a given crystallite decreases
during the experiment the apparent sublimation rate will also decrease. To minimize this problem, data was taken over a small
temperature ranges (ca. 30 °C), and overall sublimation was kept low (ca. 25% mass loss representing a surface area change of
less than 15%). In experiments where significant surface area changes occurred the values of msub deviated significantly from
linearity on a log(msub) versus 1/T plot.
3. The compound being analyzed must not decompose to any significant degree, because the mass changes due to decomposition
will cause a reduction in the apparent msub value, producing erroneous results. With a simultaneous TG/DTA system it is
possible to observe exothermic events if decomposition occurs, however the clearest indication is shown by the mass loss
versus time curves which are no longer linear but exhibit exponential decays characteristic of first or second order
decomposition processes.
Data Analysis
The basis of analyzing isothermal TGA data involves using the Clausius-Clapeyron relation between vapor pressure (p) and
temperature (T), 2.8.6 , where ∆Hsub is the enthalpy of sublimation and R is the gas constant (8.314 J/K.mol).
d ln(p) ΔHsub
  =  (2.8.6)
2
dT RT

Since msub data are obtained from TGA data, it is necessary to utilize the Langmuir equation, 2.8.7 , that relates the vapor pressure
of a solid with its sublimation rate.
2πRT 0.5
p  =  [ ] msub (2.8.7)
MW

After integrating 2.8.6 in log form, substituting 2.8.7 , and consolidating, one one obtains the useful equality, 2.8.8 .


− −0.0522(ΔHsub ) 0.0522(ΔHsub ) 1 1306
log(msub √T ) = +[ − log( )] (2.8.8)
T T 2 MW

Hence, the linear slope of a log(msubT1/2) versus 1/T plot yields ΔHsub. An example of a typical plot and the corresponding ΔHsub
value is shown in Figure 2.8.21. In addition, the y intercept of such a plot provides a value for Tsub, the calculated sublimation
temperature at atmospheric pressure.

Figure 2.8.21 Plot of log(msubT1/2) versus 1/T and the determination of the ΔHsub (112.6 kJ/mol) for Fe(acac)3 (R2 = 0.9989).
Adapted from B. D. Fahlman and A. R. Barron, Adv. Mater. Optics Electron., 2000, 10, 223.
Table 2.8.3 lists the typical results using the TGA method for a variety of metal β-diketonates, while Table 2.8.4 lists similar
values obtained for gallium chalcogenide cubane compounds.
Table 2.8.3 Selected thermodynamic data for metal β-diketonate compounds determined from thermogravimetric analysis. Data from B. D.
Fahlman and A. R. Barron, Adv. Mater. Optics Electron., 2000, 10, 223.
Calculated vapor
Compound ΔHsub (kJ/mol) ΔSsub (J/K.mol) Tsub calc. (°C)
pressure @ 150 °C (Torr)

Al(acac)3 93 220 150 3.261

2.8.14 https://chem.libretexts.org/@go/page/55853
Al(tfac)3 74 192 111 9.715

Al(hfac)3 52 152 70 29.120

Cr(acac)3 91 216 148 3.328

Cr(tfac)3 71 186 109 9.910

Cr(hfac)3 46 134 69 29.511

Fe(acac)3 112 259 161 2.781

Fe(tfac)3 96 243 121 8.340

Fe(hfac)3 60 169 81 25.021

Co(acac)3 138 311 170 1.059

Co(tfac)3 119 295 131 3.319

Co(hfac)3 73 200 90 9.132

Table 2.8.4 Selected thermodynamic data for gallium chalcogenide cubane compounds determined from thermogravimetric analysis. Data
from E. G. Gillan, S. G. Bott, and A. R. Barron, Chem. Mater., 1997, 9, 3, 796.
Calculated vapor
Compound ∆Hsub (kJ/mol) ∆Ssub (J/K. mol) Tsub calc. (°C)
pressure @ 150 °C (Torr)

[(Me3C)GaS]4 110 300 94 22.75

[(EtMe2C)GaS]4 124 330 102 18.89

[(Et2MeC)GaS]4 137 339 131 1.173

[(Et3C)GaS]4 149 333 175 0.018

[(Me3C)GaSe)]4 119 305 116 3.668

[(EtMe2C)GaSe]4 137 344 124 2.562

[(Et2MeC)GaSe]4 147 359 136 0.815

[(Et3C)GaSe]4 156 339 189 0.005

A common method used to enhance precursor volatility and corresponding efficacy for CVD applications is to incorporate partially
(Figure 2.8.16 b ) or fully (Figure 2.8.16 c) fluorinated ligands. As may be seen from Table 2.8.3 this substitution does results in
significant decrease in the ΔHsub, and thus increased volatility. The observed enhancement in volatility may be rationalized either
by an increased amount of intermolecular repulsion due to the additional lone pairs or that the reduced polarizability of fluorine
(relative to hydrogen) causes fluorinated ligands to have less intermolecular attractive interactions.
Determination of Sublimation Entropy

The entropy of sublimation is readily calculated from the ΔHsub and the calculated Tsub data, 2.8.9
ΔHsub
ΔSsub   =   (2.8.9)
Tsub

Table 2.8.3 and Table 2.8.4 show typical values for metal β-diketonate compounds and gallium chalcogenide cubane compounds,
respectively. The range observed for gallium chalcogenide cubane compounds (ΔSsub = 330 ±20 J/K.mol) is slightly larger than
values reported for the metal β-diketonates compounds (ΔSsub = 130 - 330 J/K.mol) and organic compounds (100 - 200 J/K.mol),
as would be expected for a transformation giving translational and internal degrees of freedom. For any particular chalcogenide,
i.e., [(R)GaS]4, the lowest ΔSsubare observed for the Me3C derivatives, and the largest ΔSsub for the Et2MeC derivatives, see Table
2.8.4. This is in line with the relative increase in the modes of freedom for the alkyl groups in the absence of crystal packing

forces.
Determination of Vapor Pressure
While the sublimation temperature is an important parameter to determine the suitability of a potential precursor compounds for
CVD, it is often preferable to express a compound's volatility in terms of its vapor pressure. However, while it is relatively

2.8.15 https://chem.libretexts.org/@go/page/55853
straightforward to determine the vapor pressure of a liquid or gas, measurements of solids are difficult (e.g., use of the isoteniscopic
method) and few laboratories are equipped to perform such experiments. Given that TGA apparatus are increasingly accessible, it
would therefore be desirable to have a simple method for vapor pressure determination that can be accomplished on a TGA.
Substitution of 2.8.5 into 2.8.8 allows for the calculation of the vapor pressure (p) as a function of temperature (T). For example,
Figure 2.8.22 shows the calculated temperature dependence of the vapor pressure for [(Me3C)GaS]4. The calculated vapor
pressures at 150 °C for metal β-diketonates compounds and gallium chalcogenide cubane compounds are given in Table 2.8.3 and
Table 2.8.4

Figure 2.8.22 A plot of calculated vapor pressure (Torr) against temperature (K) for [(Me3C)GaS]4. Adapted from E. G. Gillan, S.
G. Bott, and A. R. Barron, Chem. Mater., 1997, 9, 3, 796.
The TGA approach to show reasonable agreement with previous measurements. For example, while the value calculated for
Fe(acac)3(2.78 Torr @ 113 °C) is slightly higher than that measured directly by the isoteniscopic method (0.53 Torr @ 113 °C);
however, it should be noted that measurements using the sublimation bulb method obtained values much lower (8 x 10-3 Torr @
113 °C). The TGA method offers a suitable alternative to conventional (direct) measurements of vapor pressure.

Differential Scanning Calorimetry (DSC)


Differential scanning calorimetry (DSC) is a technique used to measure the difference in the heat flow rate of a sample and a
reference over a controlled temperature range. These measurements are used to create phase diagrams and gather thermoanalytical
information such as transition temperatures and enthalpies.
History
DSC was developed in 1962 by Perkin-Elmer employees Emmett Watson and Michael O’Neill and was introduced at the
Pittsburgh Conference on Analytical Chemistry and Applied Spectroscopy. The equipment for this technique was available to
purchase beginning in 1963 and has evolved to control temperatures more accurately and take measurements more precisely,
ensuring repeatability and high sensitivity.

Figure 2.8.23 An excerpt from the original DSC patent.

2.8.16 https://chem.libretexts.org/@go/page/55853
Theory
Phase Transitions
Phase transitions refer to the transformation from one state of matter to another. Solids, liquids, and gasses are changed to other
states as the thermodynamic system is altered, thereby affecting the sample and its properties. Measuring these transitions and
determining the properties of the sample is important in many industrial settings and can be used to ensure purity and determine
composition (such as with polymer ratios). Phase diagrams (Figure 2.8.23 ) can be used to clearly demonstrate the transitions in
graphical form, helping visualize the transition points and different states as the thermodynamic system is changed.

Figure 2.8.24 An example of a typical phase diagram.


Differential Thermal Analysis
Prior to DSC, differential thermal analysis (DTA) was used to gather information about transition states of materials. In DTA, the
sample and reference are heated simultaneously with the same amount of heat and the temperature of each is monitored
independently. The difference between the sample temperature and the reference temperature gives information about the
exothermic or endothermic transition occurring in the sample. This strategy was used as the foundation for DSC, which sought to
measure the difference in energy needed to keep the temperatures the same instead of measure the difference in temperature from
the same amount of energy.
Differntial Scanning Calorimeter
Instead of measuring temperature changes as heat is applied as in DTA, DSC measures the amount of heat that is needed to increase
the temperatures of the sample and reference across a temperature gradient. The sample and reference are kept at the same
temperature as it changes across the gradient, and the differing amounts of heat required to keep the temperatures synchronized are
measured. As the sample undergoes phase transitions, more or less heat is needed, which allows for phase diagrams to be created
from the data. Additionally, specific heat, glass transition temperature, crystallization temperature, melting temperature, and
oxidative/thermal stability, among other properties, can be measured using DSC.
Applications
DSC is often used in industrial manufacturing, ensuring sample purity and confirming compositional analysis. Also used in
materials research, providing information about properties and composition of unknown materials can be determined. DSC has also
been used in the food and pharmaceutical industries, providing characterization and enabling the fine-tuning of certain properties.
The stability of proteins and folding/unfolding information can also be measured with DSC experiments.
Instrumentation
Equipment
The sample and reference cells (also known as pans), each enclosing their respective materials, are contained in an insulted
adiabatic chamber (Figure 2.8.25 ). The cells can be made of a variety of materials, such as aluminum, copper, gold and platinum.
The choice of which is dictated by the necessary upper temperature limit. A variable heating element around each cell transfers
heat to the sample, causing both cells’ temperature to rise in coordination with the other cell. A temperature monitor measures the
temperatures of each cell and a microcontroller controls the variable heating elements and reports the differential power required
for heating the sample versus the reference. A typical setup, including a computer for controlling software, is shown in Figure
2.8.26.

2.8.17 https://chem.libretexts.org/@go/page/55853
Figure 2.8.25 Diagram of basic DSC equipment.

Figure 2.8.26 Picture of basic DSC setup in a laboratory.


Modes of Operations
With advancement in DSC equipment, several different modes of operations now exist that enhance the applications of DSC.
Scanning mode typically refers to conventional DSC, which uses a linear increase or decrease in temperature. An example of an
additional mode often found in newer DSC equipment is an isothermal scan mode, which keeps temperature constant while the
differential power is measured. This allows for stability studies at constant temperatures, particularly useful in shelf life studies for
pharmaceutical drugs.
Calibration
As with practically all laboratory equipment, calibration is required. Calibration substances, typically pure metals such as indium or
lead, are chosen that have clearly defined transition states to ensure that the measured transitions correlate to the literature values.
Obtaining Measurements
Sample Preparation
Sample preparation mostly consists of determining the optimal weight to analyze. There needs to be enough of the sample to
accurately represent the material, but the change in heat flow should typically be between 0.1 - 10 mW. The sample should be kept
as thin as possible and cover as much of the base of the cell as possible. It is typically better to cut a slice of the sample rather than
crush it into a thin layer. The correct reference material also needs to be determined in order to obtain useful data.
DSC Curves
DSC curves (e.g., Figure 2.8.27 ) typically consist of heat flow plotted versus the temperature. These curves can be used to
calculate the enthalpies of transitions, (ΔH), 2.8.10 , by integrating the peak of the state transition, where K is the calorimetric
constant and A is the area under the curve.
ΔH   =  KA (2.8.10)

Figure 2.8.27 An idealized DSC curve showing the shapes associated with particular phase transitions.

2.8.18 https://chem.libretexts.org/@go/page/55853
Sources of error
Common error sources apply, including user and balance errors and improper calibration. Incorrect choice of reference material and
improper quantity of sample are frequent errors. Additionally, contamination and how the sample is loaded into the cell affect the
DSC.
DSC Characterization of Polymers

Differential scanning calorimetry (DSC), at the most fundamental level, is a thermal analysis technique used to track changes in the
heat capacity of some substance. To identify this change in heat capacity, DSC measures heat flow as a function of temperature and
time within a controlled atmosphere. The measurements provide a quantitative and qualitative look into the physical and chemical
alterations of a substance related to endothermic or exothermic events.
The discussion done here will be focused on the analysis of polymers; therefore, it is important to have an understanding of
polymeric properties and how heat capacity is measured within a polymer.
Overview of Polymeric Properties
A polymer is, essentially, a chemical compound whose molecular structure is a composition of many monomer units bonded
together (Figure 2.8.28 ). The physical properties of a polymer and, in turn, its thermal properties are determined by this very
ordered arrangement of the various monomer units that compose a polymer. The ability to correctly and effectively interpret
differential scanning calorimetry data for any one polymer stems from an understanding of a polymer’s composition. As such,
some of the more essential dynamics of polymers and their structures are briefly addressed below.

Figure 2.8.28 Schematic diagram of monomer circles polymerizing to form a polymer chain. In turn, example of ethylene
monomers polymerizing to form polyethylene. Copyright FIMMTECH Inc. Used with Permission.
An aspect of the ordered arrangement of a polymer is its degree of polymerization, or, more simply, the number of repeating units
within a polymer chain. This degree of polymerization plays a role in determining the molecular weight of the polymer. The
molecular weight of the polymer, in turn, plays a role in determining various thermal properties of the polymer such as the
perceived melting temperature.
Related to the degree of polymerization is a polymer’s dispersity, i.e. the uniformity of size among the particles that compose a
polymer. The more uniform a series of molecules, the more monodisperse the polymer; however, the more non-uniform a series of
molecules, the more polydisperse the polymer. Increases in initial transition temperatures follow an increase in polydispersity. This
increase is due to higher intermolecular forces and polymer flexibility in comparison to more uniform molecules.
In continuation with the study of a polymer’s overall composition is the presence of cross-linking between chains. The ability for
rotational motion within a polymer decreases as more chains become cross-linked, meaning initial transition temperatures will
increase due to a greater level of energy needed to overcome this restriction. In turn, if a polymer is composed of stiff functional
groups, such as carbonyl groups, the flexibility of the polymer will drastically decrease, leading to higher transitional temperatures
as more energy will be required to break these bonds. The same is true if the backbone of a polymer is composed of stiff molecules,
like aromatic rings, as this also causes the flexibility of the polymer to decrease. However, if the backbone or internal structure of
the polymer is composed of flexible groups, such as aliphatic chains, then either the packing or flexibility of the polymer decreases.
Thus, transitional temperatures will be lower as less energy is needed to break apart these more flexible polymers.
Lastly, the actual bond structure (i.e. single, double, triple) and chemical properties of the monomer units will affect the transitional
temperatures. For examples, molecules more predisposed towards strong intermolecular forces, such as molecules with greater
dipole to dipole interactions, will result in the need for higher transitional temperatures to provide enough energy to break these
interactions.
In terms of the relationship between heat capacity and polymers: heat capacity is understood to be the amount of energy a unit or
system can hold before its temperature raises one degree; further, in all polymers, there is an increase in heat capacity with an
increase in temperature. This is due to the fact that as polymers are heated, the molecules of the polymer undergo greater levels of

2.8.19 https://chem.libretexts.org/@go/page/55853
rotation and vibration which, in turn, contribute to an increase in the internal energy of the system and thus an increase in the heat
capcity of the polymer.
In knowing the composition of a polymer, it becomes easier to not only pre-emptively hypothesize the results of any DSC analysis
but also troubleshoot why DSC data does not seem to corroborate with the apparent properties of a polymer.
Note, too, that there are many variations in DSC techniques and types as they relate to characterization of polymers. These
differences are discussed below.
Standard DSC (Heat Flux DSC)
The composition of a prototypical, unmodified DSC includes two pans. One is an empty reference plate and the other contains the
polymer sample. Within the DSC system is also a thermoelectric disk. Calorimetric measurements are then taken by heating both
the sample and empty reference plate at a controlled rate, say 10 °C/min, through the thermoelectric disk. A purge gas is admitted
through an orifice in the system, which is preheated by circulation through a heating block before entering the system.
Thermocouples within the thermoelectric disk then register the temperature difference between the two plates. Once a temperature
difference between the two plates is measured, the DSC system will alter the applied heat to one of the pans so as to keep the
temperature between the two pans constant. In Figure 2.8.29 is a cross-section of a common heat flux DSC instrument.

Figure 2.8.29 Schematic diagram of a heat flux DSC. Used with permission Copyright TA Instruments.
The resulting plot that is one in which the heat flow is understood to be a function of temperature and time. As such, the slope at
any given point is proportional to the heat capacity of the sample. The plot as a whole, however, is reperesentative of thermal
events within the polymer. The orientation of peaks or stepwise movements within the plot, therefore, lend themselves to
interpretation as thermal events.
To interpret these events, it is important to define the thermodynamic system of the DSC instrument. For most heat flux systems,
the thermodynamic system is understood to be only the sample. This means that when, for example, an exothermic event occurs,
heat from the polymer is released to the outside environment and a positive change is measured on the plot. As such, all exothermic
events will be positive shifts within the plot while all endothermic events will be negative shifts within the plot. However, this can
be flipped within the DSC system, so be sure to pay attention to the orientation of your plot as “exo up” or “exo down.” See Figure
2.8.30 for an example of a standard DSC plot of polymer poly(ethylene terephthalate) (PET). By understanding this relationship

within the DSC system, the ability to interpret thermal events, such as the ones described below, becomes all the more
approachable.

2.8.20 https://chem.libretexts.org/@go/page/55853
Figure 2.8.30 Standard Exo up Heat Flux DSC Spectrum of the PET polymer. Adapted from B. Demirel, A. Yaraș, and H. Elçiçek,
BAÜ Fen Bil. Enst. Derg. Cilt, 2011, 13, 26.
Heat Capacity (Cp)
As previously stated, a typical plot created via DSC will be a measure of heat flow vs temperature. If the polymer undergoes no
thermal processes, the plot of heat flow vs temperature will be zero slope. If this is the case, then the heat capacity of the polymer is
proportional to the distance between the zero-slopped line and the x-axis. However, in most instances, the heat capacity is
measured to be the slope of the resulting heat flow vs temperature plot. Note that any thermal alteration to a polymer will result in a
change in the polymer’s heat capacity; therefore, all DSC plots with a non-zero slope indicate some thermal event must have
occurred.
However, it is also possible to directly measure the heat capacity of a polymer as it undergoes some phase change. To do so, a heat
capacity vs temperature plot is to be created. In doing so it becomes easier to zero in on and analyze a weak thermal event in a
reproducible manner. To measure heat capacity as a function of increasing temperature, it is necessary to divide all values of a
standard DSC plot by the measured heating rate.
For example, say a polymer has undergone a subtle thermal event at a relatively low temperature. To confirm a thermal event is
occurring, zero in on the temperature range the event was measured to have occurred at and create a heat capacity vs temperature
plot. The thermal event becomes immediately identifiable by the presence of a change in the polymer’s heat capacity as shown in
Figure 2.8.31.

Figure 2.8.31 Direct DSC heat capacity measurement of a phase change material at the melting temperature. Adapted from P. Giri
and C. Pal, Mod. Chem. Appl., 2014, 2, 142.
Glass Transition Temperature (Tg)

As a polymer is continually heated within the DSC system, it may reach the glass transition: a temperature range under which a
polymer can undergo a reversible transition between a brittle or viscous state. The temperature at which this reversible transition
can occur is understood to be the glass transition temperature (Tg); however, make note that the transition does not occur suddenly
at one temperature but, instead, transitions slowly across a range of temperatures.
Once a polymer is heated to the glass transition temperature, it will enter a molten state. Upon cooling the polymer, it loses its
elastic properties and instead becomes brittle, like glass, due to a decrease in chain mobility. Should the polymer continue to be
heated above the glass transition temperature, it will become soft due to increased heat energy inducing different forms of
transitional and segmental motion within the polymer, promoting chain mobility. This allows the polymer to be deformed or
molded without breaking.

2.8.21 https://chem.libretexts.org/@go/page/55853
Upon reaching the glass transition range, the heat capacity of the polymer will change, typically become higher. In turn, this will
produce a change in the DSC plot. The system will begin heating the sample pan at a different rate than the reference pan to
accommodate this change in the polymer’s heat capacity. Figure 2.8.32 is an example of the glass transition as measured by DSC.
The glass transition has been highlighted, and the glass transition temperature is understood to be the mid-point of the transitional
range.

Figure 2.8.32 Standard Exo up Heat Flux DSC Spectrum of the PET polymer. Zoomed in on Glass Transition. Adapted from B.
Demirel, A. Yaraș, and H. Elçiçek, BAÜ Fen Bil. Enst. Derg. Cilt, 2011, 13, 26.
While the DSC instrument will capture a glass transition, the glass transition temperature cannot, in actuality, be exactly defined
with a standard DSC. The glass transition is a property that is completely dependent on the extent that the polymer is heated or
cooled. As such, the glass transition is dependent on the applied heating or cooling rate of the DSC system. Therefore, the glass
transition of the same polymer can have different values when measured on separate occasions. For example, if the applied cooling
rate is lower during a second trial, then the measured glass transition temperature will also be lower.
However, in having a general knowledge of the glass transition temperature, it becomes possible to hypothesize the polymers chain
length and structure. For example, the chain length of a polymer will affect the number of Van der Waal or entangling chain
interactions that occur. These interactions will in turn determine just how resistant the polymer is to increasing heat. Therefore, the
temperature at which Tg occurs is correlated to the magnitude of chain interactions. In turn, if the glass transition of a polymer is
consistently shown to occur quickly at lower temperatures, it may be possible to infer that the polymer has flexible functional
groups that promote chain mobility.
Crystallization (Tc)
Should a polymer sample continue to be heated beyond the glass transition temperature range, it becomes possible to observe
crystallization of the polymer sample. Crystallization is understood to be the process by which polymer chains form ordered
arrangements with one another, thereby creating crystalline structures.
Essentially, before the glass transition range, the polymer does not have enough energy from the applied heat to induce mobility
within the polymer chains; however, as heat is continually added, the polymer chains begin to have greater and greater mobility.
The chains eventually undergo transitional, rotational, and segmental motion as well as stretching, disentangling, and unfolding.
Finally, a peak temperature is reached and enough heat energy has been applied to the polymer that the chains are mobile enough to
move into very ordered parallel, linear arrangements. At this point, crystallization begins. The temperature at which crystallization
begins is the crystallization temperature (Tc).
As the polymer undergoes crystalline arrangements, it will release heat since intramolecular bonding is occurring. Because heat is
being released, the process is exothermic and the DSC system will lower the amount of heat being supplied to the sample plate in
relation to the reference plate so as to maintain a constant temperature between the two plates. As a result, a positive amount of
energy is released to the environment and an increase in heat flow is measured in an “exo up” DSC system, as seen in Figure
2.8.33. The maximum point on the curve is known to be the Tc of the polymer while the area under the curve is the latent energy of

crystallization, i.e., the change in the heat content of the system associated with the amount of heat energy released by the polymer
as it undergoes crystallization.

2.8.22 https://chem.libretexts.org/@go/page/55853
Figure 2.8.33 Standard Exo up Heat Flux DSC Spectrum of the PET polymer. Crystallization temperature is highlighted. Adapted
from B. Demirel, A. Yaraș, and H. Elçiçek, BAÜ Fen Bil. Enst. Derg. Cilt, 2011, 13, 26.

The degree to which crystallization can be measured by the DSC is dependent not only on the measured conditions but also on the
polymer itself. For example, in the case of a polymer with very random ordering, i.e., an amorphous polymer, crystallization will
not even occur.

In knowing the crystallization temperature of the polymer, it becomes possible to hypothesize on the polymer’s chain structure,
average molecular weight, tensile strength, impact strength, resistance to solvents, etc. For example, if the polymer tends to have a
lower crystallization temperature and a small latent heat of crystallization, it becomes possible to assume that the polymer may
already have a chain structure that is highly linear since not much energy is needed to induce linear crystalline arrangements.
In turn, in obtaining crystallization data via DSC, it becomes possible to determine the percentage of crystalline structures within
the polymer, or, the degree of crystallinity. To do so, compare the latent heat of crystallization, as determined by the area under the
crystallization curve, to the latent heat of a standard sample of the same polymer with a known crystallization degree.
Knowledge of the polymer sample’s degree of crystallinity also provides an avenue for hypothesizing the composition of the
polymer. For example, having a very high degree of crystallinity may suggest that the polymer contains small, brittle molecules that
are very ordered.
Melting behavior (Tm)
As the heat being applied pushes the temperature of the system beyond Tc, the polymer begins to approach a thermal transition
associated with melting. In the melting phase, the heat applied provides enough energy to, now, break apart the intramolecular
bonds holding together the crystalline structure, undoing the polymer chains’ ordered arrangements. As this occurs, the temperature
of the sample plate does not change as the applied heat is no longer being used to raise the temperature but instead to break apart
the ordered arrangements.
As the sample melts, the temperature slowly increases as less and less of the applied heat is needed to break apart crystalline
structures. Once all the polymer chains in the sample are able to move around freely, the temperature of the sample is said to reach
the melting temperature (Tm). Upon reaching the melting temperature, the applied heat begins exclusively raising the temperature
of the sample; however, the heat capacity of the polymer will have increased upon transitioning from the solid crystalline phase to
the melt phase, meaning the temperature will increase more slowly than before.
Since, during the endothermic melting process of the polymer, most of the applied heat is being absorbed by the polymer, the DSC
system must substantially increase the amount of heat applied to the sample plate so as to maintain the temperature between the
sample plate and the reference plate. Once the melting temperature is reached, however, the applied heat of the sample plate
decreases to match the applied heat of the reference plate. As such, since heat is being absorbed from the environment, the resulting
“exo up” DSC plot will have a negative curve as seen in Figure 2.8.34 where the lowest point is understood to be the melt phase
temperature. The area under the curve is, in turn, understood to be the latent heat of melting, or, more precisely, the change in the
heat content of the system associated with the amount of heat energy absorbed by the polymer to undergo melting.

2.8.23 https://chem.libretexts.org/@go/page/55853
Figure 2.8.34 Standard Exo up Heat Flux DSC Spectrum of the PET polymer. Melting temperature is highlighted. Adapted from B.
Demirel, A. Yaraș, and H. Elçiçek, BAÜ Fen Bil. Enst. Derg. Cilt, 2011, 13, 26.
Once again, in knowing the melting range of the polymer, insight can be gained on the polymer’s average molecular weight,
composition, and other properties. For example, the greater the molecular weight or the stronger the intramolecular attraction
between functional groups within crosslinked polymer chains, the more heat energy that will be needed to induce melting in the
polymer.
Modulated DSC: an Overview
While standard DSC is useful in characterization of polymers across a broad temperature range in a relatively quick manner and
has user-friendly software, it still has a series of limitations with the main limitation being that it is highly operator dependent.
These limitations can, at times, reduce the accuracy of analysis regarding the measurements of Tg, Tc and Tm, as described in the
previous section. For example, when using a synthesized polymer that is composed of multiple blends of different monomer
compounds, it can become difficult to interpret the various transitions of the polymer due to overlap. In turn, some transitional
events are completely dependent on what the user decides to input for the heating or cooling rate.
To resolve some of the limitations associated with standard DSC, there exists modulated DSC (MDSC). MDSC not only uses a
linear heating rate like standard DSC, but also uses a sinusoidal, or modulated, heating rate. In doing so, it is as though the MDSC
is performing two, simultaneous experiements on the sample.
What is meant by a modulated heating rate is that the MDSC system will vary the heating rate of the sample by a small range of
heat across some modulating period of time. However, while the temperature rate of change is sinusoidal, it is still ultimately
increasing acorss time as indicated in Figure 2.8.35. In turn, Figure 2.8.36 also shows the sinusoidal heating rate as a function of
time overlaying the linear heating rate of standard DSC. The linear heating rate of DSC is 2 °C/min and the modulated heating rate
of MDSC varies from roughly ~0.1 °C/min and ~3.8 °C/min across a period of time.

Figure 2.8.35 Schematic of sample temperature as a function of time with an underlying linear heating rate of Standard DSC.
Adapted from E. Verdonck, K. Schaap, and L. C. Thomas, Int. J. Pharm., 1999, 192, 3.

2.8.24 https://chem.libretexts.org/@go/page/55853
Figure 2.8.36 Heating rate as a function of time with underlying linear heating rate. Adapted from E. Verdonck, K. Schaap, and L.
C. Thomas, Int. J. Pharm., 1999, 192, 3.
By providing two heating rates, a linear and a modulated one, MDSC is able to measure more accurately how heating rates affect
the rate of heat flow within a polymer sample. As such, MDSC offers a means to eliminate the applied heating rate aspects of
operator dependency.
In turn, the MDSC instrument also performs mathematical processes that separate the standard DSC plot into reversing and a non-
reversing components. The reversing signal is representative of properties that respond to temperature modulation and heating rate,
such as glass transition and melting. On the other hand, the non-reversing component is representative of kinetic, time-dependent
process such as decomposition, crystallization, and curing. Figure 2.8.37 provides an example of such a plot using PET.

Figure 2.8.37 Modulated DSC signals of PET, split into reversing and non-reversing components as well as total heat flow,
showcasing the related transitional temperatures. Adapted from E. Verdonck, K. Schaap, and L. C. Thomas, Int. J. Pharm., 1999,
192, 3.
The mathematics behind MDSC is most simply represented by this formula: dH/dt = Cp(dT/dt) + f(T,t) where dH/dt is the total
change in heat flow that would be derived from a standard DSC. Cp is heat capacity derived from modulated heating rate, dT/dt is
representative of both the linear and modulated heating rate, and f(T,t) is representative of kinetic, time-dependent events, i.e the
non-reversing signal. When combining Cp and dT/dt, creating Cp(dT/dt), the reversing signal is produced. The non-reversing signal
is, therefore, found by simply subtracting the reversing signal from the total heat flow singal, i.e. dH/dt = Cp(dT/dt) + f(T,t)
As such, MDSC is capable of independently measuring not only total heat flow but also the heating rate and kinetic components of
said heat flow, meaning MDSC can break down complex or small transitions into their many singular components with improved
sensitivity, allowing for more accurate analysis. Below are some cases in which MDSC proved to be useful for analytics.
Modulated DSC: Advanced Analysis of Tg
Using a standard DSC, it can be difficult to ascertain the accuracy of measured transitions that are relatively weak, such as Tg, since
these transitions can be overlapped by stronger, kinetic transitions. This is quite the problem as missing a weak transition could
cause the misinterpretation of polymer to be a uniform sample as opposed to a polymer blend. To resolve this, it is useful to split
the plot into its reversing component, i.e. the portion which will contain heat dependent properties like Tg, and its non-reversing,
kinetic component.
For example, shown in the Figure 2.8.38 is the MDSC of an unknown polymer blend which, upon analysis, is composed of PET,
amorphous polycarbonate (PC), and a high density polyethylene (HDPE). Looking at the reversing signal, the Tg of polycarbonate
is around 140 °C and the Tg of PET is around 75 °C. As seen in the total heat flow signal, which is representative of a standard
DSC plot, the Tg of PC would have been more difficult to analyze and, as such, may have been incorrectly analyzed.

2.8.25 https://chem.libretexts.org/@go/page/55853
Figure 2.8.38 MDSC signals of a polymer blend composed of HDPE, PC, and PET. Adapted from E. Verdonck, K. Schaap, and L.
C. Thomas, Int. J. Pharm., 1999, 192, 3.

Modulated DSC: Advanced Analysis of Tm


Further, there are instances in which a polymer or, more likely, a polymer blend will produce two different sets of crystalline
structures. With two crystalline structures, the resulting melting peak will be poorly defined and, thus, difficult to analyze via a
standard DSC.
Using MDSC, however, it becomes possible to isolate the reversing signal, which will contain the melting curve. Through isolation
of the reversing signal, it becomes clear that there is an overlapping of two melting peaks such that the MDSC system reveals two
melting points. For example, as seen in Figure 2.8.39 the analysis of a poly(lactic acid) polymer (PLA) with 10% wt of a plasticize
(P600) reveals two melting peaks in the reversing signal not visible in the total heat flow. The presence of two melting peaks could,
in turn, suggest the formation of two crystalline structures within the polymer sample. Other interpretations are, of course, possible
via analyzing the reversing signal.

Figure 2.8.39 MDSC of PLA with varying concentrations of a plasticizer. The solid lines represent total heat flow and the dashed
lines represent reversing heat flow. Adapted from Z. Kulinski and E. Piorkowska, Polymer, 2005, 46, 10290.
Modulated DSC: Analysis of Polymer Aging
In many instances, polymers may be left to sit in refrigeration or stored at temperatures below their respective glass transition
temperatures. By leaving a polymer under such conditions, the polymer is situated to undergo physical aging. Typically, the more
flexible the chains of a polymer are, the more likely they will undergo time-related changes in storage. That is to say, the polymer
will begin to undergo molecular relaxation such that the chains will form very dense regions while they conglomerate together. As
the polymer ages, it will tend towards embrittlement and develop internal stresses. As such, it is very important to be aware if the
polymer being studied has gone through aging while in storage.
If a polymer has undergone physical aging, it will develop a new endothermic peak when undergoing thermal analysis. This occurs
because, as the polymer is being heated, the polymer chains absorb heat, increase mobility, and move to a more relaxed condition
as time goes on, transforming back to pre-aged conditions. In turn an endothermic shift, in association with this heat absorbance,
will occur just before the Tg step change. This peak is known as the enthalpy of relaxation (ΔHR).
Since the Tg and ΔHR are relatively close to one another energy-wise, they will tend to overlap, making it difficult to distinguish
the two from one another. However, ΔHR is a kinetics dependent thermal shift while Tg is a heating dependent thermal shift;
therefore, the two can be separated into a non-reversing and reversing plot via MDSC and be independently analyzed.
Figure 2.8.40 is an example of an MDSC plot of a polymer blend of PET, PC, and HDPE in which the enthalpy of relaxation of
PET is visible in the dashed non reversing signal around 75 °C. In turn, within the reversing signal, the glass transition of PET is
visible around 75 °C as well.

2.8.26 https://chem.libretexts.org/@go/page/55853
Figure 2.8.40 MDSC signals of polymer blend composed of HDPE, PC, and PET. Adapted from E. Verdonck, K. Schaap, and L. C.
Thomas, Int. J. Pharm., 1999, 192, 3.

Quasi-isothermal DSC
While MDSC is a strong step in the direction of elinating operator error, it is possible to have an even higher level of precision and
accuracy when analyzing a polymer. To do so, the DSC system must expose the sample to quasi-isothermal conditions. In creating
quasi-isothermal conditions, the polymer sample is held at a specific temperature for extended periods of time with no applied
heating rate. With the heating rate being efficticely zero, the conditions are isothermal. The temperature of the sample may change,
but the change will be derived solely from a kinetic transition that has occurred within the polymer. Once a kinetic transition has
occurred within the polymer, it will absorb or release some heat, which will raise or decrease the temperature of the system without
the application of any external heat.
In creating these conditions, issues created by the variation of the applied heating rate by operators is no longer a large concern.
Further, in subjecting a polymer sample to quasi-isothermal conditions, it becomes possible to get improved and more accurate
measurements of heat dependent thermal events, such as events typically found in the reversing signal, as a function of time.
Quasi-isothermal DSC: Improved Glass Transition
As mentioned earlier, the glass transition is volatile in the sense that it is highly dependent on the heating and cooling rate of the
DSC system as applied by the operator. An minor change in the heating or cooling rate between two experimental measurements of
the same polymer sample can result in fairly different measured glass transitions, even though the sample itself has not been
altered.
Remember also, that the glass transition is a measure of the changing Cp of the polymer sample as it crosses certain heat energy
thresholds. Therefore, it should be possible to capture a more accurate and precise glass transition under quasi-isothermal
conditions since these conditions produce highly accurate Cpmeasurements as a function of time.
By applying quasi-isothermal conditions, the polymer’s Cp can be measured in fixed-temperature steps within the apparent glass
transition range as measured via standard DSC. In measuring the polymer across a set of quasi-isothermal steps, it becomes
possible to obtain changing Cp rates that, in turn, would be nearly reflective of an exact glass transition range for a polymer.
In Figure 2.8.41 the glass transition of polystyrene is shown to vary depending on the heating or cooling rate of the DSC; however,
when applying qusi-isothermal conditions and measuring the heat capacity at temperature steps produces a very accurate glass
transition that can be used as a standard for comparison.

Figure 2.8.41 MDSC plot of polystyrene measuring the glass transition as changes in heat capacity as a function of temperature
using either a heating or cooling rate as indicated by the solid lines. The dotted line indicates a quasi-isothermal measurement of
the glass transition of polystyrene. Adapted from L. C. Thomas, A. Boller, I. Okazaki, and B. Wunderlich, Thermochim. Acta,
1997, 291, 85.

2.8.27 https://chem.libretexts.org/@go/page/55853
Low-Temperature Specific Heat Measurements for Magnetic Materials
Magnetic materials attract the attention of researchers and engineers because of their potential for application in magnetic and
electronic devices such as navigational equipment, computers, and even high-speed transportation. Perhaps more valuable still,
however, is the insight they provide into fundamental physicals. Magnetic materials provide an opportunity for studying exotic
quantum mechanical phenomena such as quantum criticality, superconductivity, and heavy fermionic behavior intrinsic to these
materials. A battery of characterization techniques exist for measuring the physical properties of these materials, among them a
method for measuring the specific heat of a material throughout a large range of temperatures. Specific heat measurments are an
important means of determining the transition temperature of magnetic materials—the temperature below which magnetic ordering
occurs. Additionally, the functionality of specific heat with temperature is characteristic of the behavior of electrons within the
material and can be used to classify materials into different categories.
Temperature-dependence of Specific Heat
The molar specific heat of a material is defined as the amount of energy required to raise the temperature of 1 mole of the material
by 1 K. This value is calculated theoretically by taking the partial derivative of the internal energy with respect to temperature. This
value is not a constant, as it is typically treated in high-school science courses: it depends on the temperature of the material.
Moreover, the temperature-dependence itself also changes based on the type of material. There are three broad families of solid
state materials defined by their specific heat behaviors. Each of these families is discussed in the following sections.
Insulators
Insulators have specific heat with the simplest dependence on temperature. According to the Debye theory of specific heat, which
models materials as phonons (lattice vibrational modes) in a potential well, the internal energy of an insulating system is given by
2.8.11 , where TD is the Debye temperature, defined as the temperature associated with the energy of the highest allowed phonon

mode of the material. In the limit that T<<TD, the energy expression reduces to 2.8.12 .
4 TD /T 3
9N kB T x
U  = ∫ dx (2.8.11)
3 x
T 0
e −1
D

4 4
3 π N kB T
U  = (2.8.12)
3
5T
D

For most magnetic materials, the Debye temperature is several orders of magnitude higher than the temperature at which magnetic
ordering occurs, making this a valid approximation of the internal energy. The specific heat derived from this expression is given
by 2.8.13
4
δU 12 π N kB 3 3
Cν   = = T = βT (2.8.13)
3
δT 5T
D

The behavior described by the Debye theory accurately matches experimental measurements of specific heat for insulators at low
temperatures. Normal insulators, then, have a T3 dependence in the specific heat that is dominated by contributions from phonon
excitations. Essentially all energy absorbed by insulating materials is stored in the vibrational modes of a solid lattice. At very low
temperatures this contribution is very small, and insulators display a high sensitivity to changes in heat energy.
Metals: Fermi Liquids
While the Debye theory of specific heat accurately describes the behavior of insulators, it does not adequately describe the
temperature dependence of the specific heat for metallic materials at low temperatures, where contributions from delocalized
conduction electrons becomes significant. The predictions made by the Debye model are corrected in the Einstein-Debye model of
specific heat, where an additional term describing the contributions from the electrons (as modeled by a free electron gas) is added
to the phonon contribution. The internal energy of a free electron gas is given by 2.8.14 ,where g(Ef) is the density of states at the
Fermi level, which is material dependent. The partial derivative of this expression with respect to temperature yields the specific
heat of the electron gas, 2.8.15 .
2
π 2
U = (kB T ) g(Ef ) + U0 (2.8.14)
6

2
π
2
Cν = k g(Ef )T = γT (2.8.15)
B
3

2.8.28 https://chem.libretexts.org/@go/page/55853
Combining this expression with the phonon contribution to specific heat gives the expression predicted by the Einstein-Debye
model, 2.8.16 .
2 4
pi 12 π N kB 3 3
2
Cν = k g(Ef )T   + T = γT   +  β T (2.8.16)
B 3
3 5T
D

This is the general expression for the specific heat of a Fermi liquid—a variation on the Fermi gas in which fermions (typically
electrons) are allowed to interact with each other and form quasiparticles—weakly bound and often short-lived composites of more
fundamental particles such as electron-hole pairs or the Cooper pairs of BCS superconductor theory.
Most metallic materials follow this behavior and are thus classified as Fermi liquids. This is easily confirmed by measuring the heat
capacity as a function of temperature and linearizing the results by plotting C/T vs. T2. The slope of this graph equal the coefficient
β, and the y-intercept is equal to γ. The ability to obtain these coefficients is important for gaining understanding of some unique
physical phenomena. For example, the compound YbRh2Si2 is a heavy fermionic material—a material with charge carriers that
have an “effective” mass much greater than the normal mass of an electron. The increased mass is due to coupling of magnetic
moments between conduction electrons and localized magnetic ions. The coefficient γ is related to the density of states at the Fermi
level, which is dependent on the carrier mass. Determination of this coefficient via specific heat measurements provides a way to
determine the effective carrier mass and the coupling strength of the quasiparticles.
Additionally, knowledge of Fermi-liquid behavior provides insight for application development. The temperature dependence of
the specific heat shows that the phonon contribution dominates at higher temperatures, where the behavior of metals and insulators
is very similar. At low temperatures, the electronic term is dominant, and metals can absorb more heat without a signficant change
in temperature. As will be discussed breifly later, this property of metals is utilized in low-temperature refrigeration systems for
heat storage at low temperatures.
Metals: non-Fermi liquids
While most metals fall under the category of Fermi liquids, there are some that show a different dependence on temperature.
Naturally, these are classified as non-Fermi liquids. Often, deviation from Fermi-liquid behavior is an indicator of some of the
interesting physical phenomena that currently garner the attention of many condensed matter researchers. For instance, non-Fermi
liquid behavior has been observed near quantum critical points. Classically, fluctuations in physical properties such as magnetic
susceptibility and resistivity occur near critical points which include phase changes or magnetic ordering transitions. Normally,
these fluctuations are suppressed at low temperatures—at absolute zero, classical systems collapse into the lowest energy state and
remain stable; However, when the critical transition temperature is lowered by the application of pressure, doping, or magnetic
field to absolute zero, the fluctuations are enhanced as the temperature approaches absolute zero, propagating throughout the whole
of the material. As this is not classically allowed, this behavior indicates a quantum mechanical effect at play that is currently not
well understood. The transition point is then called a quantum critical point. Non-fermi liquid behavior as identified by deviations
in the expected specific heat, then, is used to identify materials that can provide an experimental basis for development of a theory
that describes the physics of quantum criticality.
Determination of magnetic transition temperatures viz specific heat measurements
While analysis of the temperature dependence of specific heat is a vital tool for studying the strange physical behaviors of quantum
mechanics in solid state materials, these are studied by only a small subsection of the physics community. The utility of specific
heat measurements is not limited to a few niche subjects, however. Possibly the most important use for specific heat measurements
is the determination of critical transition temperatures. For any sort of physical state transition—phase transitions, magnetic
ordering, transitions to superconducting states—a sharp increase in the specific heat occurs during the transition. This increase in
specific heat is the reason why, for example, water does not change temperature as it changes from a liquid to a solid. These
increases are quite obvious in plots of the specific heat vs. temperature as seen in Figure 2.8.42. These transition-associated peaks
are called Schottky anomalies, as normal specific heat behavior is not followed near to the transition temperature.

2.8.29 https://chem.libretexts.org/@go/page/55853
Figure 2.8.42 A Schottky anomaly at the magnetic ordering temperature of URu2Si2 . Adapted with permission from J. C. Lashley,
M. F. Hundley, A. Migliori, J. L. Sarrao, P. G. Pagliuso, T. W. Darling, M. Jaime, J. C. Cooley, W. L. Hults, L. Morales, D. J.
Thoma, J. L. Smith, J. Boerio-Goates, B. F. Woodfield, G. R.Stewart, R. A. Fisher, and N. E. Phillips. Cryogenics, 2003, 43, 369.
Copyright: Elsevier publishing.
For the purposes of this chapter, the following sections will focus on specific heat measurements as they relate to magnetic ordering
transititions. The following sections will describe the practical aspects of measuring the specific heat of these materials.
A practical guide to low-temperature specific heat measurements
The thermal relaxation method of measurement
Specific heat is measured using a calorimeter. The design of basic calorimeters for use over a short range of temperatures is
relatively simple. They consist of a sample with a known mass and an unknown specific heat, an energy source which provides
heat energy to the sample, a heat reservoir (of known mass and specific heat) that absorbs heat from the sample, insulation to
provide adiabatic conditions inside the calorimeter, and probes for measuring the temperature of the sample and the reservoir. The
sample is heated with a pulse to a temperature higher than the heat reservoir, which decreases as energy is absorbed by the reservoir
until a thermal equilibrium is established. The total energy change is calculated using the specific heat and temperature change of
the reservoir. The specific heat of the sample is calculated by dividing the total energy change by the product of the mass of the
sample and the temperature change of the sample.
However, this method of measurement produces an average value of the specific heat over the range of the change in temperature
of the sample, and therefore, is insufficient for producing accurate measurements of the specific heat as a function of temperature.
The solution, then, is to minimize the temperature change by reducing the amount of heat added to the system; yet, this presents
another obstacle to making measurement as, in general, the temperature change of the reservoir is much smaller than that of the
sample. If the change in temperature of the sample is minimized, the temperature change of reservoir becomes too small to measure
with precision. A more direct method of measurement, then, seems to be required.
Fortunately, such a method exists: it is known as the thermal relaxation method. This method involves measurement of the specific
heat without the need for precise knowledge of temperature changes in the reservoir. In this method, solid samples are affixed to a
platform. Both the specific heat of the sample and the platform itself contribute to the measured specific heat; therefore, the
contribution from the platform must be subtracted. This contribution is determined by measuring the specific heat without a sample
present. Both the sample and the platform are in thermal contact with a heat reservoir at low temperature as depicted in Figure
2.8.43.

Figure 2.8.43 A schematic representation depicting the thermal connection between the sample and the heat reservoir.

2.8.30 https://chem.libretexts.org/@go/page/55853
A heat pulse is delivered to the sample to produce a minimal increase in the temperature of the sample. The temperature is
measured vs. time as it decays back to the temperature of the reservoir as shown in 2.8.44.

Figure 2.8.44 The low-temperature heat-pulse temperature decay for a copper standard. Reused with permission from J. S. Hwang,
K. J. Lin, and C. Tien. Rev. Sci. Instrum., 1997, 68, 94. Copyright: AIP publishing.
The temperature of the sample decays according to 2.8.17 , where T0 is the temperature of the heat reservoir, and ΔT is the
temperature difference between the initial sample temperature and the reservoir temperature. The decay time constant τ is directly
related to the specific heat of the sample by 2.8.18 , Where K is the thermal conductance of the thermal link between the sample
and the heat reservoir. In order for this to be valid, however, the thermal conductance must be sufficiently large that the energy
transfer from the heated sample to the reservoir can be treated as a single process. If the thermal conduction is poor, a two-τ
behavior arises corresponding to two separate processes with different time constants—slow heat transfer from the sample to the
platform, and fast transfer from the platform to the reservoir. Figure 2.8.45 shows a relaxation curve in which the two- τ behavior
plays a significant role.
t/τ
T = Δe   +  T0 (2.8.17)

τ   =  Cp /K (2.8.18)

Figure 2.8.45 A thermal relaxation decay graph for a sample of a Pb sample displaying the two-τ effect. Reused with permission
from J. S. Hwang, K. J. Lin, and C. Tien. Rev. Sci. Instrum., 1997, 68, 94. Copyright: AIP publishing.
The two-τ effect is generally undesireable for making measurements. It can be avoided by reducing thermal conductance between
the sample and the platform, effectively making the contribution from the heat transfer from the sample to the platform
insignificant compared to the transfer from the platform to the reservoir; however, if the conductance between the sample and the
platform is too low, the time required to reach thermal equilibrium becomes excessively long, translating into very long
measurement times. It is necessary, then, to optimize the conductance to compensate for both of these issues. This essentially
provides a limitation on the temperature range over which these effects are insignificant.
In order to measure at different temepratures, the temperature of the heat reservoir is increased stepwise from the lowest
temperature until the desired temperature range is covered. At each step, the temperature is allowed to equilibrate, and a data point
is measured.
Instrumentation
Thermal relaxation calorimeters use advanced technology to make precise measurements of the specific heat using components
made of highly specialized materials. For example, the sample platform is made of synthetic sapphire which is used as a standard
material, the grease which is applied to the sample to provide even thermal contact with the platform is a special hydrocarbon-
based material which can withstand millikelvin temperatures without creeping, cracking, or releasing vapor, and the resistance
thermometers used for ultralow temperatures are often made of treated graphite or germanium. The culmination of years of
materials science research and careful engineering has produced instrumentation with the capability for precise measurements from

2.8.31 https://chem.libretexts.org/@go/page/55853
temperatures down to the millikelvin level. There are four main systems that function to provide the proper conditions for
measurement: the reservoir temperature control, the sample temperature control, the magnetic field control, and the pressure
control system. The essential components of these systems will be discussed in more detail in the following sections with special
emphasis on the cooling systems that allow these extreme low temperatures to be achieved.
Cooling systems
The first of these is responsible for maintaining the low baseline temperature to which the sample temperature relaxes. This is
typically accomplished with the use of liquid helium cryostats or, in more recent years, so-called “cryogen-free” pulse tube coolers.
A cryostat is simply a bath of cryogenic fluid that is kept in thermal contact with the sample. The fluid bath may be static or may be
pumped through a circulation system for better cooling. The cryostat must also be thermally insulated from the external
environment in order to maintain low temperatures. Insulation is provided by a metallic vacuum dewar: The vacuum virtually
eliminates conuductive or convective heat transfer from the environment and the reflective metallic outer sleeve acts as a radiation
shield. For the low temperatures required to observe some magnetic transitions, liquid helium is generally required. 4He liquefies at
4.2 K, and the rarer (and much more expensive) isotope, 3He, liquefies at 1.8 K. For temperatures lower than 1.8 K, modern
instruments employ evaporative attachments such as a 1-K pot, 3He refrigerator, or a dilution refrigerator. The 1-K pot is so named
because it can achieve temperatures down to 1 K. It consists of a small vessel filled with liquid 4He under reduced pressure. Heat is
absorbed as the liquid evaporates and is carried away by the vapor. The 3He refrigerator utilizes a 1-K pot for liquefaction of 3He,
then evaporation of 3He provide cooling to the sample. 3He refrigerators can provide temperatures as low as 200 mK. The dilution
refrigerator works on a similar principle, however the working fluid is a mixture of 3He and 4He. Phase separation of the 3He from
the mixture provides further heat absorption as the 3He evaporates. Dilution refrigerators can achieve temperatures as low as 0.002
K (That’s cold!). Evaporative refrigerators work only on a small area in thermal contact with the sample, rather than delivering
cooling power to the entire volume of the cryostat bath.
Cryostat baths provide very high cooling power for very efficient cooling; however, they come with a major drawback: the cost of
helium is prohibitively high. The helium vapor that boils off as it provides cooling to the sample must leave the system in order to
carry the heat away and must therefore be replaced. Even when the instrument is not in use, there is some loss of helium due to the
imperfect nature of the insulating dewars. In order to get the most use out of the helium, then, cryostat systems must always be in
use. In addition, rather than allowing expensive helium to simply escape, recovery systems for helium exhaust must be installed in
order to operate in a cost-effective manner, though these systems are not 100% efficient, and the cost of operation and maintenance
of recovery systems is not small either. “Cryogen-free” coolers provide an alternative to cryostats in order to avoid the costs
associated with helium usage and recovery.
Figure 2.8.46 shows a Gifford-McMahon type pulse tube—one example of the cryogen-free coolers.

Figure 2.8.46 A schematic representation of a Gifford-McMahon type pulse tube cooler. Adapted from P. D. Nissen. Closed Cycle
Refrigerators – Pulse Tube Coolers. Retrieved from <www.nbi.dk/~nygard/DPN-Pulsetubecoolers_low2010.pdf>.
In this type of cooler, helium gas is driven through the regenerator by a compressor. As a small volume element of the gas passes
throughout the regenerator, it drops in temperature as it deposits heat into the regenerator. The regenerator must have a high
specific heat in order to effectively absorb energy from the helium gas. For higher-temperature pulse tube coolers, the regenerator
is often made of copper mesh; however, for very low temperatures, helium has a higher specific heat than most metals.
Regenerators for this temperature range are often made of porous rare earth ceramics with magnetic transitions in the low
temperature range. The increase in specific heat near the Schottky anomaly for these materials provides the necessary capacity for
heat absorption. As the gas enters the tube at a temperature TL(from the diagram above) it is compressed, raising the temperature in
accordance with the ideal gas law. At this point, the gas is at a temperature higher than TH and excess heat is exhausted through the
heat exchanger marked X3 until the temperature is in equilibrium with TH. When the rotary valve in the compressor turns, the
expansion cycle begins, and the gas cools as it expands adiabatically to a temperature below TL. It then absorbs heat from the
sample through the heat exchanger X2. This step provides the cooling power in pulse tube coolers. Afterward, it travels back
through the regenerator at a cold temperature and reabsorbs the heat that was initially stored during compression, and regains it’s

2.8.32 https://chem.libretexts.org/@go/page/55853
original temperature through the heat exchanger X1. Figure 2.8.47 illustrates the temperature cyle experienced by a volume
element of the working gas as it moves through the pulse tube.

Figure 2.8.47 Temperature cycle of a gas element moving in the pulse tube cooler. A to B: Gas initially at TH moves through the
regenerator to heat exchanger X2 dropping to TL. B to C: Gas is compressed and the temperature rises above TH. C to D: Gas is
shunted along to X3 and drops to TH. D to E: Gas is expanded adiabatically to T<TL. E to F: Cold gas is shunted to X2 and rises to
TL, absorbing heat from the sample. F to G: Gas at TL moves through the regenerator reabsorbing heat until it reaches TH at heat
exchanger X1.
Pulse tube coolers are not truly “cryogen-free” as they are advertised, but they are preferable to cryostats because there is no net
loss of the helium in the system. However, pulse tubes are not a perfect solution. They have very low efficiency over large changes
in temperature and at very low temperatures as given by 2.8.19 .
ΔT
ζ  =  1  −   (2.8.19)
TH

As a result, pulse tube coolers consume a lot of electricity to provide the necessary cooling and may take a long time to achieve the
desired temperature. Over large temperature ranges such as the 4 – 300 K range typically used in specific heat measurements, pulse
tubes can be used in stages, with one providing pre-cooling for the next, to increase the cooling power and provide a shorter
cooling time, though this tends to increase the energy consumption. The cost of running a pulse tube system is still generally less
that that of a cryostat, however, and unlike cryostats, pulse tube systems do not have to be used constantly in order to remain cost-
effective.
Sample Conditions
While the cooling system works more or less independently, the other systems—the sample temperature control, the magnetic field
control, and the pressure control systems—work together to create the proper conditions for measurement of the sample. The
sample temperature control system provides the heat pulse used to increase the temperature of the sample before relaxation occurs.
The components of this system are incorporated into the sapphire sample platform as shown in Figure 2.8.48.

Figure 2.8.48 The sample platform with important components of the sample temperature control system. Reused with
permission from R. J. Schutz. Rev. Sci. Instrum., 1974, 45, 548. Copyright: AIP publishing.
The sample is affixed to the platform over the thermometer with a small amount of grease, which also provides thermal
conductance between the heating element and the sample. The heat pulse is delivered to the sample by running a small current
pulse through the heating element, and the response is measured by a resistance thermometer. The resistance thermometer is made
of specially-treated carbon or germanium which have standardized resistances for given temperatures. The thermometer is
calibrated to these standards to provide accurate temperature readings throughout the range of temperatures used for specific heat

2.8.33 https://chem.libretexts.org/@go/page/55853
measurements. A conductive wire provides thermal connection between the sample platform and the heat reservoir. This wire must
provide high conductivity to ensure that the heat transfer from the sample to the platform is the dominant process and prevent
significant two-τ behavior. Sample preparation is also governed by the temperature control system. The sample must be in good
thermal contact with the platform, therefore, a sample with a flat face is preferable. The volume of the sample cannot be too large,
either, or the heating element will not be able to heat the sample uniformly. A temperature gradient throughout the sample skews
the measurement of the temperature made by the thermometer. Moreover, it is impossible to assign a 1:1 correspondence between
the specific heat and temperature if the specific heat values do not correspond to a singular temperature. For the best measurements,
heat capacity samples must be cut from large single-crystals or polycrystalline solids using a hard diamond saw to prevent
contamination of the sample with foreign material.
The magnetic field control system provides magnetic fields ranging from 0 to >15 T. As was mentioned previously, strong
magnetic fields can suppress the transition to magnetically ordered states to lower temepratures, which is important for studying
quantum critical behaviors. The magnetic field control consists of a high-current solenoid and regulating electronics to ensure
stable current and field outputs.
The pressure systems controls the pressure in the sample chamber, which is physically separated from the bath by a wall which
allows thermal transfer only. While the sample is installed in the chamber, the vacuum system must be able to maintain low
pressures (~10-5 torr) to ensure that no gas is present. If the vacuum system fails, water from any air present in the system can
condense inside the sample chamber, including on the sample platform, which alters thermal conductance and throws off
measurement of the specific heat. Moreover, as the temperature in the chamber drops, water can freeze and expand in the chamber
which can cause significant damage to the instrument itself.
Conclusions
Through the application of specialized materials and technology, measurements of the specific heat have become both highly
accurate and very precise. As our measurement capabilities expand toward the 0 K limit, exciting prospects arise for completion of
our understanding, discovery of new phenomena, and development of important applications of novel magnetic materials. Specific
heat measurements, then, are a vital tool for studying magnetic materials, whether as a means of exploring the strange phenomena
of quantum physics such as quantum criticality or heavy fermions, or simply as a routine method of characterizing physical
transitions between different states.

2.8: Thermal Analysis is shared under a CC BY 4.0 license and was authored, remixed, and/or curated by Pavan M. V. Raja & Andrew R. Barron
via source content that was edited to conform to the style and standards of the LibreTexts platform; a detailed edit history is available upon
request.

2.8.34 https://chem.libretexts.org/@go/page/55853
2.9: Electrical Permittivity Characterization of Aqueous Solutions
Introduction
Permittivity (in the framework of electromagnetics) is a fundamental material property that describes how a material will affect,
and be affected by, a time-varying electromagnetic field. The parameters of permittivity are often treated as a complex function of
the applied electromagnetic field as complex numbers allow for the expression of magnitude and phase. The fundamental equation
for the complex permittivity of a substance (εs) is given by 2.9.1 , where ε’ and ε’’ are the real and imaginary components,
respectively, ω is the radial frequency (rad/s) and can be easily converted to frequency (Hertz, Hz) using 2.9.2 .
′ ′′
εs = ε (ω)  −  i ε (ω) (2.9.1)

ω  =  2πf (2.9.2)

Specifically, the real and imaginary parameters defined within the complex permittivity equation describe how a material will store
electromagnetic energy and dissipate that energy as heat. The processes that influence the response of a material to a time-varying
electromagnetic field are frequency dependent and are generally classified as either ionic, dipolar, vibrational, or electronic in
nature. These processes are highlighted as a function of frequency in Figure 2.9.1. Ionic processes refer to the general case of a
charged ion moving back and forth in response a time-varying electric field, whilst dipolar processes correspond to the ‘flipping’
and ‘twisting’ of molecules, which have a permanent electric dipole moment such as that seen with a water molecule in a
microwave oven. Examples of vibrational processes include molecular vibrations (e.g. symmetric and asymmetric) and associated
vibrational-rotation states that are Infrared (IR) active. Electronic processes include optical and ultra-violet (UV) absorption and
scattering phenomenon seen across the UV-visible range.

Figure 2.9.1 A dielectric permittivity spectrum over a wide range of frequencies. ε′ and ε″ denote the real and the imaginary part of
the permittivity, respectively. Various processes are labeled on the image: ionic and dipolar relaxation, and atomic and electronic
resonances at higher energies.
The most common relationship scientists that have with permittivity is through the concept of relative permittivity: the permittivity
of a material relative to vacuum permittivity. Also known as the dielectric constant, the relative permittivity (εr) is given by 2.9.3 ,
where εs is the permittivity of the substance and ε0 is the permittivity of a vacuum (ε0 = 8.85 x 10-12 Farads/m). Although relative
permittivity is in fact dynamic and a function of frequency, the dielectric constants are most often expressed for low frequency
electric fields where the electric field is essential static in nature. Table 2.9.1 depicts the dielectric constants for a range of
materials.
εr   =  εs / ε0 (2.9.3)

Table 2.9.1 : Relative permittivities of various materials under static (i.e. non time-varying) electric fields.
Material Relative Permittivity

Vacuum 1 (by definition)

Air 1.00058986

Polytetrafluoroethylene (PTFE, Teflon) 2.1

Paper 3.85

2.9.1 https://chem.libretexts.org/@go/page/55855
Material Relative Permittivity

Diamond 5.5-10

Methanol 30

Water 80.1

Titanium dioxide (TiO2) 86-173

Strontium titanate (SrTiO3) 310

Barium titanate (BaTiO3) 1,200 - 10,000

Calcium copper titanate (CaCu3Ti4O12) >250,000

Dielectric constants may be useful for generic applications whereby the high-frequency response can be neglected, although
applications such as radio communications, microwave design, and optical system design call for a more rigorous and
comprehensive analysis. This is especially true for electrical devices such as capacitors, which are circuit elements that store and
discharge electrical charge in both a static and time-varying manner. Capacitors can be thought of as two parallel plate electrodes
that are separated by a finite distance and ‘sandwich’ together a piece of material with characteristic permittivity values. As can be
seen in Figure 2.9.2, the capacitance is a function of the permittivity of the material between the plates, which in turn is dependent
on frequency. Hence, for capacitors incorporated into the circuit design for radio communication applications, across the spectrum
8.3 kHz – 300 GHz, the frequency response would be important as this will determine the capacitors ability to charge and discharge
as well as the thermal response from electric fields dissipating their power as heat through the material.

Figure 2.9.2 Parallel plate capacitor of area, A, separated by a distance, d. The capacitance of the capacitor is directly related to the
permittivity (ε) of the material between the plates, as shown in the equation.
Evaluating the electrical characteristics of materials is become increasingly popular – especially in the field of electronics whereby
miniaturization technologies often require the use of materials with high dielectric constants. The composition and chemical
variations of materials such as solids and liquids can adopt characteristic responses, which are directly proportional to the amounts
and types of chemical species added to the material. The examples given herein are related to aqueous suspensions whereby the
electrical permittivity can be easily modulated via the addition of sodium chloride (NaCl).

Instrumentation
A common and reliable method for measuring the dielectric properties of liquid samples is to use an impedance analyzer in
conjunction with a dielectric probe. The impedance analyzer directly measures the complex impedance of the sample under test and
is then converted to permittivity using the system software. There are many methods used for measuring impedance, each of which
has their own inherent advantages and disadvantages and factors associated with that particular method. Such factors include
frequency range, measurement accuracy, and ease of operation. Common impedance measurements include bridge method,
resonant method, current-voltage (I-V) method, network analysis method, auto-balancing bridge method, and radiofrequency (RF)
I-V method. The RF I-V method used herein has several advantages over previously mentioned methods such as extended
frequency coverage, better accuracy, and a wider measured impedance range. The principle of the RF I-V method is based on the
linear relationship of the voltage-current ratio to impedance, as given by Ohm’s law (V=IZ where V is voltage, I is current, and Z is
impedance). This results in the impedance measurement sensitivity being constant regardless of measured impedance. Although a
full description of this method involves circuit theory and is outside the scope of this module (see “Impedance Measurement
Handbook” for full details) a brief schematic overview of the measurement principles is shown in Figure 2.9.3.

2.9.2 https://chem.libretexts.org/@go/page/55855
Figure 2.9.3 (a) Dielectric probe (liquids are placed on this probe). Circuit schematic of impedance measurements for (b) low and
(c) high impedance materials. Circuit symbols Osc, Zx, V, I, and R represent oscillator (i.e. frequency source), sample impedance,
voltage, current, and resistance, respectively.
As can be seen in Figure 3, the RF I-V method, which incorporates the use of a dielectric probe, essentially measures variations in
voltage and current when a sample is placed on the dielectric probe. For the low-impedance case, the impedance of the sample (Zx)
is given by 2.9.4 , for a high-impedance sample, the impedance of the sample (Zx) is given by 2.9.5 .
2R
Zx   =  V /I   = (2.9.4)
V2
  −  1
V1

R V1
Zx   =  V /I   = [ −  1] (2.9.5)
2 V2

The instrumentation and methods described herein consist of an Agilent E4991A impedance analyzer connected to an Agilent
85070E dielectric probe kit. The impedance analyzer directly measures the complex impedance of the sample under test by
measuring either the frequency-dependent voltage or current across the sample. These values are then converted to permittivity
values using the system software.

Applications
Electrical permittivity of deionized water and saline (0.9 % w/v NaCl)
In order to acquire the electrical permittivity of aqueous solutions the impedance analyzer and dielectric probe must first be
calibrated. In the first instance, the impedance analyzer unit is calibrated under open-circuit, short-circuit, 50 ohm load, and low
loss capacitance conditions by attaching the relevant probes shown in Figure 2.9.4. The dielectric probe is then attached to the
system and re-calibrated in open-air, with an attached short circuit probe, and finally with 500 μl of highly purified deionized water
(with a resistivity of 18.2 MΩ/cm at 25 °C) (Figure 2.9.5 ). The water is then removed and the system is ready for acquiring data.

Figure 2.9.4 Impedance analyzer calibration (A) Agilent E4991A impedance analyzer connected to 85070E dielectric probe. (B)
Calibrations standards (left-to-right: open circuit, short circuit, 50 ohm load, low-loss capacitor), (C) Attachment of the open
circuit, short circuit, 50 ohm load, and low-loss capacitor (left-to-right, respectively).

2.9.3 https://chem.libretexts.org/@go/page/55855
Figure 2.9.5 Dielectric probe calibration. (A) Impedance analyzer screen shot showing data line for dielectric probe in open-air. (B)
short circuit probe (C) dielectric probe (D) dielectric probe connected to impedance analyzer under open air conditions (E) short-
circuit probe attached to dielectric probe (F) 500 μl of deionized water on dielectric probe.
In order to maintain accurate calibration only the purest deionized water with a resistivity of 18.2 MΩ/cm at 25 °C should be used.
To perform an analysis simply load the dielectric probe with 500 μl of the sample and click on the ‘acquire data’ tab in the
software. The system will perform a scan across the frequency range 200 MHz – 3 GHz and acquire the real and imaginary parts of
the complex permittivity. The period with which a data point is taken as well as the scale (i.e. log or linear) can also be altered in
the software if necessary. To analyze another sample, remove the liquid and gently dry the dielectric probe with a paper towel. An
open air refresh calibration should then be performed (by pressing the relevant button in the software) as this prevents errors and
instrument drift from sample to sample. To analyze a normal saline (0.9 % NaCl w/v) solution, dissolve 8.99 g of NaCl in 1 litre of
DI water (18.2 MΩ/cm at 25 °C) to create a 154 mM NaCl solution (equivalent to a 0.9 % NaCl w/v solution). Load 500 μl of the
sample on the dielectric probe and acquire a new data set as mentioned previously.
Users should consult the “Agilent Installation and Quick Start Guide” manual for full specifics in regards to impedance analyzer
and dielectric probe calibration settings.
Data Analysis
The data files extracted from the impedance analyzer and dielectric probe setup previously described can be opened using any
standard data processing software such as Microsoft Excel. The data will appear in three columns, which will be labeled frequency
(Hz), ε', and ε" (representing the real and imaginary components of the permittivity, respectively). Any graphing software can be
used to create simple graphs of the complex permittivity versus frequency. In the example below (Figure 2.9.6 ) we have used
Prism to graph the real and complex permittivity’s versus frequency (200 MHz – 3 GHz) for the water and saline samples. For this
frequency range no error correction is needed. For the analysis of frequencies below 200 MHz down to 10 MHz, which can be
achieved using the impedance analyzer and dielectric probe configuration, error correction algorithms are needed to take into
account electrode polarization effects that skew and distort the data. Gach et al. cover these necessary algorithms that can be used if
needed.

Figure 2.9.6 Real and Imaginary components of permittivity for water (left) and saline (right) samples across the frequency range
200 MHz – 3 GHz.

2.9: Electrical Permittivity Characterization of Aqueous Solutions is shared under a CC BY 4.0 license and was authored, remixed, and/or curated
by Pavan M. V. Raja & Andrew R. Barron via source content that was edited to conform to the style and standards of the LibreTexts platform; a
detailed edit history is available upon request.

2.9.4 https://chem.libretexts.org/@go/page/55855
2.10: Dynamic Mechanical Analysis
Dynamic mechanical analysis (DMA), also known as forced oscillatory measurements and dynamic rheology, is a basic tool used
to measure the viscoelastic properties of materials (particularly polymers). To do so, DMA instrument applies an oscillating force
to a material and measures its response; from such experiments, the viscosity (the tendency to flow) and stiffness of the sample can
be calculated. These viscoelastic properties can be related to temperature, time, or frequency. As a result, DMA can also provide
information on the transitions of materials and characterize bulk properties that are important to material performance. DMA can
be applied to determine the glass transition of polymers or the response of a material to application and removal of a load, as a few
common examples. The usefulness of DMA comes from its ability to mimic operating conditions of the material, which allows
researchers to predict how the material will perform.

A Brief History
Oscillatory experiments have appeared in published literature since the early 1900s and began with rudimentary experimental
setups to analyze the deformation of metals. In an initial study, the material in question was hung from a support, and torsional
strain was applied using a turntable. Early instruments of the 1950s from manufacturers Weissenberg and Rheovibron exclusively
measured torsional stress, where force is applied in a twisting motion.
Due to its usefulness in determining polymer molecular structure and stiffness, DMA became more popular in parallel with the
increasing research on polymers. The method became integral in the analysis of polymer properties by 1961. In 1966, the
revolutionary torsional braid analysis was developed; because this technique used a fine glass substrate imbued with the material of
analysis, scientists were no longer limited to materials that could provide their own support. Using torsional braid analysis, the
transition temperatures of polymers could be determined through temperature programming. Within two decades, commercial
instruments became more accessible, and the technique became less specialized. In the early 1980s, one of the first DMAs using
axial geometries (linear rather than torsional force) was introduced.
Since the 1980s, DMA has become much more user-friendly, faster, and less costly due to competition between vendors.
Additionally, the developments in computer technology have allowed easier and more efficient data processing. Today, DMA is
offered by most vendors, and the modern instrument is detailed in the Instrumentationsection.

Basic Principles of DMA


DMA is based on two important concepts of stress and strain. Stress (σ) provides a measure of force (F) applied to area (A), 2.10.1
.
σ  =  F /A (2.10.1)

Stress to a material causes strain (γ), the deformation of the sample. Strain can be calculated by dividing the change in sample
dimensions (∆Y) by the sample’s original dimensions (Y) (2.10.2 ). This value is often given as a percentage of strain.
γ  =  ΔY /Y (2.10.2)

The modulus (E), a measure of stiffness, can be calculated from the slope of the stress-strain plot, Figure 2.10.1, as displayed in
\label{3} . This modulus is dependent on temperature and applied stress. The change of this modulus as a function of a specified
variable is key to DMA and determination of viscoelastic properties. Viscoelastic materials such as polymers display both elastic
properties characteristic of solid materials and viscous properties characteristic of liquids; as a result, the viscoelastic properties are
often a compromise between the two extremes. Ideal elastic properties can be related to Hooke’s spring, while viscous behavior is
often modeled using a dashpot, or a motion-resisting damper.
E  =  σ/y (2.10.3)

2.10.1 https://chem.libretexts.org/@go/page/55857
Figure 2.10.1 An example of a typical stress versus strain plot.

Creep-recovery
Creep-recovery testing is not a true dynamic analysis because the applied stress or strain is held constant; however, most modern
DMA instruments have the ability to run this analysis. Creep-recovery tests the deformation of a material that occurs when load
applied and removed. In the “creep” portion of this analysis, the material is placed under immediate, constant stress until the
sample equilibrates. “Recovery” then measures the stress relaxation after the stress is removed. The stress and strain are measured
as functions of time. From this method of analysis, equilibrium values for viscosity, modulus, and compliance (willingness of
materials to deform; inverse of modulus) can be determined; however, such calculations are beyond the scope of this review.
Creep-recovery tests are useful in testing materials under anticipated operation conditions and long test times. As an example,
multiple creep-recovery cycles can be applied to a sample to determine the behavior and change in properties of a material after
several cycles of stress.

Dynamic Testing
DMA instruments apply sinusoidally oscillating stress to samples and causes sinusoidal deformation. The relationship between the
oscillating stress and strain becomes important in determining viscoelastic properties of the material. To begin, the stress applied
can be described by a sine function where σo is the maximum stress applied, ω is the frequency of applied stress, and t is time.
Stress and strain can be expressed with the following 2.10.4 .

σ  =  σ0 sin(ωt + δ);  y = y0 cos(ωt) (2.10.4)

The strain of a system undergoing sinusoidally oscillating stress is also sinuisoidal, but the phase difference between strain and
stress is entirely dependent on the balance between viscous and elastic properties of the material in question. For ideal elastic
systems, the strain and stress are completely in phase, and the phase angle (δ) is equal to 0. For viscous systems, the applied stress
leads the strain by 90o. The phase angle of viscoelastic materials is somewhere in between (Figure 2.10.2 ).

Figure 2.10.2 Applied sinusoidal stress versus time (above) aligned with measured stress versus time (below). (a) The applied
stress and measured strain are in phase for an ideal elastic material. (b) The stress and strain are 90o out of phase for a purely
viscous material. (c) Viscoelastic materials have a phase lag less than 90o. Image adapted from M. Sepe, Dynamic Mechanical
Analysis for Plastics Engineering, Plastics Design Library: Norwich, NY (1998).

2.10.2 https://chem.libretexts.org/@go/page/55857
In essence, the phase angle between the stress and strain tells us a great deal about the viscoelasticity of the material. For one, a
small phase angle indicates that the material is highly elastic; a large phase angle indicates the material is highly viscous.
Furthermore, separating the properties of modulus, viscosity, compliance, or strain into two separate terms allows the analysis of
the elasticity or the viscosity of a material. The elastic response of the material is analogous to storage of energy in a spring, while
the viscosity of material can be thought of as the source of energy loss.
A few key viscoelastic terms can be calculated from dynamic analysis; their equations and significance are detailed in Table 2.10.1.
Table 2.10.1 Key viscoelastic terms that can be calculated with DMA.
Term Equation Significance

Overall modulus representing stiffness of


Complex modulus (E*) E* = E’ + iE” material; combined elastic and viscous
components
Storage modulus; measures stored energy and
Elastic modulus (E’) E’ = (σo/γo)cosδ
represents elastic portion
Loss modulus; contribution of viscous
Viscous modulus (E”) E” = (σo/γo)sinδ
component on polymer that flows under stress
Damping or index of viscoelasticity; compares
Loss tangent (tanδ) Tanδ = E”/E’
viscous and elastic moduli

Types of Dynamic Experiments


A temperature sweep is the most common DMA test used on solid materials. In this experiment, the frequency and amplitude of
oscillating stress is held constant while the temperature is increased. The temperature can be raised in a stepwise fashion, where the
sample temperature is increased by larger intervals (e.g., 5 oC) and allowed to equilibrate before measurements are taken.
Continuous heating routines can also be used (1-2 oC/minute). Typically, the results of temperature sweeps are displayed as storage
and loss moduli as well as tan delta as a function of temperature. For polymers, these results are highly indicative of polymer
structure. An example of a thermal sweep of a polymer is detailed later in this module.
In time scans, the temperature of the sample is held constant, and properties are measured as functions of time, gas changes, or
other parameters. This experiment is commonly used when studying curing of thermosets, materials that change chemically upon
heating. Data is presented graphically using modulus as a function of time; curing profiles can be derived from this information.
Frequency scans test a range of frequencies at a constant temperature to analyze the effect of change in frequency on temperature-
driven changes in material. This type of experiment is typically run on fluids or polymer melts. The results of frequency scans are
displayed as modulus and viscosity as functions of log frequency.

Instrumentation
The most common instrument for DMA is the forced resonance analyzer, which is ideal for measuring material response to
temperature sweeps. The analyzer controls deformation, temperature, sample geometry, and sample environment.
Figure 2.10.3 displays the important components of the DMA, including the motor and driveshaft used to apply torsional stress as
well as the linear variable differential transformer (LVDT) used to measure linear displacement. The carriage contains the sample
and is typically enveloped by a furnace and heat sink.

Figure 2.10.3 General schematic of DMA analyzer.


The DMA should be ideally selected to analyze the material at hand. The DMA can be either stress or strain controlled: strain-
controlled analyzers move the probe a certain distance and measure the stress applied; strain-controlled analyzers provide a

2.10.3 https://chem.libretexts.org/@go/page/55857
constant deformation of the sample (Figure 2.10.4 ) Although the two techniques are nearly equivalent when the stress-strain plot
(Figure 2.10.1 ) is linear, stress-controlled analyzers provide more accurate results.

Figure adapted from M. Sepe, Dynamic Mechanical Analysis for Plastics Engineering, Plastics Design Library: Norwich, NY
(1998).
DMA analyzers can also apply stress or strain in two manners—axial and torsional deformation (Figure 2.10.5 ) Axial deformation
applies a linear force to the sample and is typically used for solid and semisolid materials to test flex, tensile strength, and
compression. Torsional analyzers apply force in a twisting motion; this type of analysis is used for liquids and polymer melts but
can also be applied to solids. Although both types of analyzers have wide analysis range and can be used for similar samples, the
axial instrument should not be used for fluid samples with viscosities below 500 Pa-s, and torsional analyzers cannot handle
materials with high modulus.
Different fixtures can be used to hold the samples in place and should be chosen according to the type of samples analyzed. The
sample geometry affects both stress and strain and must be factored into the modulus calculations through a geometry factor. The
fixture systems are specific to the type of stress application. Axial analyzers have a greater number of fixture options; one of the
most commonly used fixtures is extension/tensile geometry used for thin films or fibers. In this method, the sample is held both
vertically and lengthwise by top and bottom clamps, and stress is applied upwards

Figure 2.10.5 Axial analyzer with DMA instrument (left) and axial analyzer with extension/tensile geometry (right).
For torsional analyzers, the simplest geometry is the use of parallel plates. The plates are separated by a distance determined by the
viscosity of the sample. Because the movement of the sample depends on its radius from the center of the plate, the stress applied is
uneven; the measured strain is an average value.

DMA of the glass transition polymers


As the temperature of a polymer increases, the material goes through a number of minor transitions (Tγ and Tβ) due to expansion;
at these transitions, the modulus also undergoes changes. The glass transition of polymers (Tg) occurs with the abrupt change of
physical properties within 140-160 oC; at some temperature within this range, the storage (elastic) modulus of the polymer drops
dramatically. As the temperature rises above the glass transition point, the material loses its structure and becomes rubbery before
finally melting. The idealized modulus transition is pictured in Figure 2.10.6.

2.10.4 https://chem.libretexts.org/@go/page/55857
Figure 2.10.6 Ideal storage modulus transitions of viscoelastic polymers. Adapted from K. P. Menard, Dynamic Mechanical
Analysis: A Practical Introduction, 2nd ed., CRC Press: Boca Raton, FL (2008).
The glass transition temperature can be determined using either the storage modulus, complex modulus, or tan δ (vs temperature)
depending on context and instrument; because these methods result in such a range of values (Figure 2.10.6 ), the method of
calculation should be noted. When using the storage modulus, the temperature at which E’ begins to decline is used as the Tg. Tan δ
and loss modulus E” show peaks at the glass transition; either onset or peak values can be used in determining Tg. These different
methods of measurement are depicted graphically in Figure 2.10.7.

Figure 2.10.7 Different industrial methods of calculating glass transition temperature (Tg). Copyright 2014, TA Instruments. Used
with permission.

Advantages and limitations of DMA


Dynamic mechanical analysis is an essential analytical technique for determining the viscoelastic properties of polymers. Unlike
many comparable methods, DMA can provide information on major and minor transitions of materials; it is also more sensitive to
changes after the glass transition temperature of polymers. Due to its use of oscillating stress, this method is able to quickly scan
and calculate the modulus for a range of temperatures. As a result, it is the only technique that can determine the basic structure of
a polymer system while providing data on the modulus as a function of temperature. Finally, the environment of DMA tests can be
controlled to mimic real-world operating conditions, so this analytical method is able to accurately predict the performance of
materials in use.
DMA does possess limitations that lead to calculation inaccuracies. The modulus value is very dependent on sample dimensions,
which means large inaccuracies are introduced if dimensional measurements of samples are slightly inaccurate. Additionally,
overcoming the inertia of the instrument used to apply oscillating stress converts mechanical energy to heat and changes the
temperature of the sample. Since maintaining exact temperatures is important in temperature scans, this also introduces
inaccuracies. Because data processing of DMA is largely automated, the final source of measurement uncertainty comes from
computer error.

2.10: Dynamic Mechanical Analysis is shared under a CC BY 4.0 license and was authored, remixed, and/or curated by Pavan M. V. Raja &
Andrew R. Barron via source content that was edited to conform to the style and standards of the LibreTexts platform; a detailed edit history is
available upon request.

2.10.5 https://chem.libretexts.org/@go/page/55857
2.11: Finding a Representative Lithology
From sediment to sample
Sample sediments are typically sent in a large plastic bag inside a brown paper bag labeled with the company or organization name,
drill site name and number, and the depth the sediment was taken (in meters).
The first step in determining a lithology is to prepare a sample from your bulk sediment. To do this, you will need to crush some of
the bulk rocks of your sediment into finer grains (Figure 2.11.1 ). You will need a hard surface, a hammer or mallet, and your
sediment. An improvised container such as the cardboard one shown in Figure 2.11.2 may be useful in containing fragments that
try to escape the hard surface during vigorous hammering. Remove the plastic sediment bag from the brown mailer bag. Empty
approximately 10-20 g of bulk sediment onto the hard surface. Repeatedly strike the larger rock sized portions of the sediment until
the larger units are broken into grains that are approximately the size of a grain of rice.

Figure 2.11.1 A hammer and hard surface for crushing. The makeshift cardboard shield on the left can be placed around the hard
surface to control fragmentation.
Some samples will give off oily or noxious odors when crushed. This is because of trapped hydrocarbons or sulfurous compounds
and is normal. The next step in the process, washing, will take care of these impurities and the smell.
Once the sample has been appropriately crushed on the macro scale, a micro uniformity in grain size can be achieved through the
use of a pulverizing micro mill machine such as the Planetary Mills Pulverisette 7 in Figure 2.11.2.

Figure 2.11.2 A Pullsette micro mill, milling cup removed. The mill is set to 520 rotations per minute and a five minute run time.
To use the mill, load your crushed sample into the milling cup (Figure 2.11.3 ) along with milling stones of 15 mm diameter. Set
your rotational speed and time using the machine interface. A speed of 500-600 rpm and mill time of 3-5 minutes is suggested.
Using higher speeds or longer times can result in loss of sample as dust. Load the milling cup into the mill and press start; make

2.11.1 https://chem.libretexts.org/@go/page/55858
sure to lower the mill hood. Once the mill has completed its cycle, retrieve the sample and dump it into a plastic cup labelled with
the drill site name and depth in order to prepare it for washing. Be sure to wash and dry the mill cup and mill stones between
samples if multiple samples are being tested.

Figure 2.11.3 A milling cup with mill stones and the crushed sample before milling.
Washing the Sample
If your sample is dirty, as in contaminated with hydrocarbons such as crude oil, it will need to be washed. To wash your sample you
will need your sample cup, a washbasin, a spoon, a 150-300 µm sieve, household dish detergent, and a porcelain ramekin if a
drying oven is available (Figure 2.11.4 ).

Figure 2.11.4 A washbasin, with detergent in a squirt bottle and the sample in a cup for washing (sieve not pictured).
Take your sample cup to the wash basin and fill the cup halfway with water, adding a squirt of dish detergent. Vigorously stir the
cup with the spoon for 20 seconds, ensuring each grain is coated with the detergent water. Pour your sample into the sieve and turn
on the faucet. Run water over the sample to allow the detergent and dust particles to wash through the sieve. Continue to wash the
sample this way until all the detergent is washed from the sample. Once clean, empty the sieve onto a surface to leave to dry
overnight, or into a ramekin if a drying oven is available. Place ramekin into drying oven set to at least 100 °C for a minimum of 2
hours to allow thorough drying (Figure 2.11.5 ). Once dry, the sample is ready to be picked.

Figure 2.11.5 A drying oven with temperature set above the temperature of evaporation for water at 105 °C
Picking the Sample
Picking the sample is arguably the most important step in determining the lithology (Figure 2.11.6 ).

2.11.2 https://chem.libretexts.org/@go/page/55858
During this step you will create a sample uniformity to eliminate random minerals, macro contaminates such as wood, and
dropstones that dropped into your sediment depth when the sediment was drilled. You will also be able to get a general judgment as
to the lithology after picking, though further analysis is needed if chemical composition is desired. Remove sample from drying
oven. Take a piece of weighing paper and weigh out 5-10 g of sample. Use a light microscope to determine whether most of the
sample is either silt, clay, silty-clay, or sand.
Clay grains will have a gray coloration with large flat sub-surfaces and less angulation. Clay will easily deform under pressure
from forceps.
Silt grains will be darker than clay and will have specks that shine when the grain is rotated. Texture is long pieces with jagged
edges. Silt is harder in consistency.
Silty clay is a heterogenous mixture (half and half mixture) of the above.
Sand is defined as larger grain size, lighter and varied coloration, and many crystalline substructures. Sand is hard to deform
with the forceps.

Figure 2.11.6 A light microscope being used to 'pick' the sample. The sample is being separated according to the dominant
lithology in preparation for chemical analysis.
Pelleting the Sample
To prepare your sample for X-ray fluorescence (XRF) analysis you will need to prepare a sample pellet. To pellet your sample you
will need a mortar and pestle, pellet binder such as Cerox, a scapula to remove binder, a micro scale, a pellet press with housing,
and a pellet tin cup. Measure out and pour 2-4 g of sample into your mortar. Measure out and add 50% of your sample weight of
pellet binder. For example, if your sample weight was 2 g, add 1 g of binder. Grind the sample into a fine, uniform powder,
ensuring that all of the binder is thoroughly mixed with the sample (Figure 2.11.7 ).

Figure 2.11.7 A mortar and pestle being used to grind the sample to a powder for pelleting in the pellet press. A binding agent
(Cereox) is also added using the scapula.
Drop a sample of tin foil into the press housing. Pour sample into the tin foil, and then gently tap the housing against a hard surface
two to three times to ensure sample settles into the tin. Place the top press disk into the channel. Place the press housing into the
press, oriented directly under the pressing arm. Crank the lever on the press until the pressure gauge reads 15 tons (Figure 2.11.8 ).
Wait for one minute, then twist the pressure release valve and remove the press housing from the press. Reverse the press and apply
the removal cap to the bottom of the press. Place the housing into the press bottom side up and manually apply pressure by turning
the crank on top of the press until the sample pops out of the housing. Retrieve the pelleted sample (Figure 2.11.9 ). The pelleted
sample is now ready for X-ray fluorescence analysis (XRF).

2.11.3 https://chem.libretexts.org/@go/page/55858
Figure 2.11.8 A pellet press being pressurized to 15 tons.

Figure 2.11.9 A completed pellet after pressing.


XRF Analysis
Place the sample pellet into the XRF (Figure 2.11.10 and Figure 2.11.11 ) and close the XRF hood. The XRF obtain the spectrum
from the associated computer.

Figure 2.11.10 A Spectro XEPOS X-Ray fluorescence spectrometer.

2.11.4 https://chem.libretexts.org/@go/page/55858
Figure 2.11.11 The inside of the spectrometer where the sample pellets are placed for analysis
The XRF spectrum is a plot of energy and intensity. The software equipped with the XRF will be pre-programmed to recognize the
characteristic energies associated with the X-ray emissions of the elements. The XRF functions by shooting a beam of high energy
photons that are absorbed by the atoms of the sample. The inner shell electrons of sample atoms are ejected. This leaves the atom in
an excited state, with a vacancy in the inner shell. Outer shell electrons then fall into the vacancy, emitting photons with energy
equal to the energy difference between these two energy levels. Each element has a unique set of energy levels, therefore each
element emits a pattern of X-rays characteristic of that element. The intensity of these characteristic X-rays increases with the
concentration of the corresponding element leading to higher counts and higher peaks on the spectrum (Figure 2.11.12 ).

Figure 2.11.12 The XRF spectrum showing the chemical composition of the sample.

2.11: Finding a Representative Lithology is shared under a CC BY 4.0 license and was authored, remixed, and/or curated by Pavan M. V. Raja &
Andrew R. Barron via source content that was edited to conform to the style and standards of the LibreTexts platform; a detailed edit history is
available upon request.

2.11.5 https://chem.libretexts.org/@go/page/55858
CHAPTER OVERVIEW
3: Principles of Gas Chromatography
3.1: Principles of Gas Chromatography
3.2: High Performance Liquid chromatography
3.3: Basic Principles of Supercritical Fluid Chromatography and Supercrtical Fluid Extraction
3.4: Supercritical Fluid Chromatography
3.5: Ion Chromatography
3.6: Capillary Electrophoresis

Thumbnail: A gas chromatography oven, open to show a capillary column. (CC BY-SA 4.0; Polimerek)

3: Principles of Gas Chromatography is shared under a CC BY 4.0 license and was authored, remixed, and/or curated by Pavan M. V. Raja &
Andrew R. Barron via source content that was edited to conform to the style and standards of the LibreTexts platform; a detailed edit history is
available upon request.

1
3.1: Principles of Gas Chromatography
Archer J.P. Martin (Figure 3.1.1 ) and Anthony T. James (Figure 3.1.2 ) introduced liquid-gas partition chromatography in 1950 at
the meeting of the Biochemical Society held in London, a few months before submitting three fundamental papers to the
Biochemical Journal. It was this work that provided the foundation for the development of gas chromatography. In fact, Martin
envisioned gas chromatography almost ten years before, while working with R. L. M. Synge (Figure 3.1.3 ) on partition
chromatography. Martin and Synge, who were awarded the chemistry Nobel prize in 1941, suggested that separation of volatile
compounds could be achieved by using a vapor as the mobile phase instead of a liquid.

Figure 3.1.1 British chemist Archer J. P. Martin, FRS (1910-2002) shared the Nobel Prize in 1952 for partition chromatography.

Figure 3.1.2 British chemist Anthony T. James (1922-2006).

3.1.1 https://chem.libretexts.org/@go/page/55860
Figure 3.1.3 British biochemist Richard L. M. Synge, FRS (1914-1994) shared the Nobel Prize in 1952 for partition
chromatography.
Gas chromatography quickly gained general acceptance because it was introduced at the time when improved analytical controls
were required in the petrochemical industries, and new techniques were needed in order to overcome the limitations of old
laboratory methods. Nowadays, gas chromatography is a mature technique, widely used worldwide for the analysis of almost every
type of organic compound, even those that are not volatile in their original state but can be converted to volatile derivatives.

The Chromatographic Process


Gas chromatography is a separation technique in which the components of a sample partition between two phases:
1. The stationary phase.
2. The mobile gas phase.
According to the state of the stationary phase, gas chromatography can be classified in gas-solid chromatography (GSC), where the
stationary phase is a solid, and gas-liquid chromatography (GLC) that uses a liquid as stationary phase. GLC is to a great extent
more widely used than GSC.
During a GC separation, the sample is vaporized and carried by the mobile gas phase (i.e., the carrier gas) through the column.
Separation of the different components is achieved based on their relative vapor pressure and affinities for the stationary phase. The
affinity of a substance towards the stationary phase can be described in chemical terms as an equilibrium constant called the
distribution constant Kc, also known as the partition coefficient, 3.1.1 , where [A]s is the concentration of compound A in the
stationary phase and [A]m is the concentration of compound A in the mobile phase.
Kc = [A]s /[A]m (3.1.1)

The distribution constant (Kc) controls the movement of the different compounds through the column, therefore differences in the
distribution constant allow for the chromatographic separation. Figure 3.1.4 shows a schematic representation of the
chromatographic process. Kc is temperature dependent, and also depends on the chemical nature of the stationary phase. Thus,
temperature can be used as a way to improve the separation of different compounds through the column, or a different stationary
phase.

3.1.2 https://chem.libretexts.org/@go/page/55860
Figure 3.1.4 Schematic representation of the chromatographic process. Adapted from Harold M. McNair, James M. Miller, Basic
Gas Chromatography, John Wiley & Sons, New York,1998. Reproduced courtesy of John Wiley & Sons, Inc.
A Typical Chromatogram
Figure 3.1.5 shows a chromatogram of the analysis of residual methanol in biodiesel, which is one of the required properties that
must be measured to ensure the quality of the product at the time and place of delivery.

Figure 3.1.5 Chromatogram of the analysis of methanol in B100 biodiesel, following EN 14110 methodology. Reproduced
courtesy of PerkinElmer Inc. (http://www.perkinelmer.com/)
Chromatogram (Figure 3.1.5 a) shows a standard solution of methanol with 2-propanol as the internal standard. From the figure it
can be seen that methanol has a higher affinity for the mobile phase (lower Kc) than 2-propanol (iso-propanol), and therefore elutes
first. Chromatograms (Figure 3.1.5 b and c) show two samples of biodiesel, one with methanol (Figure 3.1.5 b) and another with
no methanol detection. The internal standard was added to both samples for quantitation purposes.
Instrument Overview
Components of a Gas Chromatograph System
Figure 3.1.6 shows a schematic diagram of the components of a typical gas chromatograph, while Figure 3.1.7 shows a photograph
of a typical gas chromatograph coupled to a mass spectrometer (GC/MS).

Figure 3.1.6 Schematic diagram of the components of a typical gas chromatograph. Adapted from
http://en.Wikipedia.org/wiki/Gas_chromatography

3.1.3 https://chem.libretexts.org/@go/page/55860
Figure 3.1.7 Image of a Perkin Elmer Clarus SQ 8S GC/MS. Reproduced courtesy of PerkinElmer Inc.
(http://www.perkinelmer.com/).
Carrier Gas
The role of the carrier gas -GC mobile phase- is to carry the sample molecules along the column while they are not dissolved in or
adsorbed on the stationary phase. The carrier gas is inert and does not interact with the sample, and thus GC separation's selectivity
can be attributed to the stationary phase alone. However, the choice of carrier gas is important to maintain high efficiency. The
effect of different carrier gases on column efficiency is represented by the van Deemter (packed columns) and the Golay equation
(capillary columns). The van Deemter equation, 3.1.2 , describes the three main effects that contribute to band broadening in
packed columns and, as a consequence, to a reduced efficiency in the separation process.
B
H EP T   =  A + + Cu (3.1.2)
u

These three factors are:


1. the eddy diffusion (the A-term), which results from the fact that in packed columns spaces between particles along the column
are not uniform. Therefore, some molecules take longer pathways than others, and there are also variations in the velocity of the
mobile phase.
2. the longitudinal molecular diffusion (the B-term) which is a consequence of having regions with different analyte
concentrations.
3. the mass transfer in the stationary liquid phase (the C-term)
The broadening is described in terms of the height equivalent to a theoretical plate, HEPT, as a function of the average linear gas
velocity, u. A small HEPT value indicates a narrow peak and a higher efficiency.
Since capillary columns do not have any packing, the Golay equation, 3.1.3 , does not have an A-term. The Golay equation has 2
C-terms, one for mass transfer in then stationary phase (Cs) and one for mass transfer in the mobile phase (CM).
B
H EP T   =     +  (Cs   +  CM )u (3.1.3)
u

High purity hydrogen, helium and nitrogen are commonly used for gas chromatography. Also, depending on the type of detector
used, different gases are preferred.
Injector
This is the place where the sample is volatilized and quantitatively introduced into the carrier gas stream. Usually a syringe is used
for injecting the sample into the injection port. Samples can be injected manually or automatically with mechanical devices that are
often placed on top of the gas chromatograph: the auto-samplers.
Column
The gas chromatographic column may be considered the heart of the GC system, where the separation of sample components takes
place. Columns are classified as either packed or capillary columns. A general comparison of packed and capillary columns is
shown in Table 3.1.1. Images of packed columns are shown in Figure 3.1.8 and Figure 3.1.9.
Table 3.1.1 A summary of the differences between a packed and a capillary column.
Column Type Packed Column Capillary Column

Modern technology. Today most GC


History First type of GC column used applications are developed using capillary
columns

3.1.4 https://chem.libretexts.org/@go/page/55860
Not packed with particulate material. Made of
Packed with silica particles onto which the
Composition chemically treated silica covered with thin,
stationary phase is coated.
uniform liquid phase films.

Efficiency Low High

Outside diameter 2-4 mm 0.4 mm

Column length 2-4 meters 15-60 meters

Advantages Lower cost, larger samples Faster, better for complex mixtures

Figure 3.1.8 A typical capillary GC column. Adapted from F. M. Dunnivant and J. W. Ginsbach, Gas Chromatography, Liquid
Chromatography, Capillary Electrophoresis – Mass Spectrometry. A Basic Introduction, Copyright Dunnivant & Ginsbach (2008).

Figure 3.1.9 A Glass Packed GC Column. Adapted from F. M. Dunnivant and J. W. Ginsbach, Gas Chromatography, Liquid
Chromatography, Capillary Electrophoresis – Mass Spectrometry. A Basic Introduction, Copyright Dunnivant & Ginsbach (2008).
Since most common applications employed nowadays use capillary columns, we will focus on this type of columns. To define a
capillary column, four parameters must be specified:
1. The stationary phase is the parameter that will determine the final resolution obtained, and will influence other selection
parameters. Changing the stationary phase is the most powerful way to alter selectivity in GC analysis.
2. The length is related to the overall efficiency of the column and to overall analysis time. A longer column will increase the peak
efficiency and the quality of the separation, but it will also increase analysis time. One of the classical trade-offs in gas
chromatography (GC) separations lies between speed of analysis and peak resolution.
3. The column internal diameter (ID) can influence column efficiency (and therefore resolution) and also column capacity. By
decreasing the column internal diameter, better separations can be achieved, but column overload and peak broadening may
become an issue.
4. The sample capacity of the column will also depend on film thickness. Moreover, the retention of sample components will be
affected by the thickness of the film, and therefore its retention time. A shorter run time and higher resolution can be achieved
using thin films, however these films offer lower capacity.
Detector
The detector senses a physicochemical property of the analyte and provides a response which is amplified and converted into an
electronic signal to produce a chromatogram. Most of the detectors used in GC were invented specifically for this technique, except
for the thermal conductivity detector (TCD) and the mass spectrometer. In total, approximately 60 detectors have been used in GC.
Detectors that exhibit an enhanced response to certain analyte types are known as "selective detectors".
During the last 10 years there had been an increasing use of GC in combination with mass spectrometry (MS). The mass
spectrometer has become a standard detector that allows for lower detection limits and does not require the separation of all
components present in the sample. Mass spectroscopy is one of the types of detection that provides the most information with only
micrograms of sample. Qualitative identification of unknown compounds as well as quantitative analysis of samples is possible

3.1.5 https://chem.libretexts.org/@go/page/55860
using GC-MS. When GC is coupled to a mass spectrometer, the compounds that elute from the GC column are ionized by using
electrons (EI, electron ionization) or a chemical reagent (CI, chemical ionization). Charged fragments are focused and accelerated
into a mass analyzer: typically a quadrupole mass analyzer. Fragments with different mass to charge ratios will generate different
signals, so any compound that produces ions within the mass range of the mass analyzer will be detected. Detection limits of 1-10
ng or even lower values (e.g., 10 pg) can be achieved selecting the appropriate scanning mode.
Sample Preparation Techniques
Derivatization
Gas chromatography is primarily used for the analysis of thermally stable volatile compounds. However, when dealing with non-
volatile samples, chemical reactions can be performed on the sample to increase the volatility of the compounds. Compounds that
contain functional groups such as OH, NH, CO2H, and SH are difficult to analyze by GC because they are not sufficiently volatile,
can be too strongly attracted to the stationary phase or are thermally unstable. Most common derivatization reactions used for GC
can be divided into three types:
1. Silylation.
2. Acylation.
3. Alkylation & Esterification.
Samples are derivatized before being analyzed to:
Increase volatility and decrease polarity of the compound
Reduce thermal degradation
Increase sensitivity by incorporating functional groups that lead to higher detector signals
Improve separation and reduce tailing
Advantages and Disadvantages
GC is the premier analytical technique for the separation of volatile compounds. Several features such as speed of analysis, ease of
operation, excellent quantitative results, and moderate costs had helped GC to become one of the most popular techniques
worldwide.
Advantages of GC
Due to its high efficiency, GC allows the separation of the components of complex mixtures in a reasonable time.
Accurate quantitation (usually sharp reproducible peaks are obtained)
Mature technique with many applications notes available for users.
Multiple detectors with high sensitivity (ppb) are available, which can also be used in series with a mass spectrometer since MS
is a non-destructive technique.
Disadvantages of GC
Limited to thermally stable and volatile compounds.
Most GC detectors are destructive, except for MS.
Gas Chromatography Versus High Performance Liquid Chromatography (HPLC)

Unlike gas chromatography, which is unsuitable for nonvolatile and thermally fragile molecules, liquid chromatography can safely
separate a very wide range of organic compounds, from small-molecule drug metabolites to peptides and proteins.
Table 3.1.2 Relative advantages and disadvantages of GC versus HPLC.
GC HPLC

Volatility is not important, however solubility in the mobile phase


Sample must be volatile or derivatized previous to GC analysis
becomes critical for the analysis.

Most analytes have a molecular weight (MW) below 500 Da (due to There is no upper molecular weight limit as far as the sample can be
volatility issues) dissolved in the appropriate mobile phase

Can be coupled to MS. Several mass spectral libraries are available if Methods must be adapted before using an MS detector (non-volatile
using electron ionization (e.g., http://chemdata.nist.gov/) buffers cannot be used)

For some detectors the solvent must be an issue. When changing


Can be coupled to several detectors depending on the application
detectors some methods will require prior modification

3.1.6 https://chem.libretexts.org/@go/page/55860
3.1: Principles of Gas Chromatography is shared under a CC BY 4.0 license and was authored, remixed, and/or curated by Pavan M. V. Raja &
Andrew R. Barron via source content that was edited to conform to the style and standards of the LibreTexts platform; a detailed edit history is
available upon request.

3.1.7 https://chem.libretexts.org/@go/page/55860
3.2: High Performance Liquid chromatography
High-performance liquid chromatography (HPLC) is a technique in analytical chemistry used to separate the components in a
mixture, and to identify and quantify each component. It was initially discovered as an analytical technique in the early twentieth
century and was first used to separate colored compounds. The word chromatography means color writing. It was the botanist M.
S. Tswett (Figure 3.2.1 ) who invented this method in around 1900 to study leaf pigments (mainly chlorophyll). He separated the
pigments based on their interaction with a stationary phase. In 1906 Tswett published two fundamental papers describing the
various aspects of liquid-adsorption chromatography in detail. He also pointed out that in spite of its name, other substances also
could be separated by chromatography. The modern high performance liquid chromatography has developed from this separation;
the separation efficiency, versatility and speed have been improved significantly.

Figure 3.2.1 Russian born Italian botanist Mikhail Semyonovich Tswett (1872-1919).
The molecular species subjected to separation exist in a sample that is made of analytes and matrix. The analytes are the molecular
species of interest, and the matrix is the rest of the components in the sample. For chromatographic separation, the sample is
introduced in a flowing mobile phase that passes a stationary phase. Mobile phase is a moving liquid, and is characterized by its
composition, solubility, UV transparency, viscosity, and miscibility with other solvents. Stationary phase is a stationary medium,
which can be a stagnant bulk liquid, a liquid layer on the solid phase, or an interfacial layer between liquid and solid. In HPLC, the
stationary phase is typically in the form of a column packed with very small porous particles and the liquid mobile phase is moved
through the column by a pump. The development of HPLC is mainly the development of the new columns, which requires new
particles, new stationary phases (particle coatings), and improved procedures for packing the column. A picture of modern HPLC is
shown in Figure 3.2.2.

Figure 3.2.2 A picture of modern HPLC instrument.

Instrumentation
The major components of a HPLC are shown in Figure 3.2.3. The role of a pump is to force a liquid (mobile phase) through at a
specific flow rate (milliliters per minute). The injector serves to introduce the liquid sample into the flow stream of the mobile
phase. Column is the most central and important component of HPLC, and the column’s stationary phase separates the sample

3.2.1 https://chem.libretexts.org/@go/page/55862
components of interest using various physical and chemical parameters. The detector is to detect the individual molecules that elute
from the column. The computer usually functions as the data system, and the computer not only controls all the modules of the
HPLC instrument but it takes the signal from the detector and uses it to determine the retention time, the sample components, and
quantitative analysis.

Figure 3.2.3 Schematic representation of a HPLC system: (1) solvent, (2) gradient valve, (3) high-pressure pump, (4) sample
injection loop, (5) analytical column, (6) detector, and (7) computer.
Columns
Different separation mechanisms were used based on different property of the stationary phase of the column. The major types
include normal phase chromatography, reverse phase chromatography, ion exchange, size exclusion chromatography, and affinity
chromatography.
Normal-phase Chromatography
In this method the columns are packed with polar, inorganic particles and a nonpolar mobile phase is used to run through the
stationary phase (Table 3.2.1 ). Normal phase chromatography is mainly used for purification of crude samples, separation of very
polar samples, or analytical separations by thin layer chromatography. One problem when using this method is that, water is a
strong solvent for the normal-phase chromatography, traces of water in the mobile phase can markedly affect sample retention, and
after changing the mobile phase, the column equilibration is very slow.
Table 3.2.1 Mobile phase and stationary phase used for normal phase and reverse-phase chromatography
Stationary Phase Mobile Phase

Normal Phase Polar Non polar

Reverse Phase Non polar Polar

Reverse-phase Chromatography
In reverse-phase (RP) chromatography the stationary phase has a hydrophobic character, while the mobile phase has a polar
character. This is the reverse of the normal-phase chromatography (Table 3.2.2 ). The interactions in RP-HPLC are considered to
be the hydrophobic forces, and these forces are caused by the energies resulting from the disturbance of the dipolar structure of the
solvent. The separation is typically based on the partition of the analyte between the stationary phase and the mobile phase. The
solute molecules are in equilibrium between the hydrophobic stationary phase and partially polar mobile phase. The more
hydrophobic molecule has a longer retention time while the ionized organic compounds, inorganic ions and polar metal molecules
show little or no retention time.
Ion Exchange Chromatography
The ion exchange mechanism is based on electrostatic interactions between hydrated ions from a sample and oppositely charged
functional groups on the stationary phase. Two types of mechanisms are used for the separation: in one mechanism, the elution uses
a mobile phase that contains competing ions that would replace the analyte ions and push them off the column; another mechanism
is to add a complexing reagent in the mobile phase and to change the sample species from their initial form. This modification on
the molecules will lead them to elution. In addition to the exchange of ions, ion-exchange stationary phases are able to retain
specific neutral molecules. This process is related to the retention based on the formation of complexes, and specific ions such as
transition metals can be retained on a cation-exchange resin and can still accept lone-pair electrons from donor ligands. Thus
neutral ligand molecules can be retained on resins treated with the transitional metal ions.
The modern ion exchange is capable of quantitative applications at rather low solute concentrations, and can be used in the analysis
of aqueous samples for common inorganic anions (range 10 μg/L to 10 mg/L). Metal cations and inorganic anions are all separated
predominantly by ionic interactions with the ion exchange resin. One of the largest industrial users of ion exchange is the food and

3.2.2 https://chem.libretexts.org/@go/page/55862
beverage sector to determine the nitrogen-, sulfur-, and phosphorous- containing species as well as the halide ions. Also, ion
exchange can be used to determine the dissolved inorganic and organic ions in natural and treated waters.
Size Exclusion Chromatography
It is a chromatographic method that separate the molecules in the solutions based on the size (hydrodynamic volume). This column
is often used for the separation of macromolecules and of macromolecules from small molecules. After the analyte is injected into
the column, molecules smaller than he pore size of the stationary phase enter the porous particles during the separation and flow
through he intricate channels of the stationary phase. Thus smaller components have a longer path to traverse and elute from the
column later than the larger ones. Since the molecular volume is related to molecular weight, it is expected that retention volume
will depend to some degree on the molecular weight of the polymeric materials. The relation between the retention time and the
molecular weight is shown in Figure 3.2.4.

Figure 3.2.4 Graph showing the relationship between the retention time and molecular weight in size exclusion chromatography.
Usually the type of HPLC separation method to use depends on the chemical nature and physicochemical parameters of the
samples. Figure 3.2.5 shows a flow chart of preliminary selection for the separation method according to the properties of the
analyte.

Figure 3.2.5 Diagram showing the sample properties related to the selection of HPLC type of analysis.
Detectors
Detectors that are commonly used for liquid chromatography include ultraviolet-visible absorbance detectors, refractive index
detectors, fluorescence detectors, and mass spectrometry. Regardless of the class, a LC detector should ideally have the
characteristics of about 10-12-10-11 g/mL, and a linear dynamic range of five or six orders. The principal characteristics of the
detectors to be evaluated include dynamic range, response index or linearity, linear dynamic range, detector response, detector
sensitivity, etc.
Among these detectors, the most economical and popular methods are UV and refractive index (RI) detectors. They have rather
broad selectivity reasonable detection limits most of the time. The RI detector was the first detector available for commercial use.
This method is particularly useful in the HPLC separation according to size, and the measurement is directly proportional to the
concentration of polymer and practically independent of the molecular weight. The sensitivity of RI is 10-6 g/mL, the linear
dynamic range is from 10-6to 10-4 g/mL, and the response index is between 0.97 and 1.03.
UV detectors respond only to those substances that absorb UV light at the wavelength of the source light. A great many compounds
absorb light in the UV range (180-350 nm) including substances having one or more double bonds and substances having unshared
electrons. and the relationship between the intensity of UV light transmitted through the cell and solute concentration is given by
Beer’s law, 3.2.1 and 3.2.2 .
kcl
IT   =  I0 e (3.2.1)

ln(IT )  =  ln(I0 )(−kcl) (3.2.2)

3.2.3 https://chem.libretexts.org/@go/page/55862
Where I0 is the intensity of the light entering the cell, and IT is the light transmitted through the cell, l is the path length of the cell,
c is the concentration of the solute, and k is the molar absorption coefficient of the solute. UV detectors include fixed wavelength
UV detector and multi wavelength UV detector. The fixed wavelength UV detector has sensitivity of 5*10-8 g/mL, has linear
dynamic range between 5*10-8 and 5*10-4g/mL, and the response index is between 0.98 and 1.02. The multi-wavelength UV
detector has sensitivity of 10-7 g/mL, the linear dynamic range is between 5*10-7 and 5*10-4 g/mL, and the response index is from
0.97 to 1.03. UV detectors could be used effectively for the reverse-phase separations and ion exchange chromatography. UV
detectors have high sensitivity, are economically affordable, and easy to operate. Thus UV detector is the most common choice of
detector for HPLC.
Another method, mass spectrometry, has certain advantages over other techniques. Mass spectra could be obtained rapidly; only
small amount (sub-μg) of sample is required for analysis, and the data provided by the spectra is very informative of the molecular
structure. Mass spectrometry also has strong advantages of specificity and sensitivity compared with other detectors. The
combination of HPLC-MS is oriented towards the specific detection and potential identification of chemicals in the presence of
other chemicals. However, it is difficult to interface the liquid chromatography to a mass-spectrometer, because all the solvents
need to be removed first. The common used interface includes electrospray ionization, atmospheric pressure photoionization, and
thermospray ionization.

Parameters related to HPLC separation


Flow Rate

Flow rate shows how fast the mobile phase travels across the column, and is often used for calculation of the consumption of the
mobile phase in a given time interval. There are volumetric flow rate U and linear flow rate u. These two flow rate is related by
3.2.3 , where A is the area of the channel for the flow, 3.2.4 .

U = Au (3.2.3)

2
A  =  (1/4)πεd (3.2.4)

Retention Time
The retention time (tR) can be defined as the time from the injection of the sample to the time of compound elution, and it is taken
at the apex of the peak that belongs to the specific molecular species. The retention time is decided by several factors including the
structure of the specific molecule, the flow rate of the mobile phase, column dimension. And the dead time t0 is defined as the time
for a non-retained molecular species to elute from the column.
Retention Volume
Retention volume (VR) is defined as the volume of the mobile phase flowing from the injection time until the corresponding
retention time of a molecular species, and are related by 3.2.5 . The retention volume related to the dead time is known as dead
volume V0.
VR   =  UtR (3.2.5)

Migration Rate
The migration rate can be defined as the velocity at which the species moves through the column. And the migration rate (UR) is
inversely proportional to the retention times. If only a fraction of molecules that are present in the mobile phase are moving. The
value of migration rate is then given by 3.2.6 .

uR   =  u ∗ Vmo /(Vmo + Vst ) (3.2.6)

Capacity Factor

Capacity factor (k) is the ratio of reduced retention time and the dead time, 3.2.7 .
K  =  (tR − t0 )/ t0   =  (vR − v0 )/ v0 (3.2.7)

Equilibrium Constant and Phase Ratio


In the separation, the molecules running through the column can also be considered as being in a continuous equilibrium between
the mobile phase and the stationary phase. This equilibrium could be governed by an equilibrium constant K, defined as 3.2.8 , in
which Cmo is the molar concentration of the molecules in the mobile phase, and Cst is the molar concentration of the molecules in
the stationary phase. The equilibrium constant K can also be written as 3.2.9 .

3.2.4 https://chem.libretexts.org/@go/page/55862
K  =  Cst / Cmo (3.2.8)

K  =  k(V0 / Vst ) (3.2.9)

Advantage of HPLC
The most important aspect of HPLC is the high separation capacity which enables the batch analysis of multiple components. Even
if the sample consists of a mixture, HPLC will allows the target components to be separated, detected, and quantified. Also, under
appropriate condition, it is possible to attain a high level of reproducibility with a coefficient of variation not exceeding 1%. Also, it
has a high sensitivity while a low sample consumption. HPLC has one advantage over GC column that analysis is possible for any
sample can be stably dissolved in the eluent and need not to be vaporized.With this reason, HPLC is used much more frequently in
the field of biochemistry and pharmaceutical than the GC column.

3.2: High Performance Liquid chromatography is shared under a CC BY 4.0 license and was authored, remixed, and/or curated by Pavan M. V.
Raja & Andrew R. Barron via source content that was edited to conform to the style and standards of the LibreTexts platform; a detailed edit
history is available upon request.

3.2.5 https://chem.libretexts.org/@go/page/55862
3.3: Basic Principles of Supercritical Fluid Chromatography and Supercrtical Fluid
Extraction
The discovery of supercritical fluids led to novel analytical applications in the fields of chromatography and extraction known as
supercritical fluid chromatography (SFC) and supercritical fluid extraction (SFE). Supercritical fluid chromatography is accepted
as a column chromatography methods along with gas chromatography (GC) and high-performance liquid chromatography (HPLC).
Due to to the properties of supercritical fluids, SFC combines each of the advantages of both GC and HPLC in one method. In
addition, supercritical fluid extraction is an advanced analytical technique.

Definition and Formation of Supercritical Fluids


A supercritical fluid is the phase of a material at critical temperature and critical pressure of the material. Critical temperature is the
temperature at which a gas cannot become liquid as long as there is no extra pressure; and, critical pressure is the minimum amount
of pressure to liquefy a gas at its critical temperature. Supercritical fluids combine useful properties of gas and liquid phases, as it
can behave like both a gas and a liquid in terms of different aspects. A supercritical fluid provides a gas-like characteristic when it
fills a container and it takes the shape of the container. The motion of the molecules are quite similar to gas molecules. On the other
hand, a supercritical fluid behaves like a liquid because its density property is near liquid and, thus, a supercritical fluid shows a
similarity to the dissolving effect of a liquid.
The characteristic properties of a supercritical fluid are density, diffusivity and viscosity. Supercritical values for these features take
place between liquids and gases. Table 3.3.1 demonstrates numerical values of properties for gas, supercritical fluid and liquid.
Table 3.3.1 Supercritical fluid properties compared to liquids and gases
Gas Supercritical fluid Liquid

Density (g/cm3) 0.6 x 10-3-2.0 x 10-3 0.2-0.5 0.6-2.0

Diffusivity (cm2/s) 0.1-0.4 10-3-10-4 0.2 x 10-5-2.0 x 10-5

Viscosity (cm/s) 1 x 10-4-3 x 10-4 1 x 10-4-3 x 10-4 0.2 x 10-2-3.0 x 10-2

The formation of a supercritical fluid is the result of a dynamic equilibrium. When a material is heated to its specific critical
temperature in a closed system, at constant pressure, a dynamic equilibrium is generated. This equilibrium includes the same
number of molecules coming out of liquid phase to gas phase by gaining energy and going in to liquid phase from gas phase by
losing energy. At this particular point, the phase curve between liquid and gas phases disappears and supercritical material appears.
In order to understand the definition of SF better, a simple phase diagram can be used. Figure 3.3.1 displays an ideal phase
diagram. For a pure material, a phase diagram shows the fields where the material is in the form of solid, liquid, and gas in terms of
different temperature and pressure values. Curves, where two phases (solid-gas, solid-liquid and liquid-gas) exist together, defines
the boundaries of the phase regions. These curves, for example, include sublimation for solid-gas boundary, melting for solid-liquid
boundary, and vaporization for liquid-gas boundary. Other than these binary existence curves, there is a point where all three phases
are present together in equilibrium; the triple point (TP).

Figure 3.3.1 Schematic representation of an idealized phase diagram.


There is another characteristic point in the phase diagram, the critical point (CP). This point is obtained at critical temperature (Tc)
and critical pressure (Pc). After the CP, no matter how much pressure or temperature is increased, the material cannot transform
from gas to liquid or from liquid to gas phase. This form is the supercritical fluid form. Increasing temperature cannot result in

3.3.1 https://chem.libretexts.org/@go/page/55864
turning to gas, and increasing pressure cannot result in turning to liquid at this point. In the phase diagram, the field above Tc and
Pc values is defined as the supercritical region.
In theory, the supercritical region can be reached in two ways:
Increasing the pressure above the Pc value of the material while keeping the temperature stable and then increasing the
temperature above Tc value at a stable pressure value.
Increasing the temperature first above Tc value and then increasing the pressure above Pc value.
The critical point is characteristic for each material, resulting from the characteristic Tc and Pc values for each substance.

Physical Properties of Supercritical Fluids


As mentioned above, SF shares some common features with both gases and liquids. This enables us to take advantage of a correct
combination of the properties.

Density
Density characteristic of a supercritical fluid is between that of a gas and a liquid, but closer to that of a liquid. In the supercritical
region, density of a supercritical fluid increases with increased pressure (at constant temperature). When pressure is constant,
density of the material decreases with increasing temperature. The dissolving effect of a supercritical fluid is dependent on its
density value. Supercritical fluids are also better carriers than gases thanks to their higher density. Therefore, density is an essential
parameter for analytical techniques using supercritical fluids as solvents.

Diffusivity
Diffusivity of a supercritical fluid can be 100 x that of a liquid and 1/1,000 to 1/10,000 x less than a gas. Because supercritical fluids
have more diffusivity than a liquid, it stands to reason a solute can show better diffusivity in a supercritical fluid than in a liquid.
Diffusivity is parallel with temperature and contrary with pressure. Increasing pressure affects supercritical fluid molecules to
become closer to each other and decreases diffusivity in the material. The greater diffusivity gives supercritical fluids the chance to
be faster carriers for analytical applications. Hence, supercritical fluids play an important role for chromatography and extraction
methods.

Viscosity
Viscosity for a supercritical fluid is almost the same as a gas, being approximately 1/10 of that of a liquid. Thus, supercritical fluids
are less resistant than liquids towards components flowing through. The viscosity of supercritical fluids is also distinguished from
that of liquids in that temperature has a little effect on liquid viscosity, where it can dramatically influence supercritical fluid
viscosity.
These properties of viscosity, diffusivity, and density are related to each other. The change in temperature and pressure can affect
all of them in different combinations. For instance, increasing pressure causes a rise for viscosity and rising viscosity results in
declining diffusivity.

Super Fluid Chromatography (SFC)


Just like supercritical fluids combine the benefits of liquids and gases, SFC bring the advantages and strong aspects of HPLC and
GC together. SFC can be more advantageous than HPLC and GC when compounds which decompose at high temperatures with
GC and do not have functional groups to be detected by HPLC detection systems are analyzed.
There are three major qualities for column chromatographies:
Selectivity.
Efficiency.
Sensitivity.
Generally, HPLC has better selectivity that SFC owing to changeable mobile phases (especially during a particular experimental
run) and a wide range of stationary phases. Although SFC does not have the selectivity of HPLC, it has good quality in terms of
sensitivity and efficiency. SFC enables change of some properties during the chromatographic process. This tuning ability allows
the optimization of the analysis. Also, SFC has a broader range of detectors than HPLC. SFC surpasses GC for the analysis of
easily decomposable substances; these materials can be used with SFC due to its ability to work with lower temperatures than GC.

3.3.2 https://chem.libretexts.org/@go/page/55864
Instrumentation for SFC
As it can be seen in Figure 3.3.2 SFC has a similar setup to an HPLC instrument. They use similar stationary phases with similar
column types. However, there are some differences. Temperature is critical for supercritical fluids, so there should be a heat control
tool in the system similar to that of GC. Also, there should be a pressure control mechanism, a restrictor, because pressure is
another essential parameter in order for supercritical fluid materials to be kept at the required level. A microprocessor mechanism is
placed in the instrument for SFC. This unit collects data for pressure, oven temperature, and detector performance to control the
related pieces of the instrument.

Figure 3.3.2 Scheme of a supercritical fluid chromatography instrument. Adapted from D. A. Skoog and J. J. Leary, Principles of
Instrumental Analysis, Saunders College Publishing, Philadelphia (1992).

Stationary Phase
SFC columns are similar to HPLC columns in terms of coating materials. Open-tubular columns and packed columns are the two
most common types used in SFC. Open-tubular ones are preferred and they have similarities to HPLC fused-silica columns. This
type of column contains an internal coating of a cross-linked siloxane material as a stationary phase. The thickness of the coating
can be 0.05-1.0 μm. The length of the column can range from of 10 to 20 m.

Mobile Phases
There is a wide variety of materials used as mobile phase in SFC. The mobile phase can be selected from the solvent groups of
inorganic solvents, hydrocarbons, alcohols, ethers, halides; or can be acetone, acetonitrile, pyridine, etc. The most common
supercritical fluid which is used in SFC is carbon dioxide because its critical temperature and pressure are easy to reach.
Additionally, carbon dioxide is low-cost, easy to obtain, inert towards UV, non-poisonous and a good solvent for non-polar
molecules. Other than carbon dioxide, ethane, n-butane, N2O, dichlorodifluoromethane, diethyl ether, ammonia, tetrahydrofuran
can be used. Table 3.3.2 shows select solvents and their Tc and Pc values.
Table 3.3.2 Properties of some solvents as mobile phase at the critical point.
Solvent Critical Temperature (°C) Critical Pressure (bar)

Carbon dioxide (CO2) 31.1 72

Nitrous oxide (N2O) 36.5 70.6

Ammonia (NH3) 132.5 109.8

Ethane (C2H6) 32.3 47.6

n-Butane (C4H10) 152 70.6

Diethyl ether (Et2O) 193.6 63.8

Tetrahydrofuran (THF, C4H8O) 267 50.5

Dichlorodifluoromethane (CCl2F2) 111.7 109.8

Detectors
One of the biggest advantage of SFC over HPLC is the range of detectors. Flame ionization detector (FID), which is normally
present in GC setup, can also be applied to SFC. Such a detector can contribute to the quality of analyses of SFC since FID is a
highly sensitive detector. SFC can also be coupled with a mass spectrometer, an UV-visible spectrometer, or an IR spectrometer

3.3.3 https://chem.libretexts.org/@go/page/55864
more easily than can be done with an HPLC. Some other detectors which are used with HPLC can be attached to SFC such as
fluorescence emission spectrometer or thermionic detectors.

Advantages of working with SFC


The physical properties of supercritical fluids between liquids and gases enables the SFC technique to combine with the best
aspects of HPLC and GC, as lower viscosity of supercritical fluids makes SFC a faster method than HPLC. Lower viscosity leads
to high flow speed for the mobile phase.
Thanks to the critical pressure of supercritical fluids, some fragile materials that are sensitive to high temperature can be analyzed
through SFC. These materials can be compounds which decompose at high temperatures or materials which have low vapor
pressure/volatility such as polymers and large biological molecules. High pressure conditions provide a chance to work with lower
temperature than normally needed. Hence, the temperature-sensitive components can be analyzed via SFC. In addition, the
diffusion of the components flowing through a supercritical fluid is higher than observed in HPLC due to the higher diffusivity of
supercritical fluids over traditional liquids mobile phases. This results in better distribution into the mobile phase and better
separation.

Applications of SFC
The applications of SFC range from food to environmental to pharmaceutical industries. In this manner, pesticides, herbicides,
polymers, explosives and fossil fuels are all classes of compounds that can be analyzed. SFC can be used to analyze a wide variety
of drug compounds such as antibiotics, prostaglandins, steroids, taxol, vitamins, barbiturates, non-steroidal anti-inflammatory
agents, etc. Chiral separations can be performed for many pharmaceutical compounds. SFC is dominantly used for non-polar
compounds because of the low efficiency of carbon dioxide, which is the most common supercritical fluid mobile phase, for
dissolving polar solutes. SFC is used in the petroleum industry for the determination of total aromatic content analysis as well as
other hydrocarbon separations.

Supercritical Fluid Extraction (SFE)


The unique physical properties of supercritical fluids, having values for density, diffusivity and viscosity values between liquids
and gases, enables supercritical fluid extraction to be used for the extraction processes which cannot be done by liquids due to their
high density and low diffusivity and by gases due to their inadequate density in order to extract and carry the components out.
Complicated mixtures containing many components should be subject to an extraction process before they are separated via
chromatography. An ideal extraction procedure should be fast, simple, and inexpensive. In addition, sample loss or decomposition
should not be experienced at the end of the extraction. Following extraction, there should be a quantitative collection of each
component. Ideally, the amount of unwanted materials coming from the extraction should be kept to a minimum and be easily
disposable; the waste should not be harmful for environment. Unfortunately, traditional extraction methods often do not meet these
requirements. In this regard, SFE has several advantages in comparison with traditional techniques.
The extraction speed is dependent on the viscosity and diffusivity of the mobile phase. With a low viscosity and high diffusivity,
the component which is to be extracted can pass through the mobile phase easily. The higher diffusivity and lower viscosity of
supercritical fluids, as compared to regular extraction liquids, help the components to be extracted faster than other techniques.
Thus, an extraction process can take just 10-60 minutes with SFE, while it would take hours or even days with classical methods.
The dissolving efficiency of a supercritical fluid can be altered by temperature and pressure. In contrast, liquids are not affected by
temperature and pressure changes as much. Therefore, SFE has the potential to be optimized to provide a better dissolving capacity.
In classical methods, heating is required to get rid of the extraction liquid. However, this step causes the temperature-sensitive
materials to decompose. For SFE, when the critical pressure is removed, a supercritical fluid transforms to gas phase. Because
supercritical fluid solvents are chemically inert, harmless and inexpensive; they can be released to atmosphere without leaving any
waste. Through this, extracted components can be obtained much more easily and sample loss is minimized.

Instrumentation of SFE
The necessary apparatus for a SFE setup is simple. Figure 3.3.3 depicts the basic elements of a SFE instrument, which is composed
of a reservoir of supercritical fluid, a pressure tuning injection unit, two pumps (to take the components in the mobile phase in and
to send them out of the extraction cell), and a collection chamber.

3.3.4 https://chem.libretexts.org/@go/page/55864
Figure 3.3.3 Scheme of an idealized supercritical fluid extraction instrument.
There are two principle modes to run the instrument:
Static extraction.
Dynamic extraction.
In dynamic extraction, the second pump sending the materials out to the collection chamber is always open during the extraction
process. Thus, the mobile phase reaches the extraction cell and extracts components in order to take them out consistently.
In the static extraction experiment, there are two distinct steps in the process:
1. The mobile phase fills the extraction cell and interacts with the sample.
2. The second pump is opened and the extracted substances are taken out at once.
In order to choose the mobile phase for SFE, parameters taken into consideration include the polarity and solubility of the samples
in the mobile phase. Carbon dioxide is the most common mobile phase for SFE. It has a capability to dissolve non-polar materials
like alkanes. For semi-polar compounds (such as polycyclic aromatic hydrocarbons, aldehydes, esters, alcohols, etc.) carbon
dioxide can be used as a single component mobile phase. However, for compounds which have polar characteristic, supercritical
carbon dioxide must be modified by addition of polar solvents like methanol (CH3OH). These extra solvents can be introduced into
the system through a separate injection pump.

Extraction Modes
There are two modes in terms of collecting and detecting the components:
Off-line extraction.
On-line extraction.
Off-line extraction is done by taking the mobile phase out with the extracted components and directing them towards the collection
chamber. At this point, supercritical fluid phase is evaporated and released to atmosphere and the components are captured in a
solution or a convenient adsorption surface. Then the extracted fragments are processed and prepared for a separation method. This
extra manipulation step between extractor and chromatography instrument can cause errors. The on-line method is more sensitive
because it directly transfers all extracted materials to a separation unit, mostly a chromatography instrument, without taking them
out of the mobile phase. In this extraction/detection type, there is no extra sample preparation after extraction for separation
process. This minimizes the errors coming from manipulation steps. Additionally, sample loss does not occur and sensitivity
increases.

Applications of SFE
SFE can be applied to a broad range of materials such as polymers, oils and lipids, carbonhydrates, pesticides, organic pollutants,
volatile toxins, polyaromatic hydrocarbons, biomolecules, foods, flavors, pharmaceutical metabolites, explosives, and
organometallics, etc. Common industrial applications include the pharmaceutical and biochemical industry, the polymer industry,
industrial synthesis and extraction, natural product chemistry, and the food industry.
Examples of materials analyzed in environmental applications: oils and fats, pesticides, alkanes, organic pollutants, volatile toxins,
herbicides, nicotin, phenanthrene, fatty acids, aromatic surfactants in samples from clay to petroleum waste, from soil to river
sediments. In food analyses: caffeine, peroxides, oils, acids, cholesterol, etc. are extracted from samples such as coffee, olive oil,
lemon, cereals, wheat, potatoes and dog feed. Through industrial applications, the extracted materials vary from additives to
different oligomers, and from petroleum fractions to stabilizers. Samples analyzed are plastics, PVC, paper, wood etc. Drug
metabolites, enzymes, steroids are extracted from plasma, urine, serum or animal tissues in biochemical applications.

3.3.5 https://chem.libretexts.org/@go/page/55864
Summary
Supercritical fluid chromatography and supercritical fluid extraction are techniques that take advantage of the unique properties of
supercritical fluids. As such, they provide advantages over other related methods in both chromatography and extraction.
Sometimes they are used as alternative analytical techniques, while other times they are used as complementary partners for binary
systems. Both SFC and SFE demonstrate their versatility through the wide array of applications in many distinct domains in an
advantageous way.

3.3: Basic Principles of Supercritical Fluid Chromatography and Supercrtical Fluid Extraction is shared under a CC BY 4.0 license and was
authored, remixed, and/or curated by Pavan M. V. Raja & Andrew R. Barron via source content that was edited to conform to the style and
standards of the LibreTexts platform; a detailed edit history is available upon request.

3.3.6 https://chem.libretexts.org/@go/page/55864
3.4: Supercritical Fluid Chromatography
A popular and powerful tool in the chemical world, chromatography separates mixtures based on chemical properties – even some
than were previously thought inseparable. It combines a multitude of pieces, concepts, and chemicals to form an instrument suited
to specific separation. One form of chromatography that is often overlooked is that of supercritical fluid chromatography.

History
Supercritical fluid chromatography (SFC) begins its history in 1962 under the name “high pressure gas chromatography”. It started
off slow and was quickly overshadowed by the development of high performance liquid chromatography (HPLC) and the already
developed gas chromatography. SFC was not a popular method of chromatography until the late 1980s, when more publications
began exemplifying its uses and techniques.
SFC was first reported by Klesper et al. They succeeded in separating thermally labile porphyrin mixtures on polyethylene glycol
stationary phase with two mobile phase units: dichlorodifluoromethane (CCl2F2) and monochlorodifluoromethane (CHCl2F), as
shown in Figure 3.4.1. Their results proved that supercritical fluids’ low viscosity but high diffusivity functions well as a mobile
phase.

Figure 3.4.1 Thermally labile porphyrins (a) nickel etioporphyrin II and (b) nickel mesoporphyrin IX.
After Klesper’s paper detailing his separation procedure, subsequent scientists aimed to find the perfect mobile phase and the
possible uses for SFC. Using gases such as He, N2, CO2, and NH3, they examined purines, nucleotides, steroids, sugars, terpenes,
amino acids, proteins, and many more substances for their retention behavior. They discovered that CO2 was an ideal supercritical
fluid due to its low critical temperature of 31 °C and relatively low critical pressure of 72.8 atm. Extra advantages of CO2 included
it being cheap, non-flammable, and non-toxic. CO2 is now the standard mobile phase for SFC.
In the development of SFC over the years, the technique underwent multiple trial-and-error phases. Open tubular capillary column
SFC had the advantage of independently and cooperatively changing all three parameters (pressure, temperature, and modifier
content) to a certain extent. Like any chromatography method, however, it had its drawbacks. Changing the pressure, the most
important parameter, often required changing the flow velocity due to the constant diameter of the capillaries. Additionally, CO2,
the ideal mobile phase, is non-polar, and its polarity could not be altered easily or with a gradient.
Over the years, many uses were discovered for SFC. It was identified as a useful tool in the separation of chiral compounds, drugs,
natural products, and organometallics (see below for more detail). Most SFCs currently are involved a silica (or silica + modifier)
packed column with a CO2 (or CO2 + modifier) mobile phase. Mass spectrometry is the most common tool used to analyze the
separated samples.

Supercritical Fluids
What is a Supercritical Fluid?
As mentioned previously, the advantage to supercritical fluids is the combination of the useful properties from two phases: liquids
and gases. Supercritical fluids are gas-like in the ways of expanding to fill a given volume, and the motions of the particles are
close to that of a gas. On the side of liquid properties, supercritical fluids have densities near that of liquids and thus dissolve and
interact with other particles, as you would expect of a liquid. To visualize phase changes in relation to pressure and temperature,
phase diagrams are used as shown in Figure 3.4.2

3.4.1 https://chem.libretexts.org/@go/page/55866
Figure 3.4.2 A generic phase diagram (with relevant points labeled).
Figure 3.4.2 shows the stark differences between two phases in relation to the surrounding conditions. There exist two ambiguous
regions. One of these is the point at which all three lines intersect: the triple point. This is the temperature and pressure at which all
three states can exist in a dynamic equilibrium. The second ambiguous point comes at the end of the liquid/gas line, where it just
ends. At this temperature and pressure, the pure substance has reached a point where it will no longer exist as just one phase or the
other: it exists as a hybrid phase – a liquid and gas dynamic equilibrium.

Unique Properties of Supercritical Fluids


As a result of the dynamic liquid-gas equilibrium, supercritical fluids possess three unique qualities: increased density (on the scale
of a liquid), increased diffusivity (similar to that of a gas), and lowered viscosity (on the scale of a gas). Table 3.4.1 shows the
similarities in each of these properties. Remember, each of these explains a part of why SFC is an advantageous method of
chemical separation.
Table 3.4.1 : Typical properties of gas, liquid, and supercritical fluid of typical organic compounds (order of magnitude).
Density (g/mL) Diffusivity (cm2/s) Dynamic Viscosity (g/cm s)

Gas 1 x 10-3 1 x 10-1 1 x 10-2

Liquid 1.0 5 x 10-6 1 x 10-4

Supercritical Fluid 3 x 10-1 1 x 10-3 1 x 10-2

Applying the Properties of Supercritical Fluids to Chromatography


How are these properties useful? An ideal mobile phase and solvent will do three things well: interact with other particles, carry the
sample through the column, and quickly (but accurately) elute it.
Density, as a concept, is simple: the denser something is, the more likely that it will interact with particles it moves through.
Affected by an increase in pressure (given constant temperature), density is largely affected by a substance entering the
supercritical fluid zone. Supercritical fluids are characterized with densities comparable to those of liquids, meaning they have a
better dissolving effect and act as a better carrier gas. High densities among supercritical fluids are imperative for both their effect
as solvents and their effect as carrier gases.
Diffusivity refers to how fast the substance can spread among a volume. With increased pressure comes decreased diffusivity (an
inverse relationship) but with increased temperature comes increased diffusivity (a direct relationship related to their kinetic
energy). Because supercritical fluids have diffusivity values between a gas and liquid, they carry the advantage of a liquid’s density,
but the diffusivity closer to that of a gas. Because of this, they can quickly carry and elute a sample, making for an efficient mobile
phase.
Finally, dynamic viscosity can be viewed as the resistance to other components flowing through, or intercalating themselves, in the
supercritical fluid. Dynamic viscosity is hardly affected by temperature or pressure for liquids, whereas it can be greatly affected
for supercritical fluids. With the ability to alter dynamic viscosity through temperature and pressure, the operator can determine
how resistant their supercritical fluid should be.

3.4.2 https://chem.libretexts.org/@go/page/55866
Supercritical Properties of CO2
Because of its widespread use in SFC, it’s important to discuss what makes CO2 an ideal supercritical fluid. One of the biggest
limitations to most mobile phases in SFC is getting them to reach the critical point. This means extremely high temperatures and
pressures, which is not easily attainable. The best gases for this are ones that can achieve a critical point at relatively low
temperatures and pressures.
As seen from Figure 3.4.3, CO2 has a critical temperature of approximately 31 °C and a critical pressure of around 73 atm. These
are both relatively low numbers and are thus ideal for SFC. Of course, with every upside there exists a downside. In this case, CO2
lacks polarity, which makes it difficult to use its mobile phase properties to elute polar samples. This is readily fixed with a
modifier, which will be discussed later.

Figure 3.4.3 Phase diagram of CO2

The Instrument
SFC has a similar instrument setup to most other chromatography machines, notably HPLC. The functions of the parts are very
similar, but it is important to understand them for the purposes of understanding the technique. Figure 3.4.4 shows a schematic
representation of a typical apparatus.

Figure 3.4.4 Box diagram of a SFC machine.

Columns
There are two main types of columns used with SFC: open tubular and packed, as seen below. The columns themselves are near
identical to HPLC columns in terms of material and coatings. Open tubular columns are most used and are coated with a cross-
linked silica material (powdered quartz, SiO2) for a stationary phase. Column lengths range, but usually fall between 10 and 20
meters and are coated with less than 1 µm of silica stationary phase. Figure 3.4.5 demonstrates the differences in the packing of the
two columns.

3.4.3 https://chem.libretexts.org/@go/page/55866
Figure 3.4.5 Schematic visualization of the difference between (a) open tubular and (b) packed column.

Injector
Injectors act as the main site for the insertion of samples. There are many different kinds of injectors that depend on a multitude of
factors. For packed columns, the sample must be small and the exact amount depends on the column diameter. For open tubular
columns, larger volumes can be used. In both cases, there are specific injectors that are used depending on how the sample needs to
be placed in the instrument. A loop injector is used mainly for preliminary testing. The sample is fed into a chamber that is then
flushed with the supercritical fluid and pushed down the column. It uses a low-pressure pump before proceeding with the full
elution at higher pressures. An inline injector allows for easy control of sample volume. A high-pressure pump forces the
(specifically measured) sample into a stream of eluent, which proceeds to carry the sample through the column. This method allows
for specific dilutions and greater flexibility. For samples requiring no dilution or immediate interaction with the eluent, an in-
column injector is useful. This allows the sample to be transferred directly into the packed column and the mobile phase to then
pass through the column.

Pump
The existence of a supercritical fluid, as discussed previously, depends on high temperatures and high pressures. The pump is
responsible for delivering the high pressures. By pressurizing the gas (or liquid), it can cause the substance to become dense
enough to exhibit signs of the desired supercritical fluid. Because pressure couples with heat to create the supercritical fluid, the
two are usually very close together on the instrument.

Oven
The oven, as referenced before, exists to heat the mobile phase to its desired temperature. In the case of SFC, the desired
temperature is always the critical temperature of the supercritical fluid. These ovens are precisely controlled and standard across
SFC, HPLC, and GC.

Detector
So far, there has been one largely overlooked component of the SFC machine: the detector. Technically not a part of the
chromatographic separation process, the detector still plays an important role: identifying the components of the solution. While the
SFC aims to separate components with good resolution (high purity, no other components mixed in), the detector aims to define
what each of these components is made of.
The two detectors most often found on SFC instruments are either flame ionization detectors (FID) or mass spectrometers (MS):
FIDs operate through ionizing the sample in a hydrogen-powered flame. By doing so, they produce charged particles, which hit
electrodes, and the particles are subsequently quantified and identified.
MS operates through creating an ionized spray of the sample, and then separating the ions based on a mass/charge ratio. The
mass/charge ratio is plotted against ion abundance and creates a “fingerprint” for the chemical identified. This chemical
fingerprint is then matched against a database to isolate which compound it was. This can be done for each unique elution,
rendering the SFC even more useful than if it were standing alone.

Sample
Generally speaking, samples need little preparation. The only major requirement is that it dissolves in a solvent less polar than
methanol: it must have a dielectric constant lower than 33, since CO2 has a low polarity and cannot easily elute polar samples. To
combat this, modifiers are added to the mobile phase.

Stationary Phase
The stationary phase is a neutral compound that acts as a source of “friction” for certain molecules in the sample as they slide
through the column. Silica attracts polar molecules and thus the molecules attach strongly, holding until enough of the mobile

3.4.4 https://chem.libretexts.org/@go/page/55866
phase has passed through to attract them away. The combination of the properties in the stationary phase and the mobile phase help
determine the resolution and speed of the experiment.

Mobile Phase
The mobile phase (the supercritical fluid) pushes the sample through the column and elutes separate, pure, samples. This is where
the supercritical fluid’s properties of high density, high diffusivity, and low viscosity come into play. With these three properties,
the mobile phase is able to adequately interact with the sample, quickly push through it, and strongly plow through the sample to
separate it out. The mobile phase also partly determines how it separates out: it will first carry out similar molecules, ones with
similar polarities, and follow gradually with molecules with larger polarities.

Modifiers
Modifiers are added to the mobile phase to play with its properties. As mentioned a few times previously, CO2supercritical fluid
lacks polarity. In order to add polarity to the fluid (without causing reactivity), a polar modifier will often be added. Modifiers
usually raise the critical pressure and temperature of the mobile phase a little, but in return add polarity to the phase and result in a
fully resolved sample. Unfortunately, with too much modifier, higher temperatures and pressures are needed and reactivity
increases (which is dangerous and bad for the operator). Modifiers, such as ethanol or methanol, are used in small amounts as
needed for the mobile phase in order to create a more polar fluid.

Advantages of Supercritical Fluid Chromatography


Clearly, SFC possesses some extraordinary potential as far as chromatography techniques go. It has some incredible capabilities
that allow efficient and accurate resolution of mixtures. Below is a summary of its advantages and disadvantages stacked against
other conventional (competing) chromatography methods.

Advantages over HPLC


Because supercritical fluids have low viscosities the analysis is faster, there is a much lower pressure drop across the column,
and open tubular columns can be used.
Shorter column lengths are needed (10-20 m for SFC versus 15-60 m for HPLC) due to the high diffusivity of the supercritical
fluid. More interactions can occur in a shorter span of time/distance.
Resolving power is much greater (5x) than HPLC due to the high diffusivity of the supercritical fluid. More interactions result
in better separation of the components in a shorter amount of time.

Advantages over GC
Able to analyze many solutes with no derivatization since there is no need to convert most polar groups into nonpolar ones.
Can analyze thermally labile compounds more easily with high resolution since it can provide faster analysis at lower
temperatures.
Can analyze solutes with high molecular weight due to their greater solubizing power.

General Disadvantages
Cannot analyze extremely polar solutes due to relatively nonpolar mobile phase, CO2.

Applications
While the use of SFC has been mainly organic-oriented, there are still a few ways that inorganic compound mixtures are separated
using the method. The two main ones, separation of chiral compounds (mainly metal-ligand complexes) and organometallics are
discussed here.

Chiral Compounds
For chiral molecules, the procedures and choice of column in SFC are very similar to those used in HPLC. Packed with cellulose
type chiral stationary phase (or some other chiral stationary phase), the sample flows through the chiral compound and only
molecules with a matching chirality will stick to the column. By running a pure CO2 supercritical fluid mobile phase, the non-
sticking enantiomer will elute first, followed eventually (but slowly) with the other one.
In the field of inorganic chemistry, a racemic mixture of Co(acac)3, both isomers shown in Figure 3.4.6 has been resolved using a
cellulose-based chiral stationary phase. The SFC method was one of the best and most efficient instruments in analyzing the chiral

3.4.5 https://chem.libretexts.org/@go/page/55866
compound. While SFC easily separates coordinate covalent compounds, it is not necessary to use such an extensive instrument to
separate mixtures of it since there are many simpler techniques.

Figure 3.4.6 The two isomers of Co(acac)3 in a racemic mixture which were resolved by SFC.

Organometallics
Many d-block organometallics are highly reactive and easily decompose in air. SFC offers a way to chromatograph mixtures of
large, unusual organometallic compounds. Large cobalt and rhodium based organometallic compound mixtures have been
separated using SFC (Figure 3.4.7 ) without exposing the compounds to air.

Figure 3.4.7 Examples of cobalt and rhodium based organometallic compound mixtures separated by SFC. Adapted from
Compounds by I Bruheim, E Fooladi, E. Lundanes, T. Greibrokk, J. Microcolumn Sep., 2001, 13, 156.
By using a stationary phase of siloxanes, oxygen-linked silicon particles with different substituents attached, the organometallics
were resolved based on size and charge. Thanks to the non-polar, highly diffusive, and high viscosity properties of a 100% CO2
supercritical fluid, the mixture was resolved and analyzed with a flame ionization detector. It was determined that the method was
sensitive enough to detect impurities of 1%. Because the efficiency of SFC is so impressive, the potential for it in the
organometallic field is huge. Identifying impurities down to 1% shows promise for not only preliminary data in experiments, but
quality control as well.

Conclusion
While it may have its drawbacks, SFC remains an untapped resource in the ways of chromatography. The advantages to using
supercritical fluids as mobile phases demonstrate how resolution can be increased without sacrificing time or increasing column

3.4.6 https://chem.libretexts.org/@go/page/55866
length. Nonetheless, it is still a well-utilized resource in the organic, biomedical, and pharmaceutical industries. SFC shows
promise as a reliable way of separating and analyzing mixtures.

3.4: Supercritical Fluid Chromatography is shared under a CC BY 4.0 license and was authored, remixed, and/or curated by Pavan M. V. Raja &
Andrew R. Barron via source content that was edited to conform to the style and standards of the LibreTexts platform; a detailed edit history is
available upon request.

3.4.7 https://chem.libretexts.org/@go/page/55866
3.5: Ion Chromatography
Ion Chromatography is a method of separating ions based on their distinct retention rates in a given solid phase packing material.
Given different retention rates for two anions or two cations, the elution time of each ion will differ, allowing for detection and
separation of one ion before the other. Detection methods are separated between electrochemical methods and spectroscopic
methods. This guide will cover the principles of retention rates for anions and cations, as well as describing the various types of
solid-state packing materials and eluents that can be used.

Principles of Ion Chromatography


Retention Models in Anion Chromatography
The retention model for anionic chromatography can be split into two distinct models, one for describing eluents with a single
anion, and the other for describing eluents with complexing agents present. Given an eluent anion or an analyte anion, two phases
are observed, the stationary phase (denoted by S) and the mobile phase (denoted by M). As such, there is equilibrium between the
two phases for both the eluent anions and the analyte anions that can be described by Equation 3.5.1.
x− y− x− y−
y ∗ [A ]  +  x ∗ [ E ]  ⇔  y ∗ [ A ]  +  x ∗ [ E ] (3.5.1)
M S S M

This yields an equilibrium constant as given in Equation 3.5.2 .


x− y y− x y x
[A ] [E ] γ x−
γ y−
S M A E
S S

KA,E = (3.5.2)
x− y− y x
y x
[A ] [E ] γ x− γ y−
M S A E
M S

Given the activity of the two ions cannot be found in the stationary or mobile phases, the activity coefficients are set to 1. Two new
quantities are then introduced. The first is the distribution coefficient, DA, which is the ratio of analyte concentrations in the
stationary phase to the mobile phase, Equation 3.5.3 . The second is the retention factor, k1A, which is the distribution coefficient
times the ratio of volume between the two phases, Equation 3.5.4 .
[ AS ]
DA   =   (3.5.3)
[ AM ]

VS
1
k   =  DA ∗ (3.5.4)
A
VM

Substituting the two quantities from Equation 3.5.3 and Equation 3.5.4 into Equation 3.5.2 , the equilibrium constant can be
written as Equation 3.5.5
y−
VM [E ]
1 y M x
KA,E   = (k ) ∗( ) (3.5.5)
A y−
VS [E ]
S

Given there is usually a large difference in concentrations between the eluent and the analyte (with magnitudes of 10 greater
eluent), equation 4 can be re-written under the assumption that all the solid phase packing material’s functional groups are taken up
by Ey-. As such, the stationary Ey- can be substituted with the exchange capacity divided by the charge of Ey-. This yields Equation
3.5.6

VM y
Q −x y−
1
KA,E   = (k ) ∗( ) [E ] (3.5.6)
A M
VS γ

Solving for the retention factor Equation 3.5.7 is developed.


x− z− x− z−
z ∗ [A ]  +  x ∗ [ B ] ⇔ z ∗ [A ]  +  x ∗ [ B ] (3.5.7)
M S S M

Equation 3.5.8 shows the relationship between retention factor and parameters like eluent concentration and the exchange capacity,
which allows parameters of the ion chromatography to be manipulated and the retention factors to be determined. Equation 3.5.9
only works for a single analyte present, but a relationship for the selectivity between two analytes [A] and [B] can easily be
determined.
First the equilibrium between the two analytes is determined as Equation 3.5.8

3.5.1 https://chem.libretexts.org/@go/page/55868
x− z z− x
[A ] [B ]
S M
KA,B   = (3.5.8)
x− z z− x
[A ] [B ]
M S

The equilibrium constant can be written as Equation 3.5.9 (ignoring activity):


x− z−
[A ][ B ]
S M
αA,B   = (3.5.9)
x− z−
[A ][ B ]
M S

The selectivity can then be determined to be Equation 3.5.10


x− z−
[A ][ B ]
S M
αA,B   = (3.5.10)
x− z−
[A ][ B ]
M S

Equation 3.5.10 can then be simplified into a logarithmic form as the following two equations:
1
1 x −z k VM
A
logαA,B = logKA,B   + log (3.5.11)
z z VS

1
1 x −z k VM
A
logαA,B = logKA,B   + log (3.5.12)
x z VS

When the two charges are the same, it can be seen that the selectivity is only a factor of the selectivity coefficients and the charges.
When the two charges are different, it can be seen that the two retention factors are dependent upon each other.
In situations with a polyatomic eluent, three models are used to account for the multiple anions in the eluent. The first is the
dominant equilibrium model, in which one anion is so dominant in concentration; the other eluent anions are ignored. The
dominant equilibrium model works best for multivalence analytes. The second is the effective charge model, where an effective
charge of the eluent anions is found, and a relationship similar to EQ is found with the effective charge. The effective charge
models works best with monovalent analytes. The third is the multiple eluent species model, where Equation 3.5.13 describes the
retention factor:
X1 X2 X3
1
logK   =  C3 − ( + + ) −  logCP (3.5.13)
A
a b c

C3 is a constant that includes the phase volume ratio between stationary, the equilibrium constant, and mobile and the exchange
capacity. Cp is the total concentration of the eluent species. X1, X2, X3, correspond to the shares of a particular eluent anion in the
retention of the analyte.

Retention Models of Cation Chromatography


For eluents with a single cation and analytes that are alkaline earth metals, heavy metals or transition metals, a complexing agent is
used to bind with the metal during chromatography. This introduces the quantity A(m) to the retention rate calculations, where
A(m) is the ratio of free metal ion to the total concentration of metal. Following a similar derivation to the single anion case,
Equation 3.5.14 is found.
1
k Q
A y −x y+ x
KA,E =  ( ) ∗( ) [E ] (3.5.14)
M
αM ϕ γ

Solving for the retention coefficient, Equation 3.5.15 is found.


1
Q x
y+ −
x
1 γ
k = αM ϕ ∗ K ( ) y
([ E ] y
(3.5.15)
A A,E M
γ

From this expression, the retention rate of the cation can be determined from eluent concentration and the ratio of free metal ions to
the total concentration of the metal, which itself is depended on the equilibrium of the metal ion with the complexing agent.

Solid Phase Packing Materials


The solid phase packing material used in the chromatography column is important to the exchange capacity of the anion or cation.
There are many types of packing material, but all share a functional group that can bind either the anion or the cation complex. The
functional group is mounted on a polymer surface or sphere, allowing large surface area for interaction.

3.5.2 https://chem.libretexts.org/@go/page/55868
Packing Material for Anion Chromatography
The primary functional group used for anion chromatography is the ammonium group. Amine groups are mounted on the polymer
surface, and the pH is lowered to produce ammonium groups. As such, the exchange capacity is depended on the pH of the eluent.
To reduce the pH dependency, the protons on the ammonium are successively replaced with alkyl groups until the all the protons
are replaced and the functional group is still positively charged, but pH independent. The two packing materials used in almost all
anion chromatography are trimethylamine (NMe3, Figure 3.5.1 ) and dimethylanolamine (Figure 3.5.2 ).

Figure 3.5.1 A trimethylamine mounted on a polymer used as a solid phase packing material.

Figure 3.5.2 A dimethylethanolamine mounted on a polymer used as solid phase packing material.

Packing Material for Cation Chromatography


Cation chromatography allows for the use of both organic polymer based and silica gel based packing material. In the silica gel
based packing material, the most common packing material is a polymer-coated silica gel. The silicate is coated in polymer, which
is held together by cross-linking of the polymer. Polybutadiene maleic acid (Figure 3.5.3 ) is then used to create a weakly acidic
material, allowing the analyte to diffuse through the polymer and exchange. Silica gel based packing material is limited by the pH
dependent solubility of the silica gel and the pH dependent linking of the silica gel and the functionalized polymer. However, silica
gel based packing material is suitable for separation of alkali metals and alkali earth metals.

Figure 3.5.3 A polybutadiene maleic acid polymer used as a cation solid phase packing material.
Organic polymer based packing material is not limited by pH like the silica gel materials are, but are not suitable for separation of
alkali metals and alkali earth metals. The most common functional group is the sulfonic acid group (Figure 3.5.4 ) attached with a
spacer between the polymer and the sulfonic acid group.

Figure 3.5.4 A sulfonic acid group used as a cation solid phase packing material functional group.

Detection Methods
Spectroscopic Detection Methods
Photometric detection in the UV region of the spectrum is a common method of detection in ion chromatography. Photometric
methods limit the eluent possibilities, as the analyte must have a unique absorbance wavelength to be detectable. Cations that do
not have a unique absorbance wavelength, i.e. the eluent and other contaminants have similar UV visible spectra can be complexed
to for UV visible compounds. This allows detection of the cation without interference from eluents.
Coupling the chromatography with various types of spectroscopy such as Mass spectroscopy or IR spectroscopy can be a useful
method of detection. Inductively coupled plasma atomic emission spectroscopy is a commonly used method.

3.5.3 https://chem.libretexts.org/@go/page/55868
Direct Conductivity Methods
Direct conductivity methods take advantage of the change in conductivity that an analyte produces in the eluent, which can be
modeled by Equation 3.5.16 where equivalent conductivity is defined as Equation 3.5.17 .
(ΛA   −  Λg ) ∗ Cs
ΔK  = (3.5.16)
1000

L 1
Λ  = ∗ (3.5.17)
A∗R C

With L being the distance between two electrodes of area A and R being the resistance the ion creates. C is the concentration of the
ion. The conductivity can be plotted over time, and the peaks that appear represent different ions coming through the column as
described by Equation 3.5.18

Kpeak   =  (ΛA   −  Λg ) ∗ CA (3.5.18)

The values of Equivalent conductivity of the analyte and of the eluent common ions can be found in Table 3.5.1
Table 3.5.1
Cations Λ
+ 2
(S cm eq
−1
) Anions Λ
− 2
(S cm eq
−1
)

H
+
350 OH

198

Li
+
39 F

54

Na
+
50 Cl

76

K
+
74 Br

78

NH
4+
73 I

77

1/2M g
2+
53 NO

2
72

1/2Ca
2+
60 NO

3
71

1/2Sr
2+
59 HCO

3
45

1/2Ba
2+
64 1/2CO
2−
3
72

1/2Zn
2+
52 H2 P O

4
33

1/2Hg
2+
53 1/2HP O

4
57

1/2Cu
2+
55 1/3P O

4
69

1/2P b
2+
71 1/2SO
2−
4
80

1/2Co
2+
53 CN

82

1/3F e
3+
70 SCN

66

N (Et)
4+
33 Acetate 41

1/2 Phthalate 38

Propionate 36

Benzoate 32

Salicylate 30

1/2 Oxalate 74

Eluents
The choice of eluent depends on many factors, namely, pH, buffer capacity, the concentration of the eluent, and the nature of the
eluent’s reaction with the column and the packing material.

3.5.4 https://chem.libretexts.org/@go/page/55868
Eluents in Anion Chromatography
In non-suppressed anion chromatography, where the eluent and analyte are not altered between the column and the detector, there is
a wide range of eluents to be used. In the non-suppressed case, the only issue that could arise is if the eluent impaired the detection
ability (absorbing in a similar place in a UV-spectra as the analyte for instance). As such, there are a number of commonly used
eluents. Aromatic carboxylic acids are used in conductivity detection because of their low self-conductivity. Aliphatic carboxylic
acids are used for UV/visible detection because they are UV transparent. Inorganic acids can only be used in photometric detection.
In suppressed anion chromatography, where the eluent and analyte are treated between the column and detection, fewer eluents can
be used. The suppressor modifies the eluent and the analyte, reducing the self-conductivity of the eluent and possibly increasing the
self-conductivity of the analyte. Only alkali hydroxides and carbonates, borates, hydrogen carbonates, and amino acids can be used
as eluents.

Eluents in Cation Chromatography


The primary eluents used in cation chromatography of alkali metals and ammoniums are mineral acids such as HNO3. When the
cation is multivalent, organic bases such as ethylenediamine (Figure 3.5.5 ) serve as the main eluents. If both alkali metals and
alkali earth metals are present, hydrochloric acid or 2,3-diaminopropionic acid (Figure 3.5.6 ) is used in combination with a pH
variation. If the chromatography is unsuppressed, the direct conductivity measurement of the analyte will show up as a negative
peak due to the high conductivity of the H+ in the eluent, but simple inversion of the data can be used to rectify this discrepancy.

Figure 3.5.5 Ethylenediamine, a commonly used eluent in cation chromatography.

Figure 3.5.6 2,3-diaminopropionic acid, a primary eluent for cation chromatography of alkali and alkali earth metal combinations.
If transition metals or H+ are the analytes in question, complexing carboxylic acids are used to suppress the charge of the analyte
and to create photometrically detectable complexes, forgoing the need for direct conductivity as the detection method.

3.5: Ion Chromatography is shared under a CC BY 4.0 license and was authored, remixed, and/or curated by Pavan M. V. Raja & Andrew R.
Barron via source content that was edited to conform to the style and standards of the LibreTexts platform; a detailed edit history is available upon
request.

3.5.5 https://chem.libretexts.org/@go/page/55868
3.6: Capillary Electrophoresis
Capillary electrophoresis (CE) encompasses a family of electrokinetic separation techniques that uses an applied electric field to
separate out analytes based on their charge and size. The basic principle is hinged upon that of electrophoresis, which is the motion
of particles relative to a fluid (electrolyte) under the influence of an electric field. The founding father of electrophoresis, Arne W.
K. Tiselius (Figure 3.6.1a ), first used electrophoresis to separate proteins, and he went on to win a Nobel Prize in Chemistry in
1948 for his work on both electrophoresis and adsorption analysis. However, it was Stellan Hjerten (Figure 3.6.1b ) who worked
under Arne W. K. Tiselius, who pioneered work in CE in 1967, although CE was not well recognized until 1980 when James W.
Jorgenson (Figure 3.6.1c ) and Krynn D. Lukacs published a series of papers describing this new technique.

Figure 3.6.1 : (left) Swedish chemist Arne W. K. Tiselius (1902–1971) who was the founding father of electrophoresis. (center)
Swedish chemist Stellan Hjerten (1928–present) who worked under Arne W. K. Tiselius that pioneered work in CE. (right) James
W. Jorgensen (1952-present).

Instrument Overview
The main components of CE are shown in Figure 3.6.2. The electric circuit of the CE is the heart of the instrument.

Figure 3.6.2 A schematic diagram of the components of a typical capillary electrophoresis setup and the capillary column.
Injection Methods
The samples that are studied in CE are mainly liquid samples. A typical capillary column has an inner diameter of 50 μm and a
length of 25 cm. Because the column can only contain a minimal amount of running buffer, only small sample volumes can be
tested (nL to μL). The samples are introduced mainly by two injection methods: hydrodynamic and electrokinetic injection. The
two methods are displayed in Table 3.6.1 A disadvantage of electrokinetic injection is that the composition of the injected sample
may not be the same as the composition of the original sample. This is because the injection method is dependent on the
electrophoretic and electroosmotic mobility of the species in the sample. However, both injection methods depend on the
temperature and the viscosity of the solution. Hence, it is important to control both parameters when a reproducible volume of
sample injections is desired. It is advisable to use internal standards instead of external standards when performing quantitative
analysis on the samples as it is hard to control both the temperature and viscosity of the solution.
Table 3.6.1 The working principle of the two injection methods used in CE.
Injection Methods Working Principle

3.6.1 https://chem.libretexts.org/@go/page/143373
Injection Methods Working Principle

The sample vial is enclosed in a chamber with one end of fixed


capillary column immersed in it. Pressure is then applied to the chamber
Hydrodynamic Injection for a fixed period so that the sample can enter the capillary. After the
sample, has been introduced, the capillary is withdrawn and then re-
immersed into the source reservoir and separation takes place.
The sample is enclosed in a chamber with one end of capillary column
immersed in it with an electrode present. The electric field is applied,
Electrokinetic Injection and the samples enter the capillary. After the sample, has been
introduced, the capillary is withdrawn and then re-immersed into the
source reservoir and separation takes place.

Column
After the samples have been injected, the capillary column is used as the main medium to separate the components. The capillary
column used in CE shares the same characteristics as the capillary column used in gas chromatography (GC); however, the most
critical components of the CE column are:
the inner diameter of the capillary,
the total length of the capillary,
the length of the column from the injector to the detector.
Solvent Buffer
The solvent buffer carries the sample through the column. It is crucial to employ a good buffer as a successful CE experiment is
hinged upon this. CE is based on the separation of charges in an electric field. Therefore, the buffer should either sustain the pre-
existing charge on the analyte or enable the analyte to obtain a charge, and it is important to consider the pH of the buffer before
using it.
Applied Voltage (kV)

The applied voltage is important in the separation of the analytes as it drives the movement of the analyte. It is important that it is
not too high as it may become a safety concern.
Detectors
Analytes that have been separated after the applying the voltage can be detected by many detection methods. The most common
method is UV-visible absorbance. The detection takes place across the capillary with a small portion of the capillary acting as the
detection cell. The on-tube detection cell is usually made optically transparent by scraping off the polyimide coating and coating it
with another optically transparent material so that the capillary would not break easily. For species that do not have a chromophore,
a chromophore can be added to the buffer solution. When the analyte passes by, there would be a decrease in signal. This decreased
signal will correspond to the amount of analyte present. Other common detection techniques employable in CE are fluorescence
and mass spectrometry (MS).

Theory
In CE, the sample is introduced into the capillary by the above-mentioned methods. A high voltage is then applied causing the ions
of the sample to migrate towards the electrode in the destination reservoir, in this case, the cathode. Sample components migration
and separation are determined by two factors, electrophoretic mobility and electroosmotic mobility.
Electrophoretic Mobility
The electrophoretic mobility, μ , is inherently dependent on the properties of the solute and the medium in which the solute is
ep

moving. Essentially, it is a constant value, that can be calculated as given by 3.6.1 where q is the solute`s charge, η is the buffer
viscosity and r is the solute radius.
q
μep = (3.6.1)
6πηr

The electrophoretic velocity, v , is dependent on the electrophoretic mobility and the applied electric field, E (3.6.2).
ep

νep = μep E (3.6.2)

3.6.2 https://chem.libretexts.org/@go/page/143373
Thus, when solutes have a larger charge to size ratio the electrophoretic mobility and velocity will increase. Cations and the anion
would move in opposing directions corresponding to the sign of the electrophoretic mobility with is a result of their charge. Thus,
neutral species that have no charge do not have an electrophoretic mobility.
Electroosmotic Mobility
The second factor that controls the migration of the solute is the electroosmotic flow. With zero charge, it is expected that the
neutral species should remain stationary. However, under normal conditions, the buffer solution moves towards the cathode as well.
The cause of the electroosmotic flow is the electric double layer that develops at the silica solution interface.
At pH more than 3, the abundant silanol (-OH) groups present on the inner surface of the silica capillary, de-protonate to form
negatively charged silanate ions (-SiO-). The cations present in the buffer solution will be attracted to the silanate ions and some of
them will bind strongly to it forming a fixed layer. The formation of the fixed layer only partially neutralizes the negative charge on
the capillary walls. Hence, more cations than anions will be present in the layer adjacent to the fixed layer, forming the diffuse
layer. The combination of the fixed layer and diffuse layer is known as the double layer as shown in Figure 3.6.3. The cations
present in the diffuse layer will migrate towards the cathode, as these cations are solvated the solution will also flow with it,
producing the electroosmotic flow. The anions present in the diffuse layer are solvated and will move towards the anode. However,
as there are more cations than anions the cations will push the anions together with it in the direction of the cathode. Hence, the
electroosmotic flow moves in the direction of the cathode.

Figure 3.6.3 : An illustration of the electric double layer and movement of the species in solution. Adapted from D. Harvey,
Analytical Chemistry 2.0(e-textbook), 851.
The electroosmotic mobility, μeof, is described by 3.6.3 where ξ is the zeta potential, ε is the buffer dielectric constant and η is the
buffer viscosity. The electroosmotic velocity, veof, is the rate at which the buffer moves through the capillary is given by 3.6.4 .
ζε
μeof   =   (3.6.3)
4πη

νeof   =  μeof E (3.6.4)

Zeta Potential
The zeta potential, ξ, also known as the electrokinetic potential is the electric potential at the interface of the double layer. Hence, in
our case, it is the potential of the diffuse layer that is at a finite distance from the capillary wall. Zeta potential is mainly affected
and directly proportional to two factors:
1. The thickness of the double layer. A higher concentration of cations possibly due to an increase in the buffer`s ionic strength
would lead to a decrease in the thickness of the double layer. As the thickness of the double layer decreases, the zeta potential
would decrease that results in the decrease of the electroosmotic flow.
2. The charge on the capillary walls. A greater density of the silanate ions corresponds to a larger zeta potential. The formation of
silanate ions is pH dependent. Hence, at pH less than 2 there is a decrease in the zeta potential and the electroosmotic flow as
the silanol exists in its protonated form. However, as the pH increases, there are more silanate ions formed causing an increase
in zeta potential and hence, the electroosmotic flow.
Order of Elution
Electroosmotic flow of the buffer is generally greater than the electrophoretic flow of the analytes. Hence, even the anions would
move to the cathode as illustrated in Figure 3.6.4 Small, highly charged cations would be the first to elute before larger cations
with lower charge. This is followed by the neutral species which elutes as one band in the middle. The larger anions with low

3.6.3 https://chem.libretexts.org/@go/page/143373
charge elute next and lastly, the highly charged small anion would have the longest elution time. This is clearly portrayed in the
electropherogram in Figure 3.6.5

Figure 3.6.4 An illustration of the order of elution of the charged species. Adapted from D. A. Skoog, D. M. West, F. J. Holler
and S. R. Crouch, Fundamentals of Analytical Chemistry, Copyright Brooks Cole (2013).

Figure 3.6.5 A typical electropherogram demonstrating the order of elution of cations and anions. Adapted from J. Sáiz, I. J.
Koenka, T. Duc Mai, P. C. Hauser, C. García-Ruiz, TrAC, 2014, 62. 162.

Optimizing the CE Experiment


There are several components that can be varied to optimize the electropherogram obtained from CE. Hence, for any given setup
certain parameters should be known:
the total length of the capillary (L),
the length the solutes travel from the start to the detector (l),
the applied voltage (V).
Reduction in Migration Time, tmn
To shorten the analysis time, a higher voltage can be used or a shorter capillary tube can be used. However, it is important to note
that the voltage cannot be arbitrarily high as it will lead to joule heating. Another possibility is to increase μeof by increasing pH or
decreasing the ionic strength of the buffer, 3.6.5 .
1 L
tmn   =   (3.6.5)
(μep   +  μeof )V

3.6.4 https://chem.libretexts.org/@go/page/143373
Efficiency
In chromatography, the efficiency is given by the number of theoretical plates, N. In CE, there exist a similar parameter, 3.6.6
where D is the solute`s diffusion coefficient. Efficiency increase s with an increase in voltage applied as the solute spends less time
in the capillary there will be less time for the solute to diffuse. Generally, for CE, N will be very large.
2
1 μtot V l
N  = = (3.6.6)
2Dtmn 2DL

Resolution Between Two Peaks


The resolution between two peaks, R, is defined by 3.6.7 where Δv is the difference in velocity of two solutes and ṽ is the average
velocity of two solutes.


√N Δv
R = × (3.6.7)
~
4 ν

Substituting the equation by N gives 3.6.8


−−−−−−−−−−−−
V
R  = 0.177(μep,1   −  μep,2 )√ (3.6.8)
D(νav + μeof )

Therefore, increasing the applied voltage, V, will increase the resolution. However, it is not very effective as a 4-fold increase in
applied voltage would only give a 2-fold increase in resolution. In addition, increase in N, the number of theoretical plates would
result in better resolution.
Selectivity
In chromatography, selectivity, α, is defined as the ratio of the two retention factors of the solute. This is the same for CE, 3.6.9 ,
where t2 and t1 are the retention times for the two solutes such that, α is more than 1.
t2
α = (3.6.9)
t1

Selectivity can be improved by adjusting the pH of the buffer solution. The purpose is to change the charge of the species being
eluted.

Comparison Between CE and HPLC


CE unlike High-performance liquid chromatography (HPLC) accommodates many samples and tends to have a better resolution
and efficiency. A comparison between the two methods is given in Table 3.6.2.
Table 3.6.2 Advantages and disadvantages of CE versus HPLC.
CE HPLC

Wider selection of analyte to be analyzed Limited by the solubility of the sample

Higher efficiency, no stationary mass transfer term as there is no Efficiency is lowered due to the stationary mass transfer term
stationary phase (equilibration between the stationary and mobile phase)

Electroosmotic flow profile in the capillary is flat as a result no band Rounded laminar flow profile that is common in pressure driven
broadening. Better peak resolution and sharper peaks systems such as HPLC. Resulting in broader peaks and lower resolution

Some detectors require the solvent to be changed and prior modification


Can be coupled to most detectors depending on application
of the sample before analysis

Greater peak capacity as it uses a very large number of theoretical


The peak capacity is lowered as N is not as large
plates, N

High voltages are used when carrying out the experiment No need for high voltage

Micellar Electrokinetic Chromatography


CE allows the separation of charged particles, and it is mainly compared to ion chromatography. However, no separation takes
place for neutral species in CE. Thus, a modified CE technique named micellar electrokinetic chromatography (MEKC) can be

3.6.5 https://chem.libretexts.org/@go/page/143373
used to separate neutrals based on its size and its affinity to the micelle. In MEKC, surfactant species is added to the buffer solution
at a concentration at which micelles will form. An example of a surfactant is sodium dodecyl sulfate (SDS) as seen in Figure 3.6.6

Figure 3.6.6 : The structure of sodium dodecyl sulfate and its representation. An illustration of a cross section of a formed micelle.
Adapted from D. Harvey, Analytical Chemistry 2.0(e-textbook), 851.
Neutral molecules are in dynamic equilibrium between the bulk solution and interior of the micelle. In the absence of the micelle
the neutral species would reach the detector at t0 but in the presence of the micelle, it reaches the detector at tmc, where tmc is
greater than t0. The longer the neutral molecule remains in the micelle, the longer it's migration time. Thus small, non-polar neutral
species that favor interaction with the interior of the micelle would take a longer time to reach the detector than a large, polar
species. Anionic, cationic and zwitter ionic surfactants can be added to change the partition coefficient of the neutral species.
Cationic surfactants would result in positive micelles that would move in the direction of electroosmotic flow. This enables it to
move faster towards the cathode. However, due to the fast migration, it is possible that insufficient time is given for the neutral
species to interact with the micelle resulting in poor separation. Thus, all factors must be considered before choosing the right
surfactant to be used. The mechanism of separation between MEKC and liquid chromatography is the same. Both are dependent on
the partition coefficient of the species between the mobile phase and stationary phase. The main difference lies in the pseudo
stationary phase in MEKC, the micelles. The micelle which can be considered the stationary phase in MEKC moves at a slower
rate than the mobile ions.

Case Study: The Use of CE in Separation of Quantum Dots


Quantum dots (QD) are semiconductor nanocrystals that lie in the size range of 1-10 nm, and they have different
electrophoretic mobility due to their varying sizes and surface charge. CE can be used to separate and characterize such
species, and a method to characterize and separate CdSe QD in the aqueous medium has been developed. The QDs were
synthesized with an outer layer of trioctylphosphine (TOP, Figure 3.6.7a) and trioctylphosphine oxide (TOPO, Figure 3.6.7b),
making the surface of the QD hydrophobic. The background electrolyte solution used was SDS, in order to make the QDs
soluble in water and form a QD-TOPO/TOP-SDS complex. Different sizes of CdSe were used and the separation was with
respect to the charge-to-mass ratio of the complexes. It was concluded from the study that the larger the CdSe core (i.e., the
larger the charge-to-mass ratio) eluted out last. The electropherogram from the study is shown in Figure 3.6.8 from which it is
visible that good separation had taken place by using CE. Laser-induced fluorescence detection was used, the buffer system
was SDS, and the pH of the system set up was fixed at 6.5. The pH is highly important in this case as the stability of the system
and the separation is dependent on it.

Figure 3.6.7 A. The structure of trioctylphosphine (TOP). B. The structure of trioctylphosphine oxide (TOPO).

3.6.6 https://chem.libretexts.org/@go/page/143373
Figure 3.6.8 Electropherogram for a mixture of four different CdSe-TOPO/TOP-SDS complexes. Reproduced from C.
Carrillo-Carrión, Y. Moliner-Martínez, B. M. Simonet, and M. Valcárcel, Anal. Chem., 2011, 83, 2807

3.6: Capillary Electrophoresis is shared under a CC BY 4.0 license and was authored, remixed, and/or curated by Pavan M. V. Raja & Andrew R.
Barron via source content that was edited to conform to the style and standards of the LibreTexts platform; a detailed edit history is available upon
request.

3.6.7 https://chem.libretexts.org/@go/page/143373
CHAPTER OVERVIEW
4: Chemical Speciation
4.1: Magnetism
4.2: IR Spectroscopy
4.3: Raman Spectroscopy
4.4: UV-Visible Spectroscopy
4.5: Photoluminescence, Phosphorescence, and Fluorescence Spectroscopy
4.6: Mössbauer Spectroscopy
4.7: NMR Spectroscopy
4.8: EPR Spectroscopy
4.9: X-ray Photoelectron Spectroscopy
4.10: ESI-QTOF-MS Coupled to HPLC and its Application for Food Safety
4.11: Mass Spectrometry

4: Chemical Speciation is shared under a CC BY 4.0 license and was authored, remixed, and/or curated by Pavan M. V. Raja & Andrew R. Barron
via source content that was edited to conform to the style and standards of the LibreTexts platform; a detailed edit history is available upon
request.

1
4.1: Magnetism
Magnetics
Magnetic Moments
The magnetic moment of a material is the incomplete cancelation of the atomic magnetic moments in that material. Electron spin
and orbital motion both have magnetic moments associated with them (Figure 4.1.1 ) but in most atoms the electronic moments are
oriented usually randomly so that overall in the material they cancel each other out (Figure 4.1.2 ) this is called diamagnetism.

Figure 4.1.1 Orbital Magnetic Moment.

Figure 4.1.2 Magnetic moments in a diamagnetic sample.


If the cancelation of the moments is incomplete then the atom has a net magnetic moment. There are many subclasses of magnetic
ordering such as para-, superpara-, ferro-, antiferro- or ferromagnetism which can be displayed in a material and which usually
depends, upon the strength and type of magnetic interactions and external parameters such as temperature and crystal structure
atomic content and the magnetic environment which a material is placed in.
eh −23
μB   =     =  9.72 × 10 J/T (4.1.1)
4πm

The magnetic moments of atoms, molecules or formula units are often quoted in terms of the Bohr magneton, which is equal to the
magnetic moment due to electron spin
Magnetization
The magnetism of a material, the extent that which a material is magnetic, is not a static quantity, but varies compared to the
environment that a material is placed in. It is similar to the temperature of a material. For example if a material is placed in an oven
it will heat up to a temperature similar to that of the ovens. However the speed of heating of that material, and also that of cooling
are determined by the atomic structure of the material. The magnetization of a material is similar. When a material is placed in a
magnetic field it maybe become magnetized to an extent and retain that magnetization after it is removed from the field. The extent
of magnetization, and type of magnetization and the length of time that a material remains magnetized, depends again on the
atomic makeup of the material.

4.1.1 https://chem.libretexts.org/@go/page/55872
Measuring a materials magnetism can be done on a micro or macro scale. Magnetism is measured over two parameters direction
and strength. Thus magnetization has a vector quantity. The simplest form of a magnetometer is a compass. It measures the
direction of a magnetic field. However more sophisticated instruments have been developed which give a greater insight into a
materials magnetism.
So what exactly are you reading when you observe the output from a magnetometer?
The magnetism of a sample is called the magnetic moment of that sample and will be called that from now on. The single value of
magnetic moment for the sample, is a combination of the magnetic moments on the atoms within the sample ( Figure 4.1.3 ), it is
also the type and level of magnetic ordering and the physical dimensions of the sample itself.

Figure 4.1.3 Schematic representations of the net magnetic moment in a diamagnetic sample.
The "intensity of magnetization", M, is a measure of the magnetization of a body. It is defined as the magnetic moment per unit
volume or
M   =  m/V (4.1.2)

3
with units of Am (emucm in cgs notation).
A material contains many atoms and their arrangement affects the magnetization of that material. In Figure 4.1.4 (a) a magnetic
moment m is contained in unit volume. This has a magnetization of m Am. Figure 4.1.4 (b) shows two such units, with the
moments aligned parallel. The vector sum of moments is 2m in this case, but as the both the moment and volume are doubled M
remains the same. In Figure 4.1.4 (c) the moments are aligned antiparallel. The vector sum of moments is now 0 and hence the
magnetization is 0 Am.

Figure 4.1.4 Effect of moment alignment on magnetization: (a) Single magnetic moment, (b) two identical moments aligned
parallel and (c) antiparallel to each other. Adapted from J. Bland Thesis M. Phys (Hons)., 'A Mossbauer spectroscopy and
magnetometry study of magnetic multilayers and oxides.' Oliver Lodge Labs, Dept. Physics, University of Liverpool
Scenarios (b) and (c) are a simple representation of ferro- and antiferromagnetic ordering. Hence we would expect a large
magnetization in a ferromagnetic material such as pure iron and a small magnetization in an antiferromagnet such as γ-Fe2O3
Magnetic Response
When a material is passed through a magnetic field it is affected in two ways:
1. Through its susceptibility.
2. Through its permeability

4.1.2 https://chem.libretexts.org/@go/page/55872
Magnetic Susceptibility
The concept of magnetic moment is the starting point when discussing the behavior of magnetic materials within a field. If you
place a bar magnet in a field it will experience a torque or moment tending to align its axis in the direction of the field. A compass
needle behaves in the same way. This torque increases with the strength of the poles and their distance apart. So the value of
magnetic moment tells you, in effect, 'how big a magnet' you have.

Figure 4.1.5 Schematic representation of the torque or moment that a magnet experiences when it is placed in a magnetic field. The
magnetic will try to align with the magnetic field.
If you place a material in a weak magnetic field, the magnetic field may not overcome the binding energies that keep the material in
a non magnetic state. This is because it is energetically more favorable for the material to stay exactly the same. However, if the
strength of the magnetic moment is increased, the torque acting on the smaller moments in the material, it may become
energetically more preferable for the material to become magnetic. The reasons that the material becomes magnetic depends on
factors such as crystal structure the temperature of the material and the strength of the field that it is in. However a simple
explanation of this is that as the magnetic moment strength increases it becomes more favorable for the small fields to align
themselves along the path of the magnetic field, instead of being opposed to the system. For this to occur the material must
rearrange its magnetic makeup at the atomic level to lower the energy of the system and restore a balance.
It is important to remember that when we consider the magnetic susceptibility and take into account how a material changes on the
atomic level when it is placed in a magnetic field with a certain moment. The moment that we are measuring with our
magnetometer is the total moment of that sample.
M
χ  =   (4.1.3)
H

where X = susceptibility, M = variation of magnetization, and H = applied field.

Magnetic Permeability
Magnetic permeability is the ability of a material to conduct an electric field. In the same way that materials conduct or resist
electricity, materials also conduct or resist a magnetic flux or the flow of magnetic lines of force (Figure 4.1.6 ).

Figure 4.1.6 Magnetic ordering in a ferromagnetic material.


Ferromagnetic materials are usually highly permeable to magnetic fields. Just as electrical conductivity is defined as the ratio of the
current density to the electric field strength, so the magnetic permeability, μ, of a particular material is defined as the ratio of flux
density to magnetic field strength. However unlike in electrical conductivity magnetic permeability is nonlinear.

4.1.3 https://chem.libretexts.org/@go/page/55872
μ  =  B/H (4.1.4)

Permeability, where μ is written without a subscript, is known as absolute permeability. Instead a variant is used called relative
permeability.

μ  =  μ0 × μr (4.1.5)

Absolute permeability is a variation upon 'straight' or absolute permeability, μ, but is more useful as it makes clearer how the
presence of a particular material affects the relationship between flux density and field strength. The term 'relative' arises because
this permeability is defined in relation to the permeability of a vacuum, μ0.

μr   =  μ/ μ0 (4.1.6)

For example, if you use a material for which μr = 3 then you know that the flux density will be three times as great as it would be if
we just applied the same field strength to a vacuum.

Initial Permeability
Initial permeability describes the relative permeability of a material at low values of B (below 0.1 T). The maximum value for μ in
a material is frequently a factor of between 2 and 5 or more above its initial value.
Low flux has the advantage that every ferrite can be measured at that density without risk of saturation. This consistency means
that comparison between different ferrites is easy. Also, if you measure the inductance with a normal component bridge then you
are doing so with respect to the initial permeability.

Permeability of a Vacuum in the SI


The permeability of a vacuum has a finite value - about 1.257 × 10-6 H m-1 - and is denoted by the symbol μ0. Note that this value
is constant with field strength and temperature. Contrast this with the situation in ferromagnetic materials where μ is strongly
dependent upon both. Also, for practical purposes, most non-ferromagnetic substances (such as wood, plastic, glass, bone, copper
aluminum, air and water) have permeability almost equal to μ0; that is, their relative permeability is 1.0.
The permeability, μ, the variation of magnetic induction, with applied field,
μ  =  B/H (4.1.7)

Background Contributions
A single measurement of a sample's magnetization is relatively easy to obtain, especially with modern technology. Often it is
simply a case of loading the sample into the magnetometer in the correct manner and performing a single measurement. This value
is, however, the sum total of the sample, any substrate or backing and the sample mount. A sample substrate can produce a
substantial contribution to the sample total.
For substrates that are diamagnetic, under zero applied field, this means it has no effect on the measurement of magnetization.
Under applied fields its contribution is linear and temperature independent. The diamagnetic contribution can be calculated from
knowledge of the volume and properties of the substrate and subtracted as a constant linear term to produce the signal from the
sample alone. The diamagnetic background can also be seen clearly at high fields where the sample has reached saturation: the
sample saturates but the linear background from the substrate continues to increase with field. The gradient of this background can
be recorded and subtracted from the readings if the substrate properties are not known accurately.
Hysteresis

When a material exhibits hysteresis, it means that the material responds to a force and has a history of that force contained within
it. Consider if you press on something until it depresses. When you release that pressure, if the material remains depressed and
doesn’t spring back then it is said to exhibit some type of hysteresis. It remembers a history of what happened to it, and may exhibit
that history in some way. Consider a piece of iron that is brought into a magnetic field, it retains some magnetization, even after the
external magnetic field is removed. Once magnetized, the iron will stay magnetized indefinitely. To demagnetize the iron, it is
necessary to apply a magnetic field in the opposite direction. This is the basis of memory in a hard disk drive.

4.1.4 https://chem.libretexts.org/@go/page/55872
The response of a material to an applied field and its magnetic hysteresis is an essential tool of magnetometry. Paramagnetic and
diamagnetic materials can easily be recognized, soft and hard ferromagnetic materials give different types of hysteresis curves and
from these curves values such as saturation magnetization, remnant magnetization and coercivity are readily observed. More
detailed curves can give indications of the type of magnetic interactions within the sample.
Diamagnetism and Paramagnetizm
The intensity of magnetization depends upon both the magnetic moments in the sample and the way that they are oriented with
respect to each other, known as the magnetic ordering.
Diamagnetic materials, which have no atomic magnetic moments, have no magnetization in zero field. When a field is applied a
small, negative moment is induced on the diamagnetic atoms proportional to the applied field strength. As the field is reduced the
induced moment is reduced.

Figure 4.1.7 Typical effect on the magnetization, M, of an applied magnetic field, H, on (a) a paramagnetic system and (b) a
diamagnetic system. Adapted from J. Bland Thesis M. Phys (Hons)., 'A Mossbauer spectroscopy and magnetometry study of
magnetic multilayers and oxides.' Oliver Lodge Labs, Dept. Physics, University of Liverpool
In a paramagnet the atoms have a net magnetic moment but are oriented randomly throughout the sample due to thermal agitation,
giving zero magnetization. As a field is applied the moments tend towards alignment along the field, giving a net magnetization
which increases with applied field as the moments become more ordered. As the field is reduced the moments become disordered
again by their thermal agitation. The figure shows the linear response M v H where μH << kT.
Ferromagnetism

The hysteresis curves for a ferromagnetic material are more complex than those for diamagnets or paramagnets. Below diagram
shows the main features of such a curve for a simple ferromagnet.

Figure 4.1.8 Schematic of a magnetization hysteresis loop in a ferromagnetic material showing the saturation magnetization, Ms,
coercive field, Hc, and remnant magnetization, Mr. Virgin curves are shown dashed for nucleation (1) and pinning (2) type magnets.
Adapted from J. Bland Thesis M. Phys (Hons)., 'A Mossbauer spectroscopy and magnetometry study of magnetic multilayers and
oxides.' Oliver Lodge Labs, Dept. Physics, University of Liverpool
In the virgin material (point 0) there is no magnetization. The process of magnetization, leading from point 0 to saturation at M =
Ms, is outlined below. Although the material is ordered ferromagnetically it consists of a number of ordered domains arranged
randomly giving no net magnetization. This is shown in below (a) with two domains whose individual saturation moments, Ms, lie
antiparallel to each other.

4.1.5 https://chem.libretexts.org/@go/page/55872
Figure 4.1.9 The process of magnetization in a demagnetized ferromagnet. Adaped from J. Bland Thesis M. Phys (Hons)., 'A
Mossbauer spectroscopy and magnetometry study of magnetic multilayers and oxides.' Oliver Lodge Labs, Dept. Physics,
University of Liverpool
As the magnetic field, H, is applied, (b), those domains which are more energetically favorable increase in size at the expense of
those whose moment lies more antiparallel to H. There is now a net magnetization; M. Eventually a field is reached where all of the
material is a single domain with a moment aligned parallel, or close to parallel, with H. The magnetization is now M = MsCosΘ
where Θ is the angle between Ms along the easy magnetic axis and H. Finally Ms is rotated parallel to H and the ferromagnet is
saturated with a magnetization M = Ms.
The process of domain wall motion affects the shape of the virgin curve. There are two qualitatively different modes of behavior
known as nucleation and pinning, shown in Figure 4.1.10 as curves 1 and 2, respectively.
In a nucleation-type magnet saturation is reached quickly at a field much lower than the coercive field. This shows that the domain
walls are easily moved and are not pinned significantly. Once the domain structure has been removed the formation of reversed
domains becomes difficult, giving high coercivity. In a pinning-type magnet fields close to the coercive field are necessary to reach
saturation magnetization. Here the domain walls are substantially pinned and this mechanism also gives high coercivity.
Remnance
As the applied field is reduced to 0 after the sample has reached saturation the sample can still possess a remnant magnetization,
Mr. The magnitude of this remnant magnetization is a product of the saturation magnetization, the number and orientation of easy
axes and the type of anisotropy symmetry. If the axis of anisotropy or magnetic easy axis is perfectly aligned with the field then Mr
= Ms, and if perpendicular Mr= 0.
At saturation the angular distribution of domain magnetizations is closely aligned to H. As the field is removed they turn to the
nearest easy magnetic axis. In a cubic crystal with a positive anisotropy constant, K1, the easy directions are <100>. At remnance
the domain magnetizations will lie along one of the three <100> directions. The maximum deviation from H occurs when H is
along the <111> axis, giving a cone of distribution of 55o around the axis. Averaging the saturation magnetization over this angle
gives a remnant magnetization of 0.832 Ms.
Coercivity
The coercive field, Hc, is the field at which the remnant magnetization is reduced to zero. This can vary from a few Am for soft
magnets to 107Am for hard magnets. It is the point of magnetization reversal in the sample, where the barrier between the two
states of magnetization is reduced to zero by the applied field allowing the system to make a Barkhausen jump to a lower energy. It
is a general indicator of the energy gradients in the sample which oppose large changes of magnetization.
The reversal of magnetization can come about as a rotation of the magnetization in a large volume or through the movement of
domain walls under the pressure of the applied field. In general materials with few or no domains have a high coercivity whilst
those with many domains have a low coercivity. However, domain wall pinning by physical defects such as vacancies, dislocations
and grain boundaries can increase the coercivity.

4.1.6 https://chem.libretexts.org/@go/page/55872
Figure 4.1.10 Shape of hysteresis loop as a function of Θ H, the angle between anisotropy axis and applied field H, for: (a) ΘH, =
0°, (b) 45° and (c) 90°. Adaped from J. Bland Thesis M. Phys (Hons)., 'A Mossbauer spectroscopy and magnetometry study of
magnetic multilayers and oxides.' Oliver Lodge Labs, Dept. Physics, University of Liverpool
The loop illustrated in Figure 4.1.10 is indicative of a simple bi-stable system. There are two energy minima: one with
magnetization in the positive direction, and another in the negative direction. The depth of these minima is influenced by the
material and its geometry and is a further parameter in the strength of the coercive field. Another is the angle, ΘH, between the
anisotropy axis and the applied field. The above fig shows how the shape of the hysteresis loop and the magnitude of Hc varies
with ΘH. This effect shows the importance of how samples with strong anisotropy are mounted in a magnetometer when comparing
loops.
Temperature Dependence
A hysteresis curve gives information about a magnetic system by varying the applied field but important information can also be
gleaned by varying the temperature. As well as indicating transition temperatures, all of the main groups of magnetic ordering have
characteristic temperature/magnetization curves. These are summarized in Figure 4.1.11 and Figure 4.1.12.
At all temperatures a diamagnet displays only any magnetization induced by the applied field and a small, negative susceptibility.
The curve shown for a paramagnet (Figure 4.1.11 ) is for one obeying the Curie law,
c
χ  =   (4.1.8)
t

and so intercepts the axis at T = 0. This is a subset of the Curie-Weiss law,


C
χ  =   (4.1.9)
T −Θ

where θ is a specific temperature for a particular substance (equal to 0 for paramagnets).

Figure 4.1.11 Variation of reciprocal susceptibility with temperature for: (a) antiferromagnetic, (b) paramagnetic and (c)
diamagnetic ordering. Adaped from J. Bland Thesis M. Phys (Hons)., 'A Mossbauer spectroscopy and magnetometry study of
magnetic multilayers and oxides.' Oliver Lodge Labs, Dept. Physics, University of Liverpool

4.1.7 https://chem.libretexts.org/@go/page/55872
Figure 4.1.12 Variation of saturation magnetization below, and reciprocal susceptibility above Tc for: (a) ferromagnetic and (b)
ferrimagnetic ordering. Adaped from J. Bland Thesis M. Phys (Hons)., 'A Mossbauer spectroscopy and magnetometry study of
magnetic multilayers and oxides.' Oliver Lodge Labs, Dept. Physics, University of Liverpool
Above TN and TC both antiferromagnets and ferromagnets behave as paramagnets with 1/χ linearly proportional to temperature.
They can be distinguished by their intercept on the temperature axis, T = Θ. Ferromagnetics have a large, positive Θ, indicative of
their strong interactions. For paramagnetics Θ = 0 and antiferromagnetics have a negative Θ.
The net magnetic moment per atom can be calculated from the gradient of the straight line graph of 1/χ versus temperature for a
paramagnetic ion, rearranging Curie's law to give 4.1.10 .
−−−−
3Ak
μ  = √ (4.1.10)
NX

where A is the atomic mass, k is Boltzmann's constant, N is the number of atoms per unit volume and x is the gradient.
Ferromagnets below TC display spontaneous magnetization. Their susceptibility above TC in the paramagnetic region is given by
the Curie-Weiss law
where g is the gyromagnetic constant. In the ferromagnetic phase with T greater than TC the magnetization M (T) can be simplified
to a power law, for example the magnetization as a function of temperature can be given by 4.1.11 .
β
M (T ) ≈ (TC   −  T ) (4.1.11)

where the term β is typically in the region of 0.33 for magnetic ordering in three dimensions.
The susceptibility of an antiferromagnet increases to a maximum at TN as temperature is reduced, then decreases again below TN. In
the presence of crystal anisotropy in the system this change in susceptibility depends on the orientation of the spin axes: χ
(parallel)decreases with temperature whilst χ (perpendicular) is constant. These can be expressed as 4.1.12 .
C
χ ⊥= (4.1.12)

where C is the Curie constant and Θ is the total change in angle of the two sublattice magnetizations away from the spin axis, and
4.1.13

2 ′ ′
2 ng μ B (J, a ) C
H 0
χ ∥   =  ⊥  = (4.1.13)
2 ′ ′
2kT   +  ng μ γρB (J, a ) 2Θ
H 0

where ng is the number of magnetic atoms per gramme, B’ is the derivative of the Brillouin function with respect to its argument a’,
evaluated at a’0, μH is the magnetic moment per atom and γ is the molecular field coefficient.

Theory of a Superconducting Quantum Interference Device (SQUID)


One of the most sensitive forms of magnetometry is SQUID magnetometry. This uses technique uses a combination of
superconducting materials and Josephson junctions to measure magnetic fields with resolutions up to ~10-14 kG or greater. In the
proceeding pages we will describe how a SQUID actually works.
Electron-pair Waves
In superconductors the resistanceless current is carried by pairs of electrons, known as Cooper Pairs. A Cooper Pair is a pair of
electrons. Each electron has a quantized wavelength. With a Cooper pair each electrons wave couples with its opposite number
over a large distances. This phenomenon is a result of the very low temperatures at which many materials will superconduct.

4.1.8 https://chem.libretexts.org/@go/page/55872
What exactly is superconductance? When a material is at very low temperatures, its crystal lattice behaves differently than when it
higher temperatures. Usually at higher temperatures a material will have large vibrations called in the crystal lattice. These
vibrations scatter electrons as they pass through this lattice (Figure 4.1.13 ), and this is the basis for bad conductance.

Figure 4.1.13 Schematic representation of the scattering of electrons as they pass through a vibrating lattice.
With a superconductor the material is designed to have very small vibrations, these vibrations are lessened even more by cooling
the material to extremely low temperatures. With no vibrations there is no scattering of the electrons and this allows the material to
superconduct.
The origin of a Cooper pair is that as the electron passes through a crystal lattice at superconducting temperatures it negative charge
pulls on the positive charge of the nuclei in the lattice through coulombic interactions producing a ripple. An electron traveling in
the opposite direction is attracted by this ripple. This is the origin of the coupling in a Cooper pair (Figure 4.1.14 ).

Figure 4.1.14 Schematic representation of the Cooper pair coupling model.


A passing electron attracts the lattice, causing a slight ripple toward its path. Another electron passing in the opposite direction is
attracted to that displacement (Figure 4.1.15 ).

Figure 4.1.15 Schematic representation of Cooper pair coupling

Figure 4.1.16 Schematic representation of the condensation of the wavelengths of a Cooper pairs
Each pair can be treated as a single particle with a whole spin, not half a spin such as is usually the case with electrons. This is
important, as an electron which is classed in a group of matter called Fermions are governed by the Fermi exclusion principle
which states that anything with a spin of one half cannot occupy the same space as something with the same spin of one half. This
turns the electron means that a Cooper pair is in fact a Boson the opposite of a Fermion and this allows the Coopers pairs to
condensate into one wave packet. Each Coopers pair has a mass and charge twice that of a single electron, whose velocity is that of
the center of mass of the pair. This coupling can only happen in extremely cold conditions as thermal vibrations become greater
than the force that an electron can exert on a lattice. And thus scattering occurs.

4.1.9 https://chem.libretexts.org/@go/page/55872
Each pair can be represented by a wavefunction of the form
where P is the net momentum of the pair whose center of mass is at r. However, all the Cooper pairs in a superconductor can be
described by a single wavefunction yet again due to the fact that the electrons are in a Coopers pair state and are thus Bosons in the
absence of a current because all the pairs have the same phase - they are said to be "phase coherent"
This electron-pair wave retains its phase coherence over long distances, and essentially produces a standing wave over the device
circuit. In a SQUID there are two paths which form a circle and are made with the same standing wave (Figure 4.1.17 ). The wave
is split in two sent off along different paths, and then recombined to record an interference pattern by adding the difference between
the two.

Figure 4.1.17 Schematic representation of a standing wave across a SQUID circuit.


This allows measurement at any phase differences between the two components, which if there is no interference will be exactly
the same, but if there is a difference in their path lengths or in some interaction that the waves encounters such as a magnetic field it
will correspond in a phase difference at the end of each path length.
A good example to use is of two water waves emanating from the same point. They will stay in phase if they travel the same
distance, but will fall out of phase if one of them has to deviate around an obstruction such as a rock. Measuring the phase
difference between the two waves then provides information about the obstruction.
Phase and Coherence
Another implication of this long range coherence is the ability to calculate phase and amplitude at any point on the wave's path
from the knowledge of its phase and amplitude at any single point, combined with its wavelength and frequency. The wavefunction
of the electron-pair wave in the above eqn. can be rewritten in the form of a one-dimensional wave as
χ
ψp   =  ψ sin(2π)(   − ν t) (4.1.14)
λ

If we take the wave frequency, V, as being related to the kinetic energy of the Cooper pair with a wavelength, λ, being related to the
momentum of the pair by the relation λ = h/p then it is possible to evaluate the phase difference between two points in a current
carrying superconductor.
If a resistanceless current flows between points X and Y on a superconductor there will be a phase difference between these points
that is constant in time.
Effect of a Magnetic Field
The parameters of a standing wave are dependent on a current passing through the circuit; they are also strongly affected by an
applied magnetic field. In the presence of a magnetic field the momentum, p, of a particle with charge q in the presence of a
magnetic field becomes mV + qA where A is the magnetic vector potential. For electron-pairs in an applied field their moment P is
now equal to 2mV+2eA.
In an applied magnetic field the phase difference between points X and Y is now a combination of that due to the supercurrent and
that due to the applied field.
The Fluxoid
One effect of the long range phase coherence is the quantization of magnetic flux in a superconducting ring. This can either be a
ring, or a superconductor surrounding a non-superconducting region. Such an arrangement can be seen in Figure 4.1.18 where
region N has a flux density B within it due to supercurrents flowing around it in the superconducting region S.

4.1.10 https://chem.libretexts.org/@go/page/55872
Figure 4.1.18 Superconductor enclosing a non-superconducting region. Adaped from J. Bland Thesis M. Phys (Hons)., 'A
Mossbauer spectroscopy and magnetometry study of magnetic multilayers and oxides.' Oliver Lodge Labs, Dept. Physics,
University of Liverpool.

In the closed path XYZ encircling the non-superconducting region there will be a phase difference of the electron-pair wave
between any two points, such as X and Y, on the curve due to the field and the circulating current.
If the superelectrons are represented by a single wave then at any point on XYZX it can only have one value of phase and
amplitude. Due to the long range coherence the phase is single valued also called quantized meaning around the circumference of
the ring Δφ must equal 2πn where n is any integer. Due to the wave only having a single value the fluxoid can only exist in
quantized units. This quantum is termed the fluxon, φ0, given by 4.1.15 .
h −15
Φ0   =     =  2.07x 10 Wb (4.1.15)
2e

Josephson Tunneling
If two superconducting regions are kept totally isolated from each other the phases of the electron-pairs in the two regions will be
unrelated. If the two regions are brought together then as they come close electron-pairs will be able to tunnel across the gap and
the two electron-pair waves will become coupled. As the separation decreases, the strength of the coupling increases. The tunneling
of the electron-pairs across the gap carries with it a superconducting current as predicted by B.D. Josephson and is called
"Josephson tunneling" with the junction between the two superconductors called a "Josephson junction" (Figure 4.1.16 ).

Figure 4.1.19 Schematic representation of the tunneling of Cooper pairs across a Josephson junction.
The Josephson tunneling junction is a special case of a more general type of weak link between two superconductors. Other forms
include constrictions and point contacts but the general form is of a region between two superconductors which has a much lower
critical current and through which a magnetic field can penetrate.
Superconducting Quantum Interference Device (SQUID)

A superconducting quantum interference device (SQUID) uses the properties of electron-pair wave coherence and Josephson
Junctions to detect very small magnetic fields. The central element of a SQUID is a ring of superconducting material with one or
more weak links called Josephesons Junctions. An example is shown in the below. With weak-links at points W and X whose
critical current, ic, is much less than the critical current of the main ring. This produces a very low current density making the
momentum of the electron-pairs small. The wavelength of the electron-pairs is thus very long leading to little difference in phase
between any parts of the ring.

4.1.11 https://chem.libretexts.org/@go/page/55872
Figure 4.1.20 Superconducting quantum interference device (SQUID) as a simple magnetometer. Adaped from J. Bland Thesis M.
Phys (Hons)., 'A Mossbauer spectroscopy and magnetometry study of magnetic multilayers and oxides.' Oliver Lodge Labs, Dept.
Physics, University of Liverpool.
If a magnetic field, Ba , is applied perpendicular to the plane of the ring (Figure 4.1.21, a phase difference is produced in the
electron-pair wave along the path XYW and WZX. One of the features of a superconducting loop is that the magnetic flux, Φ,
passing through it which is the product of the magnetic field and the area of the loop and is quantized in units of Φ0 = h/ (2e),
where h is Planck’s constant, 2e is the charge of the Cooper pair of electrons, and Φ0 has a value of 2 × 10–15 tesla m2. If there are
no obstacles in the loop, then the superconducting current will compensate for the presence of an arbitrary magnetic field so that
the total flux through the loop (due to the external field plus the field generated by the current) is a multiple of Φ0.

Figure 4.1.21 Schematic representation of a SQUID placed in a magnetic field.


Josephson predicted that a superconducting current can be sustained in the loop, even if its path is interrupted by an insulating
barrier or a normal metal. The SQUID has two such barriers or ‘Josephson junctions’. Both junctions introduce the same phase
difference when the magnetic flux through the loop is 0, Φ0, 2Φ0 and so on, which results in constructive interference, and they
introduce opposite phase difference when the flux is Φ0/2, 3Φ0/2 and so on, which leads to destructive interference. This
interference causes the critical current density, which is the maximum current that the device can carry without dissipation, to vary.
The critical current is so sensitive to the magnetic flux through the superconducting loop that even tiny magnetic moments can be
measured. The critical current is usually obtained by measuring the voltage drop across the junction as a function of the total
current through the device. Commercial SQUIDs transform the modulation in the critical current to a voltage modulation, which is
much easier to measure.
An applied magnetic field produces a phase change around a ring, which in this case is equal
Φa
ΔΦ(B)  =  2π (4.1.16)
Φ0

where Φa is the flux produced in the ring by the applied magnetic field. The magnitude of the critical measuring current is
dependent upon the critical current of the weak-links and the limit of the phase change around the ring being an integral multiple of
2π. For the whole ring to be superconducting the following condition must be met

4.1.12 https://chem.libretexts.org/@go/page/55872
Φa
α  +  β  +  2π (4.1.17)
Φ0

where α and β are the phase changes produced by currents across the weak-links and 2πΦa/Φo is the phase change due to the
applied magnetic field.
When the measuring current is applied α and β are no longer equal, although their sum must remain constant. The phase changes
can be written as 4.1.18
Φa Φa
α = π[n − ]  −  δβ  =  π [n  − ] +δ (4.1.18)
Φ0 Φ0

where δ is related to the measuring current I. Using the relation between current and phase from the above Eqn. and rearranging to
eliminate i we obtain an expression for I, 4.1.19
Φa
Ic   =  2 ic |cosπ , sinδ| (4.1.19)
Φ0

As sinδ cannot be greater than unity we can obtain the critical measuring current, Ic from the above 4.1.20
Φa
Ic   =  2 ic |cosπ | (4.1.20)
Φ0

which gives a periodic dependence on the magnitude of the magnetic field, with a maximum when this field is an integer number of
fluxons and a minimum at half integer values as shown in the below figure.

Figure 4.1.22 Critical measuring current, Ic, as a function of applied magnetic field. Adaped from J. Bland Thesis M. Phys (Hons).,
'A Mossbauer spectroscopy and magnetometry study of magnetic multilayers and oxides.' Oliver Lodge Labs, Dept. Physics,
University of Liverpool.

Practical Guide to Using a Superconducting Quantum Interference Device


SQUIDs offer the ability to measure at sensitivities unachievable by other magnetic sensing methodologies. However, their
sensitivity requires proper attention to cryogenics and environmental noise. SQUIDs should only be used when no other sensor is
adequate for the task. There are many exotic uses for SQUID however we are just concerned with the laboratory applications of
SQUID.
In most physical and chemical laboratories a device called a MPMS (Figure 4.1.23 )
is used to measure the magnetic moment of a sample by reading the output of the SQUID detector. In a MPMS the sample moves
upward through the electronic pick up coils called gradiometers. One upward movement is one whole scan. Multiple scans are used
and added together to improve measurement resolution. After collecting the raw voltages, there is computation of the magnetic
moments of the sample.
The MPMS measures the moment of a sample by moving it through a liquid Helium cooled, superconducting sensing coil. Many
different measurements can be carried out using an MPMS however we will discuss just a few.

4.1.13 https://chem.libretexts.org/@go/page/55872
Figure 4.1.23 A MPMS work station.
Using an Magnetic Property Measurement System (MPMS)

DC Magentization
DC magnetization is the magnetic per unit volume (M) of a sample. If the sample doesn’t have a permanent magnetic moment, a
field is applied to induce one. The sample is then stepped through a superconducting detection array and the SQUID’s output
voltage is processed and the sample moment computed. Systems can be configured to measure hysteresis loops, relaxation times,
magnetic field, and temperature dependence of the magnetic moment.
A DC field can be used to magnetize samples. Typically, the field is fixed and the sample is moved into the detection coil’s region
of sensitivity. The change in detected magnetization is directly proportional to the magnetic moment of the sample. Commonly
referred to as SQUID magnetometers, these systems are properly called SQUID susceptometers (Figure 4.1.24 ).
They have a homogeneous superconducting magnet to create a very uniform field over the entire sample measuring region and the
superconducting pickup loops. The magnet induces a moment allowing a measurement of magnetic susceptibility. The
superconducting detection loop array is rigidly mounted in the center of the magnet. This array is configured as a gradient coil to
reject external noise sources. The detection coil geometry determines what mathematical algorithm is used to calculate the net
magnetization.
An important feature of SQUIDs is that the induced current is independent of the rate of flux change. This provides uniform
response at all frequencies i.e., true dc response and allows the sample to be moved slowly without degrading performance. As the
sample passes through a coil, it changes the flux in that coil by an amount proportional to the magnetic moment M of the sample.
The peak-to-peak signal from a complete cycle is thus proportional to twice M. The SQUID sensor shielded inside a niobium can is
located where the fringe fields generated by the magnet are less than 10 mT. The detection coil circuitry is typically constructed
using NbTi (Figure 4.1.25 ). This allows measurements in applied fields of 9 T while maintaining sensitivities of 10−8 emu.
Thermal insulation not shown is placed between the detection coils and the sample tube to allow the sample temperature to be
varied.

4.1.14 https://chem.libretexts.org/@go/page/55872
Figure 4.1.24 Schematic diagram of a MPMSR. Adapted from L. Fagaly, Review of Scientific Instruments, 2006, 77, 101101.
The use of a variable temperature insert can allow measurements to be made over a wide range 1.8–400 K. Typically, the sample
temperature is controlled by helium gas flowing slowly past the sample. The temperature of this gas is regulated using a heater
located below the sample measuring region and a thermometer located above the sample region. This arrangement ensures that the
entire region has reached thermal equilibrium prior to data acquisition. The helium gas is obtained from normal evaporation in the
Dewar, and its flow rate is controlled by a precision regulating valve.

Figure 4.1.25 Signal output of an MPMS. Adapted from L. Fagaly, Review of Scientific Instruments, 2006, 77, 101101.

Procedures when using an MPMS

Calibration
The magnetic moment calibration for the SQUID is determined by measuring a palladium standard over a range of magnetic fields
and then by adjusting to obtain the correct moment for the standard. The palladium standard samples are effectively point sources
with an accuracy of approximately 0.1%.

Sample mounting considerations


The type, size and geometry of a sample is usually sufficient to determine the method you use to attach it to the sample. However
mostly for MPMS measurements a plastic straw is used. This is due to the straw having minimal magnetic susceptibility.
However there are a few important considerations for the sample holder design when mounting a sample for measurement in a
magnetometer. The sample holder can be a major contributor to the background signal. Its contribution can be minimized by

4.1.15 https://chem.libretexts.org/@go/page/55872
choosing materials with low magnetic susceptibility and by keeping the mass to a minimum such as a plastic straw mentioned
above.
The materials used to hold a sample must perform well over the temperature range to be used. In a MPMS, the geometric
arrangement of the background and sample is critical when their magnetic susceptibilities will be of similar magnitude. Thus, the
sample holder should optimize the sample’s positioning in the magnetometer. A sample should be mounted rigidly in order to avoid
excess sample motion during measurement. A sample holder should also allow easy access for mounting the sample, and its
background contribution should be easy to measure. This advisory introduces some mounting methods and discusses some of the
more important considerations when mounting samples for the MPMS magnetometer. Keep in mind that these are only
recommendations, not guaranteed procedures. The researcher is responsible for assuring that the methods and materials used will
meet experimental requirements.
Sample Mounts

Platform Mounting
For many types of samples, mounting to a platform is the most convenient method. The platform’s mass and susceptibility should
be as small as possible in order to minimize its background contribution and signal distortion.

Plastic Disc
A plastic disc about 2 mm thick with an outside diameter equivalent to the pliable plastic tube’s diameter (a clear drinking straw is
suitable) is inserted and twisted into place. The platform should be fairly rigid. Mount samples onto this platform with glue. Place a
second disc, with a diameter slightly less than the inside diameter of the tube and with the same mass, on top of the sample to help
provide the desired symmetry. Pour powdered samples onto the platform and place a second disc on top. The powders will be able
to align with the field. Make sure the sample tube is capped and ventilated.

Crossed Threads
Make one of the lowest mass sample platforms by threading a cross of white cotton thread (colored dyes can be magnetic). Using a
needle made of a nonmagneticmetal, or at least carefully cleaned, thread some white cotton sewingthread through the tube walls
and tie a secure knot so that the thread platform isrigid. Glue a sample to this platform or use the platform as asupport for a sample
in a container. Use an additional thread cross on top to holdthe container in place.

Gelatin Capsule
Gelatin capsules can be very useful for containing and mounting samples. Many aspects of using gelatin capsules have been
mentioned in the section, Containing the Sample. It is best if the sample is mounted near the capsule’s center, or if it completely
fills the capsule. Use extra capsule parts to produce mirror symmetry. The thread cross is an excellent way of holding a capsule in
place.

Thread Mounting
Another method of sample mounting is attaching the sample to a thread that runs through the sample tube. The thread can be
attached to the sample holder at the ends of the sample tube with tape, for example. This method can be very useful with flat
samples, such as those on substrates, particularly when the field is in the plane of the film. Be sure to close the sample tube with
caps.
Mounting with a disc platform.
Mounting on crossed threads.
Long thread mounting.
Steps for Inserting the Sample

4.1.16 https://chem.libretexts.org/@go/page/55872
1. Cut off a small section of a clear plastic drinking straw. The section must be small enough to fit inside the straw.
2. Weigh and measure the sample.
3. Use plastic tweezers to place the sample inside the small straw segment. It is important to use plastic tweezers not metallic ones
as these will contaminate the sample.
4. Place the small straw segment inside the larger one. It should be approximately in the middle of the large drinking straw.
5. Attach the straw to the sample rod which is used to insert the sample into the SQUID machine.
6. Insert the sample rod with the attached straw into the vertical insertion hole on top of the SQUID.

Center the Sample


The sample must be centered in the SQUID pickup coils to ensure that all coils sense the magnetic moment of the sample. If the
sample is not centered, the coils read only part of the magnetic moment.
During a centering measurement the MPMS scans the entire length of the samples vertical travel path, and the MPMS reads the
maximum number of data points. During centering there are a number of terms which need to be understood.
1. A scan length is the length of a scan of a particular sample which should usually try and be the maximum of the sample.
2. A sample is centered when it is in the middle of a scan length. The data points are individual voltage readings plotting response
curves in centering scan data files.
3. Autotracking is the adjustment of a sample position to keep a sample centered in SQUID coils. Autotracking compensates for
thermal expansion and contraction in a sample rod.
As soon as a centering measurement is initiated, the sample transport moves upward, carrying the sample through the pickup coils.
While the sample moves through the coils, the MPMS measures the SQUID’s response to the magnetic moment of the sample and
saves all the data from the centering measurement.
After a centering plot is performed the plot is examined to determine whether the sample is centered in the SQUID pickup coils.
The sample is centered when the part of the large, middle curve is within 5cm of the half-way point of the scan length.
The shape of the plot is a function of the geometry of the coils. The coils are wound in a way which strongly rejects interference
from nearby magnetic sources and lets the MPMS function without a superconducting shield around the pickup coils.

Geometric Considerations
To minimize background noise and stray field effects, the MPMS magnetometer pick-up coil takes the form of a second-order
gradiometer. An important feature of this gradiometer is that moving a long, homogeneous sample through it produces no signal as
long as the sample extends well beyond the ends of the coil during measurement.
As a sample holder is moved through the gradiometer pickup coil, changes in thickness, mass, density, or magnetic susceptibility
produce a signal. Ideally, only the sample to be measured produces this change. A homogeneous sample that extends well beyond
the pick-up coils does not produce a signal, yet a small sample does produce a signal. There must be a crossover between these two
limits. The sample length (along the field direction) should not exceed 10 mm. In order to obtain the most accurate measurements,
it is important to keep the sample susceptibility constant over its length; otherwise distortions in the SQUID signal (deviations from
a dipole signal) can result. It is also important to keep the sample close to the magnetometer centerline to get the most accurate
measurements. When the sample holder background contribution is similar in magnitude to the sample signal, the relative positions
of the sample and the materials producing the background are important. If there is a spatial offset between the two along the
magnet axis, the signal produced by the combined sample and background can be highly distorted and will not be characteristic of
the dipole moment being measured.
Even if the signal looks good at one temperature, a problem can occur if either of the contributions are temperature dependent.
Careful sample positioning and a sample holder with a center, or plane, of symmetry at the sample (i.e. materials distributed
symmetrically about the sample, or along the principal axis for a symmetry plane) helps eliminate problems associated with spatial
offsets.

Containing the Sample

4.1.17 https://chem.libretexts.org/@go/page/55872
Keep the sample space of the MPMS magnetometer clean and free of contamination with foreign materials. Avoid accidental
sample loss into the sample space by properly containing the sample in an appropriate sample holder. In all cases it is important to
close the sample holder tube with caps in order to contain a sample that might become unmounted. This helps avoid sample loss
and subsequent damage during the otherwise unnecessary recovery procedure. Position caps well out of the sample-measuring
region and introduce proper venting.

Sample Preparation Workspace


Work area cleanliness and avoiding sample contamination are very important concerns. There are many possible sources of
contamination in a laboratory. Use diamond tools when cutting hard materials. Avoid carbide tools because of potential
contamination by the cobalt binder found in many carbide materials. The best tools for preparing samples and sample holders are
made of plastic, titanium, brass, and beryllium copper (which also has a small amount of cobalt). Tools labeled non-magnetic can
actually be made of steel and often be made "magnetic" from exposure to magnetic fields. However, the main concern from these
"non-magnetic" tools is contamination by the iron and other ferrous metals in the tool. It is important to have a clean white-papered
workspace and a set of tools dedicated to mounting your own samples. In many cases, the materials and tools used can be washed
in dilute acid to remove ferrous metal impurities. Follow any acid washes with careful rinsing with deionized water.
Powdered samples pose a special contamination threat, and special precautions must be taken to contain them. If the sample is
highly magnetic, it is often advantageous to embed it in a low susceptibility epoxy matrix like Duco cement. This is usually done
by mixing a small amount of diluted glue with the powder in a suitable container such as a gelatin capsule. Potting the sample in
this way can keep the sample from shifting or aligning with the magnetic field. In the case of weaker magnetic samples, measure
the mass of the glue after drying and making a background measurement. If the powdered sample is not potted, seal it into a
container, and watch it carefully as it is cycled in the airlock chamber.

Pressure Equalization
The sample space of the MPMS has a helium atmosphere maintained at low pressure of a few torr. An airlock chamber is provided
to avoid contamination of the sample space with air when introducing samples into the sample space. By pushing the purge button,
the airlock is cycled between vacuum and helium gas three times, then pumped down to its working pressure. During the cycling, it
is possible for samples to be displaced in their holders, sealed capsules to explode, and sample holders to be deformed. Many of
these problems can be avoided if the sample holder is properly ventilated. This requires placing holes in the sample holder, out of
the measuring region that will allow any closed spaces to be opened to the interlock chamber.

Air-sensitive Samples and Liquid Samples


When working with highly air-sensitive samples or liquid samples it is best to first seal the sample into a glass tube. NMR and EPR
tubes make good sample holders since they are usually made of a high-quality, low-susceptibility glass or fused silica. When the
sample has a high susceptibility, the tube with the sample can be placed onto a platform like those described earlier. When dealing
with a low susceptibility sample, it is useful to rest the bottom of the sample tube on a length of the same type of glass tubing. By
producing near mirror symmetry, this method gives a nearly constant background with position and provides an easy method for
background measurement (i.e., measure the empty tube first, then measure with a sample). Be sure that the tube ends are well out
of the measuring region.
When going to low temperatures, check to make sure that the sample tube will not break due to differential thermal expansion.
Samples that will go above room temperature should be sealed with a reduced pressure in the tube and be checked by taking the
sample to the maximum experimental temperature prior to loading it into the magnetometer. These checks are especially important
when the sample may be corrosive, reactive, or valuable.

Oxygen Contamination
This application note describes potential sources for oxygen contamination in the sample chamber and discusses its possible
effects. Molecular oxygen, which undergoes an antiferromagnetic transition at about 43 K, is strongly paramagnetic above this

4.1.18 https://chem.libretexts.org/@go/page/55872
temperature. The MPMS system can easily detect the presence of a small amount of condensed oxygen on the sample, which when
in the sample chamber can interfere significantly with sensitive magnetic measurements. Oxygen contamination in the sample
chamber is usually the result of leaks in the system due to faulty seals, improper operation of the airlock valve, outgassing from the
sample, or cold samples being loaded.

4.1: Magnetism is shared under a CC BY 4.0 license and was authored, remixed, and/or curated by Pavan M. V. Raja & Andrew R. Barron via
source content that was edited to conform to the style and standards of the LibreTexts platform; a detailed edit history is available upon request.

4.1.19 https://chem.libretexts.org/@go/page/55872
4.2: IR Spectroscopy
IR Sample Preparation: A Practical Guide
Infrared spectroscopy is based on molecular vibrations caused by the oscillation of molecular dipoles. Bonds have characteristic
vibrations depending on the atoms in the bond, the number of bonds and the orientation of those bonds with respect to the rest of
the molecule. Thus, different molecules have specific spectra that can be collected for use in distinguishing products or identifying
an unknown substance (to an extent.)
Collecting spectra through this method goes about one of three general ways. Nujol mulls and pressed pellets are typically used for
collecting spectra of solids, while thin-film cells are used for solution-phase IR spectroscopy. Spectra of gases can also be obtained
but will not be discussed in this guide.

Infrared Optical Materials and Handling


While it is all well and wonderful that substances can be characterized in this fashion one still has to be able to hold the substances
inside of the instrument and properly prepare the samples. In an infrared spectrometer (Figure 4.2.1 )
the sample to be analyzed is held in front of an infrared laser beam, in order to do this, the sample must be contained in something,
consequently this means that the very container the sample is in will absorb some of the infrared beam.

Figure 4.2.1 An example of a modern benchtop FT-IR spectrometer (Varian Corp.)


This is made somewhat complicated by the fact that all materials have some sort of vibration associated with them. Thus, if the
sample holder has an optical window made of something that absorbs near where your sample does, the sample might not be
distinguishable from the optical window of the sample holder. The range that is not blocked by a strong absorbance is known as a
window (not to be confused with the optical materials of the cell.)
Windows are an important factor to consider when choosing the method to perform an analysis, as seen in Table 4.2.1 there are a
number of different materials each with their own characteristic absorption spectra and chemical properties. Keep these factors in
mind when performing analyses and precious sample will be saved. For most organic compounds NaCl works well though it is
susceptible to attack from moisture. For metal coordination complexes KBr, or CsI typically work well due to their large windows.
If money is not a problem then diamond or sapphire can be used for plates.
Table 4.2.1 Various IR-transparent materials and their solubilities and other notes. M. R. Derrick, D. Stulik, and J. M. Landry, in Scientific
Tools in Conservation: Infrared Spectroscopy in Conservation Science. Getty Conservation Institute (1999).
Material Transparent Ranges (cm -1) Solubility Notes

NaCl 40,000 - 625 H2O Easy to polish, hygroscopic

Silica glass 55,000-3,000 HF Attacked by HF

Quartz 40,000-2,500 HF Attacked by HF

Sapphire 20,000-1,780 - Strong

Very strong, expensive, hard,


Diamond 40,000-2,500 and 1,800-200 -
useless for pellets
Attacked by acids, avoid
CaF2 70,000-1,110 Acids
ammonium salts

BaF2 65,000-700 - Avoid ammonium salts

ZnSe 10,000 - 550 Acids Brittle, attacked by acids

4.2.1 https://chem.libretexts.org/@go/page/55873
Material Transparent Ranges (cm -1) Solubility Notes

AgCl 25,000-400 - Soft, sensitive to light.

Hygroscopic, soft, easily polished,


KCl 40,000-500 H2O, Et2O, acetone
commonly used in making pellets.
Hygroscopic, soft, easily polished,
KBr 40,000-400 H2O, EtOH
commonly used in making pellets.

CsBr 10,000-250 H2O, EtOH, acetone Hygroscopic soft

CsI 10,000-200 H2O, EtOH, MeOH, acetone Hygroscopic, soft.

Teflon 5,000-1,200; 1,200-900 - Inert, disposable

4,000-3,000; 2,800-1,460; 1,380 -


Polyethylene - Inert, disposable
730; 720- 30

Proper handling of these plates will ensure they have a long, useful life. Here follows a few simple pointers on how to handle
plates:
Avoid contact with solvents that the plates are soluble in.
Keep the plates in a dessicator, the less water the better, even if the plates are insoluble to water.
Handle with gloves, clean gloves.
Avoid wiping the plates to prevent scratching.
That said, these simple guidelines will likely reduce most damage that can occur to a plate by simply holding it other faults such as
dropping the plate from a sufficient height can result in more serious damage.

Preparation of Nujol Mulls


A common method of preparing solid samples for IR analysis is mulling. The principle here is by grinding the particles to below
the wavelength of incident radiation that will be passing through there should be limited scattering. To suspend those tiny particles,
an oil, often referred to as Nujol is used. IR-transparent salt plates are used to hold the sample in front of the beam in order to
acquire data. To prepare a sample for IR analysis using a salt plate, first decide what segment of the frequency band should be
studied, refer to Table 4.2.1 for the materials best suited for the sample. Figure 4.2.2 shows the materials needed for preparing a
mull.

Figure 4.2.2 In this photograph, the sample, ferrocene, two clean and polished KBr plates, an agate mortar and pestle, a mounting
card and a spatula are displayed as the base minimum requirements for preparing a sample though a Nujol mull. Of course, a small
bottle of mineral oil is also necessary.
Preparing the mull is performed by taking a small portion of sample and adding approximately 10% of the sample volume worth of
the oil and grinding this in an agate mortar and pestle as demonstrated in Figure 4.2.3. The resulting mull should be transparent
with no visible particles.

4.2.2 https://chem.libretexts.org/@go/page/55873
Figure 4.2.3 Mulling ferrocene into mineral oil with a mortar and pestle.
Another method involves dissolving the solid in a solvent and allowing it to dry in the agate pestle. If using this method ensure that
all of the solvent has evaporated since the solvent bands will appear in the spectrum. Some gentle heating may assist this process.
This method creates very fine particles that are of a relatively consistent size. After addition of the oil further mixing (or grinding)
may be necessary.
Plates should be stored in a desiccator to prevent erosion by atmospheric moisture and should appear roughly transparent. Some
materials such as silicon will not, however. Gently rinse the plates with hexanes to wash any residual material off of the plates.
Removing the plates from the desiccator and cleaning them should follow the preparation of the mull in order to maintain the
integrity of the salt plates. Of course, if the plate is not soluble in water then it is still a good idea just to prevent the threat of
mechanical trauma or a stray jet of acetone from a wash bottle.
Once the mull has been prepared, add a drop to one IR plate (Figure 4.2.4 ), place the second plate on top of the drop and give it a
quarter turn in order to evenly coat the plate surface as seen in Figure 4.2.5. Place it into the spectrometer and acquire the desired
data.
Always handle with gloves and preferably away from any sinks, faucets, or other sources of running or spraying water.

Figure 4.2.4 The prepared mull from an agate mortar and pestle being applied to a polished KBr plate.

4.2.3 https://chem.libretexts.org/@go/page/55873
Figure 4.2.5 Sandwiched KBr plates with a Nujol mull of ferrocene.
Spectra acquired by this method will have strong C-H absorption bands throughout several ranges 3,000 – 2,800 cm-1 and 1,500 –
1,300 cm-1 and may obscure signal.
Cleaning the plate is performed as previously mentioned with hexanes or chloroform can easily be performed by rinsing and
leaving them to dry in the hood. Place the salt plates back into the desiccator as soon as reasonably possible to prevent damage. It is
highly advisable to polish the plates after use, no scratches, fogging, or pits should be visible on the face of the plate. Chips, so long
as they don’t cross the center of the plate are survivable but not desired. The samples of damaged salt plates in Figure 4.2.6 show
common problems associated with use or potentially mishandling. Clouding, and to an extent, scratches can be polished out with an
iron rouge. Areas where the crystal lattice is disturbed below the surface are impossible to fix and chips cannot be reattached.

4.2.4 https://chem.libretexts.org/@go/page/55873
FIgure 4.2.6 A series of plates indicating various forms of physical damage with a comparison to a good plate (Copyright:
Colorado University-Boulder).

Preparation of Pellets
In an alternate method, this technique is along the same lines of the nujol mull except instead of the suspending medium being
mineral oil, the suspending medium is a salt. The solid is ground into a fine powder with an agate mortar and pestle with an amount
of the suspending salt. Preparing pellets with diamond for the suspending agent is somewhat illadvised considering the great
hardness of the substance. Generally speaking, an amount of KBr or CsI is used for this method since they are both soft salts. Two
approaches can be used to prepare pellets, one is somewhat more expensive but both usually yield decent results.
The first method is the use of a press. The salt is placed into a cylindrical holder and pressed together with a ram such as the one
seen in (Figure 4.2.7 ). Afterwards, the pellet, in the holder, is placed into the instrument and spectra acquired.

Figure 4.2.7 A large benchtop hydraulic press (Specac Inc.)


An alternate, and cheaper method requires the use of a large hex nut with a 0.5 inch inner diameter, two bolts, and two wrenches
such as the kit seen in Figure 4.2.8. Step-by-step instructions for loading and using the press follows:
1. Screw one of the bolts into the nut about half way.
2. Place the salt pellet mixture into the other opening of the nut and level by tapping the assembly on a countertop.
3. Screw in the second bolt and place the assembly on its side with the bolts parallel to the countertop. Place one of the wrenches
on the bolt on the right side with the handle aiming towards yourself.
4. Take the second wrench and place it on the other bolt so that it attaches with an angle from the table of about 45 degrees.
5. The second bolt is tightened with a body weight and left to rest for several minutes. Afterwards, the bolts are removed, and the
sample placed into the instrument.

4.2.5 https://chem.libretexts.org/@go/page/55873
Figure 4.2.8 A simple pellet press with cell holder. (Cole-Parmer)
Some pellet presses also have a vacuum barb such as the one seen in (Figure 4.2.8. If your pellet press has one of these, consider
using it as it will help remove air from the salt pellet as it is pressed. This ensures a more uniform pellet and removes absorbances
in the collected spectrum due to air trapped in the pellet.

Preparation of Solution Cells


Solution cells (Figure 4.2.9 ) are a handy way of acquiring infrared spectra of compounds in solution and is particularly handy for
monitoring reactions.

Figure 4.2.9 A sealed solution cell with two injection ports and a schematic of its construction (Perkin-Elmer Inc.)
A thin-film cell consists of two salt plates with a very thin space in between them (Figure 4.2.10 ). Two channels allow liquid to be
injected and then subsequently removed. The windows on these cells can be made from a variety of IR optical materials. One
particularly useful one for water-based solutions is CaF2 as it is not soluble in water.

Figure 4.2.10 A sealed solution cell with two injection ports and a schematic of its construction (Perkin-Elmer Inc.).
Cleaning these cells can be performed by removing the solution, flushing with fresh solvent and gently removing the solvent by
syringe. Do not blow air or nitrogen through the ports as this can cause mechanical deformation in the salt window if the pressure is
high enough.

Deuterated Solvent Effects


One of the other aspects to solution-phase IR is that the solvent utilized in the cell has a characteristic absorption spectra. In some
cases this can be alleviated by replacing the solvent with its deuterated sibling. The benefit here is that C-H bonds are now C-D
bonds and have lower vibrational frequencies. Compiled in Figure 4.2.11 is a set of common solvents.

4.2.6 https://chem.libretexts.org/@go/page/55873
Figure 4.2.11 IR transparencies of various solvents and their heavy counterparts. Adapted from N. L. McNiven and R. Court, Appl.
Spectrosc., 1970, 24, 296.
This effect has numerous benefits and is often applied to determining what vibrations correspond to what bond in a given molecular
sample. This is often accomplished by using isotopically labeled “heavy” reagents such as ones that contain 2H, 15N, 18O, or 13C.

Basic Troubleshooting
There are numerous problems that can arise from improperly prepared samples, this section will go through some of the common
problems and how to correct them. For this demonstration, spectra of ferrocene will be used. The molecular structure and a
photograph of the brightly colored organometallic compound are shown in Figure 4.2.12 and Figure 4.2.13.

Figure 4.2.12 Structure of ferrocene (Fe(C5H5)2).

Figure 4.2.13 Image of ferrocene powder (Fe(C5H5)2).


Figure 4.2.14 illustrates what a good sample of ferrocene looks like prepared in a KBr pellet. The peaks are well defined and sharp.
No peak is flattened at 0% transmittance and Christiansen scattering is not evident in the baseline.

4.2.7 https://chem.libretexts.org/@go/page/55873
Figure 4.2.14 A good spectrum of ferrocene in a KBr Pellet. Adapted from NIST Chemistry WebBook.
Figure 4.2.15 illustrates a sample with some peaks with intensities that are saturated and lose resolution making peak-picking
difficult. In order to correct for this problem, scrape some of the sample off of the salt plate with a rubber spatula and reseat the
opposite plate. By applying a thinner layer of sample one can improve the resolution of strongly absorbing vibrations.

Figure 4.2.15 An overly concentrated sample of ferrocene in a KBr pellet. Adapted from NIST Chemistry WebBook.
Figure 4.2.16 illustrates a sample in which too much mineral oil was added to the mull so that the C-H bonds are far more intense
than the actual sample. This can be remedied by removing the sample from the plate, grinding more sample and adding a smaller
amount of the mull to the plate. Another possible way of doing this is if the sample is insoluble in hexanes, add a little to the mull
and wick away the hexane-oil mixture to leave a dry solid sample. Apply a small portion of oil and replate.

Figure 4.2.16 A spectrum illustrating the problems of using Nujol, areas highlighted in orange are absorbances related to the
addition of Nujol to a sample. Notice how in the 1500 wavenumber region the addition of the Nujol has partially occulted the
absorbance by the ferrocene. Adapted from NIST Chemistry WebBook.
Figure 4.2.17 illustrates the result of particles being too large and scattering light. To remedy this, remove the mull and grind
further or else use the solvent deposition technique described earlier.

4.2.8 https://chem.libretexts.org/@go/page/55873
Figure 4.2.17 A sample exhibiting the Christiansen effect on ferrocene in a Nujol mull. Orange boxes indicate Nujol occult ranges.
Adapted from NIST Chemistry WebBook.

Characteristic IR Vibrational Modes for Hydrocarbon Compounds


Table 4.2.2 Stretching and bending bands for alkanes.
Functional group Mode Wavenumber range (cm-1)

CH3 Asymmetric stretch 2962±10

CH3 Symmetric stretch 2872±10

CH3 Asymmetric bend 1460±10

CH3 Symmetric bend (umbrella mode) 1375±10

CH2 Asymmetric stretch 2926±10

CH2 Symmetric stretch 2855±10

CH2 Scissors 1455±10

CH2 Rock 720±10

CH Stretch ~2900 (weak)

CH Bend ~1350 (weak)

Substitution C-H stretch (cm-1) C=C stretch (cm-1) Out of plane bend (cm-1)

Vinyl 3090-3075 1660-1630 900±5, 910±5

Vinylidine 3090-3075 1660-1630 890±5

Cis 3050-3000 1660-1630 690±10

Trans 3050-3000 1680-1665 965±5

Tri-substituted 3050-3000 1680-1665 815±25

Tetra-substituted - 1680-1665 -

Table 4.2.3 The stretching bands for alkenes.


Table 4.2.4 The stretching bands for alkynes.
Substitution C-H stretch (cm-1) C=C stretch (cm-1) C-H wag (cm-1)

Mono-substituted 3350-3250 2140-2100 700-600

Di-substituted - 2260-2190 -

Table 4.2.5 Bands for mono- and di-substituted benzene rings.


Substitution Out of plane C-H bending Ring bend (cm-1)

Mono 770-710 690±10

Ortho 810-750 -

4.2.9 https://chem.libretexts.org/@go/page/55873
Meta 770-735 690±10

Para 860-790 -

Table 4.2.6 Bands for methyl groups bonded to benzene rings.


Vibration Wavenumber (cm-1)

CH3 symmetric stretch 2925±5

CH3 bend overtone 2865±5

Fourier Transform Infrared Spectroscopy of Metal Ligand Complexes


The infrared (IR) range of the electromagnetic spectrum is usually divided into three regions:
The far-infrared is always used for rotational spectroscopy, with wavenumber range 400 – 10 cm−1 and lower energy.
The mid-infrared is suitable for a detection of the fundamental vibrations and associated rotational-vibrational structure with the
frequency range approximately 4000 – 400 cm−1.
The near-Infrared with higher energy and wave number range 14000 – 4000 cm−1, can excite overtone or higher harmonic
vibrations.
For classical light material interaction theory, if a molecule can interact with an electromagnetic field and absorb a photon of
certain frequency, the transient dipole of molecular functional group must oscillate at that frequency. Correspondingly, this
transition dipole moment must be a non-zero value, however, some special vibration can be IR inactive for the stretching motion of
a homonuclear diatomic molecule and vibrations do not affect the molecule’s dipole moment (e.g., N2).

Mechanistic Description of the Vibrations of Polyatomic Molecules


A molecule can vibrate in many ways, and each way is called a "vibrational mode". If a molecule has N atoms, linear molecules
have 3N-5 degrees of vibrational modes whereas nonlinear molecules have 3N-6 degrees of vibrational modes. Take H2O for
example; a single molecule of H2O has O-H bending mode (Figure 4.2.18 a), antisymmetric stretching mode (Figure 4.2.18 b), and
symmetric stretching mode (Figure 4.2.18 c).

Figure 4.2.18 Three types of hydroxy vibration modes. (a) bending mode; (b) antisymmetric stretching mode; (c) symmetric
stretching mode.
If a diatomic molecule has a harmonic vibration with the energy, 4.2.1 , where n+1/2 with n = 0, 1, 2 ...). The motion of the atoms
can be determined by the force equation, 4.2.2 , where k is the force constant). The vibration frequency can be described by 4.2.3 .
In which m is actually the reduced mass (mred or μ), which is determined from the mass m1 and m2 of the two atoms, 4.2.4 .
En   =   − hv (4.2.1)

F   =   − kx (4.2.2)

1/2
ω  =  (k/m) (4.2.3)

m1 m2
mred   =  μ  =   (4.2.4)
m1   +  m2

Principle of Absorption Bands


In IR spectrum, absorption information is generally presented in the form of both wavenumber and absorption intensity or percent
transmittance. The spectrum is generally showing wavenumber (cm-1) as the x-axis and absorption intensity or percent
transmittance as the y-axis.
Transmittance, "T", is the ratio of radiant power transmitted by the sample (I) to the radiant power incident on the sample (I0).
Absorbance (A) is the logarithm to the base 10 of the reciprocal of the transmittance (T). The absorption intensity of molecule
vibration can be determined by the Lambert-Beer Law, \label{5} . In this equation, the transmittance spectra ranges from 0 to
100%, and it can provide clear contrast between intensities of strong and weak bands. Absorbance ranges from infinity to zero. The

4.2.10 https://chem.libretexts.org/@go/page/55873
absorption of molecules can be determined by several components. In the absorption equation, ε is called molar extinction
coefficient, which is related to the molecule behavior itself, mainly the transition dipole moment, c is the concentration of the
sample, and l is the sample length. Line width can be determined by the interaction with surroundings.
A  =  log(1/T )  =   − log(I / I0 )  =  εcl (4.2.5)

The Infrared Spectrometer


As shown in Figure 4.2.19, there are mainly four parts for fourier transform infrared spectrometer (FTIR):
Light source. Infrared energy is emitted from a glowing black-body source as continuous radiations.
Interferometer. It contains the interferometer, the beam splitter, the fixed mirror and the moving mirror. The beam splittertakes
the incoming infrared beam and divides it into two optical beams. One beam reflects off the fixed mirror. The other beam
reflects off of the moving mirror which moves a very short distance. After the divided beams are reflected from the two mirrors,
they meet each other again at the beam splitter. Therefore, an interference pattern is generated by the changes in the relative
position of the moving mirror to the fixed mirror. The resulting beam then passes through the sample and is eventually focused
on the detector.
Sample compartment. It is the place where the beam is transmitted through the sample. In the sample compartment, specific
frequencies of energy are absorbed.
Detector. The beam finally passes to the detector for final measurement. The two most popular detectors for a FTIR
spectrometer are deuterated triglycine sulfate (pyroelectric detector) and mercury cadmium telluride (photon or quantum
detector). The measured signal is sent to the computer where the Fourier transformation takes place.

Figure 4.2.19 The main components of a fourier transform infrared (FTIR) spectrometer.

A Typical Application: the detection of metal ligand complexes


Some General Absorption peaks for common types of functional groups
It is well known that all molecules chemicals have distinct absorption regions in the IR spectrum. Table 4.2.7 shows the absorption
frequencies of common types of functional groups. For systematic evaluation, the IR spectrum is commonly divided into some sub-
regions.
In the region of 4000 - 2000 cm–1, the appearance of absorption bands usually comes from stretching vibrations between
hydrogen and other atoms. The O-H and N-H stretching frequencies range from 3700 - 3000 cm–1. If hydrogen bond forms
between O-H and other group, it generally caused peak line shape broadening and shifting to lower frequencies. The C-H
stretching bands occur in the region of 3300 - 2800 cm–1. The acetylenic C-H exhibits strong absorption at around 3300 cm–1.
Alkene and aromatic C-H stretch vibrations absorb at 3200-3000 cm–1. Generally, asymmetric vibrational stretch frequency of
alkene C-H is around 3150 cm-1, and symmetric vibrational stretch frequency is between 3100 cm-1 and 3000 cm-1. The
saturated aliphatic C-H stretching bands range from 3000 - 2850 cm–1, with absorption intensities that are proportional to the
number of C-H bonds. Aldehydes often show two sharp C-H stretching absorption bands at 2900 - 2700 cm–1. However, in
water solution, C-H vibrational stretch is much lower than in non-polar solution. It means that the strong polarity solution can
greatly reduce the transition dipole moment of C-H vibration.
Furthermore, the stretching vibrations frequencies between hydrogen and other heteroatoms are between 2600 - 2000cm-1, for
example, S-H at 2600 - 2550 cm–1, P-H at 2440 - 2275 cm–1, Si-H at 2250 - 2100 cm–1.
The absorption bands at the 2300 - 1850 cm–1 region usually present only from triple bonds, such as C≡C at 2260 - 2100 cm–1,
C≡N at 2260 - 2000 cm–1, diazonium salts –N≡N at approximately 2260 cm–1, allenes C=C=C at 2000 - 1900 cm–1. The peaks
of these groups are all have strong absorption intensities. The 1950 - 1450 cm–1 region stands for double-bonded functional
groups vibrational stretching.

4.2.11 https://chem.libretexts.org/@go/page/55873
Most carbonyl C=O stretching bands range from 1870 - 1550 cm–1, and the peak intensities are from mean to strong.
Conjugation, ring size, hydrogen bonding, and steric and electronic effects can lead to significant shifts in absorption
frequencies. Furthermore, if carbonyl links with electron-withdrawing group, such as acid chlorides and acid anhydrides, it
would give rise to IR bands at 1850 - 1750 cm–1. Ketones usually display stretching bands at 1715 cm-1.
None conjugated aliphatic C=C and C=N have absorption bands at 1690 - 1620 cm–1. Besides, around 1430 and 1370cm-1,
there are two identical peaks presenting C-H bending.
The region from 1300 - 910 cm–1 always includes the contributions from skeleton C-O and C-C vibrational stretches, giving
additional molecular structural information correlated with higher frequency areas. For example, ethyl acetate not only shows
its carbonyl stretch at 1750 - 1735 cm–1, but also exhibits its identical absorption peaks at 1300 - 1000 cm–1 from the skeleton
vibration of C-O and C-C stretches.
Table 4.2.7 The typical frequencies of functional groups.
Group Frequency (cm-1) Strength Appearance

Strong in nonpolar solvent


C-H stretch 2850-3400
Weak in polar solvent

O-H stretch, N-H stretch 3200-3700 Broad in solvent

C≡N stretch,
2050-2300 Medium or strong
R-N=C=S stretch

C≡O stretch (bound with metal) around 2000 Medium or strong

C≡C stretch 2100-2260 Weak

ca 1715 (ketone),
C=O stretch Strong
ca 1650 (amides)

C=C stretch 1450-1700 Weak to strong

C-H bend 1260 - 1470 Strong

C-O stretch 1040-1300 Medium or strong

General Introduction of Metal Ligand Complex


The metal electrons fill into the molecular orbital of ligands (CN, CO, etc.) to form complex compound. As shown in Figure
4.2.20, a simple molecular orbital diagram for CO can be used to explain the binding mechanism.

Figure 4.2.20 Molecular orbital diagram for carbon monoxide (CO).


The CO and metal can bind with three ways:
Donation of a pair of electrons from the C-O σ* orbital into an empty metal orbital (Figure 4.2.21 a).
Donation from a metal d orbital into the C-O π* orbital to form a M-to-CO π-back bond (Figure 4.2.21 b).
Under some conditions a pair of carbon π electron can donate into an empty metal d-orbital.

Figure 4.2.21 Main binding interaction types between metal and CO. (a) CO-to-metal σ bond; (b) M-to-CO π-back bond.

4.2.12 https://chem.libretexts.org/@go/page/55873
Some Factors to Include the Band Shifts and Strength
Herein, we mainly consider two properties: ligand stretch frequency and their absorption intensity. Take the ligand CO for example
again. The frequency shift of the carbonyl peaks in the IR mainly depends on the bonding mode of the CO (terminal or bridging)
and electron density on the metal. The intensity and peak numbers of the carbonyl bands depends on some factors: CO ligands
numbers, geometry of the metal ligand complex and fermi resonance.
Effect on Electron Density on Metal

As shown in Table 4.2.8, a greater charge on the metal center result in the CO stretches vibration frequency decreasing. For
example, [Ag(CO)]+show higher frequency of CO than free CO, which indicates a strengthening o
f the CO bond. σ donation removes electron density from the nonbonding HOMO of CO. From Figure, it is clear that the HOMO
has a small amount of anti-bonding property, so removal of an electron actually increases (slightly) the CO bond strength.
Therefore, the effect of charge and electronegativity depends on the amount of metal to CO π-back bonding and the CO IR
stretching frequency.
Table 4.2.8 Different types of ligands frequencies of different electron density on a metal center.
dx Complex CO stretch frequency (cm-1)

free CO 2143

d10 [Ag(CO)]+ 2204

d10 Ni(CO)4 2060

d10 [Co(CO)4]- 1890

d6 [Mn(CO)6]+ 2090

d6 Cr(CO)6 2000

d6 [V(CO)6]- 1860

If the electron density on a metal center is increasing, more π-back bonding to the CO ligand(s) will also increase, as shown in
Table 4.2.9. It means more electron density would enter into the empty carbonyl π* orbital and weaken the C-O bond. Therefore, it
makes the M-CO bond strength increasing and more double-bond-like (M=C=O).
Ligation Donation Effect
Some cases, as shown in Table 4.2.9, different ligands would bind with same metal at the same metal-ligand complex. For
example, if different electron density groups bind with Mo(CO)3 as the same form, as shown in Figure 4.2.22, the CO vibrational
frequencies would depend on the ligand donation effect. Compared with the PPh3 group, CO stretching frequency which the
complex binds the PF3 group (2090, 2055 cm-1) is higher. It indicates that the absolute amount of electron density on that metal
may have certain effect on the ability of the ligands on a metal to donate electron density to the metal center. Hence, it may be
explained by the Ligand donation effect. Ligands that are trans to a carbonyl can have a large effect on the ability of the CO ligand
to effectively π-backbond to the metal. For example, two trans π-backbonding ligands will partially compete for the same d-orbital
electron density, weakening each other’s net M-L π-backbonding. If the trans ligand is a π-donating ligand, the free metal to CO π-
backbonding can increase the M-CO bond strength (more M=C=O character). It is well known that pyridine and amines are not
those strong π-donors. However, they are even worse π-backbonding ligands. So the CO is actually easy for π-back donation
without any competition. Therefore, it naturally reduces the CO IR stretching frequencies in metal carbonyl complexes for the
ligand donation effect.
Table 4.2.9 The effect of different types of ligands on the frequency of the carbonyl ligand
Metal Ligand Complex CO Stretch Frequency (cm-1)

Mo(CO)3(PF3)3 2090, 2055

Mo(CO)3[P(OMe)3]3 1977, 1888

Mo(CO)3(PPh3)3 1934, 1835

Mo(CO)3(NCCH3)3 1915, 1783

Mo(CO)3(pyridine)3 1888, 1746

4.2.13 https://chem.libretexts.org/@go/page/55873
Figure 4.2.22 Schematic representation of competitive back-donation from a transition metal to multiple π-acceptor ligands

Geometry Effects
Some cases, metal-ligand complex can form not only terminal but also bridging geometry. As shown in Figure 4.2.23, in the
compound Fe2(CO)7(dipy), CO can act as a bridging ligand. Evidence for a bridging mode of coordination can be easily obtained
through IR spectroscopy. All the metal atoms bridged by a carbonyl can donate electron density into the π* orbital of the CO and
weaken the CO bond, lowering vibration frequency of CO. In this example, the CO frequency in terminal is around 2080 cm-1, and
in bridge, it shifts to around 1850 cm-1.

Figure 4.2.23 The structure of Fe2(CO)7(dipy)

Pump-probe Detection of Molecular Functional Group Vibrational Lifetime


The dynamics of molecular functional group plays an important role during a chemical process, chemical bond forming and
breaking, energy transfer and other dynamics happens within picoseconds domain. It is very difficult to study such fast processes
directly, for decades scientists can only learn from theoretical calculations, lacking experimental methods.
However, with the development of ultrashort pulsed laser enable experimental study of molecular functional group dynamics. With
ultrafast laser technologies, people develop a series of measuring methods, among which, pump-probe technique is widely used to
study the molecular functional group dynamics. Here we concentrate on how to use pump-probe experiment to measure functional
group vibrational lifetime. The principle, experimental setup and data analysis will be introduced.

Principles of the Pump-probe Technique


For every function group within a molecule, such as the C≡N triple bond in phenyl selenocyanate (C6H5SeCN) or the C-D single
bond in deuterated chloroform (DCCl3), they have an individual infrared vibrational mode and associated energy levels. For a
typical 3-level system (Figure 4.2.24, both the 0 to 1 and the 1 to 2 transition are near the probe pulse frequency (they don't
necessarily need to have exactly the same frequency).

Figure 4.2.24 Schematic representation of a typical three level system


In a pump-probe experiment, we use the geometry as is shown in Figure 4.2.25. Two synchronized laser beams, one of which is
called pump beam (Epu) while the other probe beam (Epr). There is a delay in time between each pulse. The laser pulses hit the
sample, the intensity of ultrafast laser (fs or ps) is strong enough to generated 3rd order polarization and produce 3rd order optical
response signal which is use to give dynamics information of molecular function groups. For the total response signals we have
\label{6} , where µ10 µ21 are transition dipole moment and E0, E1, and E2 are the energies of the three levels, and t3 is the time
delay between pump and probe beam. The delay t3 is varied and the response signal intensity is measured. The functional group
vibration life time is determined from the data.

4.2.14 https://chem.libretexts.org/@go/page/55873
Figure 4.2.25
4 −i( E1 −E0 )t3/h−Γt3
S  =  4μ e (4.2.6)
10

Typical Experimental Set-up


The optical layout of a typical pump-probe setup is schematically displayed in Figure 4.2.26. In the setup, the output of the
oscillator (500 mW at 77 MHz repetition rate, 40 nm bandwidth centered at 800 nm) is split into two beams (1:4 power ratio). Of
this, 20% of the power is to seed a femtosecond (fs) amplifier whose output is 40 fs pulses centered at 800 nm with power of ~3.4
W at 1 KHz repetition rate. The rest (80%) of the seed goes through a bandpass filter centered at 797.5nm with a width of 0.40 nm
to seed a picosecond (ps) amplifier. The power of the stretched seed before entering the ps amplifier cavity is only ~3 mW. The
output of the ps amplifier is 1ps pulses centered at 800 nm with a bandwidth ~0.6 nm. The power of the ps amplifier output is ~3
W. The fs amplifier is then to pump an optical parametric amplifier (OPA) which produces ~100 fs IR pulses with bandwidth of
~200 cm-1 that is tunable from 900 to 4000 cm-1. The power of the fs IR pulses is 7~40 mW, depending on the frequencies. The ps
amplifier is to pump a ps OPA which produces ~900 fs IR pulses with bandwidth of ~21 cm-1, tunable from 900 - 4000 cm-1. The
power of the fs IR pulses is 10 ~ 40 mW, depending on frequencies.

Figure 4.2.26 Schematic representation of the optical layout for a pump-probe experiment.
In a typical pump-probe setup, the ps IR beam is collimated and used as the pump beam. Approximately 1% of the fs IR OPA
output is used as the probe beam whose intensity is further modified by a polarizer placed before the sample. Another polarizer is
placed after the sample and before the spectrograph to select different polarizations of the signal. The signal is then sent into a
spectrograph to resolve frequency, and detected with a mercury cadmium telluride (MCT) dual array detector. Use of a pump pulse
(femtosecond, wide band) and a probe pulse (picoseconds, narrow band), scanning the delay time and reading the data from the
spectrometer, will give the lifetime of the functional group. The wide band pump and spectrometer described here is for collecting
multiple group of pump-probe combination.
Data Analysis
For a typical pump-probe curve shown in Figure 4.2.27 life time t is defined as the corresponding time value to the half intensity as
time zero.

Figure 4.2.27 A tympical pump-probe curve.


Table 4.2.10 shows the pump-probe data of the C≡N triple bond in a series of aromatic cyano compounds: n-propyl cyanide
(C3H7CN), ethyl thiocyanate (C2H5SCN), and ethyl selenocyanate (C2H5SeCN) for which the νC≡N for each compound (measured
in CCl4 solution) is 2252 cm-1), 2156 cm-1, and ~2155 cm-1, respectively.
Table 4.2.10 Pump-probe intensity data for C≡N stretching frequency in n-propyl cyanide, ethyl thiocyanate, and ethyl selenocyanate as a
function of delay (ps).

4.2.15 https://chem.libretexts.org/@go/page/55873
Delay (ps) C3H7CN C2H5SCN C2H5SeCN

0 -0.00695 -0.10918 -0.06901

0.1 -0.0074 -0.10797 -0.07093

0.2 -0.00761 -0.1071 -0.07247

0.3 -0.00768 -0.10545 -0.07346

0.4 -0.0076 -0.10487 -0.07429

0.5 -0.00778 -0.10287 -0.07282

0.6 -0.00782 -0.10286 -0.07235

0.7 -0.00803 -0.10222 -0.07089

0.8 -0.00764 -0.10182 -0.07073

0.9 -0.00776 -0.10143 -0.06861

1 -0.00781 -0.10099 -0.06867

1.1 -0.00745 -0.10013 -0.06796

1.2 -0.00702 -0.10066 -0.06773

1.3 -0.00703 -0.0989 -0.0676

1.4 -0.00676 -0.0995 -0.06638

1.5 -0.00681 -0.09757 -0.06691

1.6 -0.00639 -0.09758 -0.06696

1.7 -0.00644 -0.09717 -0.06583

1.8 -0.00619 -0.09741 -0.06598

1.9 -0.00613 -0.09723 -0.06507

2 -0.0066 -0.0962 -0.06477

2.5 -0.00574 -0.09546 -0.0639

3 -0.0052 -0.09453 -0.06382

3.5 -0.0482 -0.09353 -0.06389

4 -0.0042 -0.09294 -0.06287

4.5 -0.00387 -0.09224 -0.06197

5 -0.00351 -0.09009 -0.06189

5.5 -0.00362 -0.09084 -0.06188

6 -0.00352 -0.08938 -0.06021

6.5 -0.00269 -0.08843 -0.06028

7 -0.00225 -0.08788 -0.05961

7.5 -0.00231 -0.08694 -0.06065

8 -0.00206 -0.08598 -0.05963

8.5 -0.00233 -0.08552 -0.05993

9 -0.00177 -0.08503 -0.05902

9.5 -0.00186 -0.08508 -0.05878

10 -0.00167 -0.0842 -0.0591

11 -0.00143 -0.08295 -0.05734

4.2.16 https://chem.libretexts.org/@go/page/55873
A plot of intensity versus time for the data from TABLE is shown Figure 4.2.28. From these curves the C≡N stretch lifetimes can
be determined for C3H7CN, C2H5SCN, and C2H5SeCN as ~5.5 ps, ~84 ps, and ~282 ps, respectively.

Figure 4.2.28 The C≡N stretch lifetimes for benzyl cyanide, phenyl thiocyanate, and phenyl selenocyanate.
From what is shown above, the pump-probe method is used in detecting C≡N vibrational lifetimes in different chemicals. One
measurement only takes several second to get all the data and the lifetime, showing that pump-probe method is a powerful way to
measure functional group vibrational lifetime.

Attenuated Total Reflectace- Fourier Transform Infrared Spectroscopy


Attenuated total reflectance-Fourier transform infrared spectroscopy (ATR-FTIR) is a physical method of compositional analysis
that builds upon traditional transmission FTIR spectroscopy to minimize sample preparation and optimize reproducibility.
Condensed phase samples of relatively low refractive index are placed in close contact with a crystal of high refractive index and
the infrared (IR) absorption spectrum of the sample can be collected. Based on total internal reflection, the absorption spectra of
ATR resemble those of transmission FTIR. To learn more about transmission IR spectroscopy (FTIR) please refer to the section
further up this page titled Fourier Transform Infrared Spectroscopy of Metal Ligand Complexes.
First publicly proposed in 1959 by Jacques Fahrenfort from the Royal Dutch Shell laboratories in Amsterdam, ATR IR
spectroscopy was described as a technique to effectively measure weakly absorbing condensed phase materials. In Fahrenfort's first
article describing the technique, published in 1961, he used a hemicylindrical ATR crystal (see Experimental Conditions) to
produce single-reflection ATR (Figure 4.2.29 ). ATR IR spectroscopy was slow to become accepted as a method of characterization
due to concerns about its quantitative effectiveness and reproducibility. The main concern being the sample and ATR crystal
contact necessary to achieve decent spectral contrast. In the late 1980’s FTIR spectrometers began improving due to an increased
dynamic range, signal to noise ratio, and faster computers. As a result ATR-FTIR also started gaining traction as an efficient
spectroscopic technique. These days ATR accessories are often manufactured to work in conjunction with most FTIR
spectrometers, as can be seen in Figure 4.2.30.

Figure 4.2.29 The first ATR Infrared Spectrometer designed by Jacques Fahrenfort featuring a hemicylindrical ATR crystal.
Reproduced from J. Fahrenfort, Spectrochim. Acta, 1961, 17, 698. Copyright: Elsevier (1961).

4.2.17 https://chem.libretexts.org/@go/page/55873
Figure 4.2.30 An ATR attachment on an FTIR spectrometer.

Total Internal Reflection


For additional information on light waves and their properties please refer to the module on Vertical Scanning Interferometry (VSI)
in chapter 10.1.
When considering light propagating across an interface between two materials with different indices of refraction, the angle of
refraction can be given by Snell’s law, 4.2.7 , where none of the incident light will be transmitted.
φc   =  φmax (4.2.7)

The reflectance of the interface is total and whenever light is incident from a higher refractive index medium onto a lower
refractive index medium, the reflection is deemed internal (as opposed to external in the opposite scenario). Total internal
reflectance experiences no losses, or no transmitted light (Figure 4.2.31

Figure 4.2.31 At the interface between two materials with different indices of refraction, (a) when the angle of incident light, θ1, is
below the critical angle, θc, both reflection and transmission occur, and (b) when the angle of incident light exceeds the critical
angle, total internal reflection (TIR) occurs, spawning an evanescent wave at the interface. Adapted from M. Schnippering, S. R. T.
Neil, S. R. Mackenzie, and P. R. Unwin, Chem. Soc. Rev., 2011, 40, 207. Copyright: Royal Society of Chemistry (2011).
Supercritical internal reflection refers to angles of incidence above the critical angle of incidence allowing total internal reflectance.
It is in this angular regime where only incident and reflected waves will be present. The transmitted wave is confined to the
interface where its amplitude is at a maximum and will damp exponentially into the lower refractive index medium as a function of
distance. This wave is referred to as the evanescent wave and it extends only a very short distance beyond the interface.
To apply total internal reflection to the experimental setup in ATR, consider n2 to be the internal reflectance element or ATR crystal
(the blue trapezoid in Figure 4.2.32 )
where n2 is the material with the higher index of refraction. This should be a material that is fully transparent to the incident
infrared radiation to give a real value for the refractive index. The ATR crystal must also have a high index of refraction to allow
total internal reflection with many samples that have an index of refraction n1, where n1<n2.

Figure 4.2.32 The ATR crystal shown in blue, within which the incident IR light shown in red is totally reflecting. Above the
crystal the evanescent wave is emitted and penetrates the sample.
We can consider the sample to be absorbing in the infrared. Electromagnetic energy will pass through the crystal/sample interface
and propagate into the sample via the evanescent wave. This energy loss must be compensated with the incident IR light. Thus,
total reflectance is no longer occurring and the reflection inside the crystal is attenuated. If a sample does not absorb, the

4.2.18 https://chem.libretexts.org/@go/page/55873
reflectance at the interface shows no attenuation. Therefore if the IR light at a particular frequency does not reach the detector, the
sample must have absorbed it.
The penetration depth of the evanescent wave within the sample is on the order of 1µm. The expression of the penetration depth is
given in 4.2.8 and is dependent upon the wavelength and angle of incident light as well as the refractive indices of the ATR crystal
and sample. The effective path length is the product of the depth of penetration of the evanescent wave and the number of points
that the IR light reflects at the interface between the crystal and sample. This path length is equivalent to the path length of a
sample in a traditional transmission FTIR setup.
λ n1
2 1/2
dp = (sinω − ( ) ) (4.2.8)
2πn1 n2

Experimental Conditions
Refractive Indices of ATR Crystal and Sample
Typically an ATR attachment can be used with a traditional FTIR where the beam of incident IR light enters a horizontally
positioned crystal with a high refractive index in the range of 1.5 to 4, as can be seen in Table 4.2.11 will consist of organic
compounds, inorganic compounds, and polymers which have refractive indices below 2 and can readily be found on a database.
Table 4.2.11 A summary of popular ATR crystals. Data obtained from F. M. Mirabella, Internal reflection spectroscopy: Theory and
applications, 15, Marcel Dekker, Inc., New York (1993).
Material Refractive Index (RI) Spectral Range (cm-1)

Zinc Selenide (ZnSe) 2.4 20,000 - 650

Germanium (Ge) 4 5,500 - 870

Sapphire (Al2O3) 1.74 50,000 - 2,000

45,000 - 2,500,
Diamond (C) 2.4
1650 - 200

Single and Multiple Reflection Crystals


Multiple reflection ATR was initially more popular than single reflection ATR because of the weak absorbances associated with
single reflection ATR. More reflections increased the evanescent wave interaction with the sample, which was believed to increase
the signal to noise ratio of the spectrum. When IR spectrometers developed better spectral contrast, single reflection ATR became
more popular. The number of reflections and spectral contrast increases with the length of the crystal and decreases with the angle
of incidence as well as thickness. Within multiple reflection crystals some of the light is transmitted and some is reflected as the
light exits the crystal, resulting in some of the light going back through the crystal for a round trip. Therefore, light exiting the ATR
crystal contains components that experienced different number of reflections at the crystal-sample interface.
Angle of Incidence
It was more common in earlier instruments to allow selection of the incident angle, sometimes offering selection between 30°, 45°,
and 60°. In all cases for total internal reflection to hold, the angle of incidence must exceed the critical angle and ideally
complement the angle of the crystal edge so that the light enters at a normal angle of incidence. These days 45° is the standard
angle on most ATR-FTIR setups.
ATR Crystal Shape

For the most part ATR crystals will have a trapezoidal shape as shown in Figure 4.2.31. This shape facilitates sample preparation
and handling on the crystal surface by enabling the optical setup to be placed below the crystal. However, different crystal shapes
(Figure 4.2.33 ) may be used for particular purposes, whether it is to achieve multiple reflections or reduce the spot size. For
example, a hemispherical crystal may be used in a microsampling experiment in which the beam diameter can be reduced at no
expense to the light intensity. This allows appropriate measurement of a small sample without compromising the quality of the
resulting spectral features.

4.2.19 https://chem.libretexts.org/@go/page/55873
Figure 4.2.33 An assortment of ATR crystal shapes: a)triangular, b)hemispherical, c)parallelogram, d) trapezoidal, e) pentagonal,
f)cylindrical. Adapted from F. M. Mirabella, Internal reflection spectroscopy: Theory and applications, 15, Marcel Dekker, Inc.,
New York (1993).
Crystal-sample contact
Because the path length of the evanescent wave is confined to the interface between the ATR crystal and sample, the sample should
make firm contact with the ATR crystal (Figure 4.2.34 ). The sample sits atop the crystal and intimate contact can be ensured by
applying pressure above the sample. However, one must be mindful of the ATR crystal hardness. Too much pressure may distort
the crystal and affect the reproducibility of the resulting spectrum.

Figure 4.2.34 A close-up image of an ATR accessory attached to a Nexus 670 FTIR.
The wavelength effect expressed in \label{7} shows an increase in penetration depth at increased wavelength. In terms of
wavenumbers the relationship becomes inverse. At 4000 cm-1 penetration of the sample is 10x less than penetration at 400 cm-1
meaning the intensity of the peaks may appear higher at lower wavenumbers in the absorbance spectrum compared to the spectral
features in a transmission FTIR spectrum (if an automated correction to the ATR setup is not already in place).

Selecting an ATR Crystal


ATR functions effectively on the condition that the refractive index of the crystal is of a higher refractive index than the sample.
Several crystals are available for use and it is important to select an appropriate option for any given experiment (Table 4.2.11 ).
When selecting a material, it is important to consider reactivity, temperature, toxicity, solubility, and hardness.
The first ATR crystals in use were KRS-5, a mixture of thallium bromide and iodide, and silver halides. These materials are not
listed in the table because they are not in use any longer. While cost-effective, they are not practical due to their light sensitivity,
softness, and relatively low refractive indices. In addition KRS-5 is terribly toxic and dissolves on contact with many solvents,
including water.
At present diamond is a favorable option for its hardness, inertness and wide spectral range, but may not be a financially viable
option for some experiments. ZnSe and germanium are the most common crystal materials. ZnSe is reasonably priced, has
significant mechanical strength and a long endurance. However, the surface will become etched with exposure to chemicals on
either extreme of the pH scale. With a strong acid ZnSe will react to form toxic hydrogen selenide gas. ZnSe is also prone to
oxidation and care must be taken to avoid the formation of an IR absorbing layer of SeO2. Germanium has a higher refractive
index, which reduces the depth of penetration to 1 µm and may be preferable to ZnSe in applications involving intense sample
absorptions or for use with samples that produce strong background absorptions. Sapphire is physically robust with a wide spectral
range, but has a relatively low refractive index in terms of ATR crystals, meaning it may not be able to test as many samples as
another crystal might.

4.2.20 https://chem.libretexts.org/@go/page/55873
Sample Versatility
Solids
The versatility of ATR is reflected in the various forms and phases that a sample can assume. Solid samples need not be
compressed into a pellet, dispersed into a mull or dissolve in a solution. A ground solid sample is simply pressed to the surface of
the ATR crystal. For hard samples that may present a challenge to grind into a fine solid, the total area in contact with the crystal
may be compromised unless small ATR crystals with exceptional durability are used (e.g., 2 mm diamond). Loss of contact with
the crystal would result in decreased signal intensity because the evanescent wave may not penetrate the sample effectively. The
inherently short path length of ATR due to the short penetration depth (0.5-5 µm) enables surface-modified solid samples to be
readily characterized with ATR.
Powdered samples are often tedious to prepare for analysis with transmission spectroscopy because they typically require being
made into a KBr pellet to and ensuring the powdered sample is ground up sufficiently to reduce scattering. However, powdered
samples require no sample preparation when taking the ATR spectra. This is advantageous in terms of time and effort, but also
means the sample can easily be recovered after analysis.
Liquids

The advantage of using ATR to analyze liquid samples becomes apparent when short effective path lengths are required. The
spectral reproducibility of liquid samples is certain as long as the entire length of the crystal is in contact with the liquid sample,
ensuring the evanescent wave is interacting with the sample at the points of reflection, and the thickness of the liquid sample
exceeds the penetration depth. A small path length may be necessary for aqueous solutions in order to reduce the absorbance of
water.

Sample Preparation
ATR-FTIR has been used in fields spanning forensic analysis to pharmaceutical applications and even art preservation. Due to its
ease of use and accessibility ATR can be used to determine the purity of a compound. With only a minimal amount of sample this
researcher is able to collect a quick analysis of her sample and determine whether it has been adequately purified or requires further
processing. As can be seen in Figure 4.2.35, the sample size is minute and requires no preparation. The sample is placed in close
contact with the ATR crystal by turning a knob that will apply pressure to the sample (Figure 4.2.36 ).

Figure 4.2.35 Photograph of a small sample size is being placed on the ATR crystal.

Figure 4.2.36 Turning the knob applies pressure to the sample, ensuring good contact with the ATR crystal.

4.2.21 https://chem.libretexts.org/@go/page/55873
ATR has an added advantage in that it inherently encloses the optical path of the IR beam. In a transmission FTIR, atmospheric
compounds are constantly exposed to the IR beam and can present significant interference with the sample measurement. Of course
the transmission FTIR can be purged in a dry environment, but sample measurement may become cumbersome. In an ATR
measurement, however, light from the spectrometer is constantly in contact with the sample and exposure to the environment is
reduced to a minimum.

Application to Inorganic Chemistry


One exciting application of ATR is in the study of classical works of art. In the study of fragments of a piece of artwork, where
samples are scarce and one-of-a-kind, ATR is a suitable method of characterization because it requires only a small sample size.
Determining the compounds present in art enables proper preservation and historical insight into the pieces.
In a study examining several paint samples from a various origins, a micro-ATR was employed for analysis. This study used a
silicon crystal with a refractive index of 2.4 and a reduced beam size. Going beyond a simple surface analysis, this study explored
the localization of various organic and inorganic compounds in the samples by performing a stratigraphic analysis. The researchers
did so by embedding the samples in both KBr and a polyester resins. Two embedding techniques were compared to observe cross-
sections of the samples. The mapping of the samples took approximately 1-3 hours which may seem quite laborious to some, but
considering the precious nature of the sample, the wait time was acceptable to the researchers.
The optical microscope picture ( Figure 4.2.37 ) shows a sample of a blue painted area from the robe of a 14th century Italian
polychrome statue of a Madonna. The spectra shown in Figure 4.2.38 were acquired from the different layers pictured in the box
marked in Figure 4.2.37. All spectra were collected from the cross-sectioned sample and the false-color map on each spectrum
indicates the location of each of these compounds within the embedded sample. The spectra correspond to the inorganic
compounds listed in Table 4.2.12, which also highlights characteristic vibrational bands.

Figure 4.2.37 A paint sample from which four inorganic compounds were identified by ATR spectroscopy. The numbers indicate
different layers in the sample, composed of different inorganic compounds. The boxed area shows the region within which ATR
mapping occurred. Reproduced from R. Mazzeo, E. Joseph, S. Prati, and A. Millemaggi. Anal. Chim. Acta, 2007, 599, 107.
Copyright: Elsevier (2007).

4.2.22 https://chem.libretexts.org/@go/page/55873
Figure 4.2.37 . The images are labeled with the layer that corresponds to its location in the paint sample. Reproduced from R.
Mazzeo, E. Joseph, S. Prati, and A. Millemaggi. Anal. Chim. Acta, 2007, 599, 107. Copyright: Elsevier (2007)
Table 4.2.12 this table shows the inorganic compounds identified in the paint sample shown in 4.2.37 . Data from R. Mazzeo, E. Joseph, S.
Prati, and A. Millemaggi. Anal. Chim. Acta, 2007, 599, 107.
Compound Selected Spectral Bands Assignment

Cu3(CO3)2(OH)2 (Azurite) 1493 CO32- asymmetric stretch

Silicate based blue-pigments 1035 Si-O stretching

2PbCO3 ⋅ Pb(OH)2 (White lead) 1399 CO32- asymmetric stretch

A natural ferruginous aluminum silicate red


3697 OH stretching
pigment (Bole)

CaSO4 ⋅ (Gypsum) 1109 SO42- asymmetric stretch

The deep blue layer 3 corresponds to azurite and the light blue paint layer 2 to a mixture of silicate based blue pigments and white
lead. Although beyond the ATR crystal’s spatial resolution limit of 20 µm, the absorption of bole was detected by the characteristic
triple absorption bands of 3697, 3651, and 3619 cm-1 as seen in spectrum d of Figure 4.2.37. The white layer 0 was identified as
gypsum.
To identify the binding material, the KBr embedded sample proved to be more effective than the polyester resin. This was due in
part to the overwhelming IR absorbance of gypsum in the same spectral range (1700-1600 cm-1) as a characteristic stretch of the
binding as well as some contaminant absorption due to the polyester embedding resin.
To spatially locate specific pigments and binding media, ATR mapping was performed on the area highlighted with a box in Figure
4.2.37. The false color images alongside each spectrum in Figure 4.2.38 indicate the relative presence of the compound

corresponding to each spectrum in the boxed area. ATR mapping was achieved by taking 108 spectra across the 220x160 µm area
and selecting for each identified compound by its characteristic vibrational band.

4.2: IR Spectroscopy is shared under a CC BY 4.0 license and was authored, remixed, and/or curated by Pavan M. V. Raja & Andrew R. Barron
via source content that was edited to conform to the style and standards of the LibreTexts platform; a detailed edit history is available upon
request.

4.2.23 https://chem.libretexts.org/@go/page/55873
4.3: Raman Spectroscopy
Raman and Surface-Enhanced Raman Spectroscopy
What is Raman Spectroscopy
Raman spectroscopy is a powerful tool for determining chemical species. As with other spectroscopic techniques, Raman
spectroscopy detects certain interactions of light with matter. In particular, this technique exploits the existence of Stokes and Anti-
Stokes scattering to examine molecular structure. When radiation in the near infrared (NIR) or visible range interacts with a
molecule, several types of scattering can occur. Three of these can be seen in the energy diagram in Figure 4.3.1.

Figure 4.3.1 : Three types of scattering by a molecule excited by a photon with energy E = hν. The most common transition is
marked with bold arrows.
In all three types of scattering, an incident photon of energy hν raises the molecule from a vibrational state to one of the infinite
number of virtual states located between the ground and first electronic states. The type of scattering observed is dependent on how
the molecule relaxes after excitation.
Rayleigh Scattering
1. The molecule is excited to any virtual state.
2. The molecule relaxes back to its original state.
3. The photon is scattered elastically, leaving with its original energy.
Stokes Scattering
1. The molecule is excited to any virtual state.
2. The molecule relaxes back to a higher vibrational state than it had originally.
3. The photon leaves with energy hν-ΔE and has been scattered inelastically.
Anti-Stokes Scattering
1. The molecule begins in a vibrationally excited state.
2. The molecule is excited to any virtual state.
3. The molecule relaxes back to a lower vibrational state than it had originally.
4. The photon leaves with energy hν+ΔE, and has been scattered superelastically.
Rayleigh scattering is by far the most common transition, due to the fact that no change has to occur in the vibrational state of the
molecule. The anti-Stokes transition is the least common, as it requires the molecule to be in a vibrationally excited state before the
photon is incident upon it. Due to the lack of intensity of the anti-Stokes signal and filtering requirements that eliminate photons
with incident energy and higher, generally only Stokes scattering is used in Raman measurements. The relative intensities of
Rayleigh, Stokes and anti-Stokes scattering can be seen in Figure 4.3.2.

4.3.1 https://chem.libretexts.org/@go/page/55879
Figure 4.3.2 Location and relative intensity (indicated by peak height and width) of the Stokes and anti-Stokes scattering relative to
Rayleigh scattering.
Raman spectroscopy observes the change in energy between the incident and scattered photons associated with the Stokes and anti-
Stokes transitions. This is typically measured as the change in the wavenumber (cm-1), from the incident light source. Because
Raman measures the change in wavenumber, measurements can be taken using a source at any wavelength; however, near infrared
and visible radiation are commonly used. Photons with ultraviolet wavelengths could work as well, but tend to cause
photodecomposition of the sample.
Comparison between Raman and Infrared Spectroscopy
Raman spectroscopy sounds very much like infrared (IR) spectroscopy; however, IR examines the wavenumber at which a
functional group has a vibrational mode, while Raman observes the shift in vibration from an incident source. The Raman
frequency shift is identical to the IR peak frequency for a given molecule or functional group. As mentioned above, this shift is
independent of the excitation wavelength, giving versatility to the design and applicability of Raman instruments.
The cause of the vibration is also mechanistically different between IR and Raman. This is because the two operate on different sets
of selection rules. IR absorption requires a dipole moment or change in charge distribution to be associated with the vibrational
mode. Only then can photons of the same energy as the vibrational state of molecule interact. A schematic of this can be seen in
Figure 4.3.3.

Figure 4.3.3 A change in dipole moment is required for a vibrational mode to be IR active.
Raman signals, on the other hand, due to scattering, occur because of a molecule’s polarizability, illustrated in Figure 4.3.4. Many
molecules that are inactive or weak in the IR will have intense Raman signals. This results in often complementary techniques.

Figure 4.3.4 A change in the polarizability of a bond is required for a vibrational mode to be Raman active.
What does Raman Spectroscopy Measure?
Raman activity depends on the polarizability of a bond. This is a measure of the deformability of a bond in an electric field. This
factor essentially depends on how easy it is for the electrons in the bond to be displaced, inducing a temporary dipole. When there
is a large concentration of loosely held electrons in a bond, the polarizability is also large, and the group or molecule will have an
intense Raman signal. Because of this, Raman is typically more sensitive to the molecular framework of a molecule rather than a
specific functional group as in IR. This should not be confused with the polarity of a molecule, which is a measure of the separation
of electric charge within a molecule. Polar molecules often have very weak Raman signals due to the fact that electronegative
atoms hold electrons so closely.
Raman spectroscopy can provide information about both inorganic and organic chemical species. Many electron atoms, such as
metals in coordination compounds, tend to have many loosely bound electrons, and therefore tend to be Raman active. Raman can
provide information on the metal ligand bond, leading to knowledge of the composition, structure, and stability of these complexes.
This can be particularly useful in metal compounds that have low vibrational absorption frequencies in the IR. Raman is also very
useful for determining functional groups and fingerprints of organic molecules. Often, Raman vibrations are highly characteristic to

4.3.2 https://chem.libretexts.org/@go/page/55879
a specific molecule, due to vibrations of a molecule as a whole, not in localized groups. The groups that do appear in Raman
spectra have vibrations that are largely localized within the group, and often have multiple bonds involved.
What is Surface-Enhanced Raman Spectroscopy
Raman measurements provide useful characterization of many materials. However, the Raman signal is inherently weak (less than
0.001% of the source intensity), restricting the usefulness of this analytical tool. Placing the molecule of interest near a metal
surface can dramatically increase the Raman signal. This is the basis of surface-enhanced Raman spectroscopy (SERS). There are
several factors leading to the increase in Raman signal intensity near a metal surface
1. The distance to the metal surface.
Signal enhancement drops off with distance from the surface.
The molecule of interest must be close to the surface for signal enhancement to occur.
2. Details about the metal surface: morphology and roughness.
This determines how close and how many molecules can be near a particular surface area.
3. The properties of the metal.
Greatest enhancement occurs when the excitation wavelength is near the plasma frequency of the metal.
4. The relative orientation of the molecule to the normal of the surface.
The polarizability of the bonds within the molecule can be affected by the electrons in the surface of the metal.

Surface-Enhanced Raman Spectroscopy for the Study of Surface Chemistry


The ever-rising interest in nanotechnology involves the synthesis and application of materials with a very high surface area to
volume ratio. This places increasing importance on understanding the chemistry occurring at a surface, particularly the surface of a
nanoparticle. Slight modifications of the nanoparticle or its surrounding environment can greatly affect many properties including
the solubility, biological toxicity, and reactivity of the nanomaterial. Noble metal nanomaterials are of particular interest due to
their unique optical properties and biological inertness.
One tool employed to understand the surface chemistry of noble metal nanomaterial, particularly those composed of gold or silver
is surface-enhanced Raman spectroscopy (SERS). Replacing a metal surface with a metal nanoparticle increases the available
surface area for the adsorption of molecules. Compared to a flat metal surface, a similar sample size using nanoparticles will have a
dramatically stronger signal, since signal intensity is directly related to the concentration of the molecule of interest. Due to the
shape and size of the structure, the electrons in the nanoparticle oscillate collectively when exposed to incident electromagnetic
radiation. This is called the localized surface plasmon resonance (LSPR) of the nanoparticle. The LSPR of the nanoparticles boosts
the Raman signal intensity dramatically for molecules of interest near the surface of the nanoparticle. In order to maximize this
effect, a nanoparticle should be selected with its resonant wavelength falling in the middle of the incident and scattered
wavelengths.
The overall intensity enhancement of SERS can be as large as a factor of 106, with the surface plasmon resonance responsible for
roughly four orders of magnitude of this signal increase. The other two orders of magnitude have been attributed to chemical
enhancement mechanisms arising charge interactions between the metal particle and the adsorbate or from resonances in the
adsorbate alone, as discussed above.
Why is SERS Useful for Studying Surface Chemistry?
Traditionally, SERS uses nanoparticles made of conductive materials, such as gold, to learn more about a particular molecule.
However, of interest in many growing fields that incorporate nanotechnology is the structure and functionalization of a nanoparticle
stabilized by some surfactant or capping agent. In this case, SERS can provide valuable information regarding the stability and
surface structure of the nanoparticle. Another use of nanoparticles in SERS is to provide information about a ligand’s structure and
the nature of ligand binding. In many applications it is important to know whether a molecule is bound to the surface of the
nanoparticle or simply electrostatically interacting with it.
Sample Preparation and Instrumental Details
The standard Raman instrument is composed of three major components. First, the instrument must have an illumination system.
This is usually composed of one or more lasers. The major restriction for the illumination system is that the incident frequency of
light must not be absorbed by the sample or solvent. The next major component is the sample illumination system. This can vary
widely based on the specifics of the instrument, including whether the system is a standard macro-Raman or has micro-Raman

4.3.3 https://chem.libretexts.org/@go/page/55879
capabilities. The sample illumination system will determine the phase of material under investigation. The final necessary piece of
a Raman system is the spectrometer. This is usually placed 90° away from the incident illumination and may include a series of
filters or a monochromator. An example of a macro-Raman and micro-Raman setup can be Figure 4.3.5 and Figure 4.3.6. A
macro-Raman spectrometer has a spatial resolution anywhere from 100 μm to one millimeter while a micro-Raman spectrometer
uses a microscope to magnify its spatial resolution.

Figure 4.3.5 Schematic of a macro-Raman spectrometer.

Figure 4.3.6 Schematic of a micro-Raman spectrometer where illumination and collection are performed through microscope
objective.

Characterization of Single-Walled Carbon Nanotubes by Raman Spectroscopy


Carbon nanotubes (CNTs) have proven to be a unique system for the application of Raman spectroscopy, and at the same time
Raman spectroscopy has provided an exceedingly powerful tool useful in the study of the vibrational properties and electronic
structures of CNTs. Raman spectroscopy has been successfully applied for studying CNTs at single nanotube level.
The large van der Waals interactions between the CNTs lead to an agglomeration of the tubes in the form of bundles or ropes. This
problem can be solved by wrapping the tubes in a surfactant or functionalizing the SWNTs by attaching appropriate chemical
moieties to the sidewalls of the tube. Functionalization causes a local change in the hybridization from sp2 to sp3 of the side-wall
carbon atoms, and Raman spectroscopy can be used to determine this change. In addition information on length, diameter,
electronic type (metallic or semiconducting), and whether nanotubes are separated or in bundle can be obtained by the use of
Raman spectroscopy. Recent progress in understanding the Raman spectra of single walled carbon nanotubes (SWNT) have
stimulated Raman studies of more complicated multi-wall carbon nanotubes (MWNT), but unfortunately quantitative determination
of the latter is not possible at the present state of art.

Characterizing SWNT's
Raman spectroscopy is a single resonance process, i.e., the signals are greatly enhanced if either the incoming laser energy (Elaser)
or the scattered radiation matches an allowed electronic transition in the sample. For this process to occur, the phonon modes are
assumed to occur at the center of the Brillouin zone (q = 0). Owing to their one dimensional nature, the Π-electronic density of
states of a perfect, infinite, SWNTs form sharp singularities which are known as van Hove singularities (vHs), which are
energetically symmetrical with respect to Fermi level (Ef) of the individual SWNTs. The allowed optical transitions occur between
matching vHs of the valence and conduction band of the SWNTs, i.e., from first valence band vHs to the first conduction band vHs
(E11) or from the second vHs of the valence band to the second vHs of the conduction band (E22). Since the quantum state of an
electron (k) remains the same during the transition, it is referred to as k-selection rule.
The electronic properties, and therefore the individual transition energies in SWNTs are given by their structure, i.e., by their chiral
vector that determines the way SWNT is rolled up to form a cylinder. Figure 4.3.7 shows a SWNT having vector R making an
angle θ, known as the chiral angle, with the so-called zigzag or r1 direction.

4.3.4 https://chem.libretexts.org/@go/page/55879
Figure 4.3.7 The unrolled honeycomb lattice of a nanotube. When the sites O and A, and the sites B and C are connected, a portion
of a graphene sheet can be rolled seamlessly to form a SWNT. The vectors OA and OB define the chiral vector R of the nanotube,
respectively. The rectangle OABC defines the unit cell if the nanotube. The figure is constructed for (n,m) = (4,2) nanotube.
Adapted from M. S. Dresselhaus, G. Dresselhaus, R. Saito, and A. Jorio,Physics Reports, 2004, 2, 47.
Raman spectroscopy of an ensemble of many SWNTs having different chiral vectors is sensitive to the subset of tubes where the
condition of allowed transition is fulfilled. A ‘Kataura-Plot’ gives the allowed electronic transition energies of individual SWNTs
as a function of diameter d, hence information on which tubes are resonant for a given excitation wavelength can be inferred. Since
electronic transition energies vary roughly as 1/d, the question whether a given laser energy probes predominantly semiconducting
or metallic tubes depends on the mean diameter and diameter distribution in the SWNT ensemble. However, the transition energies
that apply to an isolated SWNT do not necessarily hold for an ensemble of interacting SWNTs owing to the mutual van der Waals
interactions.
Figure 4.3.8 shows a typical Raman spectrum from 100 to 3000 cm-1 taken of SWNTs produced by catalytic decomposition of
carbon monoxide (HiPco-process). The two dominant Raman features are the radial breathing mode (RBM) at low frequencies and
tangential (G-band) multifeature at higher frequencies. Other weak features, such as the disorder induced D-band and the G’ band
(an overtone mode) are also shown.

Figure 4.3.8 Raman spectrum of HiPco SWNTs using a laser of wavelength of λexc = 633 nm. Adapted from R. Graupner, J.
Raman Spectrosc., 2007, 38, 673.
Modes in the Raman Spectra of SWNTs
Radial Breamthing Modes (RBMs)
Out of all Raman modes observed in the spectra of SWNTs, the radial breathing modes are unique to SWNTs. They appear between
150 cm-1 < ωRBM < 300 cm-1 from the elastically scattered laser line. It corresponds to the vibration of the C atoms in the radial
direction, as if the tube is breathing (Figure 4.3.9). An important point about these modes is the fact that the energy (or
wavenumber) of these vibrational modes depends on the diameter (d) of the SWNTs, and not on the way the SWNT is rolled up to
form a cylinder, i.e., they do not depend on the θ of the tube.

4.3.5 https://chem.libretexts.org/@go/page/55879
Figure 4.3.9 Schematic picture showing vibration for RBM. Adapted from A. Jorio, M. A. Pimenta, A. G. S. Filho, R. Saito, G.
Dresselhaus, and M. S. Dresselhaus, New J. Phys., 2003, 5, 139.
These features are very useful for characterizing nanotube diameters through the relation ωRBM = A/d + B, where A and B are
constants and their variations are often attributed to environmental effects, i.e., whether the SWNTs are present as individual tubes
wrapped in a surfactant, isolated on a substrate surface, or in the form of bundles. However, for typical SWNT bundles in the
diameter range, d = 1.5 ± 0.2 nm, A = 234 cm-1 nm and B = 10 cm-1(where B is an upshift coming from tube-tube interactions). For
isolated SWNTs on an oxidized Si substrate, A= 248 cm-1 nm and B = 0. As can be seen from Figure 4.3.10, the relation ωRBM =
A/d + B holds true for the usual diameter range i.e., when d lies between 1 and 2 nm. However, for d less than 1 nm, nanotube
lattice distortions lead to chirality dependence of ωRBM and for large diameters tubes when, d is more than 2 nm the intensity of
RBM feature is weak and is hardly observable.

Figure 4.3.10 RBM frequencies ωRBM = A/d + B versus nanotube diameter for (i) A = 234 cm-1 nm and B = 10 cm-1, for SWNT
bundles (dashed curve); (ii) A = 248 cm-1 nm and B = 0, for isolated SWNTs (solid curve). Adapted from A. Jorio, M. A. Pimenta,
A. G. S. Filho, R. Saito, G. Dresselhaus, and M. S. Dresselhaus, New J. Phys., 2003, 5, 139.
Hence, a single Raman measurement gives an idea of the tubes that are in resonance with the laser line, but does not give a
complete characterization of the diameter distribution of the sample. However, by taking Raman spectra using many laser lines, a
good characterization of the diameter distributions in the sample can be obtained. Also, natural line widths observed for isolated
SWNTs are ωRBM = 3 cm-1, but as the tube diameter is increased, broadening is observed which is denoted by ΓRBM. It has been
observed that for d > 2 nm, ΓRBM > 20 cm-1. For SWNT bundles, the line width does not reflect ΓRMB, it rather reflects an
ensemble of tubes in resonance with the energy of laser.
Variation of RBM Intensities Upon Functionalization
Functionalization of SWNTs leads to variations of relative intensities of RBM compared to the starting material (unfunctionalized
SWNTs). Owing to the diameter dependence of the RBM frequency and the resonant nature of the Raman scattering process,
chemical reactions that are sensitive to the diameter as well as the electronic structure, i.e., metallic or semiconducting of the
SWNTs can be sorted out. The difference in Raman spectra is usually inferred by thermal defunctionalization, where the functional
groups are removed by annealing. The basis of using annealing for defunctionalizing SWNTs is based on the fact that annealing
restores the Raman intensities, in contrast to other treatments where a complete disintegration of the SWNTs occurs. Figure 4.3.11
shows the Raman spectra of the pristine, functionalized and annealed SWNTs. It can be observed that the absolute intensities of the
radial breathing modes is drastically reduced after functionalization. This decrease can be attributed to vHs, which themselves are a
consequence of translational symmetry of the SWNTs. Since the translational symmetry of the SWNTs is broken as a result of

4.3.6 https://chem.libretexts.org/@go/page/55879
irregular distribution of the sp3-sites due to the functionalization, these vHs are broadened and strongly reduced in intensity. As a
result, the resonant Raman cross section of all modes is strongly reduced as well.

Figure 4.3.11 Raman spectra of sidewall functionalized SWNTs of (A) pristine material, (B) functionalized SWNTs, and (C) after
annealing at 750 °C in Ar. Adapted from R. Graupner, J. Raman Spectrosc., 2007, 38, 673.
For an ensemble of functionalized SWNTs, a decrease in high wavenumber RBM intensities has been observed which leads to an
inference that destruction of small diameter SWNT takes place. Also, after prolonged treatment with nitric acid and subsequent
annealing in oxygen or vacuum, diameter enlargement of SWNTs is observed from the disappearance of RBMs from small
diameter SWNTs and the appearance of new RBMs characteristic of SWNTs with larger diameters. In addition, laser irradiation
seems to damage preferentially small diameter SWNTs. In all cases, the decrease of RBM intensities is either attributed to the
complete disintegration of SWNTs or reduction in resonance enhancement of selectively functionalized SWNTs. However, change
in RBM intensities can also have other reasons. One reason is doping induced bleaching of electronic transitions in SWNTs. When
a dopant is added, a previously occupied electronic state can be filled or emptied, as a result of which Ef in the SWNTs is shifted. If
this shift is large enough and the conduction band vHs corresponding to the respective Eiitransition that is excited by the laser light
gets occupied (n-type doping) or the valence band vHs is emptied (p-type doping), the resonant enhancement is lost as the
electronic transitions are quenched.
Sample morphology has also seen to affect the RBMs. The same unfunctionalized sample in different aggregation states gives rise
to different spectra. This is because the transition energy, Eii depends on the aggregation state of the SWNTs.
Tangential Modes (G-Band)
The tangential modes are the most intensive high-energy modes of SWNTs and form the so-called G-band, which is typically
observed at around 1600 cm-1. For this mode, the atomic displacements occur along the cicumferential direction (Figure 4.3.12).
Spectra in this frequency can be used for SWNT characterization, independent of the RBM observation. This multi-peak feature
can, for example, also be used for diameter characterization, although the information provided is less accurate than the RBM
feature, and it gives information about the metallic character of the SWNTs in resonance with laser line.

Figure 4.3.12 Schematic picture showing the atomic vibrations for the G-band. Adapted from A. Jorio, M. A. Pimenta, A. G. S.
Filho, R. Saito, G. Dresselhaus, and M. S. Dresselhaus, New J. Phys., 2003, 5, 139.

4.3.7 https://chem.libretexts.org/@go/page/55879
The tangential modes are useful in distinguishing semiconducting from metallic SWNTs. The difference is evident in the G- feature
(Figure 4.3.13 and 4.3.14) which broadens and becomes asymmetric for metallic SWNTs in comparison with the Lorentzian
lineshape for semiconducting tubes, and this broadening is related to the presence of free electrons in nanotubes with metallic
character. This broadened G-feature is usually fit using a Breit-Wigner-Fano (BWF) line that accounts for the coupling of a discrete
phonon with a continuum related to conduction electrons. This BWF line is observed in many graphite-like materials with metallic
character, such as n-doped graphite intercalation compounds (GIC), n-doped fullerenes, as well as metallic SWNTs. The intensity
of this G- mode depends on the size and number of metallic SWNTs in a bundle (Figure 4.3.15).

Figure 4.3.13 G-band for highly ordered pyrolytic graphite (HOPG), MWNT bundles, one isolated semiconducting SWNT and one
isolated metallic SWNT. The multi-peak G-band feature is not clear for MWNTs due to the large tube size. A. Jorio, M. A.
Pimenta, A. G. S. Filho, R. Saito, G. Dresselhaus, and M. S. Dresselhaus, New J. Phys., 2003, 5, 139. Copyright Institute of Physics
(2005).

Figure 4.3.14 Raman signal from three isolated semiconducting and three isolated metallic SWNTs showing the G-and D-band
profiles. SWNTs in good resonance (strong signal with low signal to noise ratio) show practically no D-band. A. Jorio, M. A.
Pimenta, A. G. S. Filho, R. Saito, G. Dresselhaus, and M. S. Dresselhaus, New J. Phys., 2003, 5, 139. Copyright Institute of Physics
(2005).

4.3.8 https://chem.libretexts.org/@go/page/55879
Figure 4.3.15 Dependence of G+ (black symbols) and G- (red symbols) frequencies as a function of diameter. Adapted from M.
Paillet, T. Michel, J. C. Meyer, V. N. Popov, L. Henrad, S. Roth, and J. L. Sauvajol, Phy. Rev. Lett., 2006, 96, 257401.
Charge of G-band Line Shape on Functionalization

Chemical treatments are found to affect the line shape of the tangential line modes. Selective functionalization of SWNTs or a
change in the ratio of metallic to semiconducting SWNTs due to selective etching is responsible for such a change. According to
Figure 4.3.16, it can be seen that an increase or decrease of the BWF line shape is observed depending on the laser wavelength. At
λexc = 633 nm, the preferentially functionalized small diameter SWNTs are semiconducting, therefore the G-band shows a decrease
in the BWG asymmetry. However, the situation is reversed at 514 nm, where small metallic tubes are probed. BWF resonance
intensity of small bundles increases with bundle thickness, so care should be taken that the effect ascribed directly to
functionalization of the SWNTs is not caused by the exfoliation of the previously bundles SWNT.

Figure 4.3.16 G-and D-band spectra of pristine (black) and ozonized (blue) SWNTs at 633 nm (left) and 514 nm (right) excitation.
Adapted from R. Graupner, J. Raman Spectrosc., 2007, 38, 673.
Disorder-Induced D-band
This is one of the most discussed modes for the characterization of functionalized SWNTs and is observed at 1300-1400 cm-1. Not
only for functionalized SWNTs, D-band is also observed for unfunctionalized SWNTs. From a large number of Raman spectra
from isolated SWNTs, about 50% exhibit observable D-band signals with weak intensity (Figure 4.3.14).
A large D-peak compared with the G-peak usually means a bad resonance condition, which indicates the presence of amorphous
carbon.
The appearance of D-peak can be interpreted due to the breakdown of the k-selection rule. It also depends on the laser energy and
diameter of the SWNTs. This behavior is interpreted as a double resonance effect, where not only one of the direct, k-conserving
electronic transitions, but also the emission of phonon is a resonant process. In contrast to single resonant Raman scattering, where
only phonons around the center of the Brillouin zone (q = 0) are excited, the phonons that provoke the D-band exhibit a non-
negligible q vector. This explains the double resonance theory for D-band in Raman spectroscopy. In few cases, the overtone of the
D-band known as the G’-band (or D*-band) is observed at 2600-2800 cm-1, and it does not require defect scattering as the two
phonons with q and –q are excited. This mode is therefore observed independent of the defect concentration.
The presence of D-band cannot be correlated to the presence of various defects (such as hetero-atoms, vacancies, heptagon-
pentagon pairs, kinks, or even the presence of impurities, etc). Following are the two main characteristics of the D-band found in
carbon nanotubes:
1. Small linewidths: ΓD values for SWNTs range from 40 cm-1 down to 7 cm-1.

4.3.9 https://chem.libretexts.org/@go/page/55879
2. Lower frequencies: D-band frequency is usually lower than the frequency of sp2-based carbons, and this downshift of frequency
shows 1/d dependence.
D-band Intensity as a Measure of Functionalization vs. Defect Density
Since D-peak appears due to the presence defects, an increase in the intensity of the band is taken as a fingerprint for successful
functionalization. But, whether D-band intensity is a measure of degree of functionalization or not is still sure. So, it is not correct
to correlate D-peak intensity or D-peak area to the degree of functionalization. From Figure 4.3.17, it can be observed that for
lower degree of functionalization, intensity of the D-band scales linearly with defect density. As the degree of functionalization is
further increased, both D and G-band area decrease, which is explained by the loss of resonance enhancement due to
functionalization. Also, normalization of the D-peak intensity to the G-band in order to correct for changes in resonance intensities
also leads to a decrease for higher densities of functional groups.

Figure 4.3.17 The left figure shows the intensity ratio ID/IG and the right figure shows D- and G-band intensity at λexc = 532 nm
with respect to degree of functionalization using diazonium reagents. Adapted from R. Graupner, J. Raman Spectrosc., 2007, 38,
673.
Limitations of Raman Spectroscopy
Though Raman spectroscopy has provides an exceedingly important tool for characterization of SWNTs, however, it suffers from
few serious limitations. One of the main limitations of Raman spectroscopy is that it does not provide any information about the
extent of functionalization in the SWNTs. The presence of D-band indicates disorder, i.e. side wall distribution, however it cannot
differentiate between the number of substituents and their distribution. Following are the two main limitations of Raman
Spectroscopy:
Quantification of Substituents
This can be illustrated by the following examples. Purified HiPco tubes may be fluorinated at 150 °C to give F-SWNTs with a C:F
ratio of approximately 2.4:1. The Raman spectra (using 780 nm excitation) for F-SWNTs shows in addition to the tangential mode
at ~1587 cm-1 an intense broad D (disorder) mode at ~ 1295 cm-1consistent with the side wall functionalization. Irrespective of the
arrangements of the fluorine substituents, thermolysis of F-SWNTs results in the loss of fluorine and the re-formation of
unfunctionalized SWNTs alnog with their cleavage into shorter length tubes. As can be seen from Figure 4.3.18, the intensity of the
D-band decreases as the thermolysis temperature increases. This is consistent with the loss of F-substituents. The G-band shows a
concomitant sharpening and increase in intensity.

Figure 4.3.18 Raman spectra of F-SWNTs (a) as prepared at 150 °C and after heating to (b) 400, (c) 450 and (d) 550 °C.
As discussed above, the presence of a significant D mode has been the primary method for determining the presence of sidewall
functionalization. It has been commonly accepted that the relative intensity of the D mode versus the tangential G mode is a

4.3.10 https://chem.libretexts.org/@go/page/55879
quantitative measure of level of substitution. However, as discussed below, the G:D ratio is also dependent on the distribution of
substituents. Using Raman spectroscopy in combination with XPS analysis of F-SWNTs that have been subjected to thermolysis at
different temperatures, a measure of the accuracy of Raman as a quantitative tool for determining substituent concentration can be
obtained. As can be seen from Figure 4.3.19, there is essentially no change in the G:D band ratio despite a doubling amount of
functional groups.Thus, at low levels of functionalization the use of Raman spectroscopy to quantify the presence of fluorine
substituents is a clearly suspect.

Figure 4.3.19 C(sp2):C-F(sp3) ratio (blue) and Raman G-band:D-band ratio (red) as a function of C:F ratio from XPS.
On the basis of above data it can be concluded that Raman spectroscopy does not provide an accurate quantification of small
differences at low levels of functionalization, whereas when a comparison between samples with high levels of functionalization or
large differences in degree of functionalization is requires Raman spectroscopy provides a good quantification.
Number vs Distribution
Fluorinated nanotubes may be readily functionalized by reaction with the appropriate amine in the presence of base according to
the scheme shown in Figure 4.3.20.

Figure 4.3.20 Synthesis of functionalized SWNTs.


When the Raman spectra of the functionalized SWNTs is taken (Figure 4.3.21), it is found out that the relative intensity of the
disorder D-band at ~1290 cm-1versus the tangential G-band (1500 - 1600 cm-1) is much higher for thiophene-SWNT than thiol-
SWNT. If the relative intensity of the D mode is the measure of the level of substitution, it can be concluded that there are more
number of thiophene groups present per C than thiol groups. However, from the TGA weight loss data the SWNT-C:substituent
ratios are calculated to be 19:1 and 17.5:1. Thus, contrary to the Raman data the TGA suggest that the number of substituents per C
(in the SWNT) is actually similar for both substituents.

Figure 4.3.21 Raman spectrum of (a) thiol-SWNT and (b)thiophene-SWNT using 780 nm excitation showing the relative intensity
of D-band at ~1300 cm-1 versus the G-band at ~1590 cm-1
This result would suggest that Raman spectroscopy is potentially unsuccessful in correctly providing the information about the
number of substituents on the SWNTs. Subsequent imaging of the functionalized SWNTs by STM showed that the distribution of

4.3.11 https://chem.libretexts.org/@go/page/55879
the functional groups was the difference between the thiol and thiphene functionalized SWNTs (Figure 4.3.22). Thus, relative ratio
of the D- and G-bands is a measure of concentration and distribution of functional groups on SWNTs.

Figure 4.3.22 Schematic representation of the functional group distribution for (a) thiol-SWNT and (b) thiophene-SWNT.

Multi-walled carbon nanotubes (MWNTs)


Most of the characteristic differences that distinguish the Raman spectra in SWNTs from the spectra of graphite are not so evident
for MWNTs. It is because the outer diameter for MWNTs is very large and the ensemble of CNTs in them varies from small to very
large. For example, the RBM Raman feature associated with a small diameter inner tube (less than 2 nm) can sometimes be
observed when a good resonance condition is established, but since the RBM signal from large diameter tubes is usually too weak
to be observable and the ensemble average of inner tube diameter broadens the signal, a good signal is not observed. However,
when hydrogen gas in the arc discharge method is used, a thin innermost nanotube within a MWNT of diameter 1 nm can be
obtained which gives strong RBM peaks in the Raman spectra.
Thereas the G+ - G- splitting is large for small diameter SWNT, the corresponding splitting of the G-band in MWNTs is both small
in intensity and smeared out due to the effect of the diameter distribution. Therefore the G-band feature predominantly exists a
weakly asymmetric characteristic lineshape, and a peak appearing close to the graphite frequency of 1582 cm-1.however for
isolated MWNTs prepared in the presence of hydrogen gas using the arc discharge method, it is possible to observe multiple G-
band splitting effects even more clearly than for the SWNTs, and this is because environmental effects become relatively small for
the innermost nanotube in a MWNT relative to the interactions occurring between SWNTs and different environments. The Raman
spectroscopy of MWNTs has not been well investigated up to now. The new directions in this field are yet to be explored.

4.3: Raman Spectroscopy is shared under a CC BY 4.0 license and was authored, remixed, and/or curated by Pavan M. V. Raja & Andrew R.
Barron via source content that was edited to conform to the style and standards of the LibreTexts platform; a detailed edit history is available upon
request.

4.3.12 https://chem.libretexts.org/@go/page/55879
4.4: UV-Visible Spectroscopy
Ultraviolet-visible (UV-vis) spectroscopy is used to obtain the absorbance spectra of a compound in solution or as a solid. What is
actually being observed spectroscopically is the absorbance of light energy or electromagnetic radiation, which excites electrons
from the ground state to the first singlet excited state of the compound or material. The UV-vis region of energy for the
electromagnetic spectrum covers 1.5 - 6.2 eV which relates to a wavelength range of 800 - 200 nm. The Beer-Lambert Law,
Equation 4.4.1 , is the principle behind absorbance spectroscopy. For a single wavelength, A is absorbance (unitless, usually seen
as arb. units or arbitrary units), ε is the molar absorptivity of the compound or molecule in solution (M-1cm-1), b is the path length
of the cuvette or sample holder (usually 1 cm), and c is the concentration of the solution (M).
A  =  εbc (4.4.1)

All of these instruments have a light source (usually a deuterium or tungsten lamp), a sample holder and a detector, but some have a
filter for selecting one wavelength at a time. The single beam instrument (Figure 4.4.1) has a filter or a monochromator between
the source and the sample to analyze one wavelength at a time. The double beam instrument (Figure 4.4.2) has a single source and
a monochromator and then there is a splitter and a series of mirrors to get the beam to a reference sample and the sample to be
analyzed, this allows for more accurate readings. In contrast, the simultaneous instrument (Figure 4.4.3) does not have a
monochromator between the sample and the source; instead, it has a diode array detector that allows the instrument to
simultaneously detect the absorbance at all wavelengths. The simultaneous instrument is usually much faster and more efficient,
but all of these types of spectrometers work well.

Figure 4.4.1 Illustration of a single beam UV-vis instrument.

Figure 4.4.2 Illustration of a double beam UV-vis instrument.

Figure 4.4.3 Illustration of a simultaneous UV-vis instrument.

What Information can be Obtained from UV-vis Spectra?


UV-vis spectroscopic data can give qualitative and quantitative information of a given compound or molecule. Irrespective of
whether quantitative or qualitative information is required it is important to use a reference cell to zero the instrument for the
solvent the compound is in. For quantitative information on the compound, calibrating the instrument using known concentrations
of the compound in question in a solution with the same solvent as the unknown sample would be required. If the information
needed is just proof that a compound is in the sample being analyzed, a calibration curve will not be necessary; however, if a
degradation study or reaction is being performed, and concentration of the compound in solution is required, thus a calibration
curve is needed.
To make a calibration curve, at least three concentrations of the compound will be needed, but five concentrations would be most
ideal for a more accurate curve. The concentrations should start at just above the estimated concentration of the unknown sample
and should go down to about an order of magnitude lower than the highest concentration. The calibration solutions should be
spaced relatively equally apart, and they should be made as accurately as possible using digital pipettes and volumetric flasks
instead of graduated cylinders and beakers. An example of absorbance spectra of calibration solutions of Rose Bengal (4,5,6,7-

4.4.1 https://chem.libretexts.org/@go/page/55881
tetrachloro-2',4',5',7'-tetraiodofluorescein, Figure 4.4.4, can be seen in Figure 4.4.5. To make a calibration curve, the value for the
absorbances of each of the spectral curves at the highest absorbing wavelength, is plotted in a graph similar to that in Figure 4.4.6
of absorbance versus concentration. The correlation coefficient of an acceptable calibration is 0.9 or better. If the correlation
coefficient is lower than that, try making the solutions again as the problem may be human error. However, if after making the
solutions a few times the calibration is still poor, something may be wrong with the instrument; for example, the lamps may be
going bad.

Figure 4.4.4 The molecular structure of Rose Bengal (4,5,6,7-tetrachloro-2',4',5',7'-tetraiodofluorescein).

Figure 4.4.5 UV-vis spectra of different concentrations of Rose Bengal.

Figure 4.4.6 Calibration curve of Rose Bengal. Equation of line: y = 0.0977x – 0.1492 (R2 = 0.996)

Limitations of UV-vis Spectroscopy


Sample
UV-vis spectroscopy works well on liquids and solutions, but if the sample is more of a suspension of solid particles in liquid, the
sample will scatter the light more than absorb the light and the data will be very skewed. Most UV-vis instruments can analyze
solid samples or suspensions with a diffraction apparatus (Figure 4.4.7), but this is not common. UV-vis instruments generally
analyze liquids and solutions most efficiently.

4.4.2 https://chem.libretexts.org/@go/page/55881
Figure 4.4.7 Schematic representation of the apparatus for collecting UV-vis spectra from solid materials.

Calibration and Reference


A blank reference will be needed at the very beginning of the analysis of the solvent to be used (water, hexanes, etc), and if
concentration analysis needs to be performed, calibration solutions need to be made accurately. If the solutions are not made
accurately enough, the actual concentration of the sample in question will not be accurately determined.

Choice of Solvent or Container


Every solvent has a UV-vis absorbance cutoff wavelength. The solvent cutoff is the wavelength below which the solvent itself
absorbs all of the light. So when choosing a solvent be aware of its absorbance cutoff and where the compound under investigation
is thought to absorb. If they are close, chose a different solvent. Table 4.4.1 provides an example of solvent cutoffs.
Table 4.4.1 : UV absorbance cutoffs of various common solvents
Solvent UV Absorbance Cutoff (nm)

Acetone 329

Benzene 278

Dimethylformamide 267

Ethanol 205

Toluene 285

Water 180

The material the cuvette (the sample holder) is made from will also have a UV-vis absorbance cutoff. Glass will absorb all of the
light higher in energy starting at about 300 nm, so if the sample absorbs in the UV, a quartz cuvette will be more practical as the
absorbance cutoff is around 160 nm for quartz (Table 4.4.2).
Table 4.4.2 : Three different types of cuvettes commonly used, with different usable wavelengths.
Material Wavelength Range (nm)

Glass 380-780

Plastic 380-780

Fused Quartz < 380

Concentration of Solution
To obtain reliable data, the peak of absorbance of a given compound needs to be at least three times higher in intensity than the
background noise of the instrument. Obviously using higher concentrations of the compound in solution can combat this. Also, if
the sample is very small and diluting it would not give an acceptable signal, there are cuvettes that hold smaller sample sizes than
the 2.5 mL of a standard cuvettes. Some cuvettes are made to hold only 100 μL, which would allow for a small sample to be
analyzed without having to dilute it to a larger volume, lowering the signal to noise ratio.

4.4.3 https://chem.libretexts.org/@go/page/55881
4.4: UV-Visible Spectroscopy is shared under a CC BY 4.0 license and was authored, remixed, and/or curated by Pavan M. V. Raja & Andrew R.
Barron via source content that was edited to conform to the style and standards of the LibreTexts platform; a detailed edit history is available upon
request.

4.4.4 https://chem.libretexts.org/@go/page/55881
4.5: Photoluminescence, Phosphorescence, and Fluorescence Spectroscopy
Photoluminescence spectroscopy is a contactless, nondestructive method of probing the electronic structure of materials. Light is
directed onto a sample, where it is absorbed and imparts excess energy into the material in a process called photo-excitation. One
way this excess energy can be dissipated by the sample is through the emission of light, or luminescence. In the case of photo-
excitation, this luminescence is called photoluminescence.
Photo-excitation causes electrons within a material to move into permissible excited states. When these electrons return to their
equilibrium states, the excess energy is released and may include the emission of light (a radiative process) or may not (a
nonradiative process). The energy of the emitted light (photoluminescence) relates to the difference in energy levels between the
two electron states involved in the transition between the excited state and the equilibrium state. The quantity of the emitted light is
related to the relative contribution of the radiative process.
In most photoluminescent systems chromophore aggregation generally quenches light emission via aggregation-caused quenching
(ACQ). This means that it is necessary to use and study fluorophores in dilute solutions or as isolated molecules. This in turn
results in poor sensitivity of devices employing fluorescence, e.g., biosensors and bioassays. However, there have recently been
examples reported in which luminogen aggregation played a constructive, instead of destructive role in the light-emitting process.
This aggregated-induced emission (AIE) is of great potential significance in particular with regard to solid state devices.
Photoluminescence spectroscopy provides a good method for the study of luminescent properties of a fluorophore.

Forms of Photoluminescence
Resonant Radiation: In resonant radiation, a photon of a particular wavelength is absorbed and an equivalent photon is
immediately emitted, through which no significant internal energy transitions of the chemical substrate between absorption and
emission are involved and the process is usually of an order of 10 nanoseconds.
Fluorescence: When the chemical substrate undergoes internal energy transitions before relaxing to its ground state by emitting
photons, some of the absorbed energy is dissipated so that the emitted light photons are of lower energy than those absorbed.
One of such most familiar phenomenon is fluorescence, which has a short lifetime (10-8 to 10-4s).
Phosphorescence: Phosphorescence is a radiational transition, in which the absorbed energy undergoes intersystem crossing
into a state with a different spin multiplicity. The lifetime of phosphorescence is usually from 10-4 - 10-2 s, much longer than
that of Fluorescence. Therefore, phosphorescence is even rarer than fluorescence, since a molecule in the triplet state has a good
chance of undergoing intersystem crossing to ground state before phosphorescence can occur.
Relation between Absorption and Emission Spectra
Fluorescence and phosphorescence come at lower energy than absorption (the excitation energy). As shown in Figure 4.5.1, in
absorption, wavelength λ0 corresponds to a transition from the ground vibrational level of S0 to the lowest vibrational level of S1.
After absorption, the vibrationally excited S1 molecule relaxes back to the lowest vibrational level of S1 prior to emitting any
radiation. The highest energy transition comes at wavelength λ0, with a series of peaks following at longer wavelength. The
absorption and emission spectra will have an approximate mirror image relation if the spacings between vibrational levels are
roughly equal and if the transition probabilities are similar. The λ0 transitions in Figure 4.5.2, do not exactly overlap. As shown in
Figure 4.5.8, a molecule absorbing radiation is initially in its electronic ground state, S0. This molecule possesses a certain
geometry and solvation. As the electronic transition is faster than the vibrational motion of atoms or the translational motion of
solvent molecules, when radiation is first absorbed, the excited S1 molecule still possesses its S0 geometry and solvation. Shortly
after excitation, the geometry and solvation change to their most favorable values for S1 state. This rearrangement lowers the
energy of excited molecule. When an S1 molecule fluoresces, it returns to the S0 state with S1 geometry and solvation. This
unstable configuration must have a higher energy than that of an S0molecule with S0 geometry and solvation. The net effect in
Figure 4.5.1 is that the λ0 emission energy is less than the λ0 excitation energy.

4.5.1 https://chem.libretexts.org/@go/page/55883
Figure 4.5.1 Energy-level diagram showing why structure is seen in the absorption and emission spectra and why the spectra are
roughly mirror images of each other. Adapted from D. C. Harris, Quantitative Chemical Analysis, 7th Ed, W. H. Freeman and
Company, New York (2006).

Figure 4.5.2 Second emission spectra. Adapted from C. M. Byron and T. C. Werner, J. Chem. Ed., 1991, 68, 433.
Instrumentation
A schematic of an emiision experiment is give in Figure 4.5.3. An excitation wavelength is selected by one monochromator, and
luminescence is observed through a second monochromator, usually positioned at 90° to the incident light to minimize the intensity
of scattered light reaching the dector. If the excitation wavelength is fixed and the emitted radiation is scanned, an emission
spectrum is produced.

Figure 4.5.3 Essentials of a luminescence experiment. The samle is irradiated at one wavelength and emission is observed over a
range of wavelengths. The excitation monochromator selects the excitation wavelength and the emission monochromator selects
one wavelength at a time to observe. Adapted from D. C. Harris, Quantitative Chemical Analysis, 7th Edition, W. H. Freeman and
Company, New York, (2006).
Relationship to UV-vis Spectroscopy
Ultraviolet-visible (UV-vis) spectroscopy or ultraviolet-visible spectrophotometry refers to absorption spectroscopy or reflectance
spectroscopy in the untraviolet-visible spectral region. The absorption or reflectance in the visible range directly affects the
perceived color of the chemicals involved. In the UV-vis spectrum, an absorbance versus wavelength graph results and it measures
transitions from the ground state to excited state, while photoluminescence deals with transitions from the excited state to the
ground state.

4.5.2 https://chem.libretexts.org/@go/page/55883
An excitation spectrum is a graph of emission intensity versus excitation wavelength. An excitation spectrum looks very much like
an absorption spectrum. The greater the absorbance is at the excitation wavelength, the more molecules are promoted to the excited
state and the more emission will be observed.
By running an UV-vis absorption spectrum, the wavelength at which the molecule absorbs energy most and is excited to a large
extent can be obtained. Using such value as the excitation wavelength can thus provide a more intense emission at a red-shifted
wavelength, which is usually within twice of the excitation wavelength.
Applications
Detection of ACQ or AIE properties
Aggregation-caused quenching (ACQ) of light emission is a general phenomenon for many aromatic compounds that fluorescence
is weakened with an increase in its solution concentration and even condensed phase. Such effect, however, comes into play in the
solid state, which has prevented many lead luminogens identified by the laboratory solution-screening process from finding real-
world applications in an engineering robust form.
Aggregation-induced emission (AIE), on the other hand, is a novel phenomenon that aggregation plays a constructive, instead of
destructive role in the light-emitting process, which is exactly opposite to the ACQ effect.
A Case Study
From the photoluminescence spectra of hexaphenylsilole (HPS, Figure 4.5.4) show in Figure 4.5.5, it can be seen that as the water
(bad solvent) fraction increases, the emission intensity of HPS increases. For BODIPY derivative Figure 4.5.6 in Figure 4.5.7, it
shows that the PL intensity peaks at 0 water content resulted from intramolecular rotation or twisting, known as twisted
intramolecular charge transfer (TICT).

Figure 4.5.4 The structure of hexaphenylsilole (HPS).

Figure 4.5.5 PL spectra of HPS solutions in acetonitrile/water mixtures. Adapted from Y. Hong, J. W. Y. Lam, and B. Z. Tang,
Chem. Commun., 2009, 4332. Copyright: The Royal Society of Chemistry (2009).

Figure 4.5.6 The structure of a triphenylamine–boradiazaindacene (BODIPY) derivative.

4.5.3 https://chem.libretexts.org/@go/page/55883
Figure 4.5.7 The structure of a triphenylamine–boradiazaindacene (BODIPY) derivative.
The emission color of an AIE luminogen is scarcely affected by solvent polarity, whereas that of a TICT luminogen typically
bathochromically shifts with increasing solvent polarity. In Figure 4.5.8, however, it shows different patterns of emission under
different excitation wavelengths. At the excitation wavelength of 372 nm, which is corresponding to the BODIPY group, the
emission intensity increases as water fraction increases. However, it decreases at the excitation wavelength of 530 nm, which is
corresponding to the TPE group. The presence of two emissions in this compound is due to the presence of two independent groups
in the compound with AIE and ACQ properties, respectively.

Figure 4.5.8 PL spectra of compound containing AIE and ACQ groups in THF/water mixtures at the excitation wavelength of 329
nm. Adapted from Y. Hong, J. W. Y. Lam, and B. Z. Tang, Chem. Commun., 2009, 4332. Copyright: The Royal Society of
Chemistry (2009).
Detection of Luminescence with Respect to Molarity
Figure 4.5.9 shows the photoluminescence spectroscopy of a BODIPY-TPE derivative of different concentrations. At the excitation
wavelength of 329 nm, as the molarity increases, the emission intensity decreases. Such compounds whose PL emission intensity
enhances at low concentration can be a good chemo-sensor for the detection of the presence of compounds with low quantity.

Figure 4.5.9 PL spectra of a BODIPY derivative solution in different concentrations in THF at excitation wavelength of 329 nm.
Other Applications

4.5.4 https://chem.libretexts.org/@go/page/55883
Apart from the detection of light emission patterns, photoluminescence spectroscopy is of great significance in other fields of
analysis, especially semiconductors.
Band Gap Determination
Band gap is the energy difference between states in the conduction and valence bands, of the radiative transition in semiconductors.
The spectral distribution of PL from a semiconductor can be analyzed to nondestructively determine the electronic band gap. This
provides a means to quantify the elemental composition of compound semiconductor and is a vitally important material parameter
influencing solar cell device efficiency.
Impurity Levels and Defect Detection
Radiative transitions in semiconductors involve localized defect levels. The photoluminescence energy associated with these levels
can be used to identify specific defects, and the amount of photoluminescence can be used to determine their concentration. The PL
spectrum at low sample temperatures often reveals spectral peaks associated with impurities contained within the host material.
Fourier transform photoluminescence microspectroscopy, which is of high sensitivity, provides the potential to identify extremely
low concentrations of intentional and unintentional impurities that can strongly affect material quality and device performance.
Recombination Mechanisms
The return to equilibrium, known as “recombination”, can involve both radiative and nonradiative processes. The quantity of PL
emitted from a material is directly related to the relative amount of radiative and nonradiative recombination rates. Nonradiative
rates are typically associated with impurities and the amount of photoluminescence and its dependence on the level of photo-
excitation and temperature are directly related to the dominant recombination process. Thus, analysis of photoluminescence can
qualitatively monitor changes in material quality as a function of growth and processing conditions and help understand the
underlying physics of the recombination mechanism.
Surface and Structure and Excited States
The widely used conventional methods such as XRD, IR and Raman spectroscopy, are very often not sensitive enough for
supported oxide catalysts with low metal oxide concentrations. Photoluminescence, however, is very sensitive to surface effects or
adsorbed species of semiconductor particles and thus can be used as a probe of electron-hole surface processes.
Limitations of Photoluminescence Spectroscopy
Very low concentrations of optical centers can be detected using photoluminescence, but it is not generally a quantitative technique.
The main scientific limitation of photoluminescence is that many optical centers may have multiple excited states, which are not
populated at low temperature.
The disappearance of luminescence signal is another limitation of photoluminescence spectroscopy. For example, in the
characterization of photoluminescence centers of silicon no sharp-line photoluminescence from 969 meV centers was observed
when they had captured self-interstitials.

Fluorescence Characterization and DNA Detection


Luminescence is a process involving the emission of light from any substance, and occurs from electronically excited states of that
substance. Normally, luminescence is divided into two categories, fluorescence and phosphorescence, depending on the nature of
the excited state.
Fluorescence is the emission of electromagnetic radiation light by a substance that has absorbed radiation of a different wavelength.
Phosphorescence is a specific type of photoluminescence related to fluorescence. Unlike fluorescence, a phosphorescent material
does not immediately re-emit the radiation it absorbs.
The process of fluorescent absorption and emission is easily illustrated by the Jablonski diagram. A classic Jablonski diagram is
shown in Figure 4.5.10, where Sn represents the nth electronic states. There are different vibrational and rotational states in every
electronic state. After light absorption, a fluorophore is excited to a higher electronic and vibrational state from ground state (here
rotational states are not considered for simplicity). By internal conversion of energy, these excited molecules relax to lower
vibrational states in S1 (Figure 4.5.10) and then return to ground states by emitting fluorescence. Actually, excited molecules
always return to higher vibration states in S0 and followed by some thermal process to ground states in S1. It is also possible for
some molecules to undergo intersystem crossing process to T2 states (Figure 4.5.10). After internal conversion and relaxing to T1,
these molecules can emit phosphorescence and return to ground states.

4.5.5 https://chem.libretexts.org/@go/page/55883
Figure 4.5.10 Jablonski diagram where, A = absorbance, F = fluorescence, P = phosphorescence, S = single state, T = triplet state,
IC = internal conversion, ISC = intersystem crossing.
The Stokes shift, the excited state lifetime and quantum yield are the three most important characteristics of fluorescence emission.
Stokes shift is the difference between positions of the band maxima of the absorption and emission spectra of the same electronic
transition. According to mechanism discussed above, an emission spectrum must have lower energy or longer wavelength than
absorption light. The quantum yield is a measure of the intensity of fluorescence, as defined by the ratio of emitted photons over
absorbed photons. Excited state lifetime is a measure of the decay times of the fluorescence.

Instrumentation of Fluorescence Spectroscopy


Spectrofluorometers
Most spectrofluorometers can record both excitation and emission spectra. An emission spectrum is the wavelength distribution of
an emission measured at a single constant excitation wavelength. In comparison, an excitation spectrum is measured at a single
emission wavelength by scanning the excitation wavelength.
Light Sources
Specific light sources are chosen depending on the application.
Arc and Incandescent Xenon Lamps
The high-pressure xenon (Xe) arc is the most versatile light source for steady-state fluorometers now. It can provides a steady light
output from 250 - 700 nm (Figure 4.5.11), with only some sharp lines near 450 and 800 nm. The reason that xenon arc lamps emit
a continuous light is the recombination of electrons with ionized Xe atoms. These ions produced by collision between Xe and
electrons. Those sharp lines near 450 nm are due to the excited Xe atoms that are not ionized.

Figure 4.5.11 Spectral irradiance of arc-discharge lamps.


During fluorescence experiment, some distortion of the excitation spectra can be observed, especially the absorbance locating in
visible and ultraviolet region. Any distortion displayed in the peaks is the result of wavelength-dependent output of Xe lamps.
Therefore, we need to apply some mathematic and physical approaches for correction.
High Pressure Mercury Lamps
Compared with xenon lamps, Hg lamps have higher intensities. As shown in Figure 4.5.11 the intensity of Hg lamps is
concentrated in a series of lines, so it is a potentially better excitation light source if matched to certain fluorophorescence.
Xe-Hg Arc Lamps
High-pressure xenon-mercury lamps have been produced. They have much higher intensity in ultraviolet region than normal Xe
lamps. Also, the introduction of Xe to Hg lamps broadens the sharp-line output of Hg lamps. Although the wavelength of output is

4.5.6 https://chem.libretexts.org/@go/page/55883
still dominated by those Hg lines, these lines are broadened and fit to various fluorophores better. The Xe-Hg lamp output depends
on the operating temperature.
Low Pressure Hg and Hg-Ar Lamps
Due to their very sharp line spectra, they are primarily useful for calibration purpose. The combination of Hg and Ar improve the
output scale, from 200 - 1000 nm.
Other Light Source
There are many other light source for experimental and industrial application, such as pulsed xenon lamps, quartz-tungsten halogen
(QTH) lamps, LED light sources, etc.
Monochromators
Most of the light sources used provide only polychromatic or white light. However, what is needed for experiments are various
chromatic light with a wavelength range of 10 nm. Monocharomators help us to achieve this aim. Prisms and diffraction gratings
are the two main kinds of monochromators used, although diffraction gratings are most useful, especially in spectrofluorometers.
Dispersion, efficiency, stray light level and resolution are important parameters for monochromators. Dispersion is mainly
determined by slit width and expressed in nm/mm. It is prepared to have low stray light level. Stray light is defined as light
transmitted by the monochromator at wavelength outside the chosen range. Also, a high efficiency is required to increase the ability
to detect low light levels. Resolution depends on the slit width. There are normally two slits, entrance and exit in a fluorometers.
Light intensity that passes through the slits is proportional to the square of the slit width. Larger slits have larger signal levels, but
lower resolution, and vice verse. Therefore, it is important to balance the signal intensity and resolution with the slit width.
Optical filters
Optical filters are used in addition to monochromators, because the light passing through monochromator is rarely ideal, optical
filters are needed for further purifying light source. If the basic excitation and emission properties of a particular system under
study, then selectivity by using optical filters is better than by the use of monochromators. Two kinds of optical filter are gradually
employed: colored filters and thin-film filters.
Colored Filters
Colored filters are the most traditional filter used before thin-film filter were developed. They can be divided into two categories:
monochromatic filter and long-pass filter. The first one only pass a small range of light (about 10 - 25 nm) centered at particular
chosen wavelength. In contrast, long pass filter transmit all wavelengths above a particular wavelength. In using these bandpass
filters, special attention must be paid to the possibility of emission from the filter itself, because many filters are made up of
luminescent materials that are easily excited by UV light. In order to avoid this problem, it is better to set up the filter further away
from the sample.
Thin-film Filters
The transmission curves of colored class filter are not suitable for some application and as such they are gradually being substituted
by thin-film filters. Almost any desired transmission curve can be obtained using a thin film filter.
Detectors
The standard detector used in many spectrofluorometers is the InGaAs array, which can provides rapid and robust spectral
characterization in the near-IR. And the liquid-nitrogen cooling is applied to decrease the background noise. Normally, detectors
are connected to a controller that can transfer a digital signal to and from the computer.
Fluorophores
At present a wide range of fluorophores have been developed as fluorescence probes in bio-system. They are widely used for
clinical diagnosis, bio-tracking and labeling. The advance of fluorometers has been accompanied with developments in fluorophore
chemistry. Thousands of fluorophores have been synthesized, but herein four categories of fluorophores will be discussed with
regard their spectral properties and application.
Intrinsic or Natural Fluorophores
Tryptophan (trp), tyrosine (tyr), and phenylalanine (phe) are three natural amino acid with strong fluorescence (Figure 4.5.12 ). In
tryptophan, the indole groups absorbs excitation light as UV region and emit fluorescence.

4.5.7 https://chem.libretexts.org/@go/page/55883
Figure 4.5.12 The structure of (a) tryptophan, (b) tyrosine and (c) phenylalanine.
Green fluorescent proteins (GFP) is another natural fluorophores. GFP is composed of 238 amino acids (Figure 4.5.13), and it
exhibits a characteristic bright green fluorescence when excited. They are mainly extracted from bioluminescent jellyfish Aequorea
vicroria, and are employed as signal reporters in molecular biology.

Figure 4.5.13 Green fluorescent proteins (GFP) ribbon diagram.


Extrinsic Fluorophores
Most bio-molecules are nonfluorescent, therefore it is necessary to connect different fluorophores to enable labeling or tracking of
the biomolecules. For example, DNA is an example of a biomolecule without fluorescence. The Rhodamine (Figure 4.5.14) and
BODIPY (Figure 4.5.15) families are two kinds of well-developed organic fluorophores. They have been extensively employed in
design of molecular probes due to their excellent photophysical properties.

Figure 4.5.14 The structure of Rhodamine 123.

4.5.8 https://chem.libretexts.org/@go/page/55883
Figure 4.5.15 The structure of selected boron-dipyrromethane (BODIPY) derivatives with their characteristic emission colors.
Red and Near-infrared (NIR) dyes
With the development of fluorophores, red and near-infrared (NIR) dyes attract increasing attention since they can improve the
sensitivity of fluorescence detection. In biological system, autofluorescence always increase the ratio of signal-to-noise (S/N) and
limit the sensitivity. As the excitation wavelength turns to longer, autopfluorescence decreases accordingly, and therefore signal-to-
noise ratio increases. Cyanines are one such group of long-wavelength dyes, e.g., Cy-3, Cy-5 and Cy-7 (Figure 4.5.16), which have
emission at 555, 655 and 755 nm respectively.

4.5.9 https://chem.libretexts.org/@go/page/55883
Figure 4.5.16 The structure of (a) Cy-3-iodo acetamide, (b) Cy-5-N-hydroxysuccinimide and (c) Cy-7-isothiocyanate.
Long-lifetime Fluorophores
Almost all of the fluorophores mentioned above are organic fluorophores that have relative short lifetime from 1-10 ns. However,
there are also a few long-lifetime organic fluorophore, such as pyrene and coronene with lifetime near 400 ns and 200 ns
respectively (Figure 4.5.17). Long-lifetime is one of the important properties to fluorophores. With its help, the autofluorescence in
biological system can be removed adequately, and hence improve the detectability over background.

Figure 4.5.17 Structures of (a) pyrene and (b) coronene.


Although their emission belongs to phosphorescence, transition metal complexes are a significant class of long-lifetime
fluorophores. Ruthenium (II), iridium (III), rhenium (I), and osmium (II) are the most popular transition metals that can combine
with one to three diimine ligands to form fluorescent metal complexes. For example, iridium forms a cationic complex with two
phenyl pyridine and one diimine ligand (Figure 4.5.18). This complex has excellent quantum yield and relatively long lifetime.

Figure 4.5.18 The structure of the cationic iridium complex, (ppy)2Ir(phen).


Applications

4.5.10 https://chem.libretexts.org/@go/page/55883
With advances in fluorometers and fluorophores, fluorescence has been a dominant techonology in the medical field, such clinic
diagnosis and flow cytometry. Herein, the application of fluorescence in DNA and RNA detecition is discussed.
The low concentration of DNA and RNA sequences in cells determine that high sensitivity of the probe is required, while the
existence of various DNA and RNA with similar structures requires a high selectivity. Hence, fluorophores were introduced as the
signal group into probes, because fluorescence spectroscopy is most sensitive technology until now.
The general design of a DNA or RNA probe involves using an antisense hybridization oligonucleotide to monitor target DNA
sequence. When the oligonucleotide is connected with the target DNA, the signal groups-the fluorophores-emit designed
fluorescence. Based on fluorescence spectroscopy, signal fluorescence can be detected which help us to locate the target DNA
sequence. The selectively inherent in the hybridization between two complementary DNA/RNA sequences make this kind of DNA
probes extremely high selectivity. A molecular Beacon is one kind of DNA probes. This simple but novel design is reported by
Tyagi and Kramer in 1996 (Figure 4.5.19) and gradually developed to be one of the most common DNA/RNA probes.

Figure 4.5.19 The structure of molecular beacon and its detecting mechanism.
Generally speaking, a molecular beacon it is composed of three parts: one oligonucleotide, a fluorophore and a quencher at
different ends. In the absence of the target DNA, the molecular beacon is folded like a hairpin due to the interaction between the
two series nucleotides at opposite ends of the oligonucleotide. At this time, the fluorescence is quenched by the close quencher.
However, in the presence of the target, the probe region of the MB will hybridize to the target DNA, open the folded MB and
separate the fluorophore and quencher. Therefore, the fluorescent signal can be detected which indicate the existence of a particular
DNA.

Fluorescence Correlation Spectroscopy


Florescence correlation spectroscopy (FCS) is an experimental technique that that measures fluctuations in fluorescence intensity
caused by the Brownian motion of particles. Fluorescence is a form of luminescence that involves the emission of light by a
substance that has absorbed light or other electromagnetic radiation. Brownian motion is the random motion of particles suspended
in a fluid that results from collisions with other molecules or atoms in the fluid. The initial experimental data is presented as
intensity over time but statistical analysis of fluctuations makes it possible to determine various physical and photo-physical
properties of molecules and systems. When combined with analysis models, FCS can be used to find diffusion coefficients,
hydrodynamic radii, average concentrations, kinetic chemical reaction rates, and single-triplet state dynamics. Singlet and triplet
states are related to electron spin. Electrons can have a spin of (+1/2) or (-1/2). For a system that exists in the singlet state, all spins
are paired and the total spin for the system is ((-1/2) + (1/2)) or 0. When a system is in the triplet state, there exist two unpaired
electrons with a total spin state of 1.
History
The first scientists to be credited with the application of fluorescence to signal-correlation techniques were Douglas Magde, Elliot
L. Elson, and Walt W.Webb, therefore they are commonly referred to as the inventors of FCS. The technique was originally used to
measure the diffusion and binding of ethidium bromide (Figure 4.5.20) onto double stranded DNA.

Figure 4.5.20 Structure of ethindium bromide, the molecule used in the first experiment involving FCS.
Initially, the technique required high concentrations of fluorescent molecules and was very insensitive. Starting in 1993, large
improvements in technology and the development of confocal microscopy and two-photon microscopy were made, allowing for
great improvements in the signal to noise ratio and the ability to do single molecule detection. Recently, the applications of FCS

4.5.11 https://chem.libretexts.org/@go/page/55883
have been extended to include the use of FörsterResonance Energy Transfer (FRET), the cross-correlation between two fluorescent
channels instead of auto correlation, and the use of laser scanning. Today, FCS is mostly used for biology and biophysics.
Instrumentation
A basic FCS setup (Figure 4.5.21) consists of a laser line that is reflected into a microscope objective by a dichroic mirror. The
laser beam is focused on a sample that contains very dilute amounts of fluorescent particles so that only a few particles pass
through the observed space at any given time. When particles cross the focal volume (the observed space) they fluoresce. This light
is collected by the objective and passes through the dichroic mirror (collected light is red-shifted relative to excitation light),
reaching the detector. It is essential to use a detector with high quantum efficiency (percentage of photons hitting the detector that
produce charge carriers). Common types of detectors are a photo-multiplier tube (rarely used due to low quantum yield), an
avalanche photodiode, and a super conducting nanowire single photo detector. The detector produces an electronic signal that can
be stored as intensity over time or can be immediately auto correlated. It is common to use two detectors and cross- correlate their
outputs leading to a cross-correlation function that is similar to the auto correlation function but is free from after-pulsing (when a
photon emits two electronic pulses). As mentioned earlier, when combined with analysis models, FCS data can be used to find
diffusion coefficients, hydrodynamic radii, average concentrations, kinetic chemical reaction rates, and single-triplet dynamics.

Figure 4.5.21 Basic FCS set-up. Close up of the objective reveals how particles in the sample move in and out of the observable
range of the objective (particles move in and out of laser light in the observed volume)
Analysis
When particles pass through the observed volume and fluoresce, they can be described mathematically as point spread functions,
with the point of the source of the light being the center of the particle. A point spread function (PSF) is commonly described as an
ellipsoid with measurements in the hundreds of nanometer range (although not always the case depending on the particle). With
respect to confocal microscopy, the PSF is approximated well by a Gaussian, 4.5.1, where I0 is the peak intensity, r and z are radial
and axial position, and wxy and wzare the radial and axial radii (with wz > wxy).
2 2 2
−2r 2 −2 z / ωz
P SF (r, z)  =  I0 e / ωxy e (4.5.1)

This Gaussian is assumed with the auto-correlation with changes being applied to the equation when necessary (like the case of a
triplet state, chemical relaxation, etc.). For a Gaussian PSF, the autocorrelation function is given by 4.5.2, where 4.5.3 is the
stochastic displacement in space of a fluorophore after time T.
2 2 2
1 Δ(τ )   +  ΔY (τ ) ΔZ(τ )
G(τ )  = ⟨exp(−  −  )⟩ (4.5.2)
2 2
⟨N ⟩ wxy wz

⃗ 
ΔR(τ )  =  (ΔX(τ ), Δ(τ ), Δ(τ )) (4.5.3)

The expression is valid if the average number of particles, N, is low and if dark states can be ignored. Because of this, FCS
observes a small number of molecules (nanomolar and picomolar concentrations), in a small volume (~1μm3) and does not require
physical separation processes, as information is determined using optics. After applying the chosen autocorrelation function, it
becomes much easier to analyze the data and extract the desired information (Figure 4.5.22).

4.5.12 https://chem.libretexts.org/@go/page/55883
Figure 4.5.22 Auto-correlated spectra of spherical 100 nm dye labeled agarose beads diffusing in water. Here it can be seen that
after the autocorrelation function was applied to the raw data using mathematical software, the fluorescence exponential decay
curve was derived for the sample. From this curve it is possible to calculate the average lifetime of the dye.
Application

FCS is often seen in the context of microscopy, being used in confocal microscopy and two-photon excitation microscopy. In both
techniques, light is focused on a sample and fluorescence intensity fluctuations are measured and analyzed using temporal
autocorrelation. The magnitude of the intensity of the fluorescence and the amount of fluctuation is related to the number of
individual particles; there is an optimum measurement time when the particles are entering or exiting the observation volume.
When too many particles occupy the observed space, the overall fluctuations are small relative to the total signal and are difficult to
resolve. On the other hand, if the time between molecules passing through the observed space is too long, running an experiment
could take an unreasonable amount of time. One of the applications of FCS is that it can be used to analyze the concentration of
fluorescent molecules in solution. Here, FCS is used to analyze a very small space containing a small number of molecules and the
motion of the fluorescence particles is observed. The fluorescence intensity fluctuates based on the number of particles present;
therefore analysis can give the average number of particles present, the average diffusion time, concentration, and particle size.
This is useful because it can be done in vivo, allowing for the practical study of various parts of the cell. FCS is also a common
technique in photo-physics, as it can be used to study triplet state formation and photo-bleaching. State formation refers to the
transition between a singlet and a triplet state while photo-bleaching is when a fluorophore is photo-chemically altered such that it
permanently looses its ability to fluoresce. By far, the most popular application of FCS is its use in studying molecular binding and
unbinding often, it is not a particular molecule that is of interest but, rather, the interaction of that molecule in a system. By dye
labeling a particular molecule in a system, FCS can be used to determine the kinetics of binding and unbinding (particularly useful
in the study of assays).
Main Advantages and Limitations
Table 4.5.1 : Advantages and limitations of PCS.
Advantage Limitation

Can be used in vivo Can be noisy depending on the system

Very sensitive Does not work if concentration of dye is too high

The same instrumentation can perform various kinds of experiments Raw data does not say much, analysis models must be applied

Has been used in various studies, extensive work has been done to If system deviates substantially from the ideal, analysis models can be
establish the technique difficult to apply (making corrections hard to calculate).

It may require more calculations to approximate PSF, depending on the


A large amount of information can be extracted
particular shape.

Molecular Phosphorescence Spectroscopy


When a material that has been radiated emits light, it can do so either via incandescence, in which all atoms in the material emit
light, or via luminescence, in which only certain atoms emit light, Figure 4.5.23. There are two types of luminescence:
fluorescence and phosphorescence. Phosphorescence occurs when excited electrons of a different multiplicity from those in their
ground state return to their ground state via emission of a photon, Figure 4.5.24. It is a longer-lasting and less common type of
luminescence, as it is a spin forbidden process, but it finds applications across numerous different fields. This module will cover

4.5.13 https://chem.libretexts.org/@go/page/55883
the physical basis of phosphorescence, as well as instrumentation, sample preparation, limitations, and practical applications
relating to molecular phosphorescence spectroscopy.

Figure 4.5.23 When an electron is excited by incident light, it may release the energy via emission of a photon

Figure 4.5.24 Phosphorescence is the decay of an electron from the excited triplet state to the singlet ground state via the emission
of a photon.
Phosphorescence

Phosphorescence is the emission of energy in the form of a photon after an electron has been excited due to radiation. In order to
understand the cause of this emission, it is first important to consider the molecular electronic state of the sample. In the singlet
molecular electronic state, all electron spins are paired, meaning that their spins are antiparallel to one another. When one paired
electron is excited to a higher-energy state, it can either occupy an excited singlet state or an excited triplet state. In an excited
singlet state, the excited electron remains paired with the electron in the ground state. In the excited triplet state, however, the
electron becomes unpaired with the electron in ground state and adopts a parallel spin. When this spin conversion happens, the
electron in the excited triplet state is said to be of a different multiplicity from the electron in the ground state. Phosphorescence
occurs when electrons from the excited triplet state return to the ground singlet state, 4.5.4 - 4.5.6, where E represents an electron
in the singlet ground state, E* represent the electron in the singlet excited state, and T* represents the electron in the triplet excited
state.
E  +  hv → E∗ (4.5.4)

E∗ → T ∗ (4.5.5)


T ∗ →  E  +  hv (4.5.6)

Electrons in the triplet excited state are spin-prohibited from returning to the singlet state because they are parallel to those in the
ground state. In order to return to the ground state, they must undergo a spin conversion, which is not very probable, especially
considering that there are many other means of releasing excess energy. Because of the need for an internal spin conversion,
phosphorescence lifetimes are much longer than those of other kinds of luminescence, lasting from 10-4 to 104 seconds.
Historically, phosphorescence and fluorescence were distinguished by the amount of time after the radiation source was removed
that luminescence remained. Fluorescence was defined as short-lived chemiluminescence (< 10-5 s) because of the ease of
transition between the excited and ground singlet states, whereas phosphorescence was defined as longer-lived chemiluminescence.
However, basing the difference between the two forms of luminescence purely on time proved to be a very unreliable metric.
Fluorescence is now defined as occurring when decaying electrons have the same multiplicity as those of their ground state.
Sample Preparation
Because phosphorescence is unlikely and produces relatively weak emissions, samples using molecular phosphorescence
spectroscopy must be very carefully prepared in order to maximize the observed phosphorescence. The most common method of
phosphorescence sample preparation is to dissolve the sample in a solvent that will form a clear and colorless solid when cooled to
77 K, the temperature of liquid nitrogen. Cryogenic conditions are usually used because, at low temperatures, there is little

4.5.14 https://chem.libretexts.org/@go/page/55883
background interference from processes other than phosphorescence that contribute to loss of absorbed energy. Additionally, there
is little interference from the solvent itself under cryogenic conditions. The solvent choice is especially important; in order to form
a clear, colorless solid, the solvent must be of ultra-high purity. The polarity of the phosphorescent sample motivates the solvent
choice. Common solvents include ethanol for polar samples and EPA (a mixture of diethyl ether, isopentane, and ethanol in a 5:5:2
ratio) for non-polar samples. Once a disk has been formed from the sample and solvent, it can be analyzed using a phosphoroscope.
Room Temperature Phosphorescence
While using a rigid medium is still the predominant choice for measuring phosphorescence, there have been recent advances in
room temperature spectroscopy, which allows samples to be measured at warmer temperatures. Similar the sample preparation
using a rigid medium for detection, the most important aspect is to maximize recorded phosphorescence by avoiding other forms of
emission. Current methods for allowing good room detection of phosphorescence include absorbing the sample onto an external
support and putting the sample into a molecular enclosure, both of which will protect the triplet state involved in phosphorescence.
Instrumentation and Measurement

Phosphorescence is recorded in two distinct methods, with the distinguishing feature between the two methods being whether or
not the light source is steady or pulsed. When the light source is steady, a phosphoroscope, or an attachment to a fluorescence
spectrometer, is used. The phosphoroscope was experimentally devised by Alexandre-Edmond Becquerel, a pioneer in the field of
luminescence, in 1857, Figure 4.5.25.

Figure 4.5.25 A lithograph depicting Alexandre-Edmond Becquerel, taken by Pierre Petit.


There are two different kinds of phosphoroscopes: rotating disk phosphoroscopes and rotating can phosphoroscopes. A rotating
disk phosphoroscope, Figure 4.5.26, comprises two rotating disk with holes, in the middle of which is placed the sample to be
tested. After a light beam penetrates one of the disks, the sample is electronically excited by the light energy and can phosphoresce;
a photomultiplier records the intensity of the phosphorescence. Changing the speed of the disks’ rotation allows a decay curve to be
created, which tells the user how long phosphorescence lasts.

Figure 4.5.26 A rotating disk phosphoroscope has slots for phosphorescence measurement.

4.5.15 https://chem.libretexts.org/@go/page/55883
The second type of phosphoroscope, the rotating can phosphoroscope, employs a rotating cylinder with a window to allow passage
of light, Figure 4.5.27. The sample is placed on the outside edge of the can and, when light from the source is allowed to pass
through the window, the sample is electronically excited and phosphoresces, and the intensity is again detected via photomultiplier.
One major advantage of the rotating can phosphoroscope over the rotating disk phosphoroscope is that, at high speeds, it can
minimize other types of interferences such as fluorescence and Raman and Rayleigh scattering, the inelastic and elastic scattering
of photons, respectively.

Figure 4.5.27 A rotating can phosphoroscope has an attached crank and gears to adjust the speed of rotation.
The more modern, advanced measurement of phosphorescence uses pulsed-source time resolved spectrometry and can be measured
on a luminescence spectrometer. A luminescence spectrometer has modes for both fluorescence and phosphorescence, and the
spectrometer can measure the intensity of the wavelength with respect to either the wavelength of the emitted light or time, Figure
4.5.28.

Figure 4.5.28 A phosphorescence intensity versus time plot which shows how a gated photomultiplier measures the intensity of
phosphorescent decay under pulsed time resolved spectrometry. Reproduced with permission from H.M. Rowe, Sing Po Chan, J. N.
Demas, and B. A. DeGraff, Anal. Chem., 2002, 74, 4821.
The spectrometer employs a gated photomultiplier to measure the intensity of the phosphorescence. After the initial burst of
radiation from the light source, the gate blocks further light, and the photomultiplier measures both the peak intensity of
phosphorescence as well as the decay, as shown in Figure 4.5.29.

Figure 4.5.29 A phosphorescence intensity versus time plot which shows how a gated photomultiplier measures the intensity of
phosphorescent decay under pulsed time resolved spectrometry. Reproduced with permission from H.M. Rowe, Sing Po Chan, J. N.
Demas, and B. A. DeGraff, Anal. Chem., 2002, 74, 4821.

4.5.16 https://chem.libretexts.org/@go/page/55883
The lifetime of the phosphorescence is able to be calculated from the slope of the decay of the sample after the peak intensity. The
lifetime depends on many factors, including the wavelength of the incident radiation as well as properties arising from the sample
and the solvent used. Although background fluorescence as well as Raman and Rayleigh scattering are still present in pulsed-time
source resolved spectrometry, they are easily detected and removed from intensity versus time plots, allowing for the pure
measurement of phosphorescence.
Limitations
The biggest single limitation of molecular phosphorescence spectroscopy is the need for cryogenic conditions. This is a direct
result of the unfavorable transition from an excited triplet state to a ground singlet state, which unlikely and therefore produces
low-intensity, difficult to detect, long-lasting irradiation. Because cooling phosphorescent samples reduces the chance of other
irradiation processes, it is vital for current forms of phosphorescence spectroscopy, but this makes it somewhat impractical in
settings outside of a specialized laboratory. However, the emergence and development of room temperature spectroscopy methods
give rise to a whole new set of applications and make phosphorescence spectroscopy a more viable method.
Practical Applications
Currently, phosphorescent materials have a variety of uses, and molecular phosphorescence spectrometry is applicable across many
industries. Phosphorescent materials find use in radar screens, glow-in-the-dark toys, and in pigments, some of which are used to
make highway signs visible to drivers. Molecular phosphorescence spectroscopy is currently in use in the pharmaceutical industry,
where its high selectivity and lack of need for extensive separation or purification steps make it useful. It also shows potential in
forensic analysis because of the low sample volume requirement.

4.5: Photoluminescence, Phosphorescence, and Fluorescence Spectroscopy is shared under a CC BY 4.0 license and was authored, remixed,
and/or curated by Pavan M. V. Raja & Andrew R. Barron via source content that was edited to conform to the style and standards of the
LibreTexts platform; a detailed edit history is available upon request.

4.5.17 https://chem.libretexts.org/@go/page/55883
4.6: Mössbauer Spectroscopy
In 1957 Rudolf Mössbauer achieved the first experimental observation of the resonant absorption and recoil-free emission of
nuclear γ-rays in solids during his graduate work at the Institute for Physics of the Max Planck Institute for Medical Research in
Heidelberg Germany. Mössbauer received the 1961 Nobel Prize in Physics for his research in resonant absorption of γ-radiation
and the discovery of recoil-free emission a phenomenon that is named after him. The Mössbauer effect is the basis of Mössbauer
spectroscopy.
The Mössbauer effect can be described very simply by looking at the energy involved in the absorption or emission of a γ-ray from
a nucleus. When a free nucleus absorbs or emits a γ-ray to conserve momentum the nucleus must recoil, so in terms of energy:
Eγ−ray   =  Enuclear transition  −  Erecoil (4.6.1)

When in a solid matrix the recoil energy goes to zero because the effective mass of the nucleus is very large and momentum can be
conserved with negligible movement of the nucleus. So, for nuclei in a solid matrix:
Eγ−ray   =  Enuclear transition (4.6.2)

This is the Mössbauer effect which results in the resonant absorption/emission of γ-rays and gives us a means to probe the
hyperfine interactions of an atoms nucleus and its surroundings.
A Mössbauer spectrometer system consists of a γ-ray source that is oscillated toward and away from the sample by a “Mössbauer
drive”, a collimator to filter the γ-rays, the sample, and a detector.

Figure 4.6.1 Schematic of Mössbauer Spectrometers. A = transmission; B = backscatter set up. Adapted from M. D. Dyar, D. G.
Agresti, M. W. Schaefer, C. A. Grant, and E. C. Sklute, Annu. Rev. Earth. Planet. Sci., 2006, 34 , 83. Copyright Annual Reviews
(2006).
Figure 4.6.2 hows the two basic set ups for a Mössbauer spectrometer. The Mössbauer drive oscillates the source so that the
incident γ-rays hitting the absorber have a range of energies due to the doppler effect. The energy scale for Mössbauer spectra (x-
axis) is generally in terms of the velocity of the source in mm/s. The source shown (57Co) is used to probe 57Fe in iron containing
samples because 57Co decays to 57Fe emitting a γ-ray of the right energy to be absorbed by 57Fe. To analyze other Mössbauer
isotopes other suitable sources are used. Fe is the most common element examined with Mössbauer spectroscopy because its 57Fe
isotope is abundant enough (2.2), has a low energy γ-ray, and a long lived excited nuclear state which are the requirements for
observable Mössbauer spectrum. Other elements that have isotopes with the required parameters for Mössbauer probing are seen in
Table 4.6.1.
Table 4.6.1 Elements with known Mössbauer isotopes and most commonly examined with Mössbauer spectroscopy.
Most commonly examined elements Fe, Ru, W, Ir, Au, Sn, Sb, Te, I, W, Ir, Eu, Gd, Dy, Er, Yb, Np

K, Ni, Zn, Ge, Kr, Tc, Ag, Xe, Cs, Ba, La, Hf, Ta, Re, Os, Pt, Hg, Ce,
Elements that exhibit Mössbauer effect
Pr, Nd, Sm, Tb, Ho, Tm, Lu, Th, Pa, U, Pu, Am

4.6.1 https://chem.libretexts.org/@go/page/55885
Mössbauer Spectra
The primary characteristics looked at in Mössbauer spectra are isomer shift (IS), quadrupole splitting (QS), and magnetic splitting
(MS or hyperfine splitting). These characteristics are effects caused by interactions of the absorbing nucleus with its environment.
Isomer shift is due to slightly different nuclear energy levels in the source and absorber due to differences in the s-electron
environment of the source and absorber. The oxidation state of an absorber nucleus is one characteristic that can be determined by
the IS of a spectra. For example due to greater d electron screening Fe2+ has less s-electron density than Fe3+ at its nucleus which
results in a greater positive IS for Fe2+.
For absorbers with nuclear angular momentum quantum number I > ½ the non-spherical charge distribution results in quadrupole
splitting of the energy states. For example Fe with a transition from I=1/2 to 3/2 will exhibit doublets of individual peaks in the
Mössbauer spectra due to quadrupole splitting of the nuclear states as shown in red in Figure 4.6.2.
In the presence of a magnetic field the interaction between the nuclear spin moments with the magnetic field removes all the
degeneracy of the energy levels resulting in the splitting of energy levels with nuclear spin I into 2I + 1 sublevels. Using Fe for an
example again, magnetic splitting will result in a sextet as shown in green in Figure 4.6.2. Notice that there are 8 possible
transitions shown, but only 6 occur. Due to the selection rule ІΔmIІ = 0, 1, the transitions represented as black arrows do not occur.

Figure 4.6.2 Characteristics of Mössbauer spectra related to nuclear energy levels. Adapted from M. D. Dyar, D. G. Agresti, M. W.
Schaefer, C. A. Grant, and E. C. Sklute, Annu. Rev. Earth. Planet. Sci., 2006, 34 , 83. Copyright Annual Reviews (2006).

Synthesis of Magnetite Nanoparticles


Numerous schemes have been devised to synthesize magnetite nanoparticles (nMag). The different methods of nMag synthesis can
be generally grouped as aqueous or non-aqueous according to the solvents used. Two of the most widely used and explored
methods for nMag synthesis are the aqueous co-precipitation method and the non-aqueous thermal decomposition method.
The co-precipitation method of nMag synthesis consists of precipitation of Fe3O4 (nMag) by addition of a strong base to a solution
of Fe2+ and Fe3+ salts in water. This method is very simple, inexpensive and produces highly crystalline nMag. The general size of
nMag produced by co-precipitation is in the 15 to 50 nm range and can be controlled by reaction conditions, however a large size
distribution of nanoparticles is produced by this method. Aggregation of particles is also observed with aqueous methods.
The thermal decomposition method consists of the high temperature thermal decomposition of an iron-oleate complex derived from
an iron precursor in the presence of surfactant in a high boiling point organic solvent under an inert atmosphere. For the many
variations of this synthetic method many different solvents and surfactants are used. However, in most every method nMag is
formed through the thermal decomposition of an iron-oleate complex to form highly crystalline nMag in the 5 to 40 nm range with
a very small size distribution. The size of nMag produced is a function of reaction temperature, the iron to surfactant ratio, and the
reaction time, and various methods are used that achieve good size control by manipulation of these parameters. The nMag
synthesized by organic methods is soluble in organic solvents because the nMag is stabilized by a surfactant surface coating with
the polar head group of the surfactant attached to and the hydrophobic tail extending away from the nMag (Figure 4.6.3). An
example of a thermal decomposition method is shown in Figure 4.6.3.

4.6.2 https://chem.libretexts.org/@go/page/55885
Figure 4.6.3 Top - The reaction equation for this method shows the iron precursor = iron oxo-hydrate, surfactant = oleic acid (OA),
and solvent = 1-octadecene. The intermediate iron-oleate complex which thermally decomposes to nMag is formed upon heating
the reaction mixture to the 320 °C reaction temperature. Bottom - TEM images showing size control by reaction time (time
decreases left to right, constant molar ratio Fe:OA = 1:4 mol, and constant reaction temp T = 320 °C) and small size distribution of
nMag. Right - Cartoon of surfactant coated nMag.

Mössbauer Analysis of Iron Oxide Nanoparticles


Spectra and Formula Calculations
Due to the potential applications of magnetite nanoparticles (Fe3O4, nMag) many methods have been devised for its synthesis.
However, stoichiometric Fe3O4 is not always achieved by different synthetic methods. B-site vacancies introduced into the cubic
inverse spinel crystal structure of nMag result in nonstoichiometric iron oxide of the formula (Fe3+)A(Fe(1-3x)2+ Fe(1+2X)3+Øx)BO4
where Ø represents B-site vacancy. The magnetic susceptibility which is key to most nMag applications decreases with increased
B-site vacancy hence the extent of B-site vacancy is important. The very high sensitivity of the Mössbauer spectrum to the
oxidation state and site occupancy of Fe3+ in cubic inverse spinel iron oxides makes Mössbauer spectroscopy valuable for
addressing the issues of whether or not the product of a synthetic method is actually nMag and the extent of B-site vacancy.
As with most analysis using multiple instrumental methods in conjunction is often helpful. This is exemplified by the use of XRD
along with Mössbauer spectroscopy in the following analysis. Figure 4.6.4 shows the XRD results and Mössbauer spectra
“magnetite” samples prepared by a Fe2+/Fe3+ co-precipitation (Mt025), hematite reduction by hydrogen (MtH2) and hematite
reduction with coal(MtC). The XRD analysis shows MtH2 and MT025 exhibiting only magnetite peaks while MtC shows the
presence of magnetite, maghemite, and hematite. This information becomes very useful when fitting peaks to the Mössbauer
spectra because it gives a chemical basis for peak fitting parameters and helps to fit the peaks correctly.

Figure 4.6.4 Mössbauer spectra (left) and corresponding XRD spectra of iron oxide sample prepared by different methods.
Adapted from A. L. Andrade, D. M. Souza, M. C. Pereira, J. D. Fabris, and R. Z. Domingues. J. Nanosci. Nanotechnol., 2009, 9,
2081.
Being that the iron occupies two local environments, the A-site and B site, and two species (Fe2+ and Fe3+) occupy the B-site one
might expect the spectrum to be a combination of 3 spectra, however delocalization of electrons or electron hopping between Fe2+
and Fe3+ in the B site causes the nuclei to sense an average valence in the B site thus the spectrum are fitted with two curves
accordingly. This is most easily seen in the Mt025 spectrum. The two fitted curves correspond to Fe3+ in the A-site and mixed
valance Fe2.5+ in the B-site. The isomer shift of the fitted curves can be used to determined which curve corresponds to which
valence. The isomer shift relative to the top fitted curve is reported to be 0.661 and the bottom fitted curve is 0.274 relative to αFe
thus the top fitted curve corresponds to less s-electron dense Fe2.5+. The magnetic splitting is quite apparent. In each of the spectra,
six peaks are present due to magnetic splitting of the nuclear energy states as explained previously. Quadrupole splitting is not so

4.6.3 https://chem.libretexts.org/@go/page/55885
apparent, but actually is present in the spectra. The three peaks to the left of the center of a spectrum should be spaced the same as
those to the right due to magnetic splitting alone since the energy level spacing between sublevels is equal. This is not the case in
the above spectra, because the higher energy I = 3/2 sublevels are split unevenly due to magnetic and quadrupole splitting
interactions.
Once the peaks have been fitted appropriately, determination of the extent of B-site vacancy in (Fe3+)A(Fe(1-3x)2+ Fe(1+2X)3+Øx)BO4
is a relatively simple matter. All one has to due to determine the number of vacancies (x) is solve the equation:
RAB 2 − 6x
= (4.6.3)
RAA 1 − 5x

where RAB or A = relative area


Area A or B site curve
(4.6.4)
Area of  both curves

of the curve for the B or A site respectively


The reasoning for this equation is as follows. Taking into account that the mixed valance Fe2.5+ curve is a result of paired
interaction between Fe2+ and Fe3+ the nonstochiometric chemical formula is (Fe3+)A(Fe(1-3x)2+Fe(1+2X)3+Øx)BO4. The relative
intensity (or relative area) of the Fe-A and Fe-B curves is very sensitive to stoichiometry because vacancies in the B-site reduce the
Fe-A curve and increase Fe-B curve intensities. This is due to the unpaired Fe5x3+ adding to the intensity of the Fe-A curve rather
than the Fe-B curve. Since the relative area is directly proportional to the number of Fe contributing to the spectrum the ratio of the
relative areas is equal to stoichiometric ratio of Fe2.5+ to Fe3+, which yields the above formula.
Example Calculation:
For MtH2 RAA/RAB = 1.89
Plugging x into the nonstoichiometric iron oxide formula yeilds:
RAB 2 − 6x
= (4.6.5)
RAA 1 − 5x

solving for x yields


RAA
2−
RAB
x = (4.6.6)
RAA
5   +  6
RAB

(Fe3+)A(Fe 1.95722+ Fe0.03563+)BO4 (very close to stoichiometric)


Figure 4.6.2 : Parameters and nonstoichiometric formulas for MtC, Mt025, and MtH2
Sample RAB/RAA X Chemical Formula

MtH2 1.89 0.007 (Fe3+)A(Fe0.9792+Fe1.0143+)BO4

MtC 1.66 0.024 (Fe3+)A(Fe0.9292+Fe1.0483+)BO4

Mt 025 1.60 0.029 (Fe3+)A(Fe0.9142+Fe1.0573+)BO4

Chemical Formulas of Nonstoichiometric Iron Oxide Nanoparticles from Mössbauer Spectroscopy


Chemical Formula Determination
Magnetite (Fe3O4) nanoparticles (n-Mag) are nanometer sized, superparamagnetic, have high saturation magnetization, high
magnetic susceptibility, and low toxicity. These properties could be utilized for many possible applications; hence, n-Mag has
attracted much attention in the scientific community. Some of the potential applications include drug delivery, hyperthermia agents,
MRI contrast agents, cell labeling, and cell separation to name a few.
The crystal structure of n-Mag is cubic inverse spinel with Fe3+ cations occupying the interstitial tetrahedral sites(A) and Fe3+
along with Fe2+ occupying the interstitial octahedral sites(B) of an FCC latticed of O2-. Including the site occupation and charge of
Fe, the n-Mag chemical formula can be written (Fe3+)A(Fe2+Fe3+)BO4. Non-stoichiometric iron oxide results from B-site vacancies
in the crystal structure. To maintain balanced charge and take into account the degree of B-site vacancies the iron oxide formula is

4.6.4 https://chem.libretexts.org/@go/page/55885
written (Fe3+)A(Fe(1-3x)2+ Fe(1+2X)3+Øx)BO4 where Ø represents B-site vacancy. The extent of B-site vacancy has a significant effect
on the magnetic properties of iron oxide and in the synthesis of n-Mag stoichiometric iron oxide is not guaranteed; therefore, B-site
vacancy warrants attention in iron oxide characterization, and can be addressed using Mössbauer spectroscopy.

4.6: Mössbauer Spectroscopy is shared under a CC BY 4.0 license and was authored, remixed, and/or curated by Pavan M. V. Raja & Andrew R.
Barron via source content that was edited to conform to the style and standards of the LibreTexts platform; a detailed edit history is available upon
request.

4.6.5 https://chem.libretexts.org/@go/page/55885
4.7: NMR Spectroscopy
Nuclear magnetic resonance spectroscopy (NMR) is a widely used and powerful method that takes advantage of the magnetic
properties of certain nuclei. The basic principle behind NMR is that some nuclei exist in specific nuclear spin states when exposed
to an external magnetic field. NMR observes transitions between these spin states that are specific to the particular nuclei in
question, as well as that nuclei's chemical environment. However, this only applies to nuclei whose spin, I, is not equal to 0, so
nuclei where I = 0 are ‘invisible’ to NMR spectroscopy. These properties have led to NMR being used to identify molecular
structures, monitor reactions, study metabolism in cells, and is used in medicine, biochemistry, physics, industry, and almost every
imaginable branch of science.
Theory

The chemical theory that underlies NMR spectroscopy depends on the intrinsic spin of the nucleus involved, described by the
quantum number S. Nuclei with a non-zero spin are always associated with a non-zero magnetic moment, as described by Equation
4.7.1, where μ is the magnetic moment, S is the spin, and γ is always non-zero. It is this magnetic moment that allows for NMR to

be used; therefore nuclei whose quantum spin is zero cannot be measured using NMR. Almost all isotopes that have both an even
number of protons and neutrons have no magnetic moment, and cannot be measured using NMR.
μ =  γ ⋅ S (4.7.1)

In the presence of an external magnetic field (B) for a nuclei with a spin I = 1/2, there are two spin states present of +1/2 and -1/2.
The difference in energy between these two states at a specific external magnetic field (Bx) are given by Equation 4.7.2, and are
shown in Figure 4.7.1 where E is energy, I is the spin of the nuclei, and μ is the magnetic moment of the specific nuclei being
analyzed. The difference in energy shown is always extremely small, so for NMR strong magnetic fields are required to further
separate the two energy states. At the applied magnetic fields used for NMR, most magnetic resonance frequencies tend to fall in
the radio frequency range.
E  =  μ ⋅ Bx /I (4.7.2)

Figure 4.7.1 The difference in energy between two spin states over a varying magnetic field B.
The reason NMR can differentiate between different elements and isotopes is due to the fact that each specific nuclide will only
absorb at a very specific frequency. This specificity means that NMR can generally detect one isotope at a time, and this results in
different types of NMR: such as 1H NMR, 13C NMR, and 31P NMR, to name only a few.
The subsequent absorbed frequency of any type of nuclei is not always constant, since electrons surrounding a nucleus can result in
an effect called nuclear shielding, where the magnetic field at the nucleus is changed (usually lowered) because of the surrounding
electron environment. This differentiation of a particular nucleus based upon its electronic (chemical) environment allows NMR be
used to identify structure. Since nuclei of the same type in different electron environments will be more or less shielded than
another, the difference in their environment (as observed by a difference in the surrounding magnetic field) is defined as the
chemical shift.
Instrumentation
An example of an NMR spectrometer is given in Figure 4.7.2. NMR spectroscopy works by varying the machine’s emitted
frequency over a small range while the sample is inside a constant magnetic field. Most of the magnets used in NMR machines to
create the magnetic field range from 6 to 24 T. The sample is placed within the magnet and surrounded by superconducting coils,
and is then subjected to a frequency from the radio wave source. A detector then interprets the results and sends it to the main
console.

4.7.1 https://chem.libretexts.org/@go/page/55887
Figure 4.7.2 Diagram of NMR spectrometer.
Interpreting NMR spectra

Chemical Shift
The different local chemical environments surrounding any particular nuclei causes them to resonate at slightly different
frequencies. This is a result of a nucleus being more or less shielded than another. This is called the chemical shift (δ). One factor
that affects chemical shift is the changing of electron density from around a nucleus, such as a bond to an electronegative group.
Hydrogen bonding also changes the electron density in 1H NMR, causing a larger shift. These frequency shifts are miniscule in
comparison to the fundamental NMR frequency differences, on a scale of Hz as compared to MHz. For this reason chemical shifts
(δ) are described by the unit ppm on an NMR spectra, 4.7.3, where Href = the resonance frequency of the reference, Hsub =
resonance frequency of the substance, and Hmachine = operating frequency of the spectrometer.
Href − Hsub
6
δ  =  ( )  × 10 (4.7.3)
Hmachine

Since the chemical shift (δ in ppm) is reported as a relative difference from some reference frequency, so a reference is required. In
1
H and 13C NMR, for example, tetramethylsilane (TMS, Si(CH3)4) is used as the reference. Chemical shifts can be used to identify
structural properties in a molecule based on our understanding of different chemical environments. Some examples of where
different chemical environments fall on a 1H NMR spectra are given in Table 4.7.1.
Table 4.7.1 Representative chemical shifts for organic groups in the 1H NMR.
Functional Group Chemical Shift Range (ppm)

Alkyl (e.g. methyl-CH3) ~1

Alkyl adjacent to oxygen (-CH2-O) 3-4

Alkene (=CH2) ~6

Alkyne (C-H) ~3

Aromatic 7-8

In Figure 4.7.3, an 1H NMR spectra of ethanol, we can see a clear example of chemical shift. There are three sets of peaks that
represent the six hydrogens of ethanol (C2H6O). The presence of three sets of peaks means that there are three different chemical
environments that the hydrogens can be found in: the terminal methyl (CH3) carbon’s three hydrogens, the two hydrogens on the
methylene (CH2) carbon adjacent to the oxygen, and the single hydrogen on the oxygen of the alcohol group (OH). Once we cover
spin-spin coupling, we will have the tools available to match these groups of hydrogens to their respective peaks.

4.7.2 https://chem.libretexts.org/@go/page/55887
Figure 4.7.3 : A 1H NMR spectra of ethanol (CH3CH2OH).
Spin-spin Coupling
Another useful property that allows NMR spectra to give structural information is called spin-spin coupling, which is caused by
spin coupling between NMR active nuclei that are not chemically identical. Different spin states interact through chemical bonds in
a molecule to give rise to this coupling, which occurs when a nuclei being examined is disturbed or influenced by a nearby nuclear
spin. In NMR spectra, this effect is shown through peak splitting that can give direct information concerning the connectivity of
atoms in a molecule. Nuclei which share the same chemical shift do not form splitting peaks in an NMR spectra.
In general, neighboring NMR active nuclei three or fewer bonds away lead to this splitting. The splitting is described by the
relationship where n neighboring nuclei result in n+1 peaks, and the area distribution can be seen in Pascal’s triangle (Figure
4.7.4). However, being adjacent to a strongly electronegative group such as oxygen can prevent spin-spin coupling. For example a

doublet would have two peaks with intensity ratios of 1:1, while a quartet would have four peaks of relative intensities 1:3:3:1. The
magnitude of the observed spin splitting depends on many factors and is given by the coupling constant J, which is in units of Hz.

Figure 4.7.4 : Pascal’s triangle.


Referring again to Figure 4.7.4, we have a good example of how spin-spin coupling manifests itself in an NMR spectra. In the
spectra we have three sets of peaks: a quartet, triplet, and a singlet. If we start with the terminal carbon’s hydrogens in ethanol,
using the n+1 rule we see that they have two hydrogens within three bonds (i.e., H-C-C-H), leading us to identify the triplet as the
peaks for the terminal carbon’s hydrogens. Looking next at the two central hydrogens, they have four NMR active nuclei within
three bonds (i.e., H-C-C-H), but there is no quintet on the spectra as might be expected. This can be explained by the fact that the
single hydrogen bonded to the oxygen is shielded from spin-spin coupling, so it must be a singlet and the two central hydrogens
form the quartet. We have now interpreted the NMR spectra of ethanol by identifying which nuclei correspond to each peak.
Peak Intensity
Mainly useful for proton NMR, the size of the peaks in the NMR spectra can give information concerning the number of nuclei that
gave rise to that peak. This is done by measuring the peak’s area using integration. Yet even without using integration the size of
different peaks can still give relative information about the number of nuclei. For example a singlet associated with three hydrogen
atoms would be about 3 times larger than a singlet associated with a single hydrogen atom.
This can also be seen in the example in Figure 4.7.3. If we integrated the area under each peak, we would find that the ratios of the
areas of the quartet, singlet, and triplet are approximately 2:1:3, respectively.

4.7.3 https://chem.libretexts.org/@go/page/55887
Limitations of NMR
Despite all of its upsides, there are several limitations that can make NMR analysis difficult or impossible in certain situations. One
such issue is that the desired isotope of an element that is needed for NMR analysis may have little or no natural abundance. For
example the natural abundance of 13C, the active isotope for carbon NMR, is about 11%, which works well for analysis. However,
in the case of oxygen the active isotope for NMR is 17O, which is only 0.035% naturally abundant. This means that there are
certain elements that can essentially never be measured through NMR.
Another problem is that some elements have an extremely low magnetic moment, μ. The sensitivity of NMR machines is based on
the magnetic moment of the specific element, but if the magnetic moment is too low it can be very difficult to obtain an NMR
spectra with enough peak intensity to properly analyze.

NMR Properties of the Element


Table 4.7.1 NMR properties of selected spin 1/2 nuclei. a Other spin 1/2 also exist.
Relative Receptivity as Compared
Isotope Natural Abundance (%) Relative NMR Frequency (MHz)
to 1H
1
H 99.985 100 1.00
3
H - 106.7 -
3
He 0.00013 76.2 5.8 x 10-7
13
C 1.11 25.1 1.8 x 10-4
15
N 0.37 10.1 3.9 x 10-6
19
F 100 94.1 8.3 x 10-1
29
Si 4.7 19.9 3.7 x 10-4
31
P 100 40.5 6.6 x 10-2
57
Fe 2.2 3.2 7.4 x 10-7
77
Se 7.6 19.1 5.3 x 10-4
89
Y 100 4.9 1.2 x 10-4
103
Rh 100 3.2 3.2 x 10-5
107
Ag 51.8 4.0 3.5 x 10-5
109
Ag 48.2 4.7 4.9 x 10-5
111
Cd 12.8 21.2 1.2 x 10-3
113
Cd 12.3 22.2 1.3 x 10-3
117
Sna 7.6 35.6 3.5 x 10-3
119
Sn 8.6 37.3 4.5 x 10-3
125
Tea 7.0 31.5 2.2 x 10-3
129
Xe 26.4 27.8 5.7 x 10-3
169
Tm 100 8.3 5.7 x 10-4
171
Yb 14.3 17.6 7.8 x 10-4
183
W 14.4 4.2 1.1 x 10-5
187
Os 1.6 2.3 2.0 x 10-7
195
Pt 33.8 21.4 3.4 x 10-3
199
Hg 16.8 17.9 9.8 x 10-4
203
Ti 29.5 57.1 5.7 x 10-2
205
Ti 70.5 57.6 1.4 x 10-1

4.7.4 https://chem.libretexts.org/@go/page/55887
Relative Receptivity as Compared
Isotope Natural Abundance (%) Relative NMR Frequency (MHz)
to 1H
207
Pb 22.6 20.9 2.0 x 10-1

Table 4.7.2 NMR properties of selected quadrupolar nuclei. a A spin 1/2 isotope also exists. b Other quadrupolar nuclei exist.
Natural Abundance Relative NMR Relative Receptivity Quadropole moment
Isotope Spin
(%) Frequency (%) as Compared to 1H (10-28 m2)
2H 1 0.015 15.4 1.5 x 10-6 2.8 x 10-3
6Li 1 7.4 14.7 6.3 x 10-4 -8 x 10-4
7Li 3/ 92.6 38.9 2.7 x 10-1 -4 x 10-2
2

9Be 3/ 100 14.1 1.4 x 10-2 5 x 10-2


2

10B 3 19.6 10.7 3.9 x 10-3 8.5 x 10-2


11B 3/ 80.4 32.1 1.3 x 10-1 4.1 x 10-2
2

14Na 1 99.6 7.2 1.0 x 10-3 1 x 10-2


17O 5/ 0.037 13.6 1.1 x 10-5 -2.6 x 10-2
2

23Na 5/ 100 26.5 9.3 x 10-2 1 x 10-1


2

25Mg 5/ 10.1 6.1 2.7 x 10-4 2.2 x 10-1


2

27Al 5/ 100 26.1 2.1 x 10-1 1.5 x 10-1


2

33S 3/ 0.76 7.7 1.7 x 10-5 -5.5 x 10-2


2

35Cl 3/ 75.5 9.8 3.6 x 10-3 -1 x 10-1


2

37Cl 3/ 24.5 8.2 6.7 x 10-4 -7.9 x 10-2


2

39Kb 3/ 93.1 4.7 4.8 x 10-4 4.9 x 10-2


2

43Ca 7/ 0.15 6.7 8.7 x 10-6 2 x 10-1


2

45Sc 7/ 100 24.3 3 x 10-1 -2.2 x 10-1


2

47Ti 5/ 7.3 5.6 1.5 x 10-4 2.9 x 10-1


2

49Ti 7/ 5.5 5.6 2.1 x 10-4 2.4 x 10-1


2

51Vb 7/ 99.8 26.3 3.8 x 10-1 -5 x 10-2


2

53Cr 3/ 9.6 5.7 8.6 x 10-5 3 x 10-2


2

55Mn 5/ 100 24.7 1.8 x 10-1 4 x 10-1


2

59Co 7/ 100 23.6 2.8 x 10-1 3.8 x 10-1


2

61Ni 3/ 1.2 8.9 4.1 x 10-1 1.6 x 10-1


2

63Cu 3/ 69.1 26.5 6.5 x 10-2 -2.1 x 10-1


2

65Cu 3/ 30.9 28.4 3.6 x 10-2 -2.0 x 10-1


2

67Zn 5/ 4.1 6.3 1.2 x 10-4 1.6 x 10-1


2

69Ga 3/ 60.4 24.0 4.2 x 10-2 1.9 x 10-1


2

71Ga 3/ 39.6 30.6 5.7 x 10-2 1.2 x 10-1


2

73Ge 9/ 7.8 3.5 1.1 x 10-4 -1.8 x 10-1


2

75As 3/ 100 17.2 2.5 x 10-2 2.9 x 10-1


2

79Br 3/ 50.5 25.1 4.0 x 10-2 3.7 x 10-1


2

81Br 3/ 49.5 27.1 4.9 x 10-2 3.1 x 10-1


2

4.7.5 https://chem.libretexts.org/@go/page/55887
Natural Abundance Relative NMR Relative Receptivity Quadropole moment
Isotope Spin
(%) Frequency (%) as Compared to 1H (10-28 m2)
87Rbb 3/ 27.9 32.8 4.9 x 10-2 1.3 x 10-1
2

87Sr 9/ 7.0 4.3 1.9 x 10-4 3 x 10-1


2

91Zr 5/ 11.2 9.3 1.1 x 10-3 -2.1 x 10-1


2

93Nb 9/ 100 24.5 4.9 x 10-1 -2.2 x 10-1


2

95Mo 5/ 15.7 6.5 5.1 x 10-4 ±1.2 x 10-1


2

97Mo 5/ 9.5 6.7 3.3 x 10-4 ±1.1


2

99Ru 5/ 12.7 4.6 1.5 x 10-4 7.6 x 10-2


2

101Ru 5/ 17.1 5.2 2.8 x 10-4 4.4 x 10-1


2

105Pd 5/ 22.2 4.6 2.5 x 10-4 8 x 10-1


2

115Inb 9/ 95.7 22.0 3.4 x 10-1 8.3 x 10-1


2

121Sb 5/ 57.3 24.0 9.3 x 10-2 -2.8 x 10-1


2

123Sb 7/ 42.7 13.0 2.0 x 10-2 3.6 x 10-1


2

127I 5/ 100 20.1 9.5 x 10-2 -7.9 x 10-1


2

131Xea 3/ 21.3 8.2 5.9 x 10-4 -1.2 x 10-1


2

133Cs 7/ 100 13.2 4.8 x 10-2 -3 x 10-3


2

137Bab 3/ 11.3 11.1 7.9 x 10-4 2.8 x 10-1


2

139La 7/ 99.9 14.2 6.0 x 10-2 2.2 x 10-1


2

177Hf 7/ 18.5 4.0 2.6 x 10-4 4.5


2

179Hf 9/ 13.8 2.5 7.4 x 10-5 5.1


2

181Ta 7/ 99.99 12.0 3.7 x 10-2 3


2

185Re 5/ 37.1 22.7 5.1 x 10-2 2.3


2

187Re 5/ 62.9 22.9 8.8 x 10-2 2.2


2

189Osa 3/ 16.1 7.8 3.9 x 10-4 8 x 10-1


2

191Ir 3/ 37.3 1.7 9.8 x 10-6 1.1


2

193Ir 3/ 62.7 1.9 2.1 x 10-5 1.0


2

197Au 3/ 100 1.7 2.6 x 10-5 5.9 x 10-1


2

201Hg 3/ 13.2 6.6 1.9 x 10-4 4.4 x 10-1


2

209Bi 9/ 100 16.2 1.4 x 10-1 -3.8 x 10-1


2

NMR Spin Coupling


The Basis of Spin Coupling
Nuclear magnetic resonance (NMR) signals arise when nuclei absorb a certain radio frequency and are excited from one spin state
to another. The exact frequency of electromagnetic radiation that the nucleus absorbs depends on the magnetic environment around
the nucleus. This magnetic environment is controlled mostly by the applied field, but is also affected by the magnetic moments of
nearby nuclei. Nuclei can be in one of many spin states Figure 4.7.5, giving rise to several possible magnetic environments for the
observed nucleus to resonate in. This causes the NMR signal for a nucleus to show up as a multiplet rather than a single peak.

4.7.6 https://chem.libretexts.org/@go/page/55887
Figure 4.7.5 The different spin states of a nucleus (I = 1/2) in a magnetic field. These different states increase or decrease the
effective magnetic field experienced by a nearby nucleus, allowing for two distinct signals.
When nuclei have a spin of I = 1/2 (as with protons), they can have two possible magnetic moments and thus split a single expected
NMR signal into two signals. When more than one nucleus affects the magnetic environment of the nucleus being examined,
complex multiplets form as each nucleus splits the signal into two additional peaks. If those nuclei are magnetically equivalent to
each other, then some of the signals overlap to form peaks with different relative intensities. The multiplet pattern can be predicted
by Pascal’s triangle (Figure 4.7.6), looking at the nth row, where n = number of nuclei equivalent to each other but not equivalent to
the one being examined. In this case, the number of peaks in the multiplet is equal to n + 1

Figure 4.7.6 Pascal’s triangle predicts the number of peaks in a multiplet and their relative intensities.
When there is more than one type of nucleus splitting an NMR signal, then the signal changes from a multiplet to a group of
multiplets (Figure 4.7.7). This is caused by the different coupling constants associated with different types of nuclei. Each nucleus
splits the NMR signal by a different width, so the peaks no longer overlap to form peaks with different relative intensities.

Figure 4.7.7 The splitting tree of different types of multiplets.


1
When nuclei have I > /2, they have more than two possible magnetic moments and thus split NMR signals into more than two
peaks. The number of peaks expected is 2I + 1, corresponding to the number of possible orientations of the magnetic moment. In
reality however, some of these peaks may be obscured due to quadrupolar relaxation. As a result, most NMR focuses on I = 1/2
nuclei such as 1H, 13C, and 31P.
Multiplets are centered around the chemical shift expected for a nucleus had its signal not been split. The total area of a multiplet
corresponds to the number of nuclei resonating at the given frequency.
Spin Coupling in molecules
Looking at actual molecules raises questions about which nuclei can cause splitting to occur. First of all, it is important to realize
that only nuclei with I ≠ 0 will show up in an NMR spectrum. When I = 0, there is only one possible spin state and obviously the
nucleus cannot flip between states. Since the NMR signal is based on the absorption of radio frequency as a nucleus transitions
from one spin state to another, I = 0 nuclei do not show up on NMR. In addition, they do not cause splitting of other NMR signals
because they only have one possible magnetic moment. This simplifies NMR spectra, in particular of organic and organometallic
compounds, greatly, since the majority of carbon atoms are 12C, which have I = 0.
For a nucleus to cause splitting, it must be close enough to the nucleus being observed to affect its magnetic environment. The
splitting technically occurs through bonds, not through space, so as a general rule, only nuclei separated by three or fewer bonds
can split each other. However, even if a nucleus is close enough to another, it may not cause splitting. For splitting to occur, the

4.7.7 https://chem.libretexts.org/@go/page/55887
nuclei must also be non-equivalent. To see how these factors affect real NMR spectra, consider the spectrum for chloroethane
(Figure 4.7.8).

Figure 4.7.8 The NMR spectrum for chloroethane. Adapted from A. M. Castillo, L. Patiny, and J. Wist. J. Magn. Reson., 2010,
209, 123.
Notice that in Figure 4.7.8 there are two groups of peaks in the spectrum for chloroethane, a triplet and a quartet. These arise from
the two different types of I ≠ 0 nuclei in the molecule, the protons on the methyl and methylene groups. The multiplet
corresponding to the CH3 protons has a relative integration (peak area) of three (one for each proton) and is split by the two
methylene protons (n = 2), which results in n + 1 peaks, i.e., 3 which is a triplet. The multiplet corresponding to the CH2 protons
has an integration of two (one for each proton) and is split by the three methyl protons ((n = 3) which results in n + 1 peaks, i.e., 4
which is a quartet. Each group of nuclei splits the other, so in this way, they are coupled.
Coupling Constants
The difference (in Hz) between the peaks of a mulitplet is called the coupling constant. It is particular to the types of nuclei that
give rise to the multiplet, and is independent of the field strength of the NMR instrument used. For this reason, the coupling
constant is given in Hz, not ppm. The coupling constant for many common pairs of nuclei are known (Table 4.7.3), and this can
help when interpreting spectra.
Table 4.7.3 Typical coupling constants for various organic structural types.
Structural Type

0.5 - 3

12 - 15

12 - 18

7 - 12

0.5 - 3

3 - 11

4.7.8 https://chem.libretexts.org/@go/page/55887
Structural Type

2-3

ortho = 6 - 9; meta = 1 - 3; para = 0 - 1

Coupling constants are sometimes written nJ to denote the number of bonds (n) between the coupled nuclei. Alternatively, they are
written as J(H-H) or JHH to indicate the coupling is between two hydrogen atoms. Thus, a coupling constant between a
phosphorous atom and a hydrogen would be written as J(P-H) or JPH. Coupling constants are calculated empirically by measuring
the distance between the peaks of a multiplet, and are expressed in Hz.
Coupling constants may be calculated from spectra using frequency or chemical shift data. Consider the spectrum of chloroethane
shown in Figure 4.7.5 and the frequency of the peaks (collected on a 60 MHz spectrometer) give in Table 4.7.4.

Figure 4.7.5 1H NMR spectrum of chloroethane. Peak positions for labeled peaks are given in Table 4.7.4 .
Table 4.7.4 Chemical shift in ppm and Hz for all peaks in the 1H NMR spectrum of chloroethane. Peak labels are given in Figure 4.7.5 .
Peak Label δ (ppm) v (Hz)

a 3.7805 226.83

b 3.6628 219.77

c 3.5452 212.71

d 3.4275 205.65

e 1.3646 81.88

f 1.2470 74.82

g 1.1293 67.76

To determine the coupling constant for a multiplet (in this case, the quartet in Figure 4.7.3, the difference in frequency (ν) between
each peak is calculated and the average of this value provides the coupling constant in Hz. For example using the data from Table
4.7.4:

Frequency of peak c - frequency of peak d = 212.71 Hz - 205.65 Hz = 7.06 Hz


Frequency of peak b - frequency of peak c = 219.77 Hz – 212.71 Hz = 7.06 Hz
Frequency of peak a - frequency of peak b = 226.83 Hz – 219.77 Hz = 7.06 Hz
Average: 7.06 Hz
∴ J(H-H) = 7.06 Hz
In this case the difference in frequency between each set of peaks is the same and therefore an average determination is not strictly
necessary. In fact for 1st order spectra they should be the same. However, in some cases the peak picking programs used will result
in small variations, and thus it is necessary to take the trouble to calculate a true average.
To determine the coupling constant of the same multiplet using chemical shift data (δ), calculate the difference in ppm between
each peak and average the values. Then multiply the chemical shift by the spectrometer field strength (in this case 60 MHz), in
order to convert the value from ppm to Hz:

4.7.9 https://chem.libretexts.org/@go/page/55887
Chemical shift of peak c - chemical shift of peak d = 3.5452 ppm – 3.4275 ppm = 0.1177 ppm
Chemical shift of peak b - chemical shift of peak c = 3.6628 ppm – 3.5452 ppm = 0.1176 ppm
Chemical shift of peak a - chemical shift of peak b = 3.7805 ppm – 3.6628 ppm = 0.1177 ppm
Average: 0.1176 ppm
Average difference in ppm x frequency of the NMR spectrometer = 0.1176 ppm x 60 MHz = 7.056 Hz
∴ J(H-H) = 7.06 Hz
Calculate the coupling constant for triplet in the spectrum for chloroethane (Figure 4.7.6) using the data from Table 4.7.5.
Using frequency data:
Frequency of peak f - frequency of peak g = 74.82 Hz – 67.76 Hz = 7.06 Hz
Frequency of peak e - frequency of peak f = 81.88 Hz – 74.82 Hz = 7.06 Hz
Average = 7.06 Hz
∴ J(H-H) = 7.06 Hz
Alternatively, using chemical shift data:
Chemical shift of peak f - chemical shift of peak g = 1.2470 ppm – 1.1293 ppm = 0.1177 ppm
Chemical shift of peak e - chemical shift of peak f = 1.3646 ppm – 1.2470 ppm = 0.1176 ppm
Average = 0.11765 ppm
0.11765 ppm x 60 MHz = 7.059 Hz
∴ J(H-H) = 7.06 Hz

Notice the coupling constant for this multiplet is the same as that in the example. This is to be expected since the two multiplets are
coupled with each other.
Second-Order Coupling
When coupled nuclei have similar chemical shifts (more specifically, when Δν is similar in magnitude to J), second-order coupling
or strong coupling can occur. In its most basic form, second-order coupling results in “roofing” (Figure 4.7.6). The coupled
multiplets point to or lean toward each other, and the effect becomes more noticeable as Δν decreases. The multiplets also become
off-centered with second-order coupling. The midpoint between the peaks no longer corresponds exactly to the chemical shift.

Figure 4.7.6 Roofing can be seen in the NMR spectrum of chloroethane. Adapted from A. M. Castillo, L. Patiny, and J. Wist, J.
Magn. Reson., 2010, 209, 123.
In more drastic cases of strong coupling (when Δν ≈ J), multiplets can merge to create deceptively simple patterns. Or, if more than
two spins are involved, entirely new peaks can appear, making it difficult to interpret the spectrum manually. Second-order
coupling can often be converted into first-order coupling by using a spectrometer with a higher field strength. This works by
altering the Δν (which is dependent on the field strength), while J (which is independent of the field strength) stays the same.

P-31 NMR Spectroscopy


Phosphorus-31 nuclear magnetic resonance (31P NMR) is conceptually the same as proton (1H) NMR. The 31P nucleus is useful in
NMR spectroscopy due to its relatively high gyromagnetic ratio (17.235 MHzT-1). For comparison, the gyromagnetic ratios of 1H
and 13C are (42.576 MHz T-1) and (10.705 MHz T-1), respectively. Furthermore, 31P has a 100% natural isotopic abundance. Like
the 1H nucleus, the 31P nucleus has a nuclear spin of 1/2 which makes spectra relatively easy to interpret. 31P NMR is an excellent
technique for studying phosphorus containing compounds, such as organic compounds and metal coordination complexes.

4.7.10 https://chem.libretexts.org/@go/page/55887
Differences Between 1H and 31P NMR
There are certain significant differences between 1H and 31P NMR. While 1H NMR spectra is referenced to tetramethylsilane
[Si(CH3)4], the chemical shifts in 31P NMR are typically reported relative to 85% phosphoric acid (δ = 0 ppm), which is used as an
external standard due to its reactivity. However, trimethyl phosphite, P(OCH3)3, is also used since unlike phosphoric acid its shift
(δ = 140 ppm) is not dependent on concentration or pH. As in 1H NMR, positive chemical shifts correspond to a downfield shift
from the standard. However, prior to the mid-1970s, the convention was the opposite. As a result, older texts and papers report
shifts using the opposite sign. Chemical shifts in 31P NMR commonly depend on the concentration of the sample, the solvent used,
and the presence of other compounds. This is because the external standard does not take into account the bulk properties of the
sample. As a result, reported chemical shifts for the same compound could vary by 1 ppm or more, especially for phosphate groups
(P=O). 31P NMR spectra are often recorded with all proton signals decoupled, i.e., 31P-{1H}, as is done with 13C NMR. This gives
rise to single, sharp signals per unique 31P nucleus. Herein, we will consider both coupled and decoupled spectra.
Interpreting Spectra

As in 1H NMR, phosphorus signals occur at different frequencies depending on the electron environment of each phosphorus
nucleus Figure 4.7.7. In this section we will study a few examples of phosphorus compounds with varying chemical shifts and
coupling to other nuclei.

Figure 4.7.7 Chemical shift ranges for different types of phosphorus compounds.
Different Phosphorus Environments and their Coupling to 1H
Consider the structure of 2,6,7-trioxa-1,4-diphosphabicyclo[2.2.2]octane [Pα(OCH2)3Pβ] shown in Figure 4.7.8. The subscripts α
and β are simply used to differentiate the two phosphorus nuclei. According to Table 1, we expect the shift of Pα to be downfield of
the phosphoric acid standard, roughly around 125 ppm to 140 ppm and the shift of Pβ to be upfield of the standard, between -5 ppm
and -70 ppm. In the decoupled spectrum shown in Figure 4.7.8, we can assign the phosphorus shift at 90.0 ppm to Pα and the shift
at -67.0 ppm to Pβ.

4.7.11 https://chem.libretexts.org/@go/page/55887
Figure 4.7.8 Structure and decoupled 31P spectrum (31P-{1H}) of Pα(OCH2)3Pβ.
Figure 4.7.9 shows the coupling of the phosphorus signals to the protons in the compound. We expect a stronger coupling for Pβ
because there are only two bonds separating Pβ from H, whereas three bonds separate Pαfrom H (JPCH > JPOCH). Indeed, JPCH = 8.9
Hz and JPOCH = 2.6 Hz, corroborating our peak assignments above.

Figure 4.7.9 The 31P spin coupled spectrum of Pα(OCH2)3Pβ.


Finally, Figure 4.7.10 shows the 1H spectrum of Pα(OCH2)3Pβ (Figure ), which shows a doublet of doublets for the proton
4.7.11

signal due to coupling to the two phosphorus nuclei.

Figure 4.7.10 1H spectrum of Pα(OCH2)3Pβ and proton splitting pattern due to phosphorus.
As suggested by the data in Figure 4.7.7 we can predict and observe changes in phosphorus chemical shift by changing the
coordination of P. Thus for the series of compounds with the structure shown in Figure 4.7.11 the different chemical shifts
corresponding to different phosphorus compounds are shown in Table 4.7.3.

Figure 4.7.11 Structure of [XPα(OCH2)3PβY].


Table 4.7.5 31P chemical shifts for variable coordination of [XPα(OCH2)3PβY] (Figure 4.7.11 ). Data from K. J. Coskran and J. G. Verkade,
Inorg. Chem., 1965, 4, 1655.
X Y Pα chemical shift (ppm) Pβ chemical shift (ppm)

- - 90.0 -67.0

O O -18.1 6.4

S - 51.8 -70.6

4.7.12 https://chem.libretexts.org/@go/page/55887
Coupling to Fluorine
19F NMR is very similar to 31P NMR in that 19F has spin 1/2 and is a 100% abundant isotope. As a result, 19F NMR is a great
technique for fluorine-containing compounds and allows observance of P-F coupling. The coupled 31P and 19F NMR spectra of
ethoxybis(trifluoromethyl)phosphine, P(CF3)2(OCH2CH3), are shown in Figure 4.7.11. It is worth noting the splitting due to JPCF =
86.6 Hz.

Figure 4.7.11 Structure, 31P-{1H} spectrum (A), and 19F-{1H} spectrum (B) for P(CF3)2(OCH2CH3). Data from K. J. Packer, J.
Chem. Soc., 1963, 960.
31P - 1H Coupling
Consider the structure of dimethyl phosphonate, OPH(OCH3)2, shown in Figure 4.7.12. As the phosphorus nucleus is coupled to a
hydrogen nucleus bound directly to it, that is, a coupling separated by a single bond, we expect JPH to be very high. Indeed, the
separation is so large (715 Hz) that one could easily mistake the split peak for two peaks corresponding to two different phosphorus
nuclei.

Figure 4.7.12 Structure and 31P NMR spectrum of OPH(OCH3)2 with only the OCH3 protons decoupled.
This strong coupling could also lead us astray when we consider the 1H NMR spectrum of dimethyl phosphonate (Figure 4.7.13).
Here we observe two very small peaks corresponding to the phosphine proton. The peaks are separated by such a large distance and
are so small relative to the methoxy doublet (ratio of 1:1:12), that it would be easy to confuse them for an impurity. To assign the
small doublet, we could decouple the phosphorus signal at 11 ppm, which will cause this peak to collapse into a singlet.

4.7.13 https://chem.libretexts.org/@go/page/55887
Figure 4.7.13 1H spectrum of OPH(OCH3)2. Data from K. Moedritzer, J. Inorg. Nucl. Chem., 1961, 22, 19.
Obtaining 31P Spectra

Sample Preparation
Unlike 13C NMR, which requires high sample concentrations due to the low isotopic abundance of 13C, 31P sample preparation is
very similar to 1H sample preparation. As in other NMR experiments, a 31P NMR sample must be free of particulate matter. A
reasonable concentration is 2-10 mg of sample dissolved in 0.6-1.0 mL of solvent. If needed, the solution can be filtered through a
small glass fiber. Note that the solid will not be analyzed in the NMR experiment. Unlike 1H NMR, however, the sample does not
to be dissolved in a deuterated solvent since common solvents do not have 31P nuclei to contribute to spectra. This is true, of
course, only if a 1H NMR spectrum is not to be obtained from this sample. Being able to use non-deuterated solvents offers many
advantages to 31P NMR, such as the simplicity of assaying purity and monitoring reactions, which will be discussed later.

Instrument Operation
Instrument operation will vary according to instrumentation and software available. However, there are a few important aspects to
instrument operation relevant to 31P NMR. The instrument probe, which excites nuclear spins and detects chemical shifts, must be
set up appropriately for a 31P NMR experiment. For an instrument with a multinuclear probe, it is a simple matter to access the
NMR software and make the switch to a 31P experiment. This will select the appropriate frequency for 31P. For an instrument which
has separate probes for different nuclei, it is imperative that one be trained by an expert user in changing the probes on the
spectrometer.
Before running the NMR experiment, consider whether the 31P spectrum should include coupling to protons. Note that 31P spectra
are typically reported with all protons decoupled, i.e., 311P-{1H}. This is usually the default setting for a 31P NMR experiment. To
change the coupling setting, follow the instructions specific to your NMR instrument software.
As mentioned previously, chemical shifts in 31P NMR are reported relative to 85% phosphoric acid. This must be an external
standard due to the high reactivity of phosphoric acid. One method for standardizing an experiment uses a coaxial tube inserted into
the sample NMR tube (Figure 4.7.14). The 85% H3PO4 signal will appear as part of the sample NMR spectrum and can thus be set
to 0 ppm.

Figure 4.7.14 Diagram of NMR tube with inserted coaxial reference insert. Image Courtesy of Wilmad-LabGlass; All Rights
Reserved.

4.7.14 https://chem.libretexts.org/@go/page/55887
Another way to reference an NMR spectrum is to use a 85% H3PO4 standard sample. These can be prepared in the laboratory or
purchased commercially. To allow for long term use, these samples are typically vacuum sealed, as opposed to capped the way
NMR samples typically are. The procedure for using a separate reference is as follows.
1. Insert NMR sample tube into spectrometer.
2. Tune the 31P probe and shim the magnetic field according to your individual instrument procedure.
3. Remove NMR sample tube and insert H3PO4 reference tube into spectrometer.
4. Begin NMR experiment. As scans proceed, perform a fourier transform and set the phosphorus signal to 0 ppm. Continue to
reference spectrum until the shift stops changing.
5. Stop experiment.
6. Remove H3PO4 reference tube and insert NMR sample into spectrometer.
7. Run NMR experiment without changing the referencing of the spectrum.
31P NMR Applications

Assaying Sample Purity


31
P NMR spectroscopy gives rise to single sharp peaks that facilitate differentiating phosphorus-containing species, such as starting
materials from products. For this reason, 31P NMR is a quick and simple technique for assaying sample purity. Beware, however,
that a “clean” 31P spectrum does not necessarily suggest a pure compound, only a mixture free of phosphorus-containing
contaminants.
31
P NMR can also be used to determine the optical purity of a chiral sample. Adding an enantiomer to the chiral mixture to form
two different diastereomers will give rise to two unique chemical shifts in the 31P spectrum. The ratio of these peaks can then be
compared to determine optical purity.

Monitoring Reactions
As suggested in the previous section, 31P NMR can be used to monitor a reaction involving phosphorus compounds. Consider the
reaction between a slight excess of organic diphosphine ligand and a nickel(0) bis-cyclooctadiene, Figure 4.7.15.

Figure 4.7.15 Reaction between diphosphine ligand and nickel


The reaction can be followed by 31P NMR by simply taking a small aliquot from the reaction mixture and adding it to an NMR
tube, filtering as needed. The sample is then used to acquire a 31P NMR spectrum and the procedure can be repeated at different
reaction times. The data acquired for these experiments is found in Figure 4.7.16. The changing in 31P peak intensity can be used to
monitor the reaction, which begins with a single signal at -4.40 ppm, corresponding to the free diphosphine ligand. After an hour, a
new signal appears at 41.05 ppm, corresponding the the diphosphine nickel complex. The downfield peak grows as the reaction
proceeds relative to the upfield peak. No change is observed between four and five hours, suggesting the conclusion of the reaction.

Figure 4.7.16 31P-{1H} NMR spectra of the reaction of diphosphine ligand with nickel(0) bis-cyclooctadiene to make a
diphosphine nickel complex over time.

4.7.15 https://chem.libretexts.org/@go/page/55887
There are a number of advantages for using 31P for reaction monitoring when available as compared to 1H NMR:
There is no need for a deuterated solvent, which simplifies sample preparation and saves time and resources.
The 31P spectrum is simple and can be analyzed quickly. The corresponding 1H NMR spectra for the above reaction would
include a number of overlapping peaks for the two phosphorus species as well as peaks for both free and bound cyclooctadiene
ligand.
Purification of product is also easy assayed.
31
P NMR does not eliminate the need for 1H NMR chacterization, as impurities lacking phosphorus will not appear in a 31P
experiment. However, at the completion of the reaction, both the crude and purified products can be easily analyzed by both 1H and
31
P NMR spectroscopy.
Measuring Epoxide Content of Carbon Nanomaterials
One can measure the amount of epoxide on nanomaterials such as carbon nanotubes and fullerenes by monitoring a reaction
involving phosphorus compounds in a similar manner to that described above. This technique uses the catalytic reaction of
methyltrioxorhenium (Figure 4.7.17). An epoxide reacts with methyltrioxorhenium to form a five membered ring. In the presence
of triphenylphosphine (PPH3), the catalyst is regenerated, forming an alkene and triphenylphosphine oxide (OPPh3). The same
reaction can be applied to carbon nanostructures and used to quantify the amount of epoxide on the nanomaterial. Figure 4.7.18
illustrates the quantification of epoxide on a carbon nanotube.

Figure 4.7.17

Figure 4.7.18
Because the amount of initial PPh3 used in the reaction is known, the relative amounts of PPh3 and OPPh3can be used to
stoichiometrically determine the amount of epoxide on the nanotube. 31P NMR spectroscopy is used to determine the relative
amounts of PPh3 and OPPh3 (Figure 4.7.19).

Figure 4.7.19 31P spectrum of experiment before addition of Re complex (top) and at the completion of experiment (bottom).
The integration of the two 31P signals is used to quantify the amount of epoxide on the nanotube according to 4.7.4.
area of  OP P H3  peak
M oles of  Epoxide  =   ×  moles P P h3 (4.7.4)
area of  P P h3  peak

Thus, from a known quantity of PPh3, one can find the amount of OPPh3 formed and relate it stoichiometrically to the amount of
epoxide on the nanotube. Not only does this experiment allow for such quantification, it is also unaffected by the presence of the
many different species present in the experiment. This is because the compounds of interest, PPh3 and OPPh3, are the only ones
that are characterized by 31P NMR spectroscopy.
Conclusion
31
P NMR spectroscopy is a simple technique that can be used alongside 1H NMR to characterize phosphorus-containing
compounds. When used on its own, the biggest difference from 1H NMR is that there is no need to utilize deuterated solvents. This
advantage leads to many different applications of 31P NMR, such as assaying purity and monitoring reactions.

4.7.16 https://chem.libretexts.org/@go/page/55887
NMR Spectroscopy of Stereoisomers
Nuclear magnetic resonance (NMR) spectroscopy is a very useful tool used widely in modern organic chemistry. It exploits the
differences in the magnetic properties of different nuclei in a molecule to yield information about the chemical environment of the
nuclei, and subsequently the molecule, in question. NMR analysis lends itself to scientists more easily than say the more cryptic
data achieved form ultraviolet or infared spectra because the differences in magnetic properties lend themselves to scientists very
well. The chemical shifts that are characteristic of different chemical environments and the multiplicity of the peaks fit well with
our conception of the way molecules are structured.
Using NMR spectroscopy, we can differentiate between constitutional isomers, stereoisomers, and enantiomers. The later two of
these three classifications require close examination of the differences in NMR spectra associated with changes in chemical
environment due to symmetry differences; however, the differentiation of constitutional isomers can be easily obtained.
Constitutional Isomerism
Nuclei both posses charge and spin, or angular momentum, and from basic physics we know that a spinning charge generates a
magnetic moment. The specific nature of this magnetic moment is the main concern of NMR spectroscopy.
For proton NMR, the local chemical environment makes different protons in a molecule resonate at different frequencies. This
difference in resonance frequencies can be converted into a chemical shift (δ) for each nucleus being studied. Because each
chemical environment results in a different chemical shift, one can easily assign peaks in the NMR data to specific functional
groups based upon president. Presidents for chemical shifts can be found in any number of basic NMR text. For example, Figure
4.7.20 shows the spectra of ethyl formate and benzyl acetate. In the lower spectra, benzyl acetate, notice peaks at δ = 1.3, 4.2, and

8.0 ppm characteristic of the primary, secondary, and aromatic protons, respectively, present in the molecule. In the spectra of ethyl
formate (Figure 4.7.20 b), notice that the number of peaks is is the same as that of benzyl acetate (Figure 4.7.20 a); however, the
multiplicity of peaks and their shifts is very different.

Figure 4.7.20 1H NMR spectra of (a) ethyl formate and (b) benzyl acetate.
The difference between these two spectra is due to geminal spin-spin coupling. Spin-spin coupling is the result of magnetic
interaction between individual protons transmitted by the bonding electrons between the protons. This spin-spin coupling results in
the speak splitting we see in the NMR data. One of the benefits of NMR spectroscopy is the sensitivity to very slight changes in
chemical environment.
Stereoisomerism

Diastereomers
Based on their definition, diastereomers are stereoisomers that are not mirror images of each other and are not superimposable. In
general, diastereomers have differing reactivity and physical properties. One common example is the difference between threose
and erythrose (Figure 4.7.21.

4.7.17 https://chem.libretexts.org/@go/page/55887
Figure 4.7.21 The structures of threose and erythrose.
As one can see from Figure 4.7.22, these chemicals are very similar each having the empirical formula of C4H7O4. One may
wonder: how are these slight differences in chemical structure represented in NMR? To answer this question, we must look at the
Newman projections for a molecule of the general structure Figure 4.7.22.

Figure 4.7.22 Newman projections of a general diastereomer.


One can easily notice that the two protons represented are always located in different chemical environments. This is true because
the R group makes the proton resonance frequencies v1(I) ≠ v2(III), v2(I) ≠ v1(II), and v2(II) ≠ v1(III). Thus, diastereomers have
different vicinal proton-proton couplings and the resulting chemical shifts can be used to identify the isomeric makeup of the
sample.

Enantiomers
Enantiomers are compounds with a chiral center. In other words, they are non-superimposable mirror images. Unlike
diastereomers, the only difference between enantiomers is their interaction with polarized light. Unfortunately, this
indistinguishability of racemates includes NMR spectra. Thus, in order to differentiate between enantiomers, we must make use of
an optically active solvent also called a chiral derivatizing agent (CDA). The first CDA was (α-methoxy-α-
(trifluoromethyl)phenylacetic acid) (MTPA also known as Mosher's acid) (Figure 4.7.23).

Figure 4.7.23 The structure of the S-isomer of Mosher's Acid (S-MTPA)


Now, many CDAs exist and are readily available. It should also be noted that CDA development is a current area of active
research. In simple terms, one can think of the CDA turning an enantiomeric mixture into a mixture of diastereomeric complexes,
producing doublets where each half of the doublet corresponds to each diastereomer, which we already know how to analyze. The
resultant peak splitting in the NMR spectra due to diastereomeric interaction can easily determine optical purity. In order to do this,
one may simply integrate the peaks corresponding to the different enantiomers thus yielding optical purity of incompletely resolved
racemates. One thing of note when performing this experiment is that this interaction between the enantiomeric compounds and the
solvent, and thus the magnitude of the splitting, depends upon the asymmetry or chirality of the solvent, the intermolecular
interaction between the compound and the solvent, and thus the temperature. Thus, it is helpful to compare the spectra of the
enantiomer-CDA mixture with that of the pure enantiomer so that changes in chemical shift can be easily noted.

4.7.18 https://chem.libretexts.org/@go/page/55887
Basics of Solid-State NMR
NMR stands for nuclear magnetic resonance and functions as a powerful tool for chemical characterization. Even though NMR is
used mainly for liquids and solutions, technology has progressed to where NMR of solids can be obtained with ease. Aptly named
as solid state NMR, the expansion of usable phases has invariably increased our ability to identify chemical compounds. The
reason behind difficulties using the solid state lie in the fact that solids are never uniform. When put through a standard NMR, line
broadening interactions cannot be removed by rapid molecular motions, which results in unwieldy wide lines which provide little
to no useful information. The difference is so staggering that lines broaden by hundreds to thousands of hertz as opposed to less
than 0.1 Hz in solution when using an I = 1/2 spin nucleus.
A process known as magic angle spinning (MAS), where the sample is tilted at a specific angle, is used in order to overcome line
broadening interactions and achieve usable peak resolutions. In order to understand solid state NMR, its history, operating chemical
and mathematical principles, and distinctions from gas phase/solution NMR will be explained.
History
The first notable contribution to what we know today as NMR was Wolfgang Pauli’s (Figure 4.7.24) prediction of nuclear spin in
1926. In 1932 Otto Stern (Figure 4.7.25) used molecular beams and detected nuclear magnetic moments.

Figure 4.7.26 German physicist Otto Stern (1888 - 1969)


Four years later, Gorter performed the first NMR experiment with lithium fluoride (LiF) and hydrated potassium alum
(K[Al(SO4)2]•12H2O) at low temperatures. Unfortunately, he was unable to characterize the molecules and the first successful
NMR for a solution of water was taken in 1945 by Felix Bloch (Figure 4.7.27). In the same year, Edward Mills Purcell (Figure
1
4.7.27) managed the first successful NMR for the solid paraffin. Continuing their research, Bloch obtained the H NMR of ethanol

4.7.19 https://chem.libretexts.org/@go/page/55887
and Purcell obtained that of paraffin in 1949. In the same year, the chemical significance of chemical shifts was discovered. Finally,
high resolution solid state NMR was made possible in 1958 by the discovery of magic angle spinning.

Figure 4.7.28 American physicist Edward Mills Purcell (1912-1997).


How it Works: From Machine to Graph
NMR spectroscopy works by measuring the nuclear shielding, which can also be seen as the electron density, of a particular
element. Nuclear shielding is affected by the chemical environment, as different neighboring atoms will have different effects on
nuclear shielding, as electronegative atoms will tend to decrease shielding and vice versa. NMR requires the elements analyzed to
have a spin state greater than zero. Commonly used elements are 1H, 13C, and 29Si. Once inside the NMR machine, the presence of
a magnetic field splits the spin states (Figure 4.7.29).

Figure 4.7.29 Spin state splitting as a function of applied magnetic field.


From (Figure 4.7.29 we see that a spin state of 1/2 is split into two spin states. As spin state value increases, so does the number of
spin states. A spin of 1 will have three spin states, 3/2 will have four spin states, and so on. However, higher spin states increases the
difficulty to accurately read NMR results due to confounding peaks and decreased resolution, so spin states of ½ are generally
preferred. The E, or radiofrequency shown in (Figure 4.7.29 can be described by 4.7.5, where µ is the magnetic moment, a

4.7.20 https://chem.libretexts.org/@go/page/55887
property intrinsic to each particular element. This constant can be derived from , where ϒ is the gyromagnetic ratio, another
4.7.6

element dependent quantity, h is Planck’s constant, and I is the spin.


E  =  μB0 H0 (4.7.5)

1/2
μ  =  γh(I (I + 1)) (4.7.6)

In 4.7.5 can have E substituted for hν, leading to 4.7.7, which can solve for the NMR resonance frequency (v).
hν   =  μB0 H0 (4.7.7)

Using the frequency (v), the δ, or expected chemical shift may be computed using 4.7.8.
(νobserved − νref erence )
δ  =   (4.7.8)
νspectrometer

Delta (δ) is observed in ppm and gives the distance from a set reference. Delta is directly related to the chemical environment of the
particular atom. For a low field, or high delta, an atom is in an environment which produces induces less shielding than in a high
field, or low delta.
NMR Instrument

An NMR can be divided into three main components: the workstation computer where one operates the NMR instrument, the NMR
spectrometer console, and the NMR magnet. A standard sample is inserted through the bore tube and pneumatically lowered into
the magnet and NMR probe (Figure 4.7.30).

Figure 4.7.30 Standard NMR instrument, with main components labeled: (A) bore tube, (B) outer magnet shell, (C) NMR probe.
The first layer inside the NMR (Figure 4.7.31 is the liquid nitrogen jacket. Normally, this space is filled with liquid nitrogen at 77
K. The liquid nitrogen reservoir space is mostly above the magnet so that it can act as a less expensive refrigerant to block infrared
radiation from reaching the liquid helium jacket.

Figure 4.7.31 Diagram of the main layers inside an NMR machine.


The layer following the liquid nitrogen jacket is a 20 K radiation shield made of aluminum wrapped with alternating layers of
aluminum foil and open weave gauze. Its purpose is to block infrared radiation which the 77 K liquid nitrogen vessel was unable to
eliminate, which increases the ability for liquid helium to remain in the liquid phase due to its very low boiling point. The liquid

4.7.21 https://chem.libretexts.org/@go/page/55887
helium vessel itself, the next layer, is made of stainless steel wrapped in a single layer of aluminum foil, acting once again as an
infrared radiation shield. It is about 1.6 mm thick and kept at 4.2 K.
Inside the vessel and around the magnet is the aluminum baffle, which acts as another degree of infrared radiation protection as
well as a layer of protection for the superconducting magnet from liquid helium reservoir fluctuations, especially during liquid
helium refills. The significance is that superconducting magnets at low fields are not fully submerged in liquid helium, but higher
field superconducting magnets must maintain the superconducting solenoid fully immersed in liquid helium The vapor above the
liquid itself is actually enough to maintain superconductivity of most magnets, but if it reaches a temperature above 10 K, the
magnet quenches. During a quench, the solenoid exceeds its critical temperature for superconductivity and becomes resistive,
generating heat. This heat, in turn, boils off the liquid helium. Therefore, a small opening at the very base of the baffle exists as a
path for the liquid helium to reach the magnet surface so that during refills the magnet is protected from accidental quenching.
Problems with Solid State NMR

The most notable difference between solid samples and solution/gas in terms of NMR spectroscopy is that molecules in solution
rotate rapidly while those in a solid are fixed in a lattice. Different peak readings will be produced depending on how the molecules
are oriented in the magnetic field because chemical shielding depends upon the orientation of a molecule, causing chemical shift
anisotropy. Therefore, the effect of chemical shielding also depends upon the orientation of the molecule with respect to the
spectrometer. These counteracting forces are balanced out in gases and solutions because of their randomized molecular movement,
but become a serious issue with fixed molecules observed in solid samples. If the chemical shielding isn’t determined accurately,
neither will the chemical shifts (δ).
Another issue with solid samples are dipolar interactions which can be very large in solid samples causing linewidths of tens to
hundreds of kilohertz to be generated. Dipolar interactions are tensor quantities, which demonstrate values dependent on the
orientation and placement of a molecule in reference to its surroundings. Once again the issue goes back to the lattice structure of
solids, which are in a fixed location. Even though the molecules are fixed, this does not mean that nuclei are evenly spread apart.
Closer nuclei display greater dipolar interactions and vice versa, creating the noise seen in spectra of NMR not adapted for solid
samples. Dipolar interactions are averaged out in solution states because of randomized movement. Spin state repulsions are
averaged out by molecular motion of solutions and gases. However, in solid state, these interactions are not averaged and become a
third source of line broadening.

Magic Angle Spinning


In order to counteract chemical shift anisotropy and dipolar interactions, magic angle spinning was developed. As discussed above,
describing dipolar splitting and chemical shift aniostoropy interactions respectively, it becomes evident that both depend on the
geometric factor (3cos2θ-1).
2 2
Dipolar splitting  =  C (μ0 /8π)(γa γx / rax )(3cos θiz − 1) (4.7.9)

2
σzz   =  σ̄ + 1/3Σ σii (3cos θiz − 1) (4.7.10)

If this factor is decreased to 0, then line broadening due to chemical shift anisotropy and dipolar interactions will disappear.
Therefore, solid samples are rotated at an angle of 54.74˚, effectively allowing solid samples to behave similarly to solutions/gases
in NMR spectroscopy. Standard spinning rates range from 12 kHz to an upper limit of 35 kHz, where higher spin rates are
necessary to remove higher intermolecular interactions.
Application of Solid State NMR
The development of solid state NMR is a technique necessary to understand and classify compounds that would not work well in
solutions, such as powders and complex proteins, or study crystals too small for a different characterization method.
Solid state NMR gives information about local environment of silicon, aluminum, phosphorus, etc. in the structures, and is
therefore an important tool in determining structure of molecular sieves. The main issue frequently encountered is that crystals
large enough for X-Ray crystallography cannot be grown, so NMR is used since it determines the local environments of these
elements. Additionally, by using 13C and 15N, solid state NMR helps study amyloid fibrils, filamentous insoluble protein
aggregates related to neurodegenerative diseases such as Alzheimer’s disease, type II diabetes, Huntington’s disease, and prion
diseases.

4.7.22 https://chem.libretexts.org/@go/page/55887
Using 13-C NMR to Study Carbon Nanomaterials
Carbon Nanomaterial
There are several types of carbon nanomaterial. Members of this family are graphene, single-walled carbon nanotubes (SWNT),
multi-walled carbon nanotubes (MWNT), and fullerenes such as C60. Nano materials have been subject to various modification and
functionalizations, and it has been of interest to develop methods that could observe these changes. Herein we discuss selected
applications of 13C NMR in studying graphene and SWNTs. In addition, a discussion of how 13C NMR could be used to analyze a
thin film of amorphous carbon during a low-temperature annealing process will be presented.
13C NMR vs. 1H NMR
Since carbon is found in any organic molecule NMR that can analyze carbon could be very helpful, unfortunately the major
isotope, 12C, is not NMR active. Fortunately, 13C with a natural abundance of 1.1% is NMR active. This low natural abundance
along with lower gyromagnetic ratio for 13C causes sensitivity to decrease. Due to this lower sensitivity, obtaining a 13C NMR
spectrum with a specific signal-to-noise ratio requires averaging more spectra than the number of spectra that would be required to
average in order to get the same signal to noise ratio for a 1H NMR spectrum. Although it has a lower sensitivity, it is still highly
used as it discloses valuable information.
Peaks in a 1H NMR spectrum are split to n + 1 peak, where n is the number of hydrogen atoms on the adjacent carbon atom. The
splitting pattern in 13C NMR is different. First of all, C-C splitting is not observed, because the probability of having two adjacent
13
C is about 0.01%. Observed splitting patterns, which is due to the hydrogen atoms on the same carbon atom not on the adjacent
carbon atom, is governed by the same n + 1 rule.
In 1H NMR, the integral of the peaks are used for quantitative analysis, whereas this is problematic in 13C NMR. The long
relaxation process for carbon atoms takes longer comparing to that of hydrogen atoms, which also depends on the order of carbon
(i.e., 1°, 2°, etc.). This causes the peak heights to not be related to the quantity of the corresponding carbon atoms.
Another difference between 13C NMR and 1H NMR is the chemical shift range. The range of the chemical shifts in a typical NMR
represents the different between the minimum and maximum amount of electron density around that specific nucleus. Since
hydrogen is surrounded by fewer electrons in comparison to carbon, the maximum change in the electron density for hydrogen is
less than that for carbon. Thus, the range of chemical shift in 1H NMR is narrower than that of 13C NMR.
Solid State NMR
13
C NMR spectra could also be recorded for solid samples. The peaks for solid samples are very broad because the sample, being
solid, cannot have all anisotropic, or orientation-dependent, interactions canceled due to rapid random tumbling. However, it is still
possible to do high resolution solid state NMR by spinning the sample at 54.74° with respect to the applied magnetic field, which is
called the magic angle. In other words, the sample can be spun to artificially cancel the orientation-dependent interaction. In
general, the spinning frequency has a considerable effect on the spectrum.
13
C NMR of Carbon Nanotubes
Single-walled carbon nanotubes contain sp2 carbons. Derivatives of SWNTs contain sp3 carbons in addition. There are several
factors that affect the 13C NMR spectrum of a SWNT sample, three of which will be reviewed in this module: 13C percentage,
diameter of the nanotube, and functionalization.

13C Percentage

For sp2 carbons, there is a slight dependence of 13C NMR peaks on the percentage of 13C in the sample. Samples with lower 13C
percentage are slighted shifted downfield (higher ppm). Data are shown in Table 4.7.4. Please note that these peaks are for the sp2
carbons.
Table 4.7.4 Effects of 13C percentage on the sp2 peak. Data from S. Hayashi, F. Hoshi, T. Ishikura, M. Yumura, and S. Ohshima, Carbon,
2003, 41, 3047.
Sample δ (ppm)

SWNTs(100%) 116±1

SWNTs(1%) 118±1

4.7.23 https://chem.libretexts.org/@go/page/55887
Diameter of the Nanotubes
The peak position for SWNTs also depends on the diameter of the nanotubes. It has been reported that the chemical shift for sp2
carbons decreases as the diameter of the nanotubes increases. Figure 4.7.32 shows this correlation. Since the peak position depends
on the diameter of nanotubes, the peak broadening can be related to the diameter distribution. In other words, the narrower the peak
is, the smaller the diameter distribution of SWNTs is. This correlation is shown in Figure 4.7.33.

Figure 4.7.32 Correlation between the chemical shift of the sp2 carbon and the diameter of the nanotubes. The diameter of the
nanotubes increases from F1 to F4. Image from C. Engtrakul, V. M. Irurzun, E. L. Gjersing, J. M. Holt, B. A. Larsen, D. E.
Resasco, and J. L. Blackburn, J. Am. Chem. Soc., 2012, 134, 4850. Copyright: American Chemical Society (2012).

Figure 4.7.33 Correlation between FWHM and the standard deviation of the diameter of nanotubes. Image from C. Engtrakul, V.
M. Irurzun, E. L. Gjersing, J. M. Holt, B. A. Larsen, D. E. Resasco, and J. L. Blackburn, J. Am. Chem. Soc., 2012, 134, 4850.
Copyright: American Chemical Society (2012).

Functionalization
Solid stated 13C NMR can also be used to analyze functionalized nanotubes. As a result of functionalizing SWNTs with groups
containing a carbonyl group, a slight shift toward higher fields (lower ppm) for the sp2carbons is observed. This shift is explained
by the perturbation applied to the electronic structure of the whole nanotube as a result of the modifications on only a fraction of
the nanotube. At the same time, a new peak emerges at around 172 ppm, which is assigned to the carboxyl group of the substituent.
The peak intensities could also be used to quantify the level of functionalization. Figure 4.7.34 shows these changes, in which the
substituents are –(CH2)3COOH, –(CH2)2COOH, and –(CH2)2CONH(CH2)2NH2 for the spectra Figure 4.7.34 b, Figure 4.7.34 c,
and Figure 4.7.34 d, respectively. Note that the bond between the nanotube and the substituent is a C-C bond. Due to low
sensitivity, the peak for the sp3 carbons of the nanotube, which does not have a high quantity, is not detected. There is a small peak
around 35 ppm in Figure 4.7.34, can be assigned to the aliphatic carbons of the substituent.

4.7.24 https://chem.libretexts.org/@go/page/55887
Figure 4.7.34 13C NMR spectra for (a) pristine SWNT, (b) SWNT functionalized with –(CH2)3COOH, (c) SWNT functionalized
with –(CH2)2COOH, and (d) SWNT functionalized with –(CH2)2CONH(CH2)2NH2. Image from H. Peng, L. B. Alemany, J. L.
Margrave, and V. N. Khabashesku, J. Am. Chem. Soc., 2003, 125, 15174. Copyright: American Chemical Society (2003).
For substituents containing aliphatic carbons, a new peak around 35 ppm emerges, as was shown in Figure 4.7.34, which is due to
the aliphatic carbons. Since the quantity for the substituent carbons is low, the peak cannot be detected. Small substituents on the
sidewall of SWNTs can be chemically modified to contain more carbons, so the signal due to those carbons could be detected. This
idea, as a strategy for enhancing the signal from the substituents, can be used to analyze certain types of sidewall modifications. For
example, when Gly (–NH2CH2CO2H) was added to F-SWNTs (fluorinated SWNTs) to substitute the fluorine atoms, the 13C NMR
spectrum for the Gly-SWNTs was showing one peak for the sp2 carbons. When the aliphatic substituent was changed to 6-
aminohexanoic acid with five aliphatic carbons, the peak was detectable, and using 11-aminoundecanoic acid (ten aliphatic
carbons) the peak intensity was in the order of the size of the peak for sp2 carbons. In order to use 13C NMR to enhance the
substituent peak (for modification quantification purposes as an example), Gly-SWNTs was treated with 1-dodecanol to modify
Gly to an amino ester. This modification resulted in enhancing the aliphatic carbon peak at around 30 ppm. Similar to the results in
Figure 4.7.34, a peak at around 170 emerged which was assigned to the carbonyl carbon. The sp3 carbon of the SWNTs, which was
attached to nitrogen, produced a small peak at around 80 ppm, which is detected in a cross-polarization magic angle spinning (CP-
MAS) experiment.
F-SWNTs (fluorinated SWNTs) are reported to have a peak at around 90 ppm for the sp3 carbon of nanotube that is attached to the
fluorine. The results of this part are summarized in Figure 4.7.34 (approximate values).
Table 4.7.5 Chemical shift for different types of carbons in modified SWNTs. Note that the peak for the aliphatic carbons gets stronger if the
amino acid is esterified. Data are obtained from: H. Peng, L. B. Alemany, J. L. Margrave, and V. N. Khabashesku, J. Am. Chem. Soc., 2003,
125, 15174; L. Zeng, L. Alemany, C. Edwards, and A. Barron, Nano. Res., 2008, 1, 72; L. B. Alemany, L. Zhang, L. Zeng, C. L. Edwards,
and A. R. Barron, Chem. Mater., 2007, 19, 735.
Group δ (ppm) Intensity

sp2 carbons of SWNTs 120 Strong

–NH2(CH2)nCO2H (aliphatic carbon, n=1,5,


20-40 Depends on ‘n’
10)

–NH2(CH2)nCO2H (carboxyl carbon, n=1,5,


170 Weak
10)

sp3 carbon attached to nitrogen 80 Weak

sp3 carbon attached to fluorine 90 Weak

The peak intensities that are weak in Figure 4.7.34 depend on the level of functionalization and for highly functionalized SWNTs,
those peaks are not weak. The peak intensity for aliphatic carbons can be enhanced as the substituents get modified by attaching to
other molecules with aliphatic carbons. Thus, the peak intensities can be used to quantify the level of functionalization.
13C NMR of Functionalized Graphene
Graphene is a single layer of sp2 carbons, which exhibits a benzene-like structure. Functionalization of graphene sheets results in
converting some of the sp2 carbons to sp3. The peak for the sp2 carbons of graphene shows a peak at around 140 ppm. It has been

4.7.25 https://chem.libretexts.org/@go/page/55887
reported that fluorinated graphene produces an sp3peak at around 82 ppm. It has also been reported for graphite oxide (GO), which
contains –OH and epoxy substituents, to have peaks at around 60 and 70 ppm for the epoxy and the –OH substituents, respectively.
There are chances for similar peaks to appear for graphene oxide. Table 4.7.6 summarizes these results.
Table 4.7.6 Chemical shifts for functionalized graphene. Data are obtained from: M. Dubois, K. Guérin, J. P. Pinheiro, Z. Fawal, F. Masin,
and A. Hamwi, Carbon, 2004, 42, 1931; L. B. Casabianca, M. A. Shaibat, W. W. Cai, S. Park, R. Piner, R. S. Ruoff, and Y. Ishii, J. Am.
Chem. Soc., 2010, 132, 5672.
Type of Carbon δ (ppm)

sp2 140

sp3 attached to fluorine 80

sp3 attached to -OH (for GO) 70

sp3 attached to epoxide (for GO) 60

Analyzing Annealing Process Using 13C NMR


13C NMR spectroscopy has been used to study the effects of low-temperature annealing (at 650 °C) on thin films of amorphous
carbon. The thin films were synthesized from a 13C enriched carbon source (99%). There were two peaks in the 13C NMR spectrum
at about 69 and 142 ppm which were assigned to sp3 and sp2carbons, respectively Figure 4.7.35. The intensity of each peak was
used to find the percentage of each type of hybridization in the whole sample, and the broadening of the peaks was used to estimate
the distribution of different types of carbons in the sample. It was found that while the composition of the sample didn’t change
during the annealing process (peak intensities didn’t change, see Figure 4.7.35b), the full width at half maximum (FWHM) did
change (Figure 4.7.35a). The latter suggested that the structure became more ordered, i.e., the distribution of sp2 and sp3carbons
within the sample became more homogeneous. Thus, it was concluded that the sample turned into a more homogenous one in terms
of the distribution of carbons with different hybridization, while the fraction of sp2 and sp3 carbons remained unchanged.

Figure 4.7.35 a) Effect of the annealing process on the FWHM, which represents the change in the distribution of sp2 and sp3
carbons. b) Fractions of sp2 and sp3 carbon during the annealing process. Data are obtained from T. M. Alam, T. A. Friedmann, P.
A. Schultz, and D. Sebastiani, Phys. Rev. B., 2003, 67, 245309.
Aside from the reported results from the paper, it can be concluded that 13C NMR is a good technique to study annealing, and
possibly other similar processes, in real time, if the kinetics of the process is slow enough. For these purposes, the peak intensity
and FWHM can be used to find or estimate the fraction and distribution of each type of carbon respectively.
Summary
13C NMR can reveal important information about the structure of SWNTs and graphene. 13C NMR chemical shifts and FWHM can
be used to estimate the diameter size and diameter distribution. Though there are some limitations, it can be used to contain some
information about the substituent type, as well as be used to quantify the level of functionalization. Modifications on the substituent
can result in enhancing the substituent signal. Similar type of information can be achieved for graphene. It can also be employed to
track changes during annealing and possibly during other modifications with similar time scales. Due to low natural abundance of
13C it might be necessary to synthesize 13C-enhanced samples in order to obtain suitable spectra with a sufficient signal-to-noise

ratio. Similar principles could be used to follow the annealing process of carbon nano materials. C60will not be discussed herein.

Lanthanide Shift Reagents


Nuclear magnetic resonance spectroscopy (NMR) is the most powerful tool for organic and organometallic compound
determination. Even structures can be determined just using this technique. In general NMR gives information about the number of
magnetically distinct atoms of the specific nuclei under study, as well as information regarding the nature of the immediate
environment surrounding each nuclei. Because hydrogen and carbon are the major components of organic and organometallic
compounds, proton (1H) NMR and carbon-13 (13C) NMR are the most useful nuclei to observe.

4.7.26 https://chem.libretexts.org/@go/page/55887
Not all the protons experience resonance at the same frequency in a 1H NMR, and thus it is possible to differentiate between them.
The diversity is due to the existence of a different electronic environment around chemically different nuclei. Under an external
magnetic field (B0), the electrons in the valence shell are affected; they start to circulate generating a magnetic field, which is
apposite to the applied magnetic field. This effect is called diamagnetic shielding or diamagnetic anisotropy Figure 4.7.36.

Figure 4.7.36 Schematic representation of diamagnetic anisotropy. Adapted from D. L. Pavia, G. M. Lampman, and G. S. Kriz,
Introduction to Spectroscopy, 3th Ed., Thomson Learning, Tampa, FL, (2011).
The greater the electron density around one specific nucleus, the greater will be the induced field that opposes the applied field, and
this will result in a different resonance frequency. The identification of protons sounds simple, however, the NMR technique has a
relatively low sensitivity of proton chemical shifts to changes in the chemical and stereochemical environment; as a consequence
the resonance of chemically similar proton overlap. There are several methods that have been used to resolve this problem, such as:
the use of higher frequency spectrometers or by the use of shift reagents as aromatic solvents or lanthanide complexes. The main
issue with high frequency spectrometers is that they are very expensive, which reduces the number of institutions that can have
access to them. In contrast, shift reagents work by reducing the equivalence of nuclei by altering their magnetic environment, and
can be used on any NMR instrument. The simplest shift reagent is the one of different solvents, however problems with some
solvents is that they can react with the compound under study, and also that these solvents usually just alter the magnetic
environment of a small part of the molecule. Consequently, although there are several methods, most of the work has been done
with lanthanide complexes.
The History of Lanthanide Shift Reagents

The first significant induced chemical shift using paramagnetic ions was reported in 1969 by Conrad Hinckley (Figure 4.7.37),
where he used bispyridine adduct of tris(2,2,6,6-tetramethylhepta-3,5-dionato)europium(III) (Eu(tmhd)3), better known as
Eu(dpm)3, where dpm is the abbreviation of dipivaloyl- methanato, the chemical structure is shown in Figure 4.7.38. Hinckley
used Eu(tmhd)3 on the 1H NMR spectrum of cholesterol from 347 – 2 Hz. The development of this new chemical method to
improve the resolution of the NMR spectrum was the stepping-stone for the work of Jeremy Sanders and Dudley Williams, Figure
4.7.39 and Figure 4.7.40 respectively. They observed a significant increase in the magnitude of the induced shift after using just

the lanthanide chelate without the pyridine complex. Sugesting that the pyridine donor ligands are in competition for the active
sides of the lanthanide complex. The efficiency of Eu(tmhd)3 as a shift reagent was published by Sanders and Williams in 1970,
where they showed a significant difference in the 1H NMR spectrum of n-pentanol using the shift reagent, see Figure 4.7.41.

4.7.27 https://chem.libretexts.org/@go/page/55887
Figure 4.7.40 British chemist Dudley Williams (1937-2010).

Figure 4.7.38 Chemical Structure of Eu(tmhd)3.

4.7.28 https://chem.libretexts.org/@go/page/55887
Figure 4.7.41 1H NMR spectra of n-pentanol, (a) without the present of lanthanide reagents and (b) in the present of the lanthanide
reagent Eu(tmhd)3. Adapted from Chem Reviews, 1973, 73, 553. Copyright: American Chemical Society 1973.
Analyzing the spectra in Figure 4.7.41 it is easy to see that with the use of Eu(tmhd)3 there is any overlap between peaks. Instead,
the multiplets of each proton are perfectly clear. After these two publications the potential of lanthanide as shift reagents for NMR
studies became a popular topic. Other example is the fluorinate version of Eu(dpm)3; (tris(7,7,-dimethyl-1,1,2,2,2,3,3-
heptafluoroocta-7,7-dimethyl-4,6-dionato)europium(III), best known as Eu(fod)3, which was synthesized in 1971 by Rondeau and
Sievers. This LSR presents better solubility and greater Lewis acid character, the chemical structure is show in Figure 4.7.42.

Figure 4.7.42 Chemical structure of (tris(7,7,-dimethyl-1,1,2,2,2,3,3-heptafluoroocta-7,7-dimethyl-4,6-dionato)europium(III).


Mechanism of Inducement of Chemical Shift
Lanthanide atoms are Lewis acids, and because of that, they have the ability to cause chemical shift by the interaction with the
basic sites in the molecules. Lanthanide metals are especially effective over other metals because there is a significant
delocalization of the unpaired f electrons onto the substrate as a consequence of unpaired electrons in the f shell of the lanthanide.
The lanthanide metal in the complexes interacts with the relatively basic lone pair of electrons of aldehydes, alcohols, ketones,
amines and other functional groups within the molecule that have a relative basic lone pair of electrons, resulting in a NMR spectral
simplification.
There are two possible mechanisms by which a shift can occur: shifts by contact and shifts by pseudocontact. The first one is a
result of the transfer of electron spin density via covalent bond formation from the lanthanide metal ion to the associated nuclei.
While the magnetic effects of the unpaired electron magnetic moment causes the pseudocontact shift. Lanthanide complexes give
shifts primarily by the pseudocontact mechanism. Under this mechanism, there are several factors that influence the shift of a
specific NMR peak. The principal factor is the distance between the metal ion and the proton; the shorter the distance, the greater
the shift obtained. On the other hand, the direction of the shift depends on the lanthanide complex used. The complexes that
produce a shift to a lower field (downfield) are the ones containing erbium, europium, thulium and ytterbium, while complexes
with cerium, neodymium, holmium, praseodymium, samarium and terbium, shift resonances to higher field. Figure 6 shows the
difference betwen an NMR spectrum without the use of shift reagent versus the same spectrum in the present of a europium
complex (downfield shift) and a praseodymium complex (high-field shift).

4.7.29 https://chem.libretexts.org/@go/page/55887
Figure 4.7.43 (a) 1H NMR spectrum of n-hexanol without the present of shift reagents. (b) 1H NMR spectrum of n-hexanol in
present of 14% Pr(fod)3 and the thirt spectrum (c) is the 1H NMR spectrum of n-hexanol in the present of 6.5% Eu(fod)3. Adapted
from http://www.chem.wisc.edu/areas/reich...ech-07-lis.htm
Linewidth broadening is not desired because of loss of resolution, and lanthanide complexes unfortunately contribute extremely to
this effect when they are used in high concentrations due to their mechanism that shortens the relaxation times (T2), which in turn
increases the bandwidth. However europium and praseodymium are an extraordinary exception giving a very low shift broadening,
0.003 and 0.005 Hz/Hz respectively. Europium specially is the most used lanthanide as shift reagent because of its inefficient
nuclear spin-lattice ratio properties. It has low angular momentum quantum numbers and a diamagnetic 7F0 ground state. These
two properties contribute to a very small separation of the highest and lowest occupied metal orbitals leading to an inefficient
relaxation and a very little broadening in the NMR spectra. The excited 7F1 state will then contribute to the pseudocontact shift.
We have mentioned above that lanthanide complexes have a mechanism that influences relaxation times, and this is certainly
because paramagnetic ions have an influence in both: chemical shifts and relaxation rates. The relaxation times are of great
significant because they depend on the width of a specific resonance (peak). Changes in relaxation time could also be related with
the geometry of the complex.
Measuring the Shift
The easiest and more practical way to measure the lanthanide-induced shift (LIS) is to add aliquots of the lanthanide shift reagent
(LSR or Δvi) to the sample that has the compound of interest (substrate), and take an NMR spectra after each addition. Because the
shift of each proton will change after each addition of the LSR to lower or upper field, the LIS can me measured. With the data
collected, a plot of the LIS against the ratio of LSR: substrate will generate a straight line where the slope is representative of the
compound that is being studied. The identification of the compound by the use of chiral lanthanide shift reagents can be so precise
that it is possible to estimate the composition of enantiomers in the solution under study, see Figure 4.7.44.

Figure 4.7.45 Lanthanide induced shift of methoxyl proton resonance versus molar ratio of Eu(fod)3, for the diastereomeric MTPA
esters. δ is the normal chemical shift and δE is the chemical shift in ppm for the OMe signal in the presence of a specified molar
ratio of Eu(fod)3, in CCl4 as solvent. Adapted from S. Yamaguchi, F. Yasuhara and K. Kabuto, Tetrahedron, 1976, 32, 1363.
Now, what is the mechanism that is actually happening between the LSR and the compound under study? The LSR is a metal
complex of six coordinate sides. The LSR, in presence of substrate that contains heteroatoms with Lewis basicity character,
expands its coordination sides in solution in order to accept additional ligands. An equilibrium mixture is formed between the

4.7.30 https://chem.libretexts.org/@go/page/55887
substrate and the LSR. 4.7.11 and 4.7.12 show the equilibrium, where L is LSR, S is the substrate, and LS is the concentration of
the complex formed is solution.
K1

L  +  S ⇄  [LS] (4.7.11)

K2

[LS]  +  S ⇄ [LS2 ] (4.7.12)

The abundance of these species depends on K1 and K2, which are the binding constant. The binding constant is a special case of
equilibrium constant, but it refers with the binding and unbinding mechanism of two species. In most of the cases like, K2 is
assumed to be negligible and therefore just the first complex [LS] is assumed to be formed. The equilibrium between L + S and LS
in solution is faster than the NMR timescale, consequently a single average signal will be recorded for each nucleus.
Determination of Enantiomeric Purity
Besides the great potential of lanthanide shift reagents to improve the resolution of NMR spectrums, these complexes also have
been used to identify enantiomeric mixtures in solution. To make this possible the substrate must meet certain properties. The fist
one is that the organic compounds in the enantiomeric composition must to have a hard organic base as functional group. The shift
reagents are not effective with most of the soft bases. Though hundreds of chelates have been synthesized after Eu(dcm)3, this one
is the LSR that resulted in the most effective reagent for the resolution of enantiotopic resonances. Basically if you take an NMR of
an enantiomeric mixture sample, a big variety of peaks will appear and the hard part is to identify which of those peaks correspond
to which specific enantiomer. The differences in chemical shifts observed for enantiomeric mixtures in solution containing LSR
might arise from at least two sources: the equilibrium constants of the formation of the possible diastereometic complexes between
the substrate and the LSR, and the geometries of these complexes, which might be distinct. The enantiomeric shift differences
sometimes are defined as ΔΔδ.
In solution the exchange between substrate coordinated to the europium ion and the free substrate in solution is very fast. To be
sure that the europium complexes are binding with one or two substrate molecules, an excess of substrate is usually added.

Determination of Relaxation Parameters of Contrast Agents


Magnetic resonance imaging (MRI) (also known as nuclear magnetic resonance imaging (NMRI) or magnetic resonance
tomography (MRT)) is a powerful noninvasive diagnostic technique, which is used to generate magnetic field (B0) and interacts
with spin angular momentum of the nucleus in the tissue. Spin angular momentum depends on number of protons and neutrons of
nucleus. Nuclei with even number of protons plus neutrons are insensitive to magnetic field, so cannot be viewed by MRI.
Each nucleus can be considered as an arrow with arbitrary direction in absence of external magnetic field (Figure 4.7.46). And we
consider them to get oriented in the same direction once magnetic field applied (Figure 4.7.47). In order to get nuclei orient in
specific direction, energy is supplied, and to bring it to original position energy is emitted. All this transitions eventually lead to
changes in angular velocity, which is defined as Larmor frequency and the expression 4.7.13, where ω is the Larmor frequency, γ is
the gyromagnetic ratio, and B0 is the magnetic field. It is not easy to detect energy, which is involved in such a transition, that’s
why use of high resolution spectrometers required, those which are developed by nowadays as a most powerful MRI are close to 9
Tesla with mass approaching forty five tons. Unfortunately it is expensive tool to purchase and to operate. That’s why new
techniques should be developed, so most of the MRI spectrometers can be involved in imaging. Fortunately presence of huge
amount of nuclei in analyzed sample or body can provide with some information.

ω  =  γB0 (4.7.13)

Figure 4.7.46 Representation of nuclei in absence of magnetic field.

4.7.31 https://chem.libretexts.org/@go/page/55887
Figure 4.7.47 Representation of nuclei in presence of magnetic field.
Nuclear Magnetic Resonance Relaxometer
Each nucleus possesses microscopic magnetic spins of x, y and z. Presence of randomly distributed atoms with varying x and y
spins will lead to zero upon summation of x and y planes. But in case of z, summation of magnetic spins will not lead to
cancellation. According to Currie’s law, 4.7.14, (Mzis the resulting magnetization of z axis, C is a material specific Curie constant,
B0 is the magnetic field, and T is absolute temperature), magnetization of z axis proportional to magnetic field applied from
outside. Basically, excitation happens by passing current through coil which leads to magnetization of x, y and z axis. It is the way
of changing magnetism from z axis to x and y axis. Once external current supply is turned off, magnetization will eventually
quench. This means a change of magnetization from x and y axis to z axis, were it eventually become equilibrated and device no
more can detect the signals. Energy which is emitted from excited spin leads to development of new current inside of the same coil
recorded by detector; hence same coil can be used as detector and source of magnetic field. This process called as a relaxation and
that's why, return of magnetization to z axis called as spin-lattice relaxation or T1 relaxation (time required for magnetization to
align on z axis). Eventual result of zero magnetization on x and y axis called as spin-spin relaxation or T2 relaxation (Figure
4.7.48).

Mz   =  C B0 /T (4.7.14)

Figure 4.7.48 Magnetic spins relaxation mechanism


Contrast Agents for MRI
In MRI imaging contrast is determined according to T1, T2 or the proton density parameter. Therefor we can obtain three different
images. By changing intervals between radio frequency (RF) 90° pulses and RF 180° pulses, the desired type of image can be
obtained. There are few computational techniques available to improve contrast of image; those are repetitive scans and different
mathematical computations. Repetitive scans take a long time, therefore cannot be applied in MRI. Mathematical computation on
their own, do not provide with desired results. For that reason, in order to obtain high resolution images, contrast agents (CA) are
important part of medical imaging.

Types of Contrast Agents


There are different types of contrast agents available in markets which reduce the supremacy of T1or T2, and differentiate according
to relaxivity1 (r1) and relaxivity2 (r2) values. The relaxivity (ri) can be described as 1/Ti (s-1) of water molecules per mM
concentration of CA. Contrast agents are paramagnetic and can interact with dipole moments of water molecules, causing
fluctuations in molecules. This theory is known as Solomon-Bloembergen-Morgan (SBM) theory. Those which are efficient were
derivatives of gadolinium (e.g., gadobenic acid (Figure 4.7.49 a) and gadoxetic acid (Figure 4.7.49 b), iron (e.g.,
superparamagnetic iron oxide and ultrasmall superparamagnetic iron oxide) and manganese (e.g., manganese dipyridoxal
diphosphate). Fundamentally the role of contrast agents can be played by any paramagnetic species.

4.7.32 https://chem.libretexts.org/@go/page/55887
Figure 4.7.49 The structures of two representative commercial gadolinium MRI contrast agents; (a) gadobenic acid and (b)
gadoxetic acid.

Principal of Interactions of CA with Surrounding Media


There are two main principles of interactions of contrast agents with water molecules. One is direct interaction, which is called
inner sphere relaxation, and the other mechanism that happens in the absence of direct interaction with water molecule which is
outer sphere relaxation. If we have water molecules in the first coordination sphere of metal ion, we can consider them as the inner
sphere, and if diffusion of protons from outside happens randomly we define them as outer sphere relaxation. Another type of
relaxivity comes from already affected water molecules, which transfers their relaxivity to protons of close proximity, this type of
relaxivity called second sphere and is usually neglected or contributed as outer sphere. In inner sphere proton relaxivity there are
two main mechanisms involved in relaxation. One is dipole-dipole interactions between metal and proton and another is scalar
mechanism. Dipole-dipole interaction affects electron spin vectors and scalar mechanism usually controls water exchange. Effect of
contrast agents on T1 relaxation is much larger than on T2, since T1 is much larger for tissues than T2.
Determination of Relaxivity
Determination of relaxivity became very easy with the advancements of NMR and computer technology, where you need just to
load your sample and read values from the screen. But let’s consider in more detail what are the precautions should be taken during
sample preparation and data acquisition.

Sample Preparation
The sample to be analyzed is dissolved in water or another solvent. Generally water is used since contrast agents for medical MRI
are used in aqueous media. The amount of solution used is determined according to the internal standard volume, which is used for
calibration purposes of device and is usually provided by company producing device. A suitable sample holder is a NMR tube. It is
important to degas solvent prior measurements by bubbling gas through it (nitrogen or argon works well), so no any traces of
oxygen remains in solution, since oxygen is paramagnetic.

Data Acquisition
Before collecting data it is better to keep the sample in the device compartment for few minutes, so temperature of magnet and your
solution equilibrates. The relaxivity (ri) calculated according to (4.7.15 ), where Ti is the relaxation time in the presence of CAs,
Tid is the relaxation time in the absence of CAs, and CA is the concentration of paramagnetic CAs (mM). Having the relaxivity
values allows for a comparison of a particular compound to other known contrast agents.
ri   =  (1/ Ti   −  1/ Tid )/[C A] (4.7.15)

Two-Dimensional NMR
General Principles of Two-Dimensional Nuclear Magnetic Resonance Spectroscopy
History
Jean Jeener (Figure 4.7.50 from the Université Libre de Bruxelles first proposed 2D NMR in 1971. In 1975 Walter P. Aue, Enrico
Bartholdi, and Richard R. Ernst (Figure 4.7.51 first used Jeener’s ideas of 2D NMR to produce 2D spectra, which they published in
their paper “Two-dimensional spectroscopy, application to nuclear magnetic resonance”. Since this first publication, 2D NMR has
increasing been utilized for structure determination and elucidation of natural products, protein structure, polymers, and inorganic
compounds. With the improvement of computer hardware and stronger magnets, newly developed 2D NMR techniques can easily

4.7.33 https://chem.libretexts.org/@go/page/55887
become routine procedures. In 1991 Richard R. Ernst won the Nobel Prize in Chemistry for his contributions to Fourier Transform
NMR. Looking back on the development of NMR techniques, it is amazing that 2D NMR took so long to be developed considering
the large number of similarities that it has with the simpler 1D experiments.

Figure 4.7.50 Belgian physical chemist and physicist Jean L. C. Jeener (1931-).

Figure 4.7.51 Swiss physical chemist and Nobel Laureate Richard R. Ernst (1933-).
Why do We Need 2D NMR?
2D NMR was developed in order to address two major issues with 1D NMR. The first issue is the limited scope of a 1D spectrum.
A 2D NMR spectrum can be used to resolve peaks in a 1D spectrum and remove any overlap present. With a 1D spectrum, this is
typically performed using an NMR with higher field strength, but there is a limit to the resolution of peaks that can be obtained.
This is especially important for large molecules that result in numerous peaks as well as for molecules that have similar structural
motifs in the same molecule. The second major issue addressed is the need for more information. This could include structural or
stereochemical information. Usually to overcome this problem, 1D NMR spectra are obtained studying specific nuclei present in
the molecule (for example, this could include fluorine or phosphorus). Of course this task is limited to only nuclei that have active
spin states/spin states other than zero and it requires the use of specialized NMR probes.
2D NMR can address both of these issues in several different ways. The following four techniques are just few of the methods that
can be used for this task. The use of J-resolved spectroscopy is used to resolve highly overlapping resonances, usually seen as
complex multiplicative splitting patterns. Homonuclear correlation spectroscopy can identify spin-coupled pairs of nuclei that
overlap in 1D spectra. Heteronuclear shift-correlation spectroscopy can identify all directly bonded carbon-proton pairs, or other
combinations of nuclei pairs. Lastly, Nuclear Overhauser Effect (NOE) interactions can be used to obtain information about
through-space interactions (rather than through-bond). This technique is often used to determine stereochemistry or protein/peptide
interactions.
One-dimensional vs. Two-dimensional NMR

Similarities
The concept of 2D NMR can be considered as an extension of the concept of 1D NMR. As such there are many similarities
between the two. Since the acquisition of a 2D spectrum is almost always preceded by the acquisition of a 1D spectrum, the
standard used for reference Since 2D NMR is a more complicated experiment than 1D NMR, there are also some differences
between the two. One of the differences is in the complexity of the data obtained. A 2D spectrum often results from a change in
pulse time; therefore, it is important to set up the experiment correctly in order to obtain meaningful information. Another

4.7.34 https://chem.libretexts.org/@go/page/55887
difference arises from the fact that one spectrum is 1D while the other is 2D. As such interpreting a 2D spectrum requires a much
greater understanding of the experiment parameters. For example, one 2D experiment might investigate the specific coupling of
two protons or carbons, rather than focusing on the molecule as a whole (which is generally the target of a 1D experiment). The
specific pulse sequence used is often very helpful in interpreting the information obtained. The software used for 1D spectra is not
always compatible with 2D spectra. This is due to the fact that a 2D spectrum requires more complex processing, and the 2D
spectra generated often look quite different than 1D spectra. Some software that is commonly used to interpret 2D spectra is either
Sparky or Bruker’s TopSpin. Lastly the NMR instrument used to obtain a 2D spectrum typically generates a much larger magnetic
field (700-1000 MHz). Due to the increased cost of buying and maintaining such an instrument, 2D NMR is usually reserved for
rather complex molecules.(TMS) and the solvent used (typically CDCl3 or other deuterated solvent) are the same for both
experiments. Furthermore, 2D NMR is most often used to reveal any obscurity in a 1D spectrum (whether that is peak overlap,
splitting overlap, or something else), so the nuclei studied are the same. Most often these are 1H and 13C, but other nuclei could
also be studied.

Differences
Since 2D NMR is a more complicated experiment than 1D NMR, there are also some differences between the two. One of the
differences is in the complexity of the data obtained. A 2D spectrum often results from a change in pulse time; therefore, it is
important to set up the experiment correctly in order to obtain meaningful information. Another difference arises from the fact that
one spectrum is 1D while the other is 2D. As such interpreting a 2D spectrum requires a much greater understanding of the
experiment parameters. For example, one 2D experiment might investigate the specific coupling of two protons or carbons, rather
than focusing on the molecule as a whole (which is generally the target of a 1D experiment). The specific pulse sequence used is
often very helpful in interpreting the information obtained. The software used for 1D spectra is not always compatible with 2D
spectra. This is due to the fact that a 2D spectrum requires more complex processing, and the 2D spectra generated often look quite
different than 1D spectra. Some software that is commonly used to interpret 2D spectra is either Sparky or Bruker’s TopSpin.
Lastly the NMR instrument used to obtain a 2D spectrum typically generates a much larger magnetic field (700-1000 MHz). Due to
the increased cost of buying and maintaining such an instrument, 2D NMR is usually reserved for rather complex molecules.
The Rotating Frame and Fourier Transform
One of the central ideas that is associated with 2D NMR is the rotating frame, because it helps to visualize the changes that take
place in dimensions. Our ordinary “laboratory” frame consists of three axes (the Cartesian x, y, and z). This frame can be visualized
if one pictures the corner of a room. The intersections of the floor and the walls are the x and the y dimensions, while the
intersection of the walls is the z axis. This is usually considered the “fixed frame.” When an NMR experiment is carried out, the
frame still consists of the Cartesian coordinate system, but the x and ycoordinates rotate around the z axis. The speed with which
the x-y coordinate system rotates is directly dependent on the frequency of the NMR instrument.
When any NMR experiment is carried out, a majority of the spin states of the nucleus of interest line up with one of these three
coordinates (which we can pick to be z). Once an equilibrium of this alignment is achieved, a magnetic pulse can be exerted at a
certain angle to the z axis (usually 90° or 180°) which temporarily disrupts the equilibrium alignment of the nuclei. As the pulse is
removed, the nuclei are allowed to relax back to this equilibrium alignment with the magnetic field of the instrument. When this
relaxation takes place, the progression of the nuclei back to the equilibrium orientation is detected by a computer as a free induction
decay (FID). When a sample has different nuclei or the same nucleus in different environments, different FID can be recorded for
each individual relaxation to the equilibrium position. The FIDs of all of the individual nuclei can be recorded and superimposed.
The complex FID signal obtained can be converted to a recording of the NMR spectrum obtained by a Fourier transform(FT). The
FT is a complex mathematical concept that can be described by 4.7.16, where ω is the angular frequency.

ikωt
z(t)  =   ∑ ci e (4.7.16)

k→∞

This concept of the FT is similar for both 1D and 2D NMR. In 2D NMR a FID is obtained in one dimension first, then through the
application of a pulse a FID can be obtained in a second dimension. Both FIDs can be converted to a series of NMR spectra
through a Fourier transform, resulting in a spectrum that can be interpreted. The coupling of the two FID's in 2D NMR usually
reveals a lot more information about the specific connectivity between two atoms.

4.7.35 https://chem.libretexts.org/@go/page/55887
Four Phases and Pulse Sequence of 2D NMR
There are four general stages or time periods that are present for any 2D NMR experiment. These are preparation, evolution,
mixing, and detection. A general schematic representation is seen in Figure 4.7.53. The preparation period defines the system at the
first time phase. The evolution period allows the nuclei to precess (or move relative to the magnetic field). The mixing period
introduces a change in the way the spectra is obtained. The detection period records the FID. In obtaining a spectrum, the pulse
sequence is the most important factor that determines what data will be obtained. In general 2D experiments are a combination of
1D experiments collected by varying the timing and pulsing.

Figure 4.7.53 Visual representation of the general pulse scheme of any 2D NMR Experiment

Preparation
This is the first step in any 2D NMR experiment. It is a way to start all experiments from the same state. This state is typically
either thermal equilibrium, obeying Boltzmann statistics, or it could be a state where the spins of one nucleus are randomized in
orientation and the spins of another nucleus are in thermal equilibrium. At the end of the preparation period, the magnetizations are
usually placed perpendicular, or at a specific angle, to the magnetic field axis. This phase creates magnetizations in the x-y plane.

Evolution
The nuclei are then allowed to precess around the direction of the magnetic field. This concept is very similar to the precession of a
top in the gravitational field of the Earth. In this phase of the experiment, the rates at which different nuclei precess, as shown in
Figure 4.7.54 determine how the nuclei are reacting based on their environment. The magnetizations that are created at the end of
the preparation step are allowed to evolve or change for a certain amount of time (t1) in the environment defined by the magnetic
and radio frequency (RF) fields. In this phase, the chemical shifts of the nuclei are measured similarly to a 1D experiment, by
letting the nucleus magnetization rotate in the x-y plane. This experiment is carried out a large number of times, and then the
recorded FID is used to determine the chemical shifts.

Figure 4.7.54 Visual representation of the precession of an object.

Mixing
Once the evolution period is over, the nuclear magnetization is distributed among the spins. The spins are allowed to communicate
for a fixed period of time. This typically occurs using either magnetic pulses and/or variation in the time periods. The magnetic
pulses typically consist of a change in the rotating frame of reference relative to the original "fixed frame" that was introduced in
the preparation period, as seen in Figure 4.7.55. Experiments that only use time periods are often tailored to look at the effect of the
RF field intensity. Using either the bonds connecting the different nuclei (J-coupling) or using the small space between them (NOE
interaction), the magnetization is allowed to move from one nucleus to another. Depending on the exact experiment performed,
these changes in magnetizations are going to differ based on what information is desired. This is the step in the experiment that
determines exactly what new information would be obtained by the experiment. Depending on which chemical interactions require
suppression and which need to be intensified to reveal new information, the specific "mixing technique" can be adjusted for the
experiment.

4.7.36 https://chem.libretexts.org/@go/page/55887
Figure \PgeIndex55 Demonstration of a specific (90°) change in the frame of reference during mixing.

Detection
This is always the last period of the experiment, and it is the recording of the FID of the second nucleus studied. This phase records
the second acquisition time (t2) resulting in a spectrum, similar to the first spectrum, but typically with differences in intensity and
phase. These differences can give us information about the exact chemical and magnetic environment of the nuclei that are present.
The two different Fourier transforms are used to generate the 2D spectrum, which consists of two frequency dimensions. These two
frequencies are independent of each other, but when plotted on a single spectrum the frequency of the signal obtained in time t1 has
been converted in another coherence affected by the frequency in time t2. While the first dimension represents the chemical shifts
of the nucleus in question, the second dimension reveals new information. The overall spectrum, Figure 4.7.56, is the result of a
matrix in the two frequency domains obtained during the experiment.

Figure 4.7.56 Simple representation of a 2D spectrum, reflecting the result of two Fourier transforms.

Pulse Variation
As mentioned earlier, the pulse sequence and the mixing period are some of the most important factors that determine the type of
spectrum that will be identified. Depending on whether the magnetization is transferred through a J-coupling or NOE interaction,
different information and spectra can be obtained. Furthermore, depending on the experimental setup, the mixing period could
transfer magnetization either through a single J-coupling or through several J-couplings for nuclei that are connected together.
Similarly NOE interactions can also be controlled to specific distances. Two types of NOE interactions can be observed, positive
and negative. When the rate at which fluctuation occurs in the transverse plane of a fluctuating magnetic field matches the
frequency of double quantum transition, a positive NOE is observed. When the fluctuation is slower, a negative NOE is produced.
Obtaining a Spectrum

Sample Preparation
Sample preparation for 2D NMR is essentially the same as that for 1D NMR. Particular caution should be exercised to use clean
and dry sample tubes and use only deuterated solvents. The amount of sample used should be anywhere between 15 and 25 mg
although with sufficient time even smaller quantities may be used. The filling height of the solvent should be about 4 cm. The
solution must be clear and homogenous. Any participate needs to be filtered off prior to obtaining the spectra.

4.7.37 https://chem.libretexts.org/@go/page/55887
The Actual Experiment and Important Acquisition Parameters
The acquisition of a 2D spectrum will vary from instrument to instrument, but the process is virtually identical to obtaining a 13C
spectrum. It is important to obtain a 1D spectrum (especially 1H) before proceeding to obtain a 2D spectrum. The acquisition range
should be adjusted based on the 1D spectrum to minimize instrument time. Depending on the specific type of 2D experiment (such
as COSY or NOESY) several parameters need to be adjusted. The following 6 steps can followed to obtain almost any 2D NMR
spectrum.
1. Login to the computer system.
2. Change the sample.
3. Lock and shim the magnet.
4. Setup parameters and run the experiment. Use the 1D spectra already obtained to adjust experiment settings, paying special
attention to important acquisition parameters.
5. Process the obtained data and print the spectrum.
6. Exit and logout.
The parameters listed in Table 4.7.7 should be given special attention, as they can significantly affect the quality of the spectra
obtained.
Table 4.7.7 Some of the most important parameters for obtaining a 2D spectrum and their meaning.
Parameter Description

Acquisition Time (AQ) Data points (TD) x dwell time (DW)

Dwell Time 1/spectral width (SW)

Digital Resolution 1/AQ

Number of Scans Multiples of 8/16

TD1 Number of data points in the first time domain ( ~128-512)

SW1 Spectral Width in the first (direct) dimension

TD2 Number of data points in the second time domain (~2048-4096)

SW2 Spectral Width in the second (indirect) dimension

After Obtaining a Spectrum and Analysis


After a 2D spectrum has successfully been obtained, depending on the type of spectrum (COSY, NOESY, INEPT), it might need to
be phased. Phasing is the adjustment of the spectrum so that all of the peaks across the spectrum are in the absorptive mode
(pointing either up or down). With 2D spectra, phasing is done in both frequency dimensions. This can either be done automatically
by a software program (for simple 2D spectra with no cluster signals) or manually by the user (for more complex 2D spectra).
Sometimes, phasing can be done with the program that is used to obtain the spectrum. Afterwards the spectrum could either be
printed out or further analyzed. One example of further analysis is integrating parts of the spectrum. This could give the user
meaningful information about the relative ratio of different types of nuclei (and even quantify the ratios between two diasteriomeric
molecules).
Conclusion
Two-dimensional NMR is increasingly becoming a routine method for analyzing complex molecules, whether they are inorganic
compounds, organic natural products, proteins, or polymers. A basic understanding of 2D NMR can make it significantly easier to
analyze complex molecules and provide further confirmation for results obtained by other methods. The variation in pulse
sequences provides chemists the opportunity to analyze a large diversity of compounds. The increase in the magnetic strength of
NMR machines has allowed 2D NMR to be more often used even for simpler molecules. Furthermore, higher dimension
techniques have also been introduced, and they are slowly being integrated into the repertoire of chemists. These are essentially
simple extensions of the ideas of 2D NMR.

Two-Dimensional NMR Experiments


Since the advent of NMR, synthetic chemists have had an excellent way to characterize their synthetic products. With the arrival of
multidimensional NMR into the realm of analytical techniques, scientists have been able to study larger and more complicated

4.7.38 https://chem.libretexts.org/@go/page/55887
molecules much easier than before, due to the great amount of information 2D and 3D NMR experiments can offer. With 2D NMR,
overlapping multiplets and other complex splitting patterns seen in 1D NMR can be easily deciphered, since instead of one
frequency domain, two frequency domains are plotted and the couplings are plotted with respect to each other, which makes it
easier to determine molecular connectivity.
Spectra are obtained using a specific sequence of radiofrequency (RF) pulses that are administered to the sample, which can vary in
the angle at which the pulse is given and/or the number of pulses. Figure 4.7.57 shows a schematic diagram for a generic pulse
sequence in a 2D NMR experiment. First, a pulse is administered to the sample in what is referred to as the preparation period. This
period could be anything from a single pulse to a complex pattern of pulses. The preparation period is followed by a “wait” time
(also known as the evolution time), t1, during which no data is observed. The evolution time also can be varied to suit the needs of
the specific experiment. A second pulse is administered next during what is known as the mixing period, where the coherence at
the end of t1 is converted into an observable signal, which is recorded during the observation time, t2. Figure 4.7.58 shows a
schematic diagram of how data is converted from the time domain (depicted in the free induction decay, or FID) to a frequency
domain. The process of this transformation using Fourier Transform (FT) is the same as it is in 1D NMR, except here, it is done
twice (or three times when conducting a 3D NMR experiment).

Figure adapted from J. Keeler, Understanding NMR Spectroscopy, 2nd, Wiley, West Sussex (2010).

Figure from J. Keeler, Understanding NMR Spectroscopy, 2nd, Wiley, West Sussex (2010).
In 1D NMR, spectra are plotted with frequency (in ppm or Hz, although most commonly ppm) on the horizontal axis and with
intensity on the vertical axis. However, in 2D NMR spectra, there are two frequency domains being plotted, each on the vertical
and horizontal axes. Intensity, therefore, can be shown as a 3D plot or topographically, much like a contour map, with more contour
lines representing greater intensities, as shown in Figure 4.7.59 a. Since it is difficult to read a spectrum in a 3D plot, all spectra are
plotted as contour plots. Furthermore, since resolution in a 2D NMR spectrum is not needed as much as in a 1D spectrum, data
acquisition times are often short.
2D NMR is very advantageous for many different applications, though it is mainly used for determining structure and
stereochemistry of large molecules such as polymers and biological macromolecules, that usually exhibit higher order splitting
effects and have small, overlapping coupling constants between nuclei. Further, some 2D NMR experiments can be used to
elucidate the components of a complex mixture. This module aims to describe some of the common two-dimensional NMR
experiments used to determine qualitative information about molecular structure.
2D Experiments
COSY
COSY (COrrelation SpectroscopY) was one of the first and most popular 2D NMR experiments to be developed. It is a
homonuclear experiment that allows one to correlate different signals in the spectrum to each other. In a COSY spectrum (see
Figure 4.7.59 b), the chemical shift values of the sample’s 1D NMR spectrum are plotted along both the vertical and horizontal
axes (some 2D spectra will actually reproduce the 1D spectra along the axes, along with the frequency scale in ppm, while others

4.7.39 https://chem.libretexts.org/@go/page/55887
may simply show the scale). This allows for a collection of peaks to appear down the diagonal of the spectrum known as diagonal
peaks (shown in Figure 4.7.59 b, highlighted by the red dotted line). These diagonal peaks are simply the peaks that appear in the
normal 1D spectrum, because they show nuclei that couple to themselves. The other type of peaks appears symmetric across the
diagonal and is known as cross peaks. These peaks show which groups in the molecule that have different chemical shifts are
coupled to each other by producing a signal at the intersection of the two frequency values.

Figure 4.7.59 Example of correlation spectroscopy: (a) On the left is shown a portion of a 3D or “stacked” plot of a 2D NMR
COSY spectrum in which two frequency domains are plotted in two dimensions and intensity is plotted in the third. On the right is
shown a contour plot, where the intensities have been depicted topographically. Spectra from Acorn NMR, Inc. (b) A spectrum of
the disaccharide xylobiose (structure shown), taken from a COSY 2D NMR experiment. The red dotted line highlights the diagonal
peaks. Spectrum adapted from F. Sauriol, NMR Webcourse, Department of Chemistry, Queen’s University, Ontario,
www.chem.queensu.ca/facilities/nmr/nmr/webcourse/.
One can then determine the structure of a sample by examining what chemical shift values the cross peaks occur at in a spectrum.
Since the cross peaks are symmetric across the diagonal peaks, one can easily identify which cross peaks are real (if a certain peak
has a counterpart on the other side of the diagonal) and which are digital artifacts of the experiment. The smallest coupling that can
be detected using COSY is dependent on the linewidth of the spectrum and the signal-to-noise ratio; a maximum signal-to-noise
ratio and a minimum linewidth will allow for very small coupling constants to be detected.
Variations of COSY
Although COSY is very useful, it does have its disadvantages. First of all, because the anti-phase structure of the cross peaks,
which causes the spectral lines to cancel one another out, and the in-phase structure of the diagonal peaks, which causes
reinforcement among the peaks, there is a significant difference in intensity between the diagonal and cross peaks. This difference
in intensity makes identifying small cross peaks difficult, especially if they lie near the diagonal. Another problem is that when
processing the data for a COSY spectrum, the broad lineshapes associated with the experiment can make high-resolution work
difficult.
In one of the more popular COSY variations known as DQF COSY (Double-Quantum Filtered COSY), the pulse sequence is
altered so that all of the signals are passed through a double-quantum coherence filter, which suppresses signals with no coupling
(i.e. singlets) and allows cross peaks close to the diagonal to be clearly visible by making the spectral lines much sharper. Since
most singlet peaks are due to the solvent, DQF COSY is useful to suppress those unwanted peaks.
ECOSY (Exclusive COrrelation SpectroscopY) is another derivative of COSY that was made to detect small J-couplings,
predominantly among multiplets, usually when J ≤ 3 Hz. Also referred to as long-range COSY, this technique involves adding a
delay of about 100-400 ms to the pulse sequence. However, there is more relaxation that is occurring during this delay, which
causes a loss of magnetization, and therefore a loss of signal intensity. This experiment would be advantageous for one who would
like to further investigate whether or not a certain coupling exists that did not appear in the regular COSY spectrum.
GS-COSY (Gradient Selective COSY) is a very applied offshoot of COSY since it eliminates the need for what is known as phase
cycling. Phase cycling is a method in which the phase of the pulses is varied in such a way to eliminate unwanted signals in the
spectrum, due to the multiple ways which magnetization can be aligned or transferred, or even due to instrument hardware. In
practical terms, this means that by eliminating phase cycling, GS-COSY can produce a cleaner spectrum (less digital artifacts) in
much less time than can normal COSY.

4.7.40 https://chem.libretexts.org/@go/page/55887
Another variation of COSY is COSY-45, which administers a pulse at 45° to the sample, unlike DQF COSY which administers a
pulse perpendicular to the sample. This technique is useful because one can elucidate the sign of the coupling constant by looking
at the shape of the peak and in which direction it is oriented. Knowing the sign of the coupling constant can be useful in
discriminating between vicinal and geminal couplings. However, COSY-45 is less sensitive than other COSY experiments that use
a 90° RF pulse.
TOCSY
TOCSY (TOtal Correlation SpectroscopY) is very similar to COSY in that it is a homonuclear correlation technique. It differs from
COSY in that it not only shows nuclei that are directly coupled to each other, but also signals that are due to nuclei that are in the
same spin system, as shown in Figure 4.7.60 below. This technique is useful for interpreting large, interconnected networks of spin
couplings. The pulse sequence is arranged in such a way to allow for isotropic mixing during the sequence that transfers
magnetization across a network of atoms coupled to each other. An alternative technique to 2D TOCSY is selective 1D TOCSY,
which can excite certain regions of the spectrum by using shaped pulses. By specifying particular chemical shift values and setting
a desired excitation width, one can greatly simplify the 1D experiment. Selective 1D TOCSY is particularly useful for analyzing
polysaccharides, since each sugar subunit is an isolated spin system, which can produce its own subspectrum, as long as there is at
least one resolved multiplet. Furthermore, each 2D spectrum can be acquired with the same resolution as a normal 1D spectrum,
which allows for an accurate measurement of multiplet splittings, especially when signals from different coupled networks overlap
with one another.

Figure from F. Sauriol, NMR Webcourse, Department of Chemistry, Queen’s University, Ontario,
www.chem.queensu.ca/facilities/nmr/nmr/webcourse/.
Heteronuclear Experiments
HETCOR (Heteronuclear Correlation) refers to a 2D NMR experiment that correlates couplings between different nuclei (usually
1
H and a heteroatom, such as 13C or 15N). Heteronuclear experiments can easily be extended into three or more dimensions, which
can be thought of as experiments that correlate couplings between three or more different nuclei. Because there are two different
frequency domains, there are no diagonal peaks like there are in COSY or TOCSY. Recently, inverse-detected HETCOR
experiments have become extremely useful and commonplace, and it will be those experiments that will be covered here. Inverse-
detection refers to detecting the nucleus with the higher gyromagnetic ratio, which offers higher sensitivity. It is ideal to determine
which nucleus has the highest gyromagnetic ratio for detection and set the probe to be the most sensitive to this nucleus. In
HETCOR, the nucleus that was detected first in a 1H -13C experiment was 13C, whereas now 1H is detected first in inverse-
detection experiments, since protons are inherently more sensitive. Today, regular HETCOR experiments are not usually in
common laboratory practice.
The HMQC (Heteronuclear Multiple-Quantum Coherence) experiment acquires a spectrum (see Figure 4.7.61 a) by transferring
the proton magnetization by way of 1JCH to a heteronucleus, for example, 13C. The 13C atom then experiences its chemical shift in
the t1 time period of the pulse sequence. The magnetization then returns to the 1H for detection. HMQC detects 1JCH coupling and
can also be used to differentiate between geminal and vicinal proton couplings just as in COSY-45. HMQC is very widely used and

4.7.41 https://chem.libretexts.org/@go/page/55887
offers very good sensitivity at much shorter acquisition times than HETCOR (about 30 min as opposed to several hours with
HETCOR).
However, because it shows the 1H -1H couplings in addition to 1H -13C couplings and because the cross peaks appear as multiplets,
HMQC suffers when it comes to resolution in the 13C peaks. The HSQC (Heteronuclear Single-Quantum Coherence) experiment
can assist, as it can suppress the 1H -1H couplings and collapse the multiplets seen in the cross peaks into singlets, which greatly
enhances resolution (an example of an HSQC is shown in Figure 4.7.61 b). Figure 4.7.61 shows a side-by-side comparison of
spectra from HMQC and HSQC experiments, in which some of the peaks in the HMQC spectrum are more resolved in the HSQC
spectrum. However, HSQC administers more pulses than HMQC, which causes miss-settings and inhomogeneity between the RF
pulses, which in turn leads to loss of sensitivity. In HMBC (Heteronuclear Multiple Bond Coherence) experiments, two and three
bond couplings can be detected. This technique is particularly useful for putting smaller proposed fragments of a molecule together
to elucidate the larger overall structure. HMBC, on the other hand, cannot distinguish between 2JCH and 3JCH coupling constants.
An example spectrum is shown in Figure 4.7.61 d.

Figure 4.7.59 b) taken from a 1H-13C HMQC 2D NMR experiment. (b) A spectrum of codeine taken from an HSQC 1H-13C 2D
NMR experiment. Spectrum from Acorn NMR, Inc. c) The chemical structure of codeine. d) Another spectrum of xylobiose taken
from a 1H-13C HMBC 2D NMR experiment. Panels (a) and (d) from F. Sauriol, NMR Webcourse, Department of Chemistry,
Queen’s University, Ontario, www.chem.queensu.ca/facilities/nmr/nmr/webcourse/.

Figure 4.7.62 Side-by-side comparison of an HMQC spectrum (a) and an HSQC spectrum (b). The HSQC experiment offers better
resolution than the HMQC as well as sharper peaks. HSQC helps solve the problem of overlapping peaks, which is often seen in
NMR experiments. The sample in both spectra is codeine. Spectra from Acorn NMR, Inc.
NOESY and ROESY
NOESY (Nuclear Overhauser Effect SpectroscopY) is an NMR experiment that can detect couplings between nuclei through
spatial proximity (< 5 Å apart) rather than coupling through covalent bonds. The Nuclear Overhauser Effect (NOE) is the change in
the intensity of the resonance of a nucleus upon irradiation of a nearby nucleus (about 2.5-3.5 Å apart). For example, when an RF
pulse specifically irradiates a proton, its spin population is equalized and it can transfer its spin polarization to another proton and
alter its spin population. The overall effect is dependent on a distance of r-6. NOESY uses a mixing time without pulses to
accumulate NOEs and its counterpart ROESY (Rotating frame nuclear Overhauser Effect SpectroscopY) uses a series of pulses to
accumulate NOEs. In NOESY, NOEs are positive when generated from small molecules, are negative when generated from large
molecules (or molecules dissolved in a viscous solvent to restrict molecular tumbling), and are quite small (near zero) for medium-
sized molecules. On the contrary, ROESY peaks are always positive, regardless of molecular weight. Both experiments are useful
for determine proximity of nuclei in large biomolecules, especially proteins, where two atoms may be nearby in space, but not
necessarily through covalent connectivity. Isomers, such as ortho-, meta-, and para-substituted aromatic rings, as well as
stereochemistry, can also be distinguished through the use of an NOE experiment. Although NOESY and ROESY can generate

4.7.42 https://chem.libretexts.org/@go/page/55887
COSY and TOCSY artifacts, respectively, those unwanted signals could be minimized by variations in the pulse sequences.
Example NOESY and ROESY spectra are shown in Figure 4.7.63.

Figures (b) and (d) from E. A. Khatuntseva, V.M. Men’shov, A.S. Shashkov, Y.E. Tsvetkov, R.N. Stepanenko, R.Y. Vlasenko, E.E.
Shults, G.A. Tolstikov, T.G. Tolstikova, D.S. Baev, V.A. Kaledin, N.A. Popova, V.P. Nikolin, P.P. Laktionov, A.V. Cherepanova,
T.V. Kulakovskaya, E.V. Kulakovskaya, and N.E. Nifantiev, Beilstein J. Org. Chem. 2012, 8, 763.

How to Interpret 2D NMR Spectra


Much of the interpretation one needs to do with 2D NMR begins with focusing on the cross peaks and matching them according to
frequency, much like playing a game of Battleship®. The 1D spectrum usually will be plotted along the axes, so one can match
which couplings in one spectrum correlate to which splitting patterns in the other spectrum using the cross peaks on the 2D
spectrum (see Figure 4.7.64).

Figure 4.7.59 b). By matching up the two couplings that intersect at the cross peaks, one can easily determine which atoms are
connected to which (shown by the blue dashed lines). The diagonal peaks are highlighted by the red line for clarity – the real
COSY information is within the cross peaks.
Also, multiple 2D NMR experiments are used to elucidate the structure of a single molecule, combining different information from
the various sources. For example, one can combine homonuclear and heteronuclear experiments and piece together the information
from the two techniques, with a process known as Parallel Acquisition NMR Spectroscopy or PANSY. In the 1990s, co-variance
processing came onto the scene, which allowed scientists to process information from two separate experiments, without having to
run both experiments at the same time, which made for shorter data acquisition time. Currently, the software for co-variance
processing is available from various NMR manufacturers. There are many possible ways to interpret 2D NMR spectra, though one
common method is to label the cross peaks and make connections between the signals as they become apparent. Prof. James

4.7.43 https://chem.libretexts.org/@go/page/55887
Nowick at UC Irvine describes his method of choice for putting the pieces together when determining the structure of a sample; the
lecture in which he describes this method is posted in the links above. In this video, he provides a stepwise method to deciphering a
spectrum.
Conclusion
Within NMR spectroscopy, there are a vast variety of different methods to acquire data on molecular structure. In 1D and 2D
experiments, one can simply adjust the appearance of the spectrum by changing any one of the many parameters that are set when
running a sample, such as number of scans, relaxation delay times, the amount of pulses at various angles, etc. Many 3D and 4D
NMR experiments are actually simply multiple 2D NMR pulse sequences run in sequence, which generates more correlation
between different nuclei in a spin system. With 3D NMR experiments, three nuclei, for example 1H, 13C, and 15N can be studied
together and their connectivity can be elucidated. These techniques become invaluable when working with biological molecules
with complex 3D structures, such as proteins and polysaccharides, to analyze their structures in solution. These techniques, coupled
with ultra-fast data acquisition, leads to monitoring complex chemical reactions and/or non-covalent interactions in real time.
Through the use of these and other techniques, one can begin to supplement a characterization “toolbox” in order to continue
solving complex chemical problems.

Chemical Exchange Saturation Transfer (CEST)


Paramagnetic chemical exchange saturation transfer (PARACEST) is a powerful analytical tool that can elucidate many physical
properties of molecules and systems of interest both in vivo and in vitro through specific paramagnetic agents. Although a
relatively new imaging technique, applications for PARACEST imaging are growing as new imaging agents are being developed
with enhanced exchange properties. Current applications revolve around using these PARACEST agents for MRI imaging to
enhance contrast. However, the fundamentals of PARACEST can be used to measure properties such as temperature, pH, and
concentration of molecules and systems as we will discuss. PARACEST was developed in response to several imaging limitations
presented by diamagnetic agents. PARACEST spectral data can be easily obtained using NMR Spectroscopy while imaging can be
typically achieved with widely available clinical 1.5/4 T MRI scanners.
History

Chemical exchange saturation transfer (CEST) is a phenomenon that has been around since the 1960s. It was first discovered by
Forsén, pictured below in Figure 4.7.65, and Hoffman in 1963 and was termed magnetization transfer NMR. This technique was
limited in its applications to studying rapid chemical exchange reactions. However in 2000, Balaban, pictured below in Figure
4.7.66, revisited this topic and discovered the application of this phenomenon for imaging purposes. He termed the phenomenon

chemical exchange saturation transfer. From this seminal finding, Balaban elucidated techniques to modulate MRI contrasts to
reflect the exchange for imaging purposes.

Figure 4.7.65 Swedish physical chemist Sture Forsén (1932-).

4.7.44 https://chem.libretexts.org/@go/page/55887
Figure 4.7.66 American chemist and biologist Robert S Balaban
CEST imaging focuses on N-H, O-H, or S-H exchangeable protons. Observing these exchanges in diamagnetic molecules can be
very challenging. Several models have been developed to overcome the challenges associated with imaging with clinical scanners.
The focus of recent research has been to develop paramagnetic chemical exchange saturation transfer (PARACEST) agents. Typical
PARACEST complexes are based on lanthanide atoms. Historically, these molecules were thought to be useless for chemical
exchange due to their very fast water exchanges rates. However, recent works by Silvio Aime and Dean Sherry have shown
modified lanthanide complexes that have very slow exchange rates that make it ideal for CEST imaging. In addition to slow
exchange rates, these molecules have vastly different resonance frequencies which contributes to their enhanced contrast.
Chemical Exchange Saturation Transfer

Saturation Transfer
Chemical exchange is defined as the process of proton exchange with surrounding bulk water. Exchange can occur with non-water
exchange sites but it has been shown that its’ contribution is negligible. As stated before, CEST imaging focuses on N-H, O-H, or
S-H exchangeable protons. Molecularly every exchange proton has a very specific saturation frequency. Applying a radio-
frequency pulse that is the same as the proton’s saturation frequency results in a net loss of longitudinal magnetization.
Longitudinal magnetization exists by virtue of being in a magnet. All protons in a solution line up with the magnetic field either in
a parallel or antiparallel manner. There is a net longitudinal magnetization at equilibrium as the antiparallel state is higher in
energy. A 90° RF pulse sequence causes many of the parallel protons to move to the higher energy antiparallel state causing zero
longitudinal magnetization. This nonequilibrium state is termed as saturation, where the same amount of nuclear spins is aligned
against and with the magnetic field. These saturated protons are exchangeable and the surrounding bulk water participates in this
exchange called chemical exchange saturation transfer.
This exchange can be visualized through spectral data. The saturated proton exchange with the surrounding bulk water causes the
spectral signal from the bulk water to decrease due to decreased net longitudinal magnetization. This decrease can then be
quantified and used to measure a wide variety of properties of a molecule or a solution. In the next sub-section, we will explore the
quantification in more detail to provide a stronger conceptual understanding.

Two-system Model
Derivations of the chemical exchange saturation transfer mathematical models arise fundamentally from an understanding of the
Boltzmann equation, 4.7.17. The Boltzmann equation mathematically defines the distribution of spins of a molecule placed in a
magnetic field. There are many complex models that are used to provide a better understanding of the phenomenon. However, we
will stick with a two-system model to simplify the mathematics to focus on conceptual understanding. In this model, there are two
systems: bulk water (alpha) and an agent pool (beta). When the agent pool is saturated with a radiofrequency pulse, we make two
important assumptions. The first is that all the exchangeable protons are fully saturated and the second is that the saturation process
does not affect the bulk water protons, which retain their characteristic longitudinal magnetization.
Nhigh energy −ΔE
  =  exp( ) (4.7.17)
Nlow energy kT

To quantify the following proton exchange we shall define the equilibrium proton concentration. The Boltzmann equation gives us
the distribution of the spin states at equilibrium which is proportional to the proton concentration. As such, we shall label the two
system’s equilibrium states as M and M . Following saturation, the saturated protons of the bulk pool exchange with the agent
0
α
β
0

pool at a rate k . As such the decrease in longitudinal (Z) magnetization is given by k M . Furthermore, another effect that needs
α α
Z
α

4.7.45 https://chem.libretexts.org/@go/page/55887
to be considered is the inherent relaxation of the protons which increase the Z magnetization back to equilibrium levels, M . This 0
α

can be estimated with the following 4.7.18 where T is the longitudinal relaxation time for bulk water. Setting the two systems

equal to represent equilibrium we get the following relationship 4.7.19 that can be manipulated mathematically to yield the
generalized chemical exchange Equation 4.7.20 where τ   = k and defined as lifetime of a proton in the system and c being the
α
−1
α

concentrations of protons in their respective system. [n] represents the number of exchangeable protons per CEST molecule. In
terms of CEST calculations, the lower the ratio of Z the more prominent the CEST effect. A plot of this equation over a range of
pulse frequencies results in what is called a Z-spectra also known as a CEST spectra, shown in Figure 4.7.67. This spectrum is then
used to create CEST Images.
0 Z
Mα − Mα
(4.7.18)
T1α

0 Z
Mα − Mα
Z
kα Mα   =   (4.7.19)
T1α

Z
Mα 1
Z  = = (4.7.20)
0 Cβ [n] T1α

1  +  
Cα τα

Figure 4.7.67 Solute protons are saturated with a specific resonance frequency shown here as 8.25 ppm. This saturation is
transferred to water at an exchange rate with unsaturated protons. After a brief period, this saturation effect becomes visible on the
water signal as a decrease in proton signal. Z-spectrum is generated by measuring the normalized water saturation (Ssat/S0) as a
function of irradiation frequency. Adapted from P. C. M. Van Zijl and N. N. Yadav, Magn. Reson. Med., 2011, 65, 927.

Limitations of Diamagnetic CEST Imaging and Two-system Model


A CEST agent must have several properties to maximize the CEST effect. Maximum CEST effect is observed when the residence
lifetime of bulk water ( τ ) is as short as possible. This indirectly means that an effective CEST agent has a high exchange rate,
α

k . Furthermore, maximum effect is noted when the CEST agent concentration is high.
α

In addition to these two properties, we need to consider the fact that the two-system model’s assumptions are almost never true.
There is often a less than saturated system resulting in a decrease in the observed CEST effect. As a result, we need to consider the
power of the saturation pulses, B1. The relationship between the τ and B1 is shown in the below 4.7.21. As such, an increase in
α

saturation pulse power results in increase CEST effect. However, we cannot apply too much B1 due to in vivo limitations.
Furthermore, the ideal τ can be calculated using the above relationship.
α

1
τ  =   (4.7.21)
2πB1

Finally, another limitation that needs to be considered is the inherent only to diamagnetic CEST and provides an important
distinction between CEST and PARACEST as we will soon discuss. We assumed with the two-system model that saturation with a
radiofrequency pulse did not affect the surrounded bulk water Z-magnetization. However, this is large generalization that can only
be made for PARACEST agents as we shall soon see. Diamagnetic species, whether endogenous or exogenous, have a chemical
shift difference (Δω) between the exchangeable –NH or –OH groups and the bulk water of less than 5 ppm. This small shift
difference is a major limitation. Selective saturation often lead to partial saturation of bulk water protons. This is a more important

4.7.46 https://chem.libretexts.org/@go/page/55887
consideration where in-vivo water peak is very broad. As such, we need to maximize the shift difference between bulk water and
the contrast agent.
Paramagnetic Chemical Exchange Saturation Transfer

Strengths of PARACEST
PARACEST addresses the two complications that arise with CEST. Application of a radio frequency pulse close to the bulk water
signal will result in some off-resonance saturation of the wa
ter. This essentially limits power which enhances CEST effect. Furthermore, a slow exchange condition less than the saturation
frequency difference (Δω) means that a very slow exchange rate is required for diamagnetic CEST agents of this sort. Both
problems can be alleviated by using an agent that has a larger chemical shift separation such as paramagnetic species. Figure 4.7.68
shows the broad Δω of Eu3+complex.

Figure 4.7.68 Eu3+ complex broadens the chemical shift leading to a larger saturation frequency difference that can easily be
detected. Red spectral line represents EuDOTA-(glycine ethyl ester)4. Blue spectral line represents barbituric acid. Adapted from
A. D. Sherry and M. Woods, Annu. Rev. Biomed. Eng., 2008, 10, 391.

Selection of Lanthanide Species


Based on the criteri a established in 4.7.22, we see that only Eu3+, Tb3+, Dy3+, and Ho3+ are effective lanthanide CEST agents at
the most common MRI power level (1.5 T). However, given stronger field strengths the Table 4.7.8 suggests more CEST
efficiency. With exception of Sm3+, all other lanthanide molecules have shifts far from water peak providing a large Δω that is
desired of CEST agents. This table should be considered before design of a PARACEST experiment. Furthermore, this table eludes
the relationship between power of the saturation pulse and the observed CEST effect. Referring to the following 4.7.23, we see that
for increased saturation pulse we notice increased CEST effect. In fact, varying B1 levels changes saturation offset. The higher the
B1frequency the higher the signal intensity of the saturation offset As such, it is important to select a proper saturation pulse before
experimentation.
Table 4.7.8 The chemical shifts and proton lifetime values for various lanthanide metals in a lanthanide DOTA-4AmCE complex (Figure
4.7.68 ).

Complex Tm at 298 K (μ s) δ 1H (ppm) Δω.τα at 1.5 T Δω.τα at 4.7 T Δω.τα at 11.75 T

Pr3+ 20 -60 0.5 1.5 3.8

Nd3+ 80 -32 1.0 3.2 8.0

Sm3+ 320 -4 0.5 1.6 4.0

Eu3+ 382 50 7.7 24.0 60.0

Tb3+ 31 -600 7.5 23.4 58.5

Dy3+ 17 -720 4.9 15.4 38.5

Ho3+ 19 -360 2.8 8.6 21.5

Er3+ 9 200 0.7 2.3 5.7

Tm3+ 3 500 0.6 1.9 4.7

Yb3+ 3 200 0.2 0.5 1.9

4.7.47 https://chem.libretexts.org/@go/page/55887
Based on the criteria established in 4.7.22, we see that only Eu3+, Tb3+, Dy3+, and Ho3+ are effective lanthanide CEST agents at the
most common MRI power level (1.5 T). However, given stronger field strengths the Table 4.7.9 suggests more CEST efficiency.
With exception of Sm3+, all other lanthanide molecules have shifts far from water peak providing a large Δω that is desired of
CEST agents. This table should be considered before design of a PARACEST experiment. Furthermore, this table eludes the
relationship between power of the saturation pulse and the observed CEST effect. Referring to the following 4.7.23, we see that for
increased saturation pulse we notice increased CEST effect. In fact, varying B1 levels changes saturation offset. The higher the
B1frequency the higher the signal intensity of the saturation offset As such, it is important to select a proper saturation pulse before
experimentation.

Figure 4.7.69 Structure of lanthanide DOTA-4AmCE complex.


1
Δω ⋅ τα   =   (4.7.22)
2πB1

1
τα   =   (4.7.23)
2πB1

Running a PARACEST Experiment


Two types of experiments can be run to quantify PARACEST. The first produces quantifiable Z-spectral data and is typically run
on 400 MHz spectrometers with a B1 power between 200-1000 KHz and an irradiation time between 2 and 6 seconds based on the
lanthanide complex. Imaging experiments are typically performed on either clinical scanners are small bore MRI scanner at room
temperature using a custom surface coil. Imaging experiments usually require the followings sequence of steps:
1. Bulk water spectra are collected from PARACEST using a 2 second presaturation pulse at a desired power level based on
lanthanide complex.
2. Following base scan, the saturation frequency is stepped between ±100 ppm (relative to the bulk water frequency at 0 ppm) in 1
ppm increments. The scanning frequency can be altered to include a wider scan if lanthanide complex has a larger chemical
shift difference.
3. Following collection of data, the bulk water signal is integrated using a Matlab program. The difference between the integrated
signals measured at equivalent positive and negative saturation frequencies are plotted and subtracted using the following
4.7.24 and mapped to produce gradient images.

4. To create a CEST Image the data set is first filtered to improve signal-to-noise ratio and normalized with phantom data by
subtraction and color-coded.
5. For data tools to perform CEST Imaging analysis. Please refer to the following links for free access to open source software
tools: https://github.com/cest-sources/CEST_EVAL/ or http://www.med.upenn.edu/cmroi/software-overview.html.
Ssat(−Δω)   −  Ssat(Δω)
(4.7.24)
S0

Applications of PARACEST

Temperature Mapping
PARACEST imaging has shown to be a promising area of research in developing a noninvasive technique for temperature
mapping. Sherry et. al shows a variable-temperature dependence of a lanthanide bound water molecule resonance frequency. They
establish a linear correspondence over the range of 20-50 °C. Furthermore, they show a feasible analysis technique to locate the
chemical shift (δ) of a lanthanide in images with high spatial resolution. By developing a plot of pixel intensity versus frequency
offset they can individually identify temperature at each pixel and hence create a temperature map as shown in the Figure 4.7.70.

4.7.48 https://chem.libretexts.org/@go/page/55887
Figure 4.7.70 Temperature map of a phantom containing 1 mL of 10 mM Eu in water at pH 7.0 in degrees Celsius. Adapted from
S. Zhang, C. R. Malloy, and A. D. Sherry, J. Am. Chem. Soc., 2005, 127, 17572.

Zinc Ion Detection


Divalent zinc is an integral transition-metal that is prominent in many aqueous solutions and plays an important role in
physiological systems. The ability to detect changes in sample concentrations of Zinc ions provides valuable information regarding
a system’s. Developing specific ligands that coordinate with specific ions to enhance wate-rexchange characteristics can amplify
CEST profile. In this paper, the authors develop a Eu(dotampy) sensor shown in Figure 4.7.71 for Zn ions. This authors theorize
that the sensor coordinates with Zinc using its four pyridine donors in a square anti-prism manner as determined by NMR
Spectroscopy by observing water exchange rates and by base catalysis by observing CEST sensitivity. Authors were unable to
analyze coordination by X-ray crystallography. Following, determination of successful CEST profiles, the authors mapped in-vitro
samples of varying concentrations of Zn and were successfully able to correlate image voxel intensity with Zn concentrations as
shown in Figure 4.7.72. Furthermore, they were able to successfully demonstrate specificity of the sensor to Zn over Magnesium
and Calcium. This application proves promising as a potential detection method for Zn ions in solutions with a range of
concentrations between 5 nm to 0.12 μm.

Figure 4.7.71 Structure of Eu(dotampy) where dotampy = 1,7-bis(N,N-bis(2-pyridylmethyl) aminoethylcarbamoylmethyl)-4,10-


bis(butylcarbamoylmethyl)-1,4,7,10-tetraazacyclododecane. The four Pyridine rings are hypothesized to serve as coordinators with
Zn leading to its CEST sensitivity and specificity.

Figure 4.7.72 CEST images of phantoms with varying concentrations of Zn in mM containing 20 mM of Eu(dotampy). The CEST
images represent the intensity difference between saturation at 50 ppm and 25 ppm from bulk water. Adapted from R. Trokowski, J.
Ren, F. K. Kálmán, and A. D. Sherry, Angew. Chemie., Int. Ed., 2005, 44, 6920.

4.7: NMR Spectroscopy is shared under a CC BY 4.0 license and was authored, remixed, and/or curated by Pavan M. V. Raja & Andrew R.
Barron via source content that was edited to conform to the style and standards of the LibreTexts platform; a detailed edit history is available upon
request.

4.7.49 https://chem.libretexts.org/@go/page/55887
4.8: EPR Spectroscopy
Basic Principles for EPR Spectroscopy
Electron paramagnetic resonance spectroscopy (EPR) is a powerful tool for investigating paramagnetic species, including organic
radicals, inorganic radicals, and triplet states. The basic principles behind EPR are very similar to the more ubiquitous nuclear
magnetic resonance spectroscopy (NMR), except that EPR focuses on the interaction of an external magnetic field with the
unpaired electron(s) in a molecule, rather than the nuclei of individual atoms. EPR has been used to investigate kinetics,
mechanisms, and structures of paramagnetic species and along with general chemistry and physics, has applications in
biochemistry, polymer science, and geosciences.
The degeneracy of the electron spin states is lifted when an unpaired electron is placed in a magnetic field, creating two spin states,
ms = ± ½, where ms = - ½, the lower energy state, is aligned with the magnetic field. The spin state on the electron can flip when
electromagnetic radiation is applied. In the case of electron spin transitions, this corresponds to radiation in the microwave range.
The energy difference between the two spin states is given by the equation

ΔE  =  E+ − E− = hν = gβB (4.8.1)

-34 -1 -24 -1
where h is Planck’s constant (6.626 x 10 J s ), v is the frequency of radiation, ß is the Bohr magneton (9.274 x 10 J T ), B is
the strength of the magnetic field in Tesla, and g is known as the g-factor. The g-factor is a unitless measurement of the intrinsic
magnetic moment of the electron, and its value for a free electron is 2.0023. The value of g can vary, however, and can be
calculated by rearrangement of the above equation, i.e.,

g = (4.8.2)
βB

using the magnetic field and the frequency of the spectrometer. Since h, v, and ß should not change during an experiment, g values
decrease as B increases. The concept of g can be roughly equated to that of chemical shift in NMR.

Instrumentation
EPR spectroscopy can be carried out by either 1) varying the magnetic field and holding the frequency constant or 2) varying the
frequency and holding the magnetic field constant (as is the case for NMR spectroscopy). Commercial EPR spectrometers typically
vary the magnetic field and holding the frequency constant, opposite of NMR spectrometers. The majority of EPR spectrometers
are in the range of 8-10 GHz (X-band), though there are spectrometers which work at lower and higher fields: 1-2 GHz (L-band)
and 2-4 GHz (S-band), 35 GHz (Q-band) and 95 GHz (W-band).

Figure 4.8.1 Block diagram of a typical EPR spectrometer.


EPR spectrometers work by generating microwaves from a source (typically a klystron), sending them through an attenuator, and
passing them on to the sample, which is located in a microwave cavity (Figure 4.8.1).
Microwaves reflected back from the cavity are routed to the detector diode, and the signal comes out as a decrease in current at the
detector analogous to absorption of microwaves by the sample.

4.8.1 https://chem.libretexts.org/@go/page/55889
Samples for EPR can be gases, single crystals, solutions, powders, and frozen solutions. For solutions, solvents with high dielectric
constants are not advisable, as they will absorb microwaves. For frozen solutions, solvents that will form a glass when frozen are
preferable. Good glasses are formed from solvents with low symmetry and solvents that do not hydrogen bond. Drago provides an
extensive list of solvents that form good glasses.
EPR spectra are generally presented as the first derivative of the absorption spectra for ease of interpretation. An example is given
in Figure 4.8.2.

Figure 4.8.2 Example of first and second derivative EPR spectrum.


Magnetic field strength is generally reported in units of Gauss or mTesla. Often EPR spectra are very complicated, and analysis of
spectra through the use of computer programs is usual. There are computer programs that will predict the EPR spectra of
compounds with the input of a few parameters.

Factors that Affect EPR Spectra


Hyperfine Coupling
Hyperfine coupling in EPR is analogous to spin-spin coupling in NMR. There are two kinds of hyperfine coupling: 1) coupling of
the electron magnetic moment to the magnetic moment of its own nucleus; and 2) coupling of the electron to a nucleus of a
different atom, called super hyperfine splitting. Both types of hyperfine coupling cause a splitting of the spectral lines with
intensities following Pascal’s triangle for I = 1/2 nuclei, similar to J-coupling in NMR. A simulated spectrum of the methyl radical
is shown in Figure 4.8.3. The line is split equally by the three hydrogens giving rise to four lines of intensity 1:3:3:1 with hyperfine
coupling constant a.

Figure 4.8.3 Simulated spectrum of CH3 radical with hyperfine coupling constant a.
The hyperfine splitting constant, known as a, can be determined by measuring the distance between each of the hyperfine lines.
This value can be converted into Hz (A) using the g value in the equation:
hA  =  gβa (4.8.3)

In the specific case of Cu(II), the nuclear spin of Cu is I = 3/2, so the hyperfine splitting would result in four lines of intensity
1:1:1:1. Similarly, super hyperfine splitting of Cu(II) ligated to four symmetric I = 1 nuclei, such as 14N, would yield nine lines
with intensities would be 1:8:28:56:70:56:28:8:1.

4.8.2 https://chem.libretexts.org/@go/page/55889
Anisotropy
The g factor of many paramagnetic species, including Cu(II), is anisotropic, meaning that it depends on its orientation in the
magnetic field. The g factor for anisotropic species breaks down generally into three values of g following a Cartesian coordinate
system which is symmetric along the diagonal: gx, gy, and gz. There are four limits to this system:
i. When gx = gy = gz the spectrum is considered to be isotropic, and is not dependent on orientation in the magnetic field.
ii. When gx = gy > gz the spectrum is said to be axial, and is elongated along the z-axis. The two equivalent g values are known as
g⊥ while the singular value is known as g‖. It exhibits a small peak at low field and a large peak at high field.
iii. When gx = gy < gz the spectrum is also said to be axial, but is shortened in the xy plane. It exhibits a large peak at low field and
a small peak at high field.
iv. When gx ≠ gy ≠ gz the spectrum is said to be rhombic, and shows three large peaks corresponding to the different components
of g.
Condition ii corresponds to Cu(II) in a square planar geometry with the unpaired electron in the dx2-y2 orbital. Where there is also
hyperfine splitting involved, g is defined as being the weighted average of the lines.

Electron Paramagnetic Resonance Spectroscopy of Copper(II) Compounds


Copper(II) Compounds
Copper compounds play a valuable role in both synthetic and biological chemistry. Copper catalyzes a vast array of reactions,
primarily oxidation-reduction reactions which make use of the Cu(I)/Cu(II) redox cycle. Copper is found in the active site of many
enzymes and proteins, including the oxygen carrying proteins called hemocyanins.
Common oxidation states of copper include the less stable copper(I) state, Cu+; and the more stable copper(II) state, Cu2+. Copper
(I) has a d10 electronic configuration with no unpaired electrons, making it undetectable by EPR. The d9 configuration of Cu2+
means that its compounds are paramagnetic making EPR of Cu(II) containing species a useful tool for both structural and
mechanistic studies. Two literature examples of how EPR can provide insight into the mechanisms of reactivity of Cu(II) are
discussed herein.
Copper (II) centers typically have tetrahedral, or axially elongated octahedral geometry. Their spectra are anisotropic and generally
give signals of the axial or orthorhombic type. From EPR spectra of copper (II) compounds, the coordination geometry can be
determined. An example of a typical powder Cu(II) spectrum is shown in Figure 4.8.4.

Figure 4.8.4 Typical axial EPR spectrum for a Cu(II) compound.


The spectrum above shows four absorption-like peaks corresponding to g‖ indicating coordination to four identical atoms, most
likely nitrogen. There is also an asymmetric derivative peak corresponding to g⊥ at higher field indicating elongation along the z
axis.
Determination of an Intermediate
The reactivity and mechanism of Cu(II)-peroxy systems was investigated by studying the decomposition of the Cu(II) complex 1
with EPR as well as UV-Vis and Raman spectroscopy. The structure (Figure 4.8.5) and EPR spectrum Figure 4.8.6 of 1 are given.
It was postulated that decomposition of 1 may go through intermediates LCu(II)OOH, LCu(II)OO•, or LCu(II)O• where L =
ligand.

4.8.3 https://chem.libretexts.org/@go/page/55889
Figure 4.8.5 Structure of 1, Cu(II) compound under investigation S = CH3CN

Figure 4.8.6 EPR spectrum of 1 in CH3CN at -150 °C showing g values of g1= 2.250, g2 = 2.065, g3 = 2.030, and hyperfine
coupling constant A1 = 160 G, A2 = 7 G, and A3 = 5 G. A. Kunishita, H. Ishimaru, S. Nakashima, T. Ogura, and S. Itoh, J. Am.
Chem. Soc., 2008, 130, 4244. Copyright American Chemical Society (2008).
To determine the intermediate, a common radical trap 5,5-dimethyl-1-pyrroline-N-oxide (DMPO) was added. A 1:1 complex of
intermediate and DMPO was isolated, and given the possible structure 2 (Figure 4.8.7, which is shown along with its EPR
specturm (Figure 4.8.8).

Figure 4.8.7 Proposed structure 2, S = CH3CN.

Figure 4.8.8 EPR spectrum of 1 in CH3CN at -150 °C showing g values of g1= 2.250, g2 = 2.065, g3 = 2.045, and hyperfine
coupling constant A1 = 170 G, A2 = 25 G, and A3 = 30 G. A. Kunishita, H. Ishimaru, S. Nakashima, T. Ogura, and S. Itoh, J. Am.
Chem. Soc., 2008, 130, 4244. Copyright American Chemical Society (2008).
The EPR data show similar though different spectra for Cu(II) in each compound, indicating a similar coordination environment –
elongated axial, and most likely a LCu(II)O• intermediate.

4.8.4 https://chem.libretexts.org/@go/page/55889
Determination of a Catalytic Cycle
The mechanism of oxidizing alcohols to aldehydes using a Cu(II) catalyst, TEMPO, and O2 was investigated using EPR. A
proposed mechanism is given in Figure 4.8.9.

Figure 4.8.9 Proposed mechanism for the Cu(II) mediated oxidation of alcohols to aldehydes with TEMPO and O2. M. Contel, P.
R. Villuendas, J. Fernández-Gallardo, P. Alonso, J. M. Vincent, and R. Fish, Inorg. Chem., 2005, 44, 9771. Copyright American
Chemical Society (2005).
EPR studies were conducted during the reaction by taking aliquots at various time points and immediately freezing the samples for
EPR analysis. The resulting spectra are shown in Figure 4.8.10.

Figure 4.8.10 EPR spectra of reaction at (a) 1.2 h (b) 4 h (c) 8 h, M. Contel, P. R. Villuendas, J. Fernández-Gallardo, P. Alonso, J.
M. Vincent, and R. Fish, Inorg. Chem., 2005, 44, 9771. Copyright American Chemical Society (2005)
The EPR spectrum (a) in Figure 6, after 1.2 hours shows a signal for TEMPO at g = 2.006 as well as a signal for Cu(II) with g‖=
2.26, g⊥ = 2.06, A‖ = 520 MHz, and A⊥ < 50 MHz. After 4 hours, the signal for Cu(II) is no longer in the reaction mixture, and
the TEMPO signal has decreased significantly. Suggesting that all the Cu(II) has been reduced to Cu(I) and the majority of TEMPO
has been oxidized. After 8 hours, the signals for both Cu(II) and TEMPO have returned indicating regeneration of both species. In
this way, EPR evidence supports the proposed mechanism.

Electron-Nuclear Double Resonance Spectroscopy


Electron nuclear double resonance (ENDOR) uses magnetic resonance to simplify the electron paramagnetic resonance (EPR)
spectra of paramagnetic species (one which contains an unpaired electron). It is very powerful and advanced and it works by
probing the environment of these species. ENDOR was invented in 1956 by George Feher (Figure 4.8.11).

Figure 4.8.11 American biophysicist George Feher (1924-).

4.8.5 https://chem.libretexts.org/@go/page/55889
ENDOR: NMR Spectroscopy on an EPR Spectrometer
A transition’s metal electron spin can interact with the nuclear spins of ligands through dipolar contact interactions. This causes
shifts in the nuclear magnetic resonance (NMR) Spectrum lines caused by the ligand nuclei. The NMR technique uses these dipolar
interactions, as they correspond to the nuclear spin’s relative position to the metal atom, to give information about the nuclear
coordinates. However, a paramagnetic species (one that contains unpaired electrons) complicates the NMR spectrum by broadening
the lines considerably.
EPR is a technique used to study paramagnetic compounds. However, EPR has its limitations as it offers low resolution that result
in line broadening and line splitting. This is partly due to the electron spins coupling to surrounding nuclear spins. However, this
coupling are important to understand a paramagnetic compound and determine the coordinates of its ligands. While neither NMR
or EPR can be used to study these coupling interaction, one can use both techniques simultaneously, which is the concept behind
ENDOR. An ENDOR experiment is a double resonance experiment in which NMR resonances are detected using intensity changes
of an EPR line that is irradiated simultaneously. An important difference is that the NRM portion of an ENDOR experiment uses
microwaves rather than radiofrequencies, which results in an enhancement of the sensitivity by several orders of magnitude.

Theory
The ENDOR technique involves monitoring the effects of EPR transitions of a simultaneously driven NMR transition, which
allows for the detection of the NMR absorption with much greater sensitivity than EPR. In order to illustrate the ENDOR system, a
two-spin system is used. This involves a magnetic field (Bo) interacting with one electron (S = 1/2) and one proton (I = 1/2).
Hamiltonian Equation
The Hamiltonian equation for a two-spin system is described by 4.8.4. The equation lists four terms: the electron Zeeman
interaction (EZ), the nuclear Zeeman interaction (NZ), the hyperfine interaction (HFS), respectively. The EZ relates to the
interaction the spin of the electron and the magnetic field applied. The NZ describes the interaction of the proton’s magnetic
moment and the magnetic field. The HSF is the interaction of the coupling that occurs between spin of the electron and the nuclear
spin of the proton. ENDOR spectra contain information on all three terms of the Hamiltonian.
H0   =  HEZ   +  HN Z   +  HH F S (4.8.4)

Selection Rules
4.8.4 can be further expanded to 4.8.5. gn is the nuclear g-factor, which characterizes the magnetic moment of the nucleus. S and I
are the vector operators for the spins of the electron and nucleus, respectively. μB is the Bohr magneton (9.274 x 10-24 JT-1). μn is
the nuclear magneton (5.05 x 10-27 JT-1). h is the Plank constant (6.626 x 10-34 J s). g and A are the g and hyperfine tensors. 4.8.5
becomes 4.8.6 by assuming only isotropic interactions and the magnetic filed aligned along the Z-axis. In 4.8.6, g is the isotropic
g-factor and a is the isotropic hyperfine constant.
H   =  μB B0 gS  −  gn μn B0 I   +  hSAI (4.8.5)

H   =  gμB B0 SZ − gn μn B0 IZ   +  haSI (4.8.6)

The energy levels for the two spin systems can be calculated by ignoring second order terms in the high filed approximation by
4.8.7. This equation can be used to express the four possible energy levels of the two-spin system (S = 1/2, I = 1/2) in 4.8.8 -

4.8.11

E(MS , MI ) = gμB B0 MS − gn μn B0 MI   +  haMS MI (4.8.7)

Ea   =   − 1/2gμB B0 − 1/2 gn μn B0 − 1/4ha (4.8.8)

Eb   =   + 1/2gμB B0 − 1/2 gn μn B0 + 1/4ha (4.8.9)

Ec   =   + 1/2gμB B0 + 1/2 gn μn B0 − 1/4ha (4.8.10)

Ed   =   − 1/2gμB B0 + 1/2 gn μn B0 + 1/4ha (4.8.11)

We can apply the EPR selection rules to these energy levels (ΔMI = 0 and ΔMS = ±1) to find the two possible resonance transitions
that can occur, shown in 4.8.12 and 4.8.13. These equations can be further simplified by expressing them in frequency units, where
νe = gμnB0/h to derive 4.8.14, which defines the EPR transitions (Figure 4.8.12). In the spectrum this would give two absorption
peaks that are separated by the isotropic hyperfine splitting, a (Figure 4.8.12).
ΔEcd   =  Ec   −  Ed   =  gμB B − 1/2ha (4.8.12)

4.8.6 https://chem.libretexts.org/@go/page/55889
ΔEab   =  Eb   −  Ea   =  gμB B + 1/2ha (4.8.13)

VEP R   =  ve ± a/2 (4.8.14)

Figure 4.8.12 Energy level diagram for a two spin system (S = 1/2 and I = 1/2) in a high magnetic field for the two cases where (a)
a>0 and a/2<νn and (b) a>0 and a/2>νn. The frequency of the two resulting ENDOR lines are given by νNMR = |νn±a/2| in (a) and
νNMR = |a/2±νn| in (b).

Applications
ENDOR has advantages in both organic and inorganic paramagnetic species as it is helpful in characterizing their structure in both
solution and in the solid state. First, it enhances the resolution gained in organic radicals in solution. In ENDOR, each group of
equivalent nuclei contributes only 2 lines to the spectrum, and nonequivalent nuclei cause only an additive increase as opposed to a
multiplicative increase like in EPR. For example, the radical cation 9,10-dimethilanthracene (Figure 4.8.14) would produce 175
lines in an EPR spectrum because the spectra would include 3 sets of inequivalent protons. However ENDOR produces only three
pairs of lines (1 for each set of equivalent nuclei), which can be used to find the hyperfine couplings. This is also shown in Figure
4.8.14.

Figure 4.8.14 EPR spectrum and corresponding 1H ENDOR spectrum of the radical cation of 9,10-dimethulanthracene in fluid
solution.
ENDOR can also be used to obtain structural information from the powder EPR spectra of metal complexes. ENDOR spectroscopy
can be used to obtain the electron nuclear hyperfine interaction tensor, which is the most sensitive probe for structure
determination. A magnetic filed that assumes all possible orientations with respect to the molecular frame is applied to the
randomly oriented molecules. The resonances from this are superimposed on each other and make up the powder EPR spectrum.
ENDOR measurements are made at a selected field position in the EPR spectrum, which only contain that subset of molecules that
have orientations that contribute to the EPR intensity at the chosen value of the observing field. By selected EPR turning points at
magnetic filed values that correspond to defined molecular orientations, a “single crystal like” ENDOR spectra is obtained. This is
also called a “orientation selective” ENDOR experiment which can use simulation of the data to obtain the principal components of

4.8.7 https://chem.libretexts.org/@go/page/55889
the magnetic tensors for each interacting nucleus. This information can then be used to provide structural information about the
distance and spatial orientation of the remote nucleus. This can be especially interesting since a three dimensional structure for a
paramagnetic system where a single crystal cannot be prepared can be obtained.

4.8: EPR Spectroscopy is shared under a CC BY 4.0 license and was authored, remixed, and/or curated by Pavan M. V. Raja & Andrew R. Barron
via source content that was edited to conform to the style and standards of the LibreTexts platform; a detailed edit history is available upon
request.

4.8.8 https://chem.libretexts.org/@go/page/55889
4.9: X-ray Photoelectron Spectroscopy
XPS of Carbon Nanomaterials
X-ray photoelectron spectroscopy (XPS), also called electron spectroscopy for chemical analysis (ESCA), is a method used to
determine the elemental composition of a material’s surface. It can be further applied to determine the chemical or electronic state
of these elements.
The photoelectric effect is the ejection of electrons from the surface of a material upon exposure to electromagnetic radiation of
sufficient energy. Electrons emitted have characteristic kinetic energies proportional to the energy of the radiation, according to
4.9.1, where KE is the kinetic energy of the electron, h is Planck’s constant, ν is the frequency of the incident radiation, Eb is the

ionization, or binding, energy, and φ is the work function. The work function is a constant which is dependent upon the
spectrometer.
KE  =  hν   −  Eb   −  φ (4.9.1)

In photoelectron spectroscopy, high energy radiation is used to expel core electrons from a sample. The kinetic energies of the
resulting core electrons are measured. Using the equation with the kinetic energy and known frequency of radiation, the binding
energy of the ejected electron may be determined. By Koopman’s theorem, which states that ionization energy is equivalent to the
negative of the orbital energy, the energy of the orbital from which the electron originated is determined. These orbital energies are
characteristic of the element and its state.

Basics of XPS
Sample Preparation
As a surface technique, samples are particularly susceptible to contamination. Furthermore, XPS samples must be prepared
carefully, as any loose or volatile material could contaminate the instrument because of the ultra-high vacuum conditions. A
common method of XPS sample preparation is embedding the solid sample into a graphite tape. Samples are usually placed on 1 x
1 cm or 3 x 3 cm sheets.
Experimental Set-up
Monochromatic aluminum (hν = 1486.6 eV) or magnesium (hν = 1253.6 eV) Kα X-rays are used to eject core electrons from the
sample. The photoelectrons ejected from the material are detected and their energies measured. Ultra-high vacuum conditions are
used in order to minimize gas collisions interfering with the electrons before they reach the detector.
Measurement Specifications
XPS analyzes material between depths of 1 and 10 nm, which is equivalent to several atomic layers, and across a width of about 10
µm. Since XPS is a surface technique, the orientation of the material affects the spectrum collected.
Data Collection
X-ray photoelectron (XP) spectra provide the relative frequencies of binding energies of electrons detected, measured in electron-
volts (eV). Detectors have accuracies on the order of ±0.1 eV. The binding energies are used to identify the elements to which the
peaks correspond. XPS data is given in a plot of intensity versus binding energy. Intensity may be measured in counts per unit time
(such as counts per second, denoted c/s). Often, intensity is reported as arbitrary units (arb. units), since only relative intensities
provide relevant information. Comparing the areas under the peaks gives relative percentages of the elements detected in the
sample. Initially, a survey XP spectrum is obtained, which shows all of the detectable elements present in the sample. Elements
with low detection or with abundances near the detection limit of the spectrometer may be missed with the survey scan. Figure
4.9.1 shows a sample survey XP scan of fluorinated double-walled carbon nanotubes (DWNTs).

4.9.1 https://chem.libretexts.org/@go/page/55891
Figure 4.9.1 Survey XP spectrum of F-DWNTs (O. Kuznetsov, Rice University).
Subsequently, high resolution scans of the peaks can be obtained to give more information. Elements of the same kind in different
states and environments have slightly different characteristic binding energies. Computer software is used to fit peaks within the
elemental peak which represent different states of the same element, commonly called deconvolution of the elemental peak. Figure
4.9.2 and Figure 4.9.3 show high resolutions scans of C1s and F1s peaks, respectively, from Figure 4.9.1, along with the peak

designations.

Figure 4.9.2 econvoluted high resolution C1s spectrum of F-DWNTs (O. Kuznetsov, Rice University).

Figure 4.9.3 Deconvoluted high resolution F1s spectrum of F-DWNTs (O. Kuznetsov, Rice University).
Limitations
Both hydrogen and helium cannot be detected using XPS. For this reason, XPS can provide only relative, rather than absolute,
ratios of elements in a sample. Also, elements with relatively low atomic percentages close to that of the detection limit or low
detection by XPS may not be seen in the spectrum. Furthermore, each peak represents a distribution of observed binding energies
of ejected electrons based on the depth of the atom from which they originate, as well as the state of the atom. Electrons from
atoms deeper in the sample must travel through the above layers before being liberated and detected, which reduces their kinetic
energies and thus increases their apparent binding energies. The width of the peaks in the spectrum consequently depends on the
thickness of the sample and the depth to which the XPS can detect; therefore, the values obtained vary slightly depending on the
depth of the atom. Additionally, the depth to which XPS can analyze depends on the element being detected.

4.9.2 https://chem.libretexts.org/@go/page/55891
High resolution scans of a peak can be used to distinguish among species of the same element. However, the identification of
different species is discretionary. Computer programs are used to deconvolute the elemental peak. The peaks may then be assigned
to particular species, but the peaks may not correspond with species in the sample. As such, the data obtained must be used
cautiously, and care should be taken to avoid over-analyzing data.

XPS for Carbon Nanomaterials


Despite the aforementioned limitations, XPS is a powerful surface technique that can be used to accurately detect the presence and
relative quantities of elements in a sample. Further analysis can provide information about the state and environment of atoms in
the sample, which can be used to infer information about the surface structure of the material. This is particularly useful for carbon
nanomaterials, in which surface structure and composition greatly influence the properties of the material. There is much research
interest in modifying carbon nanomaterials to modulate their properties for use in many different applications.
Sample Preparation
Carbon nanomaterials present certain issues in regard to sample preparation. The use of graphite tape is a poor option for carbon
nanomaterials because the spectra will show peaks from the graphite tape, adding to the carbon peak and potentially skewing or
overwhelming the data. Instead, a thin indium foil (between 0.1 and 0.5 mm thick) is used as the sample substrate. The sample is
simply pressed onto a piece of the foil.

Analysis and Applications for Carbon Nanomaterials


Chemical Speciation
The XP survey scan is an effective way to determine the identity of elements present on the surface of a material, as well as the
approximate relative ratios of the elements detected. This has important implications for carbon nanomaterials, in which surface
composition is of greatest importance in their uses. XPS may be used to determine the purity of a material. For example,
nanodiamond powder is a created by detonation, which can leave nitrogenous groups and various oxygen containing groups
attached to the surface. Figure 4.9.4 shows a survey scan of a nanodiamond thin film with the relative atomic percentages of
carbon, oxygen, and nitrogen being 91.25%, 6.25%, and 1.7%, respectively. Based on the XPS data, the nanodiamond material is
approximately 91.25% pure.

Figure 4.9.4 Survey XPS of a nanodiamond thin film. Adapted from F. Y. Xie, W. G. Xie, J. Chen, X. Liu, D. Y. Lu, and W. H.
Zhang, J. Vac. Sci. Tech. B, 2008, 26, 102.
XPS is a useful method to verify the efficacy of a purification process. For example, high-pressure CO conversion single-walled
nanotubes (HiPco SWNTs) are made using iron as a catalyst, Figure 4.9.5 shows the Fe2p XP spectra for pristine and purified
HiPco SWNTs.

4.9.3 https://chem.libretexts.org/@go/page/55891
Figure 4.9.5 High resolution scan of Fe2p peak for pristine and purified HiPco SWNTs. Adapted with permission from C. M.
Yang, H. Kanoh, K. Kaneko, M. Yudasaka, and S. Iijima, J. Phys. Chem. B, 2002, 106, 8994. Copyright: American Chemical
Society (2002).
For this application, XPS is often done in conjunction with thermogravimetric analysis (TGA), which measures the weight lost
from a sample at increasing temperatures. TGA data serves to corroborate the changes observed with the XPS data by comparing
the percentage of weight loss around the region of the impurity suspected based on the XP spectra. The TGA data support the
reduction in iron content with purification suggested by the XP spectra above, for the weight loss at temperatures consistent with
iron loss decreases from 27% in pristine SWNTs to 18% in purified SWNTs. Additionally, XPS can provide information about the
nature of the impurity. In Figure 4.9.6, the Fe2p spectrum for pristine HiPco SWNTs shows two peaks characteristic of metallic
iron at 707 and 720 eV. In contrast, the Fe2p spectrum for purified HiPco SWNTs also shows two peaks at 711 and 724 eV, which
are characteristic of either Fe2O3 or Fe3O4. In general, the atomic percentage of carbon obtained from the XPS spectrum is a
measure of the purity of the carbon nanomaterials.
Bonding and Functional Groups
XP spectra give evidence of functionalization and can provide insight into the identity of the functional groups. Carbon
nanomaterials provide a versatile surface which can be functionalized to modulate their properties. For example, the sodium salt of
phenyl sulfonated SWNTs is water soluble. In the XP survey scan of the phenyl sulfonated SWNTs, there is evidence of
functionalization owing to the appearance of the S2p peak. Figure 4.9.6 shows the survey XP spectrum of phenyl sulfonated
SWNTs.

Figure 4.9.6 Survey XP spectrum of phenyl sulfonated SWNTs. Adapted with permission from F. Liang, J. M. Beach, P. K. Rai, W.
H. Guo, R. H. Hauge, M. Pasquali, R. E. Smalley, and W. E. Billups, Chem. Mater., 2006, 18, 1520. Copyright: American
Chemical Society (2006).
The survey XP spectrum of the sodium salt shows a Na1s peak (Figure 4.9.7 and the high resolution scans of Na1s and S2p show
that the relative atomic percentages of Na1s and S2p are nearly equal (Figure 4.9.8, which supports the formation of the sodium
salt.

4.9.4 https://chem.libretexts.org/@go/page/55891
Figure 4.9.7 Survey XP spectrum of phenyl sulfonated SWNTs. Adapted with permission from F. Liang, J. M. Beach, P. K. Rai, W.
H. Guo, R. H. Hauge, M. Pasquali, R. E. Smalley, and W. E. Billups, Chem. Mater., 2006, 18, 1520. Copyright: American
Chemical Society (2006).

Figure 4.9.8 High resolution S2p (left) and Na1s (right) XP spectra of phenyl sulfonated SWNTs. Adapted with permission from F.
Liang, J. M. Beach, P. K. Rai, W. H. Guo, R. H. Hauge, M. Pasquali, R. E. Smalley, and W. E. Billups, Chem. Mater., 2006, 18,
1520. Copyright: American Chemical Society (2006).
Further Characterization

High resolution scans of each of the element peaks of interest can be obtained to give more information about the material. This is a
way to determine with high accuracy the presence of elements as well as relative ratios of elements present in the sample. This can
be used to distinguish species of the same element in different chemical states and environments, such as through bonding and
hybridization, present in the material. The distinct peaks may have binding energies that differ slightly from that of the convoluted
elemental peak. Assignment of peaks can be done using XPS databases, such as that produced by NIST. The ratios of the intensities
of these peaks can be used to determine the percentage of atoms in a particular state. Discrimination between and identity of
elements in different states and environments is a strength of XPS that is of particular interest for carbon nanomaterials.
Hybridization
The hybridization of carbons influences the properties of a carbon nanomaterial and has implications in its structure. XPS can be
used to determine the hybridization of carbons on the surface of a material, such as graphite and nanodiamond. Graphite is a carbon
material consisting of sp2 carbons. Thus, theoretically the XPS of pure graphite would show a single C1s peak, with a binding
energy characteristic of sp2 carbon (around 284.2 eV). On the other hand, nanodiamond consists of sp3 bonded carbons. The XPS
of nanodiamond should show a single C1s peak, with a binding energy characteristic of sp3 carbon (around 286 eV). The ratio of
the sp2 and sp3 peaks in the C1s spectrum gives the ratio of sp2 and sp3 carbons in the nanomaterial. This ratio can be altered and
compared by collecting the C1s spectra. For example, laser treatment of graphite creates diamond-like material, with more sp3
character when a higher laser power is used. This can be observed in Figure 4.9.9, in which the C1s peak is broadened and shifted
to higher binding energies as increased laser power is applied.

Figure 4.9.9 C1s high resolution XP spectra of graphite, nanodiamond, and graphite samples with increasing laser power treatment.
Adapted from P. Merel, M. Tabbal, M. Chaker, S. Moisa, and J. Margot, Appl. Surf. Sci., 1998, 136, 105.

4.9.5 https://chem.libretexts.org/@go/page/55891
Alternatively, annealing nanodiamond thin films at very high temperatures creates graphitic layers on the nanodiamond surface,
increasing sp2 content. The extent of graphitization increases with the temperature at which the sample is annealed, as shown in
Figure 4.9.10.

Figure 4.9.10 Deconvoluted high resolution C1s XP spectra for annealed nanodiamond. Adapted from F. Y. Xie, W. G. Xie, J.
Chen, X. Liu, D. Y. Lu, and W. H. Zhang, J. Vac. Sci. Tech. B, 2008, 26, 102.
Reaction Completion
Comparing the relative intensities of various C1s peaks can be powerful in verifying that a reaction has occurred. Fluorinated
carbon materials are often used as precursors to a broad range of variously functionalized materials. Reaction of fluorinated
SWNTs (F-SWNTs) with polyethyleneimine (PEI) leads to decreases in the covalent carbon-fluoride C1s peak, as well as the
evolution of the amine C1s peak. These changes are observed in the C1s spectra of the two samples (Figure 4.9.11).

Figure 4.9.11 High resolution C1s XP spectra of F-SWNTs (top) and PEI-SWNTs (bottom). Adapted with permission from E. P.
Dillon, C. A. Crouse, and A. R. Barron, ACS Nano, 2008, 2, 156. Copyright: American Chemical Society (2008).
Nature and Extent of Functionalization
XPS can also be applied to determine the nature and extent of functionalization. In general, binding energy increases with
decreasing electron density about the atom. Species with more positive oxidation states have higher binding energies, while more
reduced species experience a greater degree of shielding, thus increasing the ease of electron removal.
The method of fluorination of carbon materials and such factors as temperature and length of fluorination affect the extent of
fluoride addition as well as the types of carbon-fluorine bonds present. A survey scan can be used to determine the amount of
fluorine compared to carbon. High resolution scans of the C1s and F1s peaks can also give information about the proportion and
types of bonds. A shift in the peaks, as well as changes in peak width and intensity, can be observed in spectra as an indication of
fluorination of graphite. Figure 4.9.12 shows the Cls and F1s spectra of samples containing varying ratios of carbon to fluorine.

4.9.6 https://chem.libretexts.org/@go/page/55891
Figure 4.9.12 C1s and F1s high resolution XP spectra for graphite fluorides. Adapted from I. Palchan, M. Crespin, H. Estrade-
Szwarckopf, and B. Rousseau. Chem. Phys. Lett., 1989, 157, 321.
Furthermore, different carbon-fluorine bonds show characteristic peaks in high resolution C1s and F1s spectra. The carbon-fluorine
interactions in a material can range from ionic to covalent. Covalent carbon-fluorine bonds show higher core electron binding
energies than bonds more ionic in character. The method of fluorination affects the nature of the fluorine bonds. Graphite
intercalation compounds are characterized by ionic carbon-fluorine bonding. Figure 4.9.13 shows the F1s spectra for two
fluorinated exfoliated graphite samples prepared with different methods.

Figure 4.9.13 High resolution F1s XP spectra of two fluorinated exfoliated graphite samples. Adapted from A. Tressaud, F.
Moguet, S. Flandrois, M. Chambon, C. Guimon, G. Nanse, E. Papirer, V. Gupta, and O.P. Bahl. J. Phys. Chem. Solids, 1996, 57,
745.
Also, the peaks for carbons attached to a single fluorine atom, two fluorine atoms, and carbons attached to fluorines have
characteristic binding energies. These peaks are seen in that C1s spectra of F- and PEI-SWNTs shown in Figure 4.9.14.

Figure 4.9.14 High resolution C1s XP spectra of F-SWNTs (top) and PEI-SWNTs (bottom). Adapted with permission from E. P.
Dillon, C. A. Crouse, and A. R. Barron, ACS Nano, 2008, 2, 156. Copyright: American Chemical Society (2008).
Table 4.9.1 lists various bonds and functionalities and the corresponding C1s binding energies, which may be useful in assigning
peaks in a C1s spectrum, and consequently in characterizing the surface of a material.
Table 4.9.1 Summary of selected C1s binding energies

4.9.7 https://chem.libretexts.org/@go/page/55891
Bond/Group Binding Energy (eV)

C-C 284.0 - 286.0

C-C (sp2) 284.3 - 284.6

C-C (sp3) 285.0 - 286.0

C-N 285.2 - 288.4

C-NR2 (amine) 285.5 - 286.4

O=C-NH (amide) 287.9 - 288.6

-C=N (nitrile) 266.3 - 266.8

C-O 286.1-290.0

O=C-OH (carboxyl) 288.0 - 290.0

-C-O (epoxy) 286.1 - 287.1

-C-OH (hydroxyl) 286.4 - 286.7

-C-O-C- (ether) 286.1 - 288.0

-C=O (aldehyde/ketone) 287.1 - 288.1

C-F 287.0-293.4

-C-F (covalent) 287.7 - 290.2

-C-F (ionic) 287.0 - 287.4

C-C-F 286.0 - 287.7

C-F2 291.6 - 292.4

C-F3 292.4 - 293.4

C-S 285.2 - 287.5

C-Cl 287.0 - 287.2

Conclusion

X-ray photoelectron spectroscopy is a facile and effective method for determining the elemental composition of a material’s
surface. As a quantitative method, it gives the relative ratios of detectable elements on the surface of the material. Additional
analysis can be done to further elucidate the surface structure. Hybridization, bonding, functionalities, and reaction progress are
among the characteristics that can be inferred using XPS. The application of XPS to carbon nanomaterials provides much
information about the material, particularly the first few atomic layers, which are most important for the properties and uses of
carbon nanomaterials.

4.9: X-ray Photoelectron Spectroscopy is shared under a CC BY 4.0 license and was authored, remixed, and/or curated by Pavan M. V. Raja &
Andrew R. Barron via source content that was edited to conform to the style and standards of the LibreTexts platform; a detailed edit history is
available upon request.

4.9.8 https://chem.libretexts.org/@go/page/55891
4.10: ESI-QTOF-MS Coupled to HPLC and its Application for Food Safety
ESI-QTOF-MS Coupled to HPLC and its Application for Food Safety
High-performance liquid chromatography (HPLC) is a very powerful separation method widely used in environmental science,
pharmaceutical industry, biological and chemical research and other fields. Generally, it can be used to purify, identify and/or
quantify one or several components in a mixture simultaneously.
Mass spectrometry (MS) is a detection technique by measuring mass-to-charge ratio of ionic species. The procedure consists of
different steps. First, a sample is injected in the instrument and then evaporated. Second, species in the sample are charged by
certain ionized methods, such as electron ionization (EI), electrospray ionization (ESI), chemical ionization (CI), matrix-assisted
laser desorption/ionization (MALDI). Finally, the ionic species wil be analyzed depending on their mass-to-charge ratio (m/z) in the
analyzer, such as quadrupole, time-of-flight (TOF), ion trap and fourier transform ion cyclotron resonance.
The mass spectrometric identification is widely used together with chromatographic separation. The most common ones are gas
chromatography-mass spectrometry (GC-MS) and liquid chromatography-mass spectrometry (LC-MS). Because of the high
sensitivity, selectivity and relatively low price of GC-MS, it has very wide applications in drug detection, environmental analysis
and so forth. For those organic chemistry research groups, it is also a daily-used and convenient equipment. However, GC-MS is
ineffective if the molecules have high boiling point and/or will be decomposed at high temperature.
In this module, we will mainly talk about liquid chromatography and electrospray ionization quadrupole time-of-flight mass
spectrometry (LC/ESI-QTOF-MS). As mentioned above, the LC has an efficient capacity of separation and MS has a high
sensitivity and strong ability of structural characterization. Furthermore, TOF-MS, has several distinctive properties on top of
regular MS, including fast acquisition rates, high accuracy in mass measurements and a large mass range. The combination of LC
and ESI-TOF-MS allow us to obtain a powerful in the quantitative and qualitative analysis of molecules in complex matrices by
reducing the matrix interferences. It may play an important role in the area of food safety.

How it Works
Generally, LC-MS has four components, including an autosampler, HPLC, ionization source and mass spectrometer, as shown in
Figure 4.10.1. Here we need to pay attention to the interface of HPLC and MS so that they can be suitable to each other and be
connected. There are specified separation column for HPLC-MS, whose inner diameter (I.D.) is usually 2.0 mm. And the flow rate,
which is 0.05 - 0.2 mL/min, is slower than typical HPLC. For the mobile phase, we use the combination of water and methanol
and/acetonitrile. And because ions will inhibit the signals in MS, if we want to modify to mobile phase, the modifier should be
volatile, such as HCO2H, CH3CO2H, [NH4][HCO2] and [NH4][CH3CO2].

Figure 4.10.1 The component of a HPCL-MS system. Adapted from W. A . Korfmacher, Drug Discov. Today, 2005, 10, 1357.
As the interface between HPLC and MS, the ionization source is also important. There are many types and ESI and atmospheric
pressure chemical ionization (APCI) are the most common ones. Both of them are working at atmospheric pressure, high voltage
and high temperature. In ESI, the column eluent as nebulized in high voltage field (3 - 5 kV). Then there will be very small charged
droplet. Finally individual ions formed in this process and goes into mass spectrometer.

Comparison of ESI-QTOF-MS and Other Mass Spectrometer Methods


There are many types of mass spectrometers which can connect with the HPLC. One of the most widely-used MS systems is single
quadrupole mass spectrometer, whichis not very expensive, shown in Figure 4.10.2. This system has two modes. One mode is total
ion monitoring (TIM) mode which can provide the total ion chromatograph. The other is selected ion monitoring (SIM) mode, in
which the user can choose to monitor some specific ions, and the latter’s sensitivity is much higher than the former’s. Further, the
mass resolution of the single quadrupole mass spectrometer is 1 Da and its largest detection mass range is 30 - 3000 Da.

4.10.1 https://chem.libretexts.org/@go/page/55893
Figure 4.10.2 Single quadrupole mass spectrometer. Adapted from W. A. Korfmacher, Using Mass Spectrometry for Drug
Metabolism Studies, 1st Edition, Taylor & Francis, Abingdon (2004).
The second MS system is the triple quadrupole MS-MS system, shown in Figure 4.10.3. Using this system, people can select the
some ions, called parent ions, and use another electron beam to collide them again to get the fragment ions, called daughter ions. In
other words, there are two steps to select the target molecules. So it reduces the matrix effect a lot. This system is very useful in the
analysis of biological samples because biological samples always have very complex matrix; however, the mass resolution is still 1
Da.

Figure 4.10.3 Triple quadrupole mass spectrometer. Adapted from W. A. Korfmacher, Using Mass Spectrometry for Drug
Metabolism Studies, 1st Edition, Taylor & Francis, Abingdon (2004).
The third system is time-of-flight (TOF) MS, shown in Figure 4.10.4, which can provide a higher mass resolution spectrum, 3 to 4
decimals of Da. Furthermore, it can detect a very large range of mass at a very fast speed. The largest detection mass range is 20 -
10000 Da. But the price of this kind of MS is very high. The last technique is a hybrid mass spectrometer, Q-TOF MS, which
combines a single quadrupole MS and a TOF MS. Using this MS, we can get high resolution chromatograph and we also can use
the MS-MS system to identify the target molecules.

Figure 4.10.4 Time-of-flight mass spectrometer. Adapted from W. A. Korfmacher, Using Mass Spectrometry for Drug Metabolism
Studies, 1st Edition, Taylor & Francis, 2004.

Application of LC/ESI-QTOF-MS in the Detection of Quinolones in Edible Animal Food


Quinolones are a family of common antibacterial veterinary medicine which can inhibit DNA-gyrase in bacterial cells. However,
the residues of quinolone in edible animal products may be directly toxic or cause resistant pathogens in humans. Therefore,
sensitive methods are required to monitor such residues possibly present in different animal-producing food, such as eggs, chicken,
milk and fish. The molecular structures of eight quinolones, ciprofloxacin (CIP), anofloxacin methanesulphonate (DAN),
enrofloxacin (ENR), difloxacin (DIF), sarafloxacin (SARA), oxolinic, acid (OXO), flumequine (FLU), ofloxacin (OFL), are shown
in Figure 4.10.5.

4.10.2 https://chem.libretexts.org/@go/page/55893
Figure 4.10.5 The molecular structure of eight quinolones. Adapted from M. M. Zheng, G. D. Ruan, and Y. Q. Feng, J.
Chromatogr. A, 2009, 1216, 7510.
LC-MS is a common detection approach in the field of food safety. But because of the complex matrix of the samples, it is always
difficult to detect those target molecules of low concentration by using single quadrupole MS. The following gives an example of
the application of LC/ESI-QTOF-MS.
Using a quaternary pump system, a Q-TOF-MS system, a C18 column (250 mm × 2.0 mm I.D., 5 µm) with a flow rate of 0.2
mL/min, and a mixture of solvents as the mobile phase comprising of 0.3% formic acid solution and acetonitrile. The gradient
phofile for mobile phase is shown in Table 4.10.1. Since at acidic pH condition, the quinolones carried a positive charge, all mass
spectra were acquired in the positive ion mode and summarizing 30,000 single spectra in the mass range of 100-500 Da.
Table 4.10.1 The gradient phofile for mobile phase
Time (min) Volume % of Formic Acid Solution Volume % of Acetonitrile

0 80 20

12 65 35

15 20 80

20 15 85

30 15 85

30.01 80 20

The optimal ionization source working parameters were as follows: capillary voltage 4.5 kV; ion energy of quadrupole 5 eV/z; dry
temperature 200 °C; nebulizer 1.2 bar; dry gas 6.0 L/min. During the experiments, HCO2Na (62 Da) was used to externally
calibrate the instrument. Because of the high mass accuracy of the TOF mass spectrometer, it can extremely reduce the matrix
effects. Three different chromatographs are shown in Figure 4.10.6. The top one is the total ion chromatograph at the window
range of 400 Da. It’s impossible to distinguish the target molecules in this chromatograph. The middle one is at one Da resolution,
which is the resolution of single quadrupole mass spectrometer. In this chromatograph, some of the molecules can be identified.
But noise intensity is still very high and there are several peaks of impurities with similar mass-to-charge ratios in the
chromatograph. The bottom one is at 0.01 Da resolution. It clearly shows the peaks of eight quinolones with very high signal to
noise ratio. In other words, due to the fast acquisition rates and high mass accuracy, LC/TOF-MS can significantly reduce the
matrix effects.

4.10.3 https://chem.libretexts.org/@go/page/55893
Figure 4.10.6 Different chromatographs of 4 ng/g eight quinolones spiked in fish samples at different mass resolutions. Peaks: 1 =
OFL; 2 = CIP; 3 =DAN; 4 = ENR; 5 = SARA; 6 = DIF; 7 =OXO; 8 = FLU. Adapted from M. M. Zheng, G. D. Ruan, and Y. Q.
Feng, J. Chromatogr. A, 2009, 1216, 7510.
The quadrupole MS can be used to further confirm the target molecules. Figure 4.10.7 shows the chromatograms obtained in the
confirmation of CIP (17.1 ng/g) in a positive milk sample and ENR (7.5 ng/g) in a positive fish sample. The chromatographs of
parent ions are shown on the left side. On the right side, they are the characteristic daughter ion mass spectra of CIP and ENR.

Figure 4.10.7 Chromatograms obtained in the confirmation of CIP (17.1 ng/g) in positive milk sample and ENR (7.5 ng/g) in
positive fish sample. Adapted from M. M. Zheng, G. D. Ruan, and Y. Q. Feng, J. Chromatogr. A, 2009, 1216, 7510.

Drawbacks of LC/Q-TOF-MS
Some of the drawbacks of LC/Q-TOF-MS are its high costs of purchase and maintenance. It is hard to apply this method in daily
detection in the area of environmental protection and food safety.
In order to reduce the matrix effect and improve the detection sensitivity, people may use some sample preparation methods, such
as liquid-liquid extraction (LLE), solid-phase extraction (SPE), distillation. But these methods would consume large amount of
samples, organic solvent, time and efforts. Nowadays, there appear some new sample preparation methods. For example, people
may use online microdialysis, supercritical fluid extraction (SFE) and pressurized liquid extraction. In the method mentioned in the

4.10.4 https://chem.libretexts.org/@go/page/55893
Application part, we use online in-tube solid-phase microextraction (SPME), which is an excellent sample preparation technique
with the features of small sample volume, simplicity solventless extraction and easy automation.

4.10: ESI-QTOF-MS Coupled to HPLC and its Application for Food Safety is shared under a CC BY 4.0 license and was authored, remixed,
and/or curated by Pavan M. V. Raja & Andrew R. Barron via source content that was edited to conform to the style and standards of the
LibreTexts platform; a detailed edit history is available upon request.

4.10.5 https://chem.libretexts.org/@go/page/55893
4.11: Mass Spectrometry
Principles of Mass Spectrometry and Modern Applications
Mass spectrometry (MS) is a powerful characterization technique used for the identification of a wide variety of chemical
compounds. At its simplest, MS is merely a tool for determining the molecular weight of the chemical species in a sample.
However, with the high resolution obtainable from modern machines, it is possible to distinguish isomers, isotopes, and even
compounds with nominally identical molecular weights. Libraries of mass spectra have been compiled which allow rapid
identification of most known compounds, including proteins as large as 100 kDa (100,000 amu).
Mass spectrometers separate compounds based on a property known as the mass-to-charge ratio. The sample to be identified is first
ionized, and then passed through some form of magnetic field. Based on parameters such as how long it takes the molecule to
travel a certain distance or the amount of deflection caused by the field, a mass can be calculated for the ion. As will be discussed
later, there are a wide variety of techniques for ionizing and detecting compounds.
Limitations of MS generally stem from compounds that are not easily ionizable, or which decompose upon ionization. Geometric
isomers can generally be distinguished easily, but differences in chirality are not easily resolved. Complications can also arise from
samples which are not easily dissolved in common solvents.
Ionization Techniques
Electron Impact (EI)
In electon impact ionization, a vaporized sample is passed through a beam of electrons. The high energy (typically 70 eV) beam
strips electrons from the sample molecules leaving a positively charged radical species. The molecular ion is typically unstable and
undergoes decomposition or rearrangement to produce fragment ions. Because of this, electron impact is classified as a “hard”
ionization technique. With regards to metal-containing compounds, fragments in EI will almost always contain the metal atom (i.e.,
[MLn]+•fragments to [MLn-1]+ + L•, not MLn-1• + L+). One of the main limitations of EI is that the sample must be volatile and
thermally stable.
Chemical Ionization (CI)
In chemical ionization, the sample is introduced to a chamber filled with excess reagent gas (such as methane). The reagent gas is
ionized by electrons, forming a plasma with species such as CH5+, which react with the sample to form the pseudomolecular ion
[M+H]+. Because CI does not involve radical reactions, fragmentation of the sample is generally much lower than that of EI. CI
can also be operated in negative mode (to generate anions) by using different reagent gases. For example, a mixture of CH4 and
NO2 will generate hydroxide ions, which can abstract protons to yield the [M-H]- species. A related technique, atmospheric
pressure chemical ionization (APCI) delivers the sample as a neutral spray, which is then ionized by corona discharge, producing
ions in a similar manner as described above. APCI is particularly suited for low molecular weight, nonpolar species that cannot be
easily analyzed by other common techniques such as ESI.
Field Ionization/Desorption
Field ionization and desorption are two closely related techniques which use quantum tunneling of electrons to generate ions.
Typically, a highly positive potential is applied to an electrode with a sharp point, resulting in a high potential gradient at the tip
Figure 4.11.1. As the sample reaches this field, electron tunneling occurs to generate the cation, which is repelled into the mass
analyzer. Field ionization utilizes gaseous samples whereas in field desorption the sample is adsorbed directly onto the electrode.
Both of these techniques are soft, resulting in low energy ions which do not easily fragment.

Figure 4.11.1 Schematic of field ionization.


Electrospray Ionization (ESI)

4.11.1 https://chem.libretexts.org/@go/page/55894
In ESI, a highly charged aerosol is generated from a sample in solution. As the droplets shrink due to evaporation, the charge
density increases until a coulombic explosion occurs, producing daughter droplets that repeat the process until individualized
sample ions are generated (Figure 4.11.2. One of the limitations of is the requirement that the sample be soluble. ESI is best
applied to charged, polar, or basic compounds.

Figure 4.11.2 Schematic of electrospray ionization.


Matrix Assisted Laser Desorption Ionization (MALDI)
Laser desorption ionization generates ions by ablation from a surface using a pulsed laser. This technique is greatly improved by
the addition of a matrix co-crystallized with the sample. As the sample is irradiated, a plume of desorbed molecules is generated. It
is believed that ionization occurs in this plume due to a variety of chemical and physical interactions between the sample and the
matrix (Figure 4.11.3). One of the major advantages of MALDI is that it produces singly charged ions almost exclusively and can
be used to volatilize extremely high molecular weight species such as polymers and proteins. A related technique, desorption
ionization on silicon (DIOS) also uses laser desorption, but the sample is immobilized on a porous silicon surface with no matrix.
This allows the study of low molecular weight compounds which may be obscured by matrix peaks in conventional MALDI.

Figure 4.11.3 Schematic of matrix assisted laser desorption ionization.


Inductively Coupled Plasma Mass Spectrometry (ICP-MS)
A plasma torch generated by electromagnetic induction is used to ionize samples. Because the effective temperature of the plasma
is about 10,000 °C, samples are broken down to ions of their constituent elements. Thus, all chemical information is lost, and the
technique is best suited for elemental analysis. ICP-MS is typically used for analysis of trace elements.
Fast Atom Bombardment (FAB) and Secondary Ion Mass Spectrometry (SIMS)
Both of these techniques involve sputtering a sample to generate individualized ions; FAB utilizes a stream of inert gas atoms
(argon or xenon) whereas SIMS uses ions such as Cs+. Ionization occurs by charge transfer between the ions and the sample or by
protonation from the matrix material (Figure 4.11.4). Both solid and liquid samples may be analyzed. A unique aspect of these
techniques for analysis of solids is the ability to do depth profiling because of the destructive nature of the ionization technique.

Figure 4.11.4 Schematic of fast atom bombardment ionization.


Choosing an Ionization Technique
Depending on the information desired from mass spectrometry analysis, different ionization techniques may be desired. For
example, a hard ionization method such as electron impact may be used for a complex molecule in order to determine the

4.11.2 https://chem.libretexts.org/@go/page/55894
component parts by fragmentation. On the other hand, a high molecular weight sample of polymer or protein may require an
ionization method such as MALDI in order to be volatilized. Often, samples may be easily analyzed using multiple ionization
methods, and the choice is simplified to choosing the most convenient method. For example, electrospray ionization may be easily
coupled to liquid chromatography systems, as no additional sample preparation is required. Table 4.11.1 provides a quick guide to
ionization techniques typically applied to various types of samples.
Table 4.11.1 Strengths of various ionization techniques
Information Desired Ionization Technique

Elemental analysis Inductively coupled plasma

Depth profiling Fast atom bombardment/secondary ion mass spectroscopy

Chemical speciation/component analysis (fragmentation desired) Electron impact

Molecular species identification of compounds soluble in common


Electrospray ionization
solvents

Molecular species identification of hydrocarbon compounds Field ionization

Molecular species identification of high molecular weight compounds Matrix assisted laser desorption ionization

Molecular species identification of halogen containing compounds Chemical ionization (negative mode)

Mass Analyzers
Sectors
A magnetic or electric field is used to deflect ions into curved trajectories depending on the m/z ratio, with heavier ions
experiencing less deflection (Figure 4.11.5). Ions are brought into focus at the detector slit by varying the field strength; a mass
spectrum is generated by scanning field strengths linearly or exponentially. Sector mass analyzers have high resolution and
sensitivity, and can detect high mass ranges, but are expensive, require large amounts of space, and are incompatible with the most
popular ionization techniques MALDI and ESI.

Figure 4.11.5 Schematic of a magnetic sector mass analyzer.


Time of Flight (TOF)
The amount of time required for an ion to travel a known distance is measured (Figure 4.11.6). A pulse of ions is accelerated
through and electric analyzer such that they have identical kinetic energies. As a result, their velocity is directly dependent on their
mass. Extremely high vacuum conditions are required to extend the mean free path of ions and avoid collisions. TOF mass
analyzers are fastest, have unlimited mass ranges, and allow simultaneous detection of all species, but are best coupled with pulsed
ionization sources such as MALDI.

Figure 4.11.6 Schematic of a time-of-flight (TOF) mass analyzer.


Quadropole
Ions are passed through four parallel rods which apply a varying voltage and radiofrequency potential (Figure 4.11.7). As the field
changes, ions respond by undergoing complex trajectories. Depending on the applied voltage and RF frequencies, only ions of a
certain m/z ratio will have stable trajectories and pass through the analyzer. All other ions will be lost by collision with the rods.
Quadrupole analyzers are relatively inexpensive, but have limited resolution and low mass range.

4.11.3 https://chem.libretexts.org/@go/page/55894
Figure 4.11.7 Schematic of a quadrupole mass analyzer.

Ion Trap
Ion traps operate under the same principle as quadrupole, but contain the ions in space. Electrodes can be manipulated to
selectively eject ions of desired m/z ratios, allowing for mass analysis. Ion traps are uniquely suited for repeated cycles of mass
spectrometry because of their ability to retain ions of desired m/z ratios. Selected fragments can be further fragmented by collision
induced dissociation with helium gas. Ion traps are compact, relatively inexpensive, and can be adapted to many hybrid
instruments.
Coupling Mass Spectrometry to Other Instruments
Mass spectrometry is a powerful tool for identification of compounds, and is frequently combined with separation techniques such
as liquid or gas chromatography for rapid identification of the compounds within a mixture. Typically, liquid chromatography
systems are paired with ESI-quadrupole mass spectrometers to take advantage of the solvated sample. GC-MS systems usually
employ electron impact ionization and quadrupole or ion trap mass analyzers to take advantage of the gas-phase molecules and
fragmentation libraries associated with EI for rapid identification.
Mass spectrometers are also often coupled in tandem to form MS-MS systems. Typically the first spectrometer utilizes a hard
ionization technique to fragment the sample. The fragments are passed on to a second mass analyzer where they may be further
fragmented and analyzed. This technique is particularly important for studying large, complex molecules such as proteins.

Fast Atom Bombardment


Fast atom bombardment (FAB) is an ionization technique for mass spectroscopy employing secondary ion mass spectroscopy
(SIMS). Before the appearance of this technique, there was only limited way to obtain the mass spectrum of the intact oligopeptide
which is not easy to be vaporized. Prior to 1970, electron ionization (EI) or chemical ionization (CI) was widely used but those
methods require the destructive vaporization of the sample. Field desorption of ions with nuclear fission overcame this problem
though due to the necessity of special technique and nuclear fission of 252Cf limits the generality of this approach. FAB became
prevalent solving those underlying problems by using bombardment of fast atom or ion which has high kinetic energy onto the
sample in matrix.
Principle
The FAB utilizes the bombardment of accelerated atom or ion beams and the ionized sample is emitted by the collision of the
beams and the sample in matrix. In this section, the detail of each step is discussed.
Atom Beam
Although ions can be accelerated by electric field relatively easily, that is not the case for the neutral atom. Therefore, in the FAB
conversion of neutral atom into ion is significant to generate the accelerated species. The fast atom such as xenon used for the
bombardment is produced through three steps (Figure 4.11.8):
1. Ionization of the atom by collision with electron.
2. Acceleration of the generated ion through high electric potential.
3. Electron transfer from the accelerated ion to another slow atom, affording the desired accelerated atom.

4.11.4 https://chem.libretexts.org/@go/page/55894
Figure 4.11.8 Process of generation of fast atom.
Ion Beam
In the same way as the atom beam, a fast ion beam also can be used. Although cesium ion (Cs+) cheaper and heavier than xenon is
often employed, they have drawback that the mass spectroscopy can be contaminated by the ions.
Bombardment
The obtained fast atom or ion is then bombarded to the sample in matrix which is a type of solvent having high boiling point,
resulting in momentum transfer and vaporization of the sample (Figure 4.11.9). The fast atom used for the bombardment is called
primary beam of atoms or ions while secondary beam of atoms or ions corresponds to the sputtered ions and neutrals. The ionized
sample is directed by ion optics, leading to the detection of those ion in mass analyzer.

Figure 4.11.9 Bombardment of accelerated atom into sample. Only sputtered species with charge is introduced into ion optics and
detected by analyzer.
Matrices
One of the crucial characteristics of FAB is using liquid matrix. For example, long-lived signal in FAB is responsible for using
matrix. Due to the high vacuum condition, usual solvent for chemistry laboratory such as water and other common organic solvent
is precluded for FAB and, therefore, solvent with high boiling point called matrix is necessary to be employed. Table 4.11.1 shows
examples of matrix.
Table 4.11.1 Typical examples of matrices. Data from C. G. Herbert and R. A. W. Johnstone, Mass Spectrometry Basics, CRC Press, New
York (2002)
Matrix Observed Ions (m/z)

Glycerol 93

Thioglycerol 109

3-Nitrobenzyl alcohol (3-NOBA) 154

n-Octyl-3-nitrophenylether (NOP) 252

Triethanolamine 150

Diethanolamine 106

Polyethylene glycol (mixtures) Dependent on the glycol used

Instrument
An image of a typical instrument for fast atom bombardment mass spectrometry is shown in Figure 4.11.10.

4.11.5 https://chem.libretexts.org/@go/page/55894
Figure 4.11.10 Instrumentation of fast atom bombardment mass spectrometry.
Spectra
The obtained spectrum by FAB has information of structure or bond nature of the compound in addition to the mass. Here, three
spectrum are shown as examples.
Glycerol
Typical FAB mass spectrum of glycerol alone is shown in Figure 4.11.11.

Figure 4.11.11 A simplified FAB mass spectrum of glycerol.


Glycerol shows signal at m/z 93 which is corresponding to the protonated glycerol with small satellite derived from isotope of
carbon (13C). At the same time, signals for cluster of protonated glycerol are also often observed at m/z 185, 277, and 369. As is
seen in this example, signal from aggregation of the sample also can be detected and this will provide the information of the
sample.
Sulfonated Azo Compound
Figure 4.11.12 shows positive FAB spectrum of sulfonated azo compound X and structure of the plausible fragments in the
spectrum. The signal of the target compound X (Mw = 409) was observed at m/z 432 and 410 as an adduct with sodium and proton,
respectively. Because of the presence of some type of relatively weak bonds, several fragmentation was observed. For example,
signal at m/z 352 and 330 resulted from the cleavage of aryl-sulfonate bond. Also, nitrogen-nitrogen bond cleavage in the azo
moiety occurred, producing the fragment signal at m/z 267 and 268. Furthermore, taking into account the fact that favorable
formation of nitrogen-nitrogen triple bond from azo moiety, aryl-nitrogen bond can be cleaved and in fact those were detected at
m/z 253 and 252. As is shown in these example, fragmentation can be used for obtaining information regarding structure and bond
nature of desired compound.

Figure 4.11.12 Positive FAB spectrum of sulfonated azo compound X. Adapted from J. J. Monaghan, M. Barber, R. S. Bordoli, R.
D. Sedgwick, and A. N. Tyler, Int. J. Mass. Spectrom., 1983, 46, 447. Copyright: Elsevier (1983)
Bradykinin Potentiator C

4.11.6 https://chem.libretexts.org/@go/page/55894
The mass spectrum of protonated molecule (MH+ = m/z 1052) of bradykinin potentiator C is shown in Figure 4.11.13. In this case
fragmentation occurs between certain amino acids, affording the information of peptide sequence. For example, signal at m/z 884 is
corresponding to the fragment as a result of scission of Gly-Leu bond. It should be noted that the pattern of fragmentation is not
only done by one type of bond cleavage. Fragmentation at the bond between Gly-Pro is a good example; two type of fragment (m/z
533 and 520) are observed. Thus, pattern of fragmentation can afford the information of sequence of peptide.

Figure 4.11.13 FAB spectrum of Bradykinin Potentiator C (above) and pattern of fragmentation (below). Adapted from J. T.
Watson, D. S. Wagner, Y.-S. Chang, J. R. Strahler, S. M. Hanash, and D. A. Gage, Int. J. Mass. Spectrom., 1991, 111, 191.
Copyright: Elsevier (1991).

Secondary Ion Mass Spectrometry (SIMS)


Secondary ion mass spectrometry (SIMS) is an analytical method which has very low detection limits, is capable of analyzing over
a broad dynamic range, has high sensitivity, and has high mass resolution. In this technique, primary ions are used to sputter a solid
(and sometimes a liquid) surface of any composition. This causes the emission of electrons, ions, and neutral species, so called
secondary particles, from the solid surface. The secondary ions are then analyzed by a mass spectrometer. Depending on the
operating mode selected, SIMS can be used for surface composition and chemical structure analysis, depth profiling, and imaging.
Theory
Of all the secondary particles that are sputtered from the sample surface, only about 1 in every 1,000 is emitted as an ion. Because
only the ions may be detected by mass spectrometry, an understanding of how these secondary ions form is important.
Sputtering Models
Sputtering can be defined as the emission of atoms, molecules, or ions from a target surface as a result of particle bombardment of
the surface. This phenomenon has been described by two different sets of models.
The first approach to describe sputtering, called linear collision cascade theory, compares the atoms to billiard balls and assumes
that atomic collisions are completely elastic. Although there are a few different types of sputtering defined by this model, the type
which is most important to SIMS is slow collisional sputtering. In this type of sputtering, the primary ion collides with the surface
of the target and causes a cascade of random collisions between the atoms in the target. Eventually, these random collisions result
in the emission of an atom from the target surface, as can be seen in Figure 4.11.14. This model does not take into account the
location of atoms- it only requires that the energy of the incoming ion be higher than the energy required to sublimate atoms from
the target surface.

4.11.7 https://chem.libretexts.org/@go/page/55894
Figure 4.11.14 A diagram that illustrates linear collision cascade theory. The primary ion collides with an atom on the surface of
the target, causing other elastic collisions to occur within the target. Eventually, a target atom or molecule is ejected from the
surface.
Despite that fact that this method makes oversimplifications regarding atomic interactions and structure, its predicted sputter yield
data is actually fairly close to the experimental data for elements such as Cu, Zn, Ag, and Au, which have high sputter yields.
However, for low sputter yield elements, the model predicts three times more sputtered ions than what is actually observed.
The second method to describe sputtering uses computer-generated three-dimensional models of the atoms and molecules in the
sample to predict the effect of particle bombardment. All models under this category describe the target solid in terms of its
constituent atoms and molecules and their interactions with one another. However, these models only take into account atomic
forces (not electronic forces) and describe atomic behavior using classical mechanics (not quantum mechanics). Two specific
examples of this type of model are:
1. The molecular dynamics model
2. The binary collision approximation.
Ionization Models
The ionization models of sputtering can be divided into two categories, theories that predict ionization outside the target and
theories that predict that they are generated inside the target. In the theories that describe ionization outside of the target, the
primary particle strikes the target, causing the emission of an excited atom or molecule from the target. This particle relaxes by
emitting an Auger electron, thus becoming an ion. Because no simple mathematical equation has been described for this theory, it
is of little practical use. For this reason, ionization inside the target models are used more often. Additionally, it has been shown
that ionization occurs more often inside the target. Although there are many models that describe ionization within the target, two
representative models of this type are the bond-breaking model and the local thermal equilibrium theory.
In the bond breaking model, the primary particle strikes the target and causes the heterolytic cleavage of a bond in the target. So,
either an anion or a cation is emitted directly from the target surface. This is an important model to mention because it has useful
implications. Stated simply, the yield of positive ions can be increased by the presence of electronegative atoms in the target, in the
primary ion beam, or in the sample chamber in general. The reverse is also true- the negative ion yield may be increased by the
presence of electropositive atoms.
The local thermal equilibrium theory can be described as an expansion of the bond breaking model. Here, the increase in yield of
positive ions when the target is in the presence of electronegative atoms is said to be the result of the high potential barrier of the
metal oxide which is formed. This results in a low probability of the secondary ion being neutralized by an electron, thus giving a
high positive ion yield.
Instrumentation
Primary Ion Sources
The primary ions in a SIMS instrument (labeled “Primary ion source” in Figure 4.11.15) are generated by one of three types of ion
guns. The first type, called an electron bombardment plasma source, uses accelerating electrons (produced from a heated filament)
to bombard an anode. If the energy of these electrons is two to three times higher than the ionization energy of the atom, ionization
occurs. Once a certain number of ions and electrons are obtained, a plasma forms. Then, an extractor is used to make a focused ion
beam from the plasma.
In the second type of source, called the liquid metal source, a liquid metal film flows over a blunt needle. When this film is
subjected to a strong electric field, electrons are ejected from the atoms in the liquid metal, leaving them ionized. An extractor then
directs the ions out of the ion gun.
The last source is called a surface ionization source. Here, atoms of low ionization energy are absorbed onto a high work function
metal. This type of system allows for the transfer of electrons from the surface atoms to the metal. When the temperature is

4.11.8 https://chem.libretexts.org/@go/page/55894
increased, more atoms (or ions) leave the surface than absorb on the surface, causing an increase in absorbed ions compared to
absorbed atoms. Eventually, nearly all of the atoms that leave the surface are ionized and can be used as an ion beam.
The type of source used depends on the type of SIMS experiment which is going to be run as well as the composition of the sample
to be analyzed. A comparison of the three different sources is given in Table 4.11.2.
Table 4.11.2 A comparison of primary ion sources. Data from J.C. Vickerman, A. Brown, N.M. Reed, Secondary ion mass spectrometry:
Principles and applications, Clarendon Press, Oxford, 1989.
Source Spot Size (µm) Brightness (A/m2Sr) Energy Speed (eV) Ion Type

Electron Bombardment
1 104-107 <10 Ar+, Xe+, O2+
Plasma

Liquid Metal 0.05 1010 >10 Ga+, In+,Cs+

Surface Ionization 0.1 107 <1 Cs+

Of the three sources, electron bombardment plasma has the largest spot size. Thus, this source has a high-diameter beam and does
not have the best spatial resolution. For this reason, this source is commonly used for bulk analysis such as depth profiling. The
liquid metal source is advantageous for imaging SIMS because it has a high spatial resolution (or low spot size). Lastly, the surface
ionization source works well for dynamic SIMS (see above)
because its very small energy spread allows for a uniform etch rate.
In addition to the ion gun type, the identity of the primary ion is also important. O2+ and Cs+ are commonly used because they
enhance the positive or negative secondary ion yield, respectively. However, use of the inert gas plasma source is advantageous
because it allows for surface studies without reacting with the surface itself. Using the O2+ plasma source allows for an increased
output of positively charged secondary ions, but it will alter the surface that is being studied. Also, a heavy primary ion allows for
better depth resolution because it does not penetrate as far into the sample as a light ion.
Sputtering
The sputter rate, or the number of secondary ions that are removed from the sample surface by bombardment by one primary ion,
depends both on the properties of the target and on the parameters of the primary beam.
There are many target factors that affect the sputter rate. A few examples are crystal structure and the topography of the target.
Specifically, hexagonal close-packed crystals and rough surfaces give the highest sputter yield. There are many other properties of
the target which effect sputtering, but they will not be discussed here.
As was discussed earlier, different primary ion sources are used for different SIMS applications. In addition to the source used, the
manner in which the source is used is also important. First, the sputter rate can be increased by increasing the energy of the beam.
For example, using a beam of energy greater than 10 keV gives a maximum of 10 sputtered particles per primary ion impact.
Second, increasing the primary ion mass will also increase the secondary ion yield. Lastly, the angle of incidence is also important.
It has been found that a maximum sputter rate can be achieved if the angle of impact is 70° relative to the surface normal.
Mass Spectrometers
The detector which measures the amount and type of secondary ions sputtered from the sample surface is a mass spectrometer. See
Figure 4.11.15 for a diagram that shows where the mass spectrometer is relative to the other instrument components. The type of
analysis one wishes to do determines which type of spectrometer is used. Both dynamic and static SIMS usually use a magnetic
sector mass analyzer because it has a high mass resolution. Static SIMS (as well as imaging SIMS) may also use a time-of-flight
system, which allows for high transmission. A description of how each of these mass spectrometers works and how the ions are
detected can be found elsewhere (see https://cnx.org/contents/kl4gTdhf@1/Principles-of-Mass-Spectrometry-and-Modern-
Applications).

Figure 4.11.15 A diagram that shows the main components of a SIMS instrument. Diagram adapted from Wilson R.G. Wilson, F.A.
Stevie, and C.W. Magee, Secondary ion mass spectrometry: A practical handbook for depth profiling and bulk impurity analysis,
John Wiley & Sons Inc., New York, 1989.

4.11.9 https://chem.libretexts.org/@go/page/55894
Samples
SIMS can be used to analyze the surface and about 30 µm below the surface of almost any solid sample and some liquid samples.
Depending on the type of SIMS analysis chosen, it is possible to obtain both qualitative and quantitative data about the sample.
Technique Selection
There are three main types of SIMS experiments: Dynamic SIMS, static SIMS, and imaging SIMS.
In dynamic SIMS analysis, the target is sputtered at a high rate. This allows for bulk analysis when the mass spectrometer is
scanned over all mass ranges to get a mass spectrum and multiple measurements in different areas of the sample are taken. If the
mass spectrometer is set to rapidly analyze individual masses sequentially as the target is eroded rapidly, it is possible to see the
depth at which specific atoms are located up to 30 µm below the sample surface. This type of analysis is called a depth profile.
Depth profiling is very useful because it is a quantitative method- it allows for the calculation of concentration as a function of
depth so long as ion-implanted standards are used and the crater depth is measured. See the previous section for more information
on ion-implants.
SIMS may also be used to obtain an image in a way similar to SEM while giving better sensitivity than SEM. Here, a finely
focused ion beam (rather than an electron beam, as in SEM) is raster-scanned over the target surface and the resulting secondary
ions are analyzed at each point. Using the identity of the ions at each analyzed spot, an image may be assembled based on the
distributions of these ions.
In static SIMS, the surface of the sample is eroded very slowly so that the ions which are emitted are from areas which have not
already been altered by the primary ion. By doing this, it is possible to identify the atoms and some of the molecules just on the
surface of the sample.
An example that shows the usefulness of SIMS is the analysis of fingerprints using this instrument. Many other forms of analysis
have been employed to characterize the chemical composition of fingerprints such as GC-MS. This is important in forensics to
determine fingerprint degradation, to detect explosives or narcotics, and to help determine age of the person who left the print by
analyzing differences in sebaceous secretions. Compared to GC-MS, SIMS is a better choice of analysis because it is relatively less
destructive. In order to do a GC-MS, the fingerprint must be dissolved. SIMS, on the other hand, is a solid state method. Also,
because SIMS only erodes through a few monolayers, the fingerprint can be kept for future analysis and for record-keeping.
Additionally, SIMS depth profiling allows the researcher to determine the order in which substances were touched. Lastly, an
image of the fingerprint can be obtained using the imaging SIMS analysis.
Sample Preparation
As with any other instrumental analysis, SIMS does require some sample preparation. First, rough samples may require polishing
because the uneven texture will be maintained as the surface is sputtered. Because surface atoms are the analyte in imaging and
static SIMS, polishing is obviously not required. However, it is required for depth profiling. Without polishing, layers beneath the
surface of the sample will appear to be mixed with the upper layer in the spectrum, as can be seen in Figure 4.11.16.

Figure 4.11.16 An example of what a poorly resolved depth profile would look like. A better depth profile would show steep
slopes at the layer transition, rather than the gradual slopes seen here.
But, polishing before analysis does not necessarily guarantee even sputtering. This is because different crystal orientations sputter
at different rates. So, if the sample is polycrystalline or has grain boundaries (this is often a problem with metal samples), the
sample may develop small cones where the sputtering is occurring, leading to an inaccurate depth profile, as is seen in Figure
4.11.17.

4.11.10 https://chem.libretexts.org/@go/page/55894
Figure 4.11.17 A diagram that shows cone formation during sputtering as a result of the polycrystalline nature of the sample. This
leads to depth resolution degredation. Diagram adapted from R.G. Wilson, F.A. Stevie, and C.W. Magee, Secondary ion mass
spectrometry: A practical handbook for depth profiling and bulk impurity analysis,John Wiley & Sons Inc., New York, 1989.
Analyzing insulators using SIMS also requires special sample preparation as a result of electrical charge buildup on the surface
(since the insulator has no conductive path to diffuse the charge through). This is a problem because it distorts the observed spectra.
To prevent surface charging, it is common practice to coat the sample with a conductive layer such as gold.
Once the sample has been prepared for analysis, it must be mounted to the sample holder. There are a few methods to doing this.
One way is to place the sample on a spring loaded sample holder which pushes the sample against a mask. This method is
advantageous because the researcher doesn’t have to worry about adjusting the sample height for different samples (see below to
find out why sample height is important). However, because the mask is on top of the sample, it is possible to accidentally sputter
the mask. Another method used to mount samples is to simply glue them to a backing plate using silver epoxy. This method
requires drying under a heat lamp to ensure that all volatiles are evaporated off the glue before analysis. Alternatively, the sample
can be pressed in a soft metal like indium. The last two methods are especially useful for mounting of insulating samples, since
they provide a conductive path to help prevent charge buildup.
When loading the mounted sample into the instrument, it is important that the sample height relative to the instrument lens is
correct. If the sample is either too close or too far away, the secondary ions will either not be detected or they will be detected at the
edge of the crater being produced by the primary ions (see Figure 4.11.18). Ideally, the secondary ions that are analyzed should be
those resulting from the center of the primary beam where the energy and intensity are most uniform.

Figure 4.11.18 A diagram showing the importance of sample height in the instrument. If it is too high or too low, the sputtered ions
will not make it through the extraction lense. Diagram adapted from R.G. Wilson, F.A. Stevie, and C.W. Magee, Secondary ion
mass spectrometry: A practical handbook for depth profiling and bulk impurity analysis, John Wiley & Sons Inc., New York, 1989.
Standards
In order to do quantitative analysis using SIMS, it is necessary to use calibration standards since the ionization rate depends on both
the atom (or molecule) and the matrix. These standards are usually in the form of ion implants which can be deposited in the
sample using an implanter or using the primary ion beam of the SIMS (if the primary ion source is mass filtered). By comparing the
known concentration of implanted ions to the number of sputtered implant ions, it is possible to calculate the relative sensitivity
factor (RSF) value for the implant ion in the particular sample. By comparing this RSF value to the value in a standard RSF table
and adjusting all the table RSF values by the difference between them, it is possible to calculate the concentrations of other atoms
in the sample. For more information on RSF values, see above.
When choosing an isotope to use for ion implantation, it is important take into consideration possible mass interferences. For
example, 11B, 16O, and 27Al have the same overall masses and will interfere with each others ion intensity in the spectra. Therefore,
one must chose an ion implant that does not have the same mass as any other species in the sample which are of interest.
Also, the depth at which the implant is deposited is also important. The implanted ion must be lower than the equilibration depth,
above which, chaotic sputtering occurs until a sputter equilibrium is reached. However, care should be taken to ensure that the

4.11.11 https://chem.libretexts.org/@go/page/55894
implanted ions do not pass the layer of interest in the sample- if the matrix changes, the implanted ions will no longer sputter at the
same rate, causing concentrations to be inaccurate.
Matrix Effects
In SIMS, matrix effects are common and originate from changes in the ionization efficiency (the number of ionized species
compared to totally number of sputtered species) and the sputtering yield. One of the main causes of matrix effects is the primary
beam. As was discussed earlier, electronegative primary ions increases the number of positively charged secondary ions, while
electropositive primary ions increases the number of negatively charged secondary ions. Matrix effects can also be caused by
species present in the sample. The consequences of these matrix effects depends on the identity of the effecting species and the
composition of the sample. To correct for matrix effects, it is necessary to use a standards and compare the results with RSFs (see
above).
Detection Limits
For most atoms, SIMS can accurately detect down to a concentration of 1ppm. For some atoms, a concentration of 10 ppb may be
achieved. The detection limit in this instrument is set by the count rate (how many ions may be counted per second) rather than by
a limitation due to the mass of the ion. So, to decrease detection limit, the sample can be sputtered at a higher rate.
Sensitivity
The sensitivity of SIMS analysis depends on the element of interest, the matrix the element is in, and what primary ion is used. The
sensitivity of SIMS towards a particular ion may easily be determined by looking at an RSF table. So, for example, looking at an
RSF table for an oxygen primary ion and positive secondary ions shows that the alkali metals have the highest sensitivity (they
have low RSF values). This makes sense, since these atoms have the lowest electron affinities and are the easiest to ionize.
Similarly, looking at the RSF table for a cesium primary ion beam and negative secondary ions shows that the halogens have the
highest sensitivity. Again, this makes sense since the halogens have the highest electron affinities and accept electrons easily.
Data Interpretation
Three types of spectra can be obtained from a SIMS analysis. From static SIMS, a mass spectrum is produced. From dynamic
SIMS, a depth profile or mass spectrum is produced. And, not surprisingly, an image is produced from imaging SIMS.
Mass Spectra
As with a typical mass spectrum, the mass to charge ratio (m/z) is compared to the ion intensity. However, because SIMS is capable
of a dynamic range of 9 orders of magnitude, the intensity of the SIMS mass spectra is displayed on a logarithmic scale. From this
data, it is possible to observe isotopic data as well as molecular ion data and their relative abundances on the sample surface.
Depth Profile
A depth profile displays the intensity of one or more ions with respect to the depth (or, equivalently, time). Caution should be taken
when interpreting this data- if ions are collected off the wall of the crater rather than from the bottom, it will appear that the layer in
question runs deeper in the sample than it actually does.

Matrix Assisted Laser Desorption Ionization (MALDI)


Development of MALDI
As alluded to in previous sections, laser desorption (LD) was originally developed to produce ions in the gas phase. This is
accomplished by pulsing a laser on the sample surface to ablate material causing ionization and vaporization of sample particles.
However, the probability of attaining a valuable mass spectrum is highly dependent on the properties of the analyte. Furthermore,
masses observed in the spectrum were products of the molecular fragmentation if the molecular weight was above 500 Da. Clearly,
this was not optimal instrumentation for analyzing large biomolecules and bioinorganic compounds that do not ionize well and
samples were degraded during the process. Matrix-assisted laser desorption ionization (MALDI) was developed and alleviated
many issues associated with LD techniques. The MALDI technique allows proteins with masses up to 300,000 Da to be detected.
This is important to bioinorganic chemistry when visualizing products resulting from catalytic reactions, metalloenzyme
modifications, and other applications.
MALDI as a process decreases the amount of damage to the sample by protected the individual analytes within a matrix (more
information of matrices later). The matrix itself absorbs much of the energy introduced by the laser during the pulsing action. Plus,
energy absorbed by the matrix in subsequently transferred to the analyte (Figure 4.11.19). Once, energized, the analyte ionized and
is released into a plume of ions containing common cations (Na+, K+, etc.), matrix ions, and analyte ions. These ions then enter the
flight tube where they are sent to the detector. Different instrumental modes adjust for differences in ion flight time (Figure

4.11.12 https://chem.libretexts.org/@go/page/55894
4.11.19). The MALDI technique is also more sensitive and universal since readjustments to match absorption frequency is not
necessary due to the matrix absorption. Many of the commonly used matrices have similar wavelength absorptions Table 4.11.3.

Figure 4.11.19 Schematic representation of the MALDI process. The laser strikes the surface of the matrix-analyte sample causing
a “plume” of matrix and analyte ions that are received by the analyzer. Yellow circles represent the matrix, red ovals represent the
sample/analyte, and the small black circles represent cations.
Table 4.11.3 Table of different small molecules used as MALDI matrices.
Matrix Wavelength Application Structure

Cyano-4-hydroxycinnamic acid UV: 337nm, 353 nm Peptides

Proteins, peptides, non-covalent


6-Aza-2-thiothymine UV: 337 nm, 353 nm
complexes

k,m,n-Di(tri)hydroxy- UV: 337 nm, Proteins, peptides, non-covalent


acetophenone 353 nm complexes

2,5-Dihydroxybenzoic acid
Proteins, peptides, carbohydrates,
(requires 10% 2-hydroxy-5- UV: 337 nm, 353 nm
synthetic polymers
methoxybenzoic acid)

Sinapinic acid UV: 337 nm, 353 nm Proteins, peptides

Proteins, peptides, adduct


Nicotinic acid UV: 266 nm
formation

Succinic acid IR: 2.94 µm, 2.79 µm Proteins, peptides

Glycerol IR: 2.94 µm, 2.79 µm Proteins, peptides

Collection of MALDI Spectra


The process of MALDI takes place in 2 steps:
1. Sample preparation.

4.11.13 https://chem.libretexts.org/@go/page/55894
2. Sample ablation
Sample Preparation
The sample for analysis is combined with a matrix (a solvent containing small organic molecules that have a strong absorbance at
the laser wavelength) and added to the MALDI plate (Figure 4.11.18). The sample is then dried to the surface of the plate before it
is analyzed, resulting in the matrix doped with the analyte of interest as a "solid solution". Figure 4.11.20 shows the loading of a
peptide in water in cyano-4-hydroxycinnamic acid matrix.

Figure 4.11.20 The addition of the sample and matrix on to a MALDI plate, the samples are left until completely dry.
Prior to insertion of the plate into the MALDI instrument, the samples must be fully dried. The MALDI plate with the dry samples
is placed on a carrier and is inserted into the vacuum chamber (Figure 4.11.21a-b). After the chamber is evacuated, it is ready to
start the step of sample ablation.

Figure 4.11.21 (a) Image of the MALDI carrier released for sample loading. (b-c) Image of the sample plate loaded into the
MALDI carrier and insertion onto the instrument.
After the sample is loaded into the instrument, the instrument camera will show activate to show a live feed from inside of the
chamber. The live feed allows the controller to view the location where the spectrum is being acquired. This becomes especially
important when the operator manually fires the laser pulses.
Collection of a Spectrum
When the sample is loaded into the vacuum chamber of the instrument, there are several options for taking a mass spectrum. First,
there are several modes for the instrument, two of which are described here: axial and reflectron modes.
Axial Mode
In the axial (or linear) mode, only a very short ion pulse is required before the ions go down the flight tube and hit the detector.
This mode is often used when exact accuracy is not required since the mass accuracy has an error of about +/- 2-5%. Sources of
these errors are found in the arrival time of different ions through the flight tube to the detector. Errors in the arrival time are caused
by the difference in initial velocity with which the ions travel based on their size. The larger ions have a lower initial velocity, thus
they reach the detector after a longer period of time. This decreases the mass detection resolution.

4.11.14 https://chem.libretexts.org/@go/page/55894
Figure 4.11.22 MALDI instrument diagram depicting the axial mode. The laser strikes the surface of the matrix-analyte sample and
a plume of ions is released into the detector. The ions with higher energy travel faster compared to those with lower energy but the
same mass. These ions will hit the detector at different times causing some loss in resolution.
Reflectron Mode
In the reflectron (“ion mirror”) mode, ions are refocused before they hit the detector. The reflectron itself is actually a set of ring
electrodes that create an electric field that is constant near the end of the flight tube. This causes the ions to slow and reverse
direction towards a separate detector. Smaller ions are then brought closer to large ions before the group of ions hit the detector.
This assists with improving detection resolution and decreases accuracy error to +/- 0.5%.

4.11.15 https://chem.libretexts.org/@go/page/55894
Figure 4.11.23 MALDI instrument diagram depicting the reflectron mode. The laser strikes the surface of the matrix-analyte
sample and a plume of ions is released into the analyzer. Higher energy ions of the same mass travel down the flight tube faster
than lower energy ion of a similar mass, however in the reflectron mode this difference in corrected. Ring electrodes are activated
and create a uniform electric field. This slows the ions and redirects them into the reflectron detector. This increases resolution by
refocusing the ions to reach the detector at a similar time.
Example of MALDI Application

While MALDI is used extensively in analyzing proteins and peptides, it is also used to analyze nanomaterials. The following
example describes the analysis of fullerene analogues synthesized for a high performance conversion system for solar power. The
fullerene C60 is a spherical carbon molecule consisting of 60 sp2carbon atoms, the properties of which may be altered through
functionalization. A series of tert-butyl-4-C61-benzoate (t-BCB) functionalized fullerenes were synthesized and isolated. MALDI
was not used extensively as a method for observing activity, but instead was used as a conformative technique to determine the
presence of desired product. Three fullerene derivatives were synthesized (Figure 4.11.24).The identity and number of functional
groups were determined using MALDI (Figure 4.11.25).

Figure 4.11.24 A series of tert-butyl-4-C61-benzoate (t-BCB) functionalized fullerenes were synthesized and isolated. These
compounds were characterized by using MALDI to confirm desired product and isolation.

Figure 4.11.25 Mass for t-BCB-B (MW = 1100) was determined by using MALDI.

Surface-Assisted Laser Desorption/Ionization Mass Spectrometry (SALDI-MS)


Surface-assisted laser desorption/ionization mass spectrometry, which is known as SALDI-MS, is a soft mass spectrometry
technique capable of analyzing all kinds of small organic molecules, polymers and large biomolecules. The essential principle of

4.11.16 https://chem.libretexts.org/@go/page/55894
this method is similar to (matrix-assisted laser desorption/ionization mass spectrometry) MALDI-MS (see
http://cnx.org/contents/925e204d-d85...3e4d60057b37@1), but the organic matrix commonly used in MALDI has been changed
into the surface of certain substrates, usually inorganic compounds. This makes SALDI a matrix-free ionization technique that
avoids the interference of matrix molecules.
SALDI is considered to be a three-step process shown in Figure 4.11.26.
Samples are mixed with the substrates that provide large surface area to support sample molecule.
The samples are irradiated with IR of UV laser pulses when the energy of laser pulses are absorbed by the substrates and
transferred to the sample molecules.
Desorption and ionization process are initiated, which produces ions that are accelerated into the analyzer.
Since the bulk of energy input goes to substrates instead of the sample molecules, it is thought to be a soft ionization technique
useful in chemistry and chemical biology fields.

Figure 4.11.26 Schematic representations of SALDI-MS using various nanoparticles and nanostructures substrates, and an affinity
probe. Reproduced from: R. Arakawa, and H. Kawasaki, Anal. Sci., 2010, 26, 1229. Copyright: Japan Society for Analytical
Chemistry.
The most important characteristic of the substrate in SALDI is a large surface areas. In the past 30 years, efforts have been made to
explore novel substrate materials that increase the sensitivity and selectivity in SALDI-MS. Depending on the substrate compounds
being used, the interaction between the substrate materials and sample molecules could be covalent, non-covalent such as
hydrophobic effect, bio-specific such as recognition between biotins and avidins, and between antigens and antibodies, or
electrostatic. With the unique characteristics stated above, SALDI is able to combine the advantages of both hard and soft
ionization techniques. On one hand, low molecular weight (LMW) molecules could be analyzed and identified in SALDI-MS,
which resembles the function of most hard ionization techniques. On the other hand, molecular or quasi-molecular ions would
dominate the spectra as what we commonly see in the spectra prepared by soft ionization techniques.
History
The SALDI technique actually emerged from its well-known rival technique, MALDI. The development of soft ionization
techniques, which mainly included MALDI and ESI, enabled chemists and chemical biologists to analyze large polymers and
biomolecules using mass spectrometry. This should be attributed to the soft ionization process which prohibited large degree of
fragmentation that complicated the spectra, and resultant ions were dominantly molecular ions or quasi-molecular ions. In other
words, tolerance of impurities would be increased since the spectra became highly simplified. While it was effective in determining
molecular weight of the analytes, the matrix peaks would also appear in low mass range, which seriously interfered with the
analysis of LMW analytes. As a result, the SALDI method emerged to resolve the problem by replacing matrix with surface that
was rather stationary.
The original idea of SALDI was raised by Tanaka (Figure 4.11.27) in 1988. Ultra-fine cobalt powders with an average diameter of
about 300 Å that were mixed in the sample were responsible of “rapid heating” due to its high photo-absorption and low heat
capacity. With a large surface area, the cobalt powders were able to conduct heat to large numbers of surrounding glycerol liquid

4.11.17 https://chem.libretexts.org/@go/page/55894
and analyte molecules, which indeed resulted in a thermal desorption/ionization mechanism. The upper mass limit was increased
up to 100 kDa, which is shown in Figure 4.11.28 for the analysis of lysozymes from chicken egg white.

Figure 4.11.27 Japanese engineer Koichi Tanaka, recipient of the Nobel Prize in Chemistry in 2002.

Figure 4.11.28 Laser ionization mass spectrum of lysozyme (enzymes that are able to damage bacteria cell walls) from chicken egg
white. Molecular weight 14306 Da.Reproduced from: K. Tanaka, H. Waki, Y. Ido, S. Akita, and Y. Yoshida, and T. Yoshida, Rapid
Commun. Mass Spectrom., 1988, 2, 151. Copyright: John Wiley and Sons.
The low mass range was not paid much attention at the beginning, and the concept of “surface-assisted” was not proposed until
Sunner (Figure 4.11.29) and co-workers reported the study on graphite SALDI in 1995. And that was the first time the term
“SALDI” was used by chemists. They achieved obtaining mass spectra of both proteins and LWM analytes by irradiating mixture
of 2-150 μm graphite particles and solutions of analytes in glycerol. Although fragmentation of the LMW glycerol molecules was
relatively complicated (Figure 4.11.30), it was still considered as a significant improvement in ionizing small molecules by soft
ionization methods.

Figure 4.11.29 American chemist Jan Sunner.

4.11.18 https://chem.libretexts.org/@go/page/55894
Figure 4.11.30 Graphite surface-assisted laser desorption mass spectrum (337 nm) of glycerol. Average of 82 laser shots. The ions
at m/z=93 is protonated glycerol. The ions at m/z= 75, 57, 45, 29, and 19 are all fragments of GI. The ions at m/z= 23, 39, and 133
are Na+, K+, and Cs+. The ions at 115, 131, and 225 are alkali ion/glycerol adducts, Na+(glycerol), K+(glycerol), and Cs+(glycerol).
Reproduced from J. Sunner, E. Dratz, and T. Chen, Anal Chem, 1995, 67, 4335. Copyright: American Chemical Society.

Despite the breakthrough mentioned above, SALDI did not widely interest chemists. Regardless of its drawbacks in upper mass
limit for the analysis of large molecules, the sensitivity was far from being satisfactory compared to hard ionization techniques in
terms of testing LMW molecules. This situation has been changed ever since nanomaterials were introduced as the substrates,
especially for the successful development of desorption/ionization on porous silicon (DIOS) shown in Figure 4.11.31. In fact,
majority of research on SALDI-MS has been focusing on exploiting novel nanomaterial substrates, aiming at further broadening
the mass range, improving the reproducibility, enhancing the sensitivity and extending the categories of compounds that were able
to be analyzed. So far, a variety of nanomaterials have been utilized in SALDI-MS, including carbon-based nanomaterials, metal-
based nanomaterials, semiconductor-based nanomaterials, etc.

Figure 4.11.31 Experimental configuration for the DIOS-MS experiments. (a) Four porous silicon plates are placed on a MALDI
plate. Each of the porous silicon plates contains photopatterned spots or grids prepared through illumination of n-type silicon with a
300-W tungsten filament through a mask and an f/50 reducing lens. (b) The silicon-based laser desorption/ionization process, in
which the sample is placed on the porous silicon plate and allowed to dry, followed by laser-induced desorption/ionization mass
spectrometry. (c) Cross-section of porous silicon, and the surface functionalities after hydrosilylation; R represents phenyl or alkyl
chains. Reproduced from J. Wei, J. M. Buriak, and G. Siuzdak Nature, 1999, 399, 243. Copyright: Nature Publishing Group.
Basic Principles of SALDI
Mechanism of Desorption and Ionization
As a soft ionization technique, SALDI is expected to produce molecular or quasi-molecular ions in the final mass spectra. Since
this requires the ionization process to be both effective and controllable, which means sufficient sample molecules could be ionized
while further fragmentation should be mostly avoided.
While the original goal mentioned above has been successfully accomplished for years, the study on desorption and ionization
mechanism in detail is still one of the most popular and controversial research areas of SALDI at present. It is mostly agreed that
the substrate material has played a significant role of both activating and protecting the analyte molecules. The schematic picture
describing the entire process is shown in Figure 4.11.33. Energy input from the pulsed laser is largely absorbed by the substrate
material, which is possibly followed by complicated energy transfer from the substrate material to the absorbed analyte molecules.

4.11.19 https://chem.libretexts.org/@go/page/55894
As a result, both thermal and non-thermal desorption could be triggered, and for different modes of SALDI experiments, the
specific desorption and ionization process greatly differs.

Figure 4.11.33 Schematic picture of surface assisted laser desorption/ionization. The gray spheres and blue spheres represent
substrate materials and analyte molecules, respectively.
The mechanism for porous silicon surface as a SALDI substrate has been widely studied by researchers. In general, the process can
be subdivided into the following steps:
1. Adsorption of neutral analyte molecules takes places by formation of hydrogen bonds with surface silanol groups;
2. Electronic excitation of the substrate under the influence of the laser pulse generates a free electron/“hole” pair. This separation
causes enrichment of positive charges near the surface layer; as a result, the acidity of the silanol groups increases and proton
transfer to analytes becomes easier;
3. Analyte ions are thermally activated and thus dissociated from the surface.
When no associated proton donor is present in the vicinity of analyte molecules, desorption might occur without ionization.
Subsequently, the desorbed analyte molecule is ionized in the gas phase by collision with incoming ions.
Signal Enhancement Factors on SALDI Substrates

Since it is the active surface responsible for adsorption, desorption and ionization of analyte molecules that features the technique,
the surface chemistry of substrate material is undoubtedly crucial for SALDI performance. But it is rather difficult to draw a
general conclusion due to the fact that the affinity between different classes of substrates and analytes is considerably versatile.
Basically, the interaction between those two components has an impact on trapping and releasing the analyte molecules, as well as
the electronic surface state of the substrate and energy transfer coefficiency.
Another important aspect is the physical properties of the substrates which could alter desorption and ionization process directly,
especially for the thermally activated pathway. This is closely related to rapid temperature increase on the substrate surface. Those
properties include optical absorption coefficiency, heat capacity and heat conductivity (or heat diffusion rate). First, higher optical
absorption coefficiency enables the substrate to absorb and generate more heat when certain amount of energy is provided by the
laser source. Moreover, a lower heat capacity usually leads to larger temperature increase upon the same amount of heat. In
addition, a lower hear conductivity helps the substrate to maintain a high temperature that will further result in a higher temperature
peak. Therefore, the thermal desorption and ionization could occur more rapidly and effectively.
Instrumentation
The instrument involved in SALDI shown in Figure 4.11.34 is similar with in MALDI to large extent. It contains a laser source
which generates pulsed laser that excites the sample mixture. There is a sample stage that places the sample mixture of substrate
materials and analytes. Usually the mass analyzer and ion detector are on the other side to let the ions go through and become
separated and detected based on different m/z value. Recent progress has been made that incorporates direct analysis in real time
(DART) ion source into SALDI-MS system which makes it possible to perform the analysis in ambient conditions. Figure 4.11.35
shows their ambient SALDI-MS method.

4.11.20 https://chem.libretexts.org/@go/page/55894
Figure 4.11.34 Schematic illistration of SALDI instrument.

Figure 4.11.35 Schematic illustration of the performance of the ambient SALDI-MS. Reproduced from J. Zhang, Z. Li, C. Zhang,
B. Feng, Z. Zhou, Y. Bai, and H. Liu, Anal. Chem., 2012, 84, 3296. Copyright: American Chemical Society, 2012.
Examples of Nanomaterials Used For Analysis of LMW Analytes in SALDI-MS
Porous Silicon as a Substrate Material
Porous silicon with large surface area could be used to trap certain analyte molecules for matrix-free desorption and ionization
process. More interestingly, a large ultraviolet absorption coefficiency was found for this porous material, which also improved the
ionization performance. It has been reported that using porous silicon as the substrate in SALDI-MS was able to work at femtomole
and attomole levels of analytes including as peptides, caffeine, an antiviral drug molecule (WIN), reserpine and N-octyl-β-D-
glucopyranoside . Compared to conventional MALDI-MS, the DIOS-MS (which was the specific type of SALDI in this research)
successfully eliminated the matrix interference and displayed much higher quasi-molecular peak (MH+), which could be observed
in Figure 4.11.36. What’s more, chemical modification of the porous silicon was able to further optimize the ionization
characteristics.

Figure 4.11.36 Analysis of WIN antiviral drug (500 fmol) using different desorption/ionization techniques. (a) MALDI-MS of
same amount of antiviral drug usinga-cyano-4-hydroxycinnamic matrix. (b) DIOS-MS spectrum of WIN. Accurate mass
measurements were obtained on WIN with a time-of-flight reflectron instrument, typically to within 10 ppm (the limit of accuracy
of this instrument in this mass range). The inset spectrum represents post-source decay fragmentation measurements on WIN.
These results are consistent with results from electrospray ionization tandem mass spectrometry experiments. Reproduced from J.
Wei, J. M. Buriak, and G. Siuzdak, Nature, 1999, 399, 243. Copyright: Nature Publishing Group, 1999.

4.11.21 https://chem.libretexts.org/@go/page/55894
Graphene as a Surface Material
Graphene is a type of popular carbon nanomaterial discovered in 2004. It has a large surface area that could effectively attach the
analyte molecules. On the other hand, the efficiency of desorption/ionization for analytes on a layer of graphene can be enhanced
for its simple monolayer structure and unique electronic properties. Polar compounds including amino acids, polyamines,
anticancer drugs, and nucleosides can be successfully analyzed. In addition, nonpolar molecules can be analyzed with high
resolution and sensitivity due to the hydrophobic nature of graphene itself. Compared with a conventional matrix, graphene
exhibited a high desorption/ionization efficiency for nonpolar compounds. The graphene substrate functions as a substrate to trap
analytes, and it transfers energy to the analytes upon laser irradiation, which allows for the analytes to be readily desorbed/ionized
and the interference of matrix to be eliminated. It has been demonstrated that the use of graphene as a substrate material avoids the
fragmentation of analytes and provides good reproducibility and a high salt tolerance, underscoring the potential application of
graphene as a matrix for MALDI-MS analysis of practical samples in complex sample matrixes. It is also proved that the use of
graphene as an adsorbent for the solid-phase extraction of squalene could improve greatly the detection limit.
Combination with GC

Gas-phase SALDI-MS analysis has a relatively high ionization efficiency, which leads to a high sensitivity. In 2009, gas
chromatography (GC) was first used with SALDI-MS, where the SALDI substrate was amorphous silicon and the analyte was N-
alkylated phenylethylamines. Detection limits were in the range of attomoles, but improvements are expected in the future. The
combination with GC is expected to expand the use of SALDI-MS even more that SALDI could be applied to separation and
identification of samples with more complexity. The instrumental setup is shown in Figure 4.11.37.

Figure 4.11.37 Instrument setup of GC-SALDI-MS. Reproduced from: S. Alimpiev, A. Grechnikov, J. Sunner, A. Borodkov, V.
Karavanskii, Y. Simanovsky, and S. Nikiforov, Anal. Chem., 2009, 81, 1255. Copyright: American Chemical Society, 2009.

Differential Electrochemical Mass Spectrometry


In the study of electrochemistry, it had always been a challenge to obtain immediate and continuous detection of electrochemical
products due to the limited formation on the surface of the electrode, until the discovery of differential electrochemical mass
spectrometry. Scientists initially tested the idea by combining porous membrane and mass spectrometry for product analysis in the
study of oxygen generation from HClO4 using porous electrode in 1971. In 1984, another similar experiment was performed using
a porous Teflon membrane with 100 μm of lacquers at the surface between the electrolytes and the vacuum system. Comparing to
previous experiment, this experiment has demonstrated a vacuum system with improved time derivative that showed nearly
immediate detection of volatile electrochemical reaction products, with high sensitivity of detecting as small as “one monolayer” at
the electrode. In summary, the experiment demonstrated in 1984 not only showed continous sample detection in mass spectrometry
but also the rates of formation, which distinguished itself from the technique performed previously in 1971. Hence, this method
was called differential electrochemical mass spectrometry (DEMS). During the past couple decades, this technique has evolved
from using classic electrode to rotating disc electrode (RDE), which provides a more homogeneous and faster transport of reaction
species to the surface of the electrode.
Described in basic terms, differential electrochemical mass spectrometry is a characterization technique that analyzes specimens
using both the electrochemical half-cell experimentation and mass spectrometry. It uses non-wetting membrane to separate the
aqueous electrolyte and gaseous electrolyte, which gaseous electrolyte will permeate through the membrane and will be ionized and
detected in the mass spectrometer using continuous, two-stage vacuum system. This analytical method can detect gaseous or
volatile electrochemical reactants, reaction products, and even reaction intermediates. The instrument consists of three major

4.11.22 https://chem.libretexts.org/@go/page/55894
components: electrochemical half-cell, PTFE (polytetrafluoroethylene) membrane interface, and quadrupole mass spectrometer
(QMS), which is a part of the vacuum system.
DEMS Operations
The entire assembly of the instrument is shown in Figure 4.11.38, which consists of three major components: an electrochemical
half-cell, a PTFE membrane interface, and the quadrupole mass spectrometer. In this section, each component will be explained
and its functionality will be explored, and additional information will be provided at the end of this section. The PTFE membrane is
micro-porous membrane that separates the aqueous electrolyte from volatile electrolyte which will be drawn to the high vacuum
portion. Using the high vacuum suction, the gaseous or volatile species will be allowed to permeate through the membrane using
differential pressure, leaving the aqueous materials on the surface due to hydrophobic nature of the membrane. The selection of the
membrane material is very important to maintain both the hydrophobicity and proper diffusion of volatile species. The species
permeated to QMS will be monitored and measured, and the kinetics of formation will be determined at the end. Depending on the
operating condition, different vacuum pumps might be required.

Figure 4.11.38 Mechanical setup of the entire differential electrochemical mass spectrometry. Adapted from Aston, S.J., “Design,
Construction and Research Application of a Differential Electrochemical Mass Spectrometer (DEMS)”, Springer, 2011, 9-27.
Electrochemical Cells
First major component of the DEMS instrument is the design of electrochemical cells. There are many different designs that have
been developed for the past several decades, depending on the types of electrochemical reactions, the types and sizes of electrodes.
However, only the classic cell will be discussed in this chapter.
DEMS method was first demonstrated using the classical method. A conventional setup of electrochemical cell is showed in Figure
4.11.39. The powdered electrode material is deposited on the porous membrane to form the working electrode, shown as Working

Electrode Material in Figure 4.11.39. In the demonstration by Wolber and Heitbaum, the electrode was prepared by having small Pt
particles deposited onto the membrane by painting a lacquer. It was later in other experimentations evolved to use sputtering
electro-catalyst layer for a more homogenous surface. The aqueous cell electrolyte is shielded with an upside down glass body with
vertical tunnel opening to the PTFE membrane. The working electrode material lies above the PTFE membrane, where it is
supported mechanically by stainless steel frit inside vacuum flange. Both the working electrode material and PTFE membrane are
compressed between vacuum castings and PTFE spacer, which is a ring that prevents the electrolyte from leakage. The counter
electrode (CE) and reference electrode (RE) made from platinum wire are placed on top of the working electrode material to create
the electrical contact. One of the main advantages of the classical design is fast respond time, with high efficiency of “0.5 for
lacquer and 0.9 with the sputter electrode”. However, this method poses certain difficulties. First, the electrolyte materials will be
absorbed on the working electrode before it permeates through the membrane. Due to the limitation of absorption rate, the
concentration on the surface of the electrode will be lower than bulk. Second, the aqueous volatile electrolyte must be absorbed
onto working electrode, and then followed by evaporation through the membrane. Therefore, the difference in rates of absorption
and evaporation will create a shift in equilibrium. Third, this method is also limited to the types of material that can be deposited on
the surface, such as single crystals or even some polycrystalline electrode surfaces. Lastly, the way RE is position could potentially
introduce impurity into the system, which will interfere with the experiment.

4.11.23 https://chem.libretexts.org/@go/page/55894
Figure adapted from J. S. Aston, Design, Construction and Research Application of a Differential Electrochemical Mass
Spectrometer (DEMS), Springer-Verlag, Berlin Heidelberg (2011).
Membrane Interface

PTFE membrane is placed between the aqueous electrolyte cell and the high vacuum system on the other end. It acts as a barrier
that prevents aqueous electrolyte from passing through, while its selectivity allows the vaporized electrochemical species to
transport to the high vacuum side, which the process is similar to vacuum membrane distillation shown in Figure 4.11.41. In order
to prevent the aqueous solution from penetrating through the membrane, the surface of the membrane must be hydrophobic, which
is a material property that repels water or aqueous fluid. Therefore, at each pore location, there is vapor and liquid interface where
the liquid will remain on the surface while the vapor will penetrate into the membrane. Then the transportation of the material in
vapor phase is triggered by the pressure difference created from the vacuum on the other end of the membrane. Therefore, the size
of the pore is crucial in controlling its hydrophobic properties and the transfer rate through the membrane. When the pore size is
less than 0.8 μm, the hydrophobic property is activated. This number is determined by calculating the surface tension of liquid, the
contact angle and the applied pressure. Therefore, a membrane with relatively small pore sizes and large pore distribution is
desired. In general membrane materials used are “typically 0.02 μm in size with thickness between 50 and 110 μm”. In terms of
materials, there are other materials such as polypropylene and polyvinylidene fluoride (PVDF)(Figure 4.11.41) have been tested;
however, PTFE material (Figure 4.11.42) as membrane has demonstrated better durability and chemical resistance to
electrochemical environment. Therefore, PTFE is shown to be the better candidate for such application, and is usually laminated
onto polypropylene for enhanced mechanical properties. Despite the hydrophobic property of PTFE material, a significant amount
aqueous material penetrates through the membrane due to the large pressure drop. Therefore, the correct sizing of the vacuum
pumps is crucial to maintain the flux of gas to be transported to the mass spectrometry at the desire pressure. More information
regarding the vacuum system will be discussed. In addition, capillary has been used in replacement of the membrane; however, this
method will not be discussed here.

Figure 4.11.41 A illustration of the vacuum membrane distillation process. Adapted from J. S. Aston, Design, Construction and
Research Application of a Differential Electrochemical Mass Spectrometer (DEMS), Springer-Verlag, Berlin Heidelberg (2011).

4.11.24 https://chem.libretexts.org/@go/page/55894
Figure 4.11.42 Chemical structures of polytetrafluoroethylene (PTFE), polypropylene and polyvinylidene fluoride (polyvinylidene
difluoride, PVDF).
Vacuum and QMS

The correctly sized vacuum system can ensure the maximum amount of vapor material to be transported across the membrane.
When the pressure drop is not adequate, part of the vapor material may be remain on the aqueous side as shown Figure 4.11.43.
However, when the pressure drop is too large, too much aqueous electrolyte will be pulled from the liquid-vapor interface,
subsequently increasing load on the vacuum pumps. In the cases of improper sized pumps can reduce pump efficiency and lower
pump life-time if such problem is not corrected immediately. In addition, in order for mass spectrometry operate properly, the gas
flux will need to maintain at a certain flow. Therefore, the vacuum pumps should provide steady flux of gas around 0.09
mbar/s.cm2 consisting mostly with gaseous or volatile species and other species that will be sent to mass spectrometry for
analyzing. In additional, due to the limitation of pump speed of single vacuum pump, vacuum system with two or more pumps will
be needed. For example, if 0.09 mbar/s.cm2 is required and pump speed of 300 s-1 that operates at 10-5 mbar, the acceptable
membrane geometrical area is 0.033 cm-2. In order to increase the membrane area, addition pumps will be required in order to
achieve the same gas flux.
Additional Information
There are several other analytical techniques such as cyclic voltammetry, potential step and galvanic step that can be combined
with DEMS experiment. Cyclic voltammetry can provide both quantitative and qualitative results using the potential dependence.
As a result, both the ion current of interested species and faradaic electrode current (the current generated by the reduction or
oxidation of some chemical substance at an electrode) will be recorded when combining cyclic voltammetry and DEMS.
Applications
The lack of commercialization of this technique has limited it to only academic research. The largest field of application of DEMS
is on electro-catalytic reactions. In addition, it is also used fuel cell research, detoxification reactions, electro-chemical gas sensors
or more fundamental relevant research such as decomposition of ionic liquids etc.
Fuel Cell Differential Electrochemical Mass Spectrometry: Ethanol Electro-oxidation
The ethanol oxidation reaction was studied using alkaline membrane electrode assemblies (MEAs), constructed using nanoparticle
Pt catalyst and alkaline polymeric membrane. DEMS will be use to study the mechanics of the ethanol oxidation reaction on the pt-
based catalysts. The relevant products of the oxidation reaction are carbon dioxide, acetaldehyde and acetic acid. However, both
carbon dioxide and acetaldehyde has the same molecular weight, which 44 g/mole. One approach is to monitor the major fragments
where ionized CO22+ at m/z = 22 and COH+ at m/z = 29 were used. Differential electrochemical mass spectrometry can detect
volatile products of the electrochemical reaction; however, detections can be varied by solubility or boiling point. CO2 is very
volatile, but also soluble in water. If KOH is present, DEMS will not detect any CO2traces. Therefore, all extra alkaline impurities
should be removed before measurements are taken. The electrochemical characteristics can also be measured under various
conditions and examples shown in Figure 4.11.43. In addition, the CCE (CO2 current efficiency) was measured under different
potentials. Using the CCE, the study concluded that the ethanol undergoes more complete oxidation using alkaline MEA than
acidic MEA.

4.11.25 https://chem.libretexts.org/@go/page/55894
Figure 4.11.43 CV for the alkaline MEA are shown with 0.1 M EtOH solution in (a) for only de-ionized water in analyte (a1) both
at 60 °C. Adapted from V. Rao, Hariyanto, C. Cremers, and U. Stimming, Fuel Cells, 2007, 5, 417.
Studies on the Decomposition of Ionic Liquids

Ionic liquids (IL) have several properties such as high ionic conductivity, low vapor pressure, high thermal and electrochemical
stability, which make them great candidate for battery electrolyte. Therefore, it is important to have better understanding of the
stability of the reaction and of the products formed during decomposition behavior. DEMS is a powerful method where it can
provide online detection of the volatile products; however, it runs into problems with high viscosity of ILs and low permeability
due to the size of the molecules. Therefore, researchers modified the traditional setup of DEMS, which the modified method made
use of the low vapor pressure of ILs and have electrochemical cell placed directly into the vacuum system. This experiment shows
that this technique can be designed for very specific application and can be modified easily.
Conclusion
DEMS technique can provide on-line detection of products for electrochemical reactions both analytically and kinetically. In
addition, the results are delivered with high sensitivity where both products and by-products can be detected as long as they are
volatile. It can be easily assembled in the laboratory environment. For the past several decades, this technique has demonstrated
advanced development and has delivered good results for many applications such as fuel cells, gas sensors etc. However, this
technique has its limitation. There are many factors that need to be considered when designing this system such as half-cell
electrochemical reaction, absorption rate and etc. Due to these constraints, the type of membrane should be selected and pump
should be sized accordingly. Therefore, this characterization method is not one size fits all and will need to be modified base on the
experimental parameters. Therefore, next step of development for DEMS is not only to improve its functions, but also to be utilized
beyond the academic laboratory.

4.11: Mass Spectrometry is shared under a CC BY 4.0 license and was authored, remixed, and/or curated by Pavan M. V. Raja & Andrew R.
Barron via source content that was edited to conform to the style and standards of the LibreTexts platform; a detailed edit history is available upon
request.

4.11.26 https://chem.libretexts.org/@go/page/55894
CHAPTER OVERVIEW
5: Reactions Kinetics and Pathways
5.1: Dynamic Headspace Gas Chromatography Analysis
5.2: Gas Chromatography Analysis of the Hydrodechlorination Reaction of Trichloroethene
5.3: Temperature-Programmed Desorption Mass Spectroscopy Applied in Surface Chemistry

5: Reactions Kinetics and Pathways is shared under a CC BY 4.0 license and was authored, remixed, and/or curated by Pavan M. V. Raja &
Andrew R. Barron via source content that was edited to conform to the style and standards of the LibreTexts platform; a detailed edit history is
available upon request.

1
5.1: Dynamic Headspace Gas Chromatography Analysis
Gas chromatography (GC) is a very commonly used chromatography in analytic chemistry for separating and analyzing
compounds that are gaseous or can be vaporized without decomposition. Because of its simplicity, sensitivity, and effectiveness in
separating components of mixtures, gas chromatography is an important tools in chemistry. It is widely used for quantitative and
qualitative analysis of mixtures, for the purification of compounds, and for the determination of such thermochemical constants as
heats of solution and vaporization, vapor pressure, and activity coefficients. Compounds are separated due to differences in their
partitioning coefficient between the stationary phase and the mobile gas phase in the column.

Physical Components of a GC System


A gas chromatograph (Figure 5.1.1 ) consists of a carrier gas system, a sampling system, a separation system, a detection system,
and a data recording system.

Figure 5.1.1 Physical components of a typical GC system. Adapted from http://en.Wikipedia.org/wiki/Gas_chromatography

An ideal separation is judged by resolution, efficiency, and symmetry of the desired peaks, as illustrated by
The carrier gas system consists of carrier gas sources, purification, and gas flow control. The carrier gas must be chemically inert.
Commonly used gases include nitrogen, helium, argon, and carbon dioxide. The choice of carrier gas often depends upon the type
of detector used. A molecular sieve is often contained in the carrier gas system to remove water and other impurities.

Auto Sampling System


An auto sampling system consists of auto sampler, and vaporization chamber. The sample to be analyzed is loaded at the injection
port via a hypodermic syringe and it will be volatilized as the injection port is heated up. Typically samples of one micro liter or
less are injected on the column. These volumes can be further reduced by using what is called a split injection system in which a
controlled fraction of the injected sample is carried away by a gas stream before entering the column.

Separation System
The separation system consists of columns and temperature controlling oven. The column is where the components of the sample
are separated, and is the crucial part of a GC system. The column is essentially a tube that contains different stationary phases have
different partition coefficients with analytes,and determine the quality of separation. There are two general types of column: packed
(Figure 5.1.2) and capillary also known as open tubular (Figure 5.1.3).
Packed columns contain a finely divided, inert, solid support material coated with liquid stationary phase. Most packed columns
are 1.5 – 10 m in length and have an internal diameter of 2 – 4 mm.
Capillary columns have an internal diameter of a few tenths of a millimeter. They can be one of two types; wall-coated open
tubular (WCOT) or support-coated open tubular (SCOT). Wall-coated columns consist of a capillary tube whose walls are
coated with liquid stationary phase. In support-coated columns, the inner wall of the capillary is lined with a thin layer of
support material such as diatomaceous earth, onto which the stationary phase has been adsorbed. SCOT columns are generally
less efficient than WCOT columns. Both types of capillary column are more efficient than packed columns.

5.1.1 https://chem.libretexts.org/@go/page/55896
Figure 5.1.2 An example of a packed GC column.

Figure 5.1.3 An example of a capillary column.

Detectors
The purpose of a detector is to monitor the carrier gas as it emerges from the column and to generate a signal in response to
variation in its composition due to eluted components. As it transmits physical signal into recordable electrical signal, it is another
crucial part of GC. The requirements of a detector for GC are listed below.
Detectors for GC must respond rapidly to minute concentration of solutes as they exit the column, i.e., they are required to have a
fast response and a high sensitivity. Other desirable properties of a detector are: linear response, good stability, ease of operation,
and uniform response to a wide variety of chemical species or, alternatively predictable and selective response to one or more
classes of solutes.

Recording Devices
GC system originally used paper chart readers, but modern system typically uses an online computer, which can track and record
the electrical signals of the separated peaks. The data can be later analyzed by software to provide the information of the gas
mixture.

How Does GC Work?


Separation Terminology
An ideal separation is judged by resolution, efficiency, and symmetry of the desired peaks, as illustrated by Figure 5.1.4.

Figure 5.1.4 Separation terminology. Adapted from www.gchelp.tk

5.1.2 https://chem.libretexts.org/@go/page/55896
Resolution (R)
Resolution can be simply expressed as the distance on the output trace between two peaks. The highest possible resolution is the
goal when developing a separation method. Resolution is defined by the R value, 5.1.1, which can be expressed mathamatically,
5.1.2, where k is capacity, α is selectivity, and N is the number of theoretical plates. An R value of 1.5 is defined as being the

minimum required for baseline separation, i.e., the two adjacent peaks are separated by the baseline. Separation for different R
values is illustrated in Figure 5.1.5.

R  =  capacity × selectivity × ef f iciency (5.1.1)

0.5
R  =  [k/(1 + k)](α −  1/α)(N /4) (5.1.2)

Figure 5.1.5 Different separation resolutions. Adapted from www.gchelp.tk

Capacity (k')
Capacity (k´) is known as the retention factor. It is a measure of retention by the stationary phase. It is calculated from 5.1.3, where
tr = retention time of analyte (substance to be analyzed), and tm = retention time of an unretained compound.

k   =  (tr − tm )/ tm (5.1.3)

Selectivity
Selectivity is related to α, the separation factor (Figure 5.1.6. The value of α should be large enough to give baseline resolution, but
minimized to prevent waste.

Figure 5.1.6 Scheme for the calculation of selectivity. Adapted from www.gchelp.tk

Efficiency
Narrow peaks have high efficiency (Figure 5.1.7), and are desired. Units of efficiency are "theoretical plates" (N) and are often
used to describe column performance. "Plates" is the current common term for N, is defined as a function of the retention time (tr)
and the full peak width at half maximum (Wb1/2), EQ.
2
N   =  5.545(tr / Wb1/2 ) (5.1.4)

Figure 5.1.7 Scheme for calculating efficiency. Adapted from www.gchelp.tk

Peak Symmetry
The symmetry of a peak is judged by the values of two half peak widths, a and b (Figure 5.1.8). When a = b, a peak is called
symmetric, which is desired. Unsymmetrical peaks are often described as "tailing" or "fronting".

5.1.3 https://chem.libretexts.org/@go/page/55896
Figure 5.1.8 Scheme for the symmetry of a peak. Adapted from www.gchelp.tk

An Ideal Separation
The attributions of an ideal separation are as follows:
Should meet baseline resolution of the compounds of interest.
Each desired peak is narrow and symmetrical.
Has no wasted dead time between peaks.
Takes a minimal amount of time to run.
The result is reproducible.
In its simplest form gas chromatography is a process whereby a sample is vaporized and injected onto the chromatographic column,
where it is separated into its many components. The elution is brought about by the flow of carrier gas (Figure 5.1.9).
The carrier gas serves as the mobile phase that elutes the components of a mixture from a column containing an immobilized
stationary phase. In contrast to most other types of chromatography, the mobile phase does not interact with molecules of the
analytes. Carrier gases, the mobile phase of GC, include helium, hydrogen and nitrogen which are chemically inert. The stationary
phase in gas-solid chromatography is a solid that has a large surface area at which adsorption of the analyte species (solutes) take
place. In gas-liquid chromatography, a stationary phase is liquid that is immobilized on the surface of a solid support by adsorption
or by chemical bonding.
Gas chromatographic separation occurs because of differences in the positions of adsorption equilibrium between the gaseous
components of the sample and the stationary phases (Figure 5.1.9). In GC the distribution ratio (ratio of the concentration of
analytes in stationary and mobile phase) is dependent on the component vapor pressure, the thermodynamic properties of the bulk
component band and affinity for the stationary phase. The equilibrium is temperature dependent. Hence the importance of the
selection the stationary phase of the column and column temperature programming in optimizing a separation.

Figure 5.1.9 Scheme for partition in mobile and stationary phases.

Choice of Method
Carrier Gas and Flow Rate
Helium, nitrogen, argon, hydrogen and air are typically used carrier gases. Which one is used is usually determined by the detector
being used, for example, a discharge ionization detection (DID) requires helium as the carrier gas. When analyzing gas samples,
however, the carrier is sometimes selected based on the sample's matrix, for example, when analyzing a mixture in argon, an argon
carrier is preferred, because the argon in the sample does not show up on the chromatogram. Safety and availability are other
factors, for example, hydrogen is flammable, and high-purity helium can be difficult to obtain in some areas of the world.
The carrier gas flow rate affects the analysis in the same way that temperature does. The higher the flow rate the faster the analysis,
but the lower the separation between analytes. Furthermore, the shape of peak will be also effected by the flow rate. The slower the
rate is, the more axial and radical diffusion are, the broader and the more asymmetric the peak is. Selecting the flow rate is
therefore the same compromise between the level of separation and length of analysis as selecting the column temperature.

5.1.4 https://chem.libretexts.org/@go/page/55896
Column Selection
Table 5.1.1 shows commonly used stationary phase in various applications.
Table 5.1.1 Some common stationary phases for gas-liquid chromatography. Adapted from www.cem.msu.edu/~cem333/Week15.pdf
Stationary Phase Common Trade Name Temperature (Celsius) Common Applications

General-purpose nonpolar phase,


Polydimethyl Siloxane OV-1, SE-30 350 hydrocarbons, polynuclear
aromatics, drugs, steroids, PCBs

Poly(phenylmethyl-dimethyl) Fatty acid methyl esters, alkaloids,


OV-3, SE-52 350
siloxane (10% phenyl) drugs, halogenated compounds

Poly(phenylmethyl) siloxane
OV-17 250 Drugs, steroids, pesticides, glycols
(50% phenyl)

Chlorinated aromatics,
Poly(trifluoropropyl-dimethyl)
OV-210 200 nitroaromatics, alkyl-substituted
siloxane
benzenes
Free acids, alcohols, ethers,
Polyethylene glycol Carbowax 20M 250
essential oils, glycols

Poly(dicyanoallyldimethyl) Polyunsaturated fatty acid, rosin


OV-275 240
siloxane acids, free acids, alcohols

Column Temperature and Temperature Program


For precise work, the column temperature must be controlled to within tenths of a degree. The optimum column temperature is
dependent upon the boiling point of the sample. As a rule of thumb, a temperature slightly above the average boiling point of the
sample results in an elution time of 2 - 30 minutes. Minimal temperatures give good resolution, but increase elution times. If a
sample has a wide boiling range, then temperature programming can be useful. The column temperature is increased (either
continuously or in steps) as separation proceeds. Another effect that temperature may have is on the shape of peak as flow rate
does. The higher the temperature is, the more intensive the diffusion is, the worse the shape is. Thus, a compromise has to be made
between goodness of separation and retention time as well as peak shape.

Detector Selection
A number of detectors are used in gas chromatography. The most common are the flame ionization detector (FID) and the thermal
conductivity detector (TCD). Both are sensitive to a wide range of components, and both work over a wide range of concentrations.
While TCDs are essentially universal and can be used to detect any component other than the carrier gas (as long as their thermal
conductivities are different from that of the carrier gas, at detector temperature), FIDs are sensitive primarily to hydrocarbons, and
are more sensitive to them than TCD. However, an FID cannot detect water. Both detectors are also quite robust. Since TCD is
non-destructive, it can be operated in-series before an FID (destructive), thus providing complementary detection of the same
analytes.For halides, nitrates, nitriles, peroxides, anhydrides and organometallics, ECD is a very sensitive detection, which can
detect up to 50 fg of those analytes. Different types of detectors are listed below in Table 5.1.2, along with their properties.
Table 5.1.2 Different types of detectors and their properties. Adapted from teaching.shu.ac.uk/hwb/chemis...m/gaschrm.html
Detector Type Support Gases Selectivity Detectability Dynamic Range

Flame Ionization Most organic


Mass flow Mass flow 100 pg 107
(FID) compounds

Thermal Conductivity
Concentration Reference Universal 1 ng 107
(TCD)

Halides, nitrates,
Electron Capture nitriles, peroxides,
Concentration Make-up 50 fg 105
(FCD) anhydrides,
organometallic

Nitrogen-Phosphorus Mass flow Hydrogen and air Nitrogen, phosphorus 10 pg 106

5.1.5 https://chem.libretexts.org/@go/page/55896
Sulphur, phosphorus,
Flame Photometric Hydrogen and air tin, boron, arsenic,
Mass flow 100 pg 103
(FPD) possibly oxygen germanium, selenium,
chromium

Aliphatics, aromatics,
ketones, esters,
Photo-ionization aldehydes, amines,
Concentration Make-up 2 pg 107
(PID) heterocyclics,
organosulphurs, some
organometallics

Hall electrolytic Halide, nitrogen,


Mass flow Hydrogen, oxygen - -
Conductivity nitrosamine, sulphur

Headspace Analysis Using GC


Most consumer products and biological samples are composed of a wide variety of compounds that differ in molecular weight,
polarity, and volatility. For complex samples like these, headspace sampling is the fastest and cleanest method for analyzing
volatile organic compounds. A headspace sample is normally prepared in a vial containing the sample, the dilution solvent, a
matrix modifier, and the headspace (Figure 5.1.10). Volatile components from complex sample mixtures can be extracted from
non-volatile sample components and isolated in the headspace or vapor portion of a sample vial. An aliquot of the vapor in the
headspace is delivered to a GC system for separation of all of the volatile components.

Figure 5.1.10 Schematic representation of the phases of the headspace in the vial. Adapted from A Technical Guide for Static
Headspace Analysis Using GC, Restek Corp. (2000).
The gas phase (G in Figure 5.1.10) is commonly referred to as the headspace and lies above the condensed sample phase. The
sample phase (S in Figure 5.1.10 contains the compound(s) of interest and is usually in the form of a liquid or solid in combination
with a dilution solvent or a matrix modifier. Once the sample phase is introduced into the vial and the vial is sealed, volatile
components diffuse into the gas phase until the headspace has reached a state of equilibrium as depicted by the arrows. The sample
is then taken from the headspace.

Basic Principles of Headspace Analysis


Partition Coefficient
Samples must be prepared to maximize the concentration of the volatile components in the headspace, and minimize unwanted
contamination from other compounds in the sample matrix. To help determine the concentration of an analyte in the headspace, you
will need to calculate the partition coefficient (K), which is defined by 5.1.5 ,where Cs is the concentration of analyte in sample
phase and Cg is the concentration of analyte in gas phase. Compounds that have low K values will tend to partition more readily
into the gas phase, and have relatively high responses and low limits of detection. K can be lowered by changing the temperature at
which the vial is equilibrated or by changing the composition of the sample matrix.
K  =  Cs / Cg (5.1.5)

5.1.6 https://chem.libretexts.org/@go/page/55896
Phase Ratio
The phase ratio (β) is defined as the relative volume of the headspace compared to volume of the sample in the sample vial, 5.1.6,
where Vs=volume of sample phase and Vg=volume of gas phase. Lower values for β (i.e., larger sample size) will yield higher
responses for volatile compounds. However, decreasing the β value will not always yield the increase in response needed to
improve sensitivity. When β is decreased by increasing the sample size, compounds with high K values partition less into the
headspace compared to compounds with low K values, and yield correspondingly smaller changes in Cg. Samples that contain
compounds with high K values need to be optimized to provide the lowest K value before changes are made in the phase ratio.

β  =  Vg / Vs (5.1.6)

5.1: Dynamic Headspace Gas Chromatography Analysis is shared under a CC BY 4.0 license and was authored, remixed, and/or curated by Pavan
M. V. Raja & Andrew R. Barron via source content that was edited to conform to the style and standards of the LibreTexts platform; a detailed
edit history is available upon request.

5.1.7 https://chem.libretexts.org/@go/page/55896
5.2: Gas Chromatography Analysis of the Hydrodechlorination Reaction of
Trichloroethene
Trichloroethene (TCE) is a widely spread environmental contaminant and a member of the class of compounds known as dense
non-aqueous phase liquids (DNAPLs). Pd/Al2O3 catalyst has shown activity for the hydrodechlorination (HDC) of chlorinated
compounds.
To quantify the reaction rate, a 250 mL screw-cap bottle with 77 mL of headspace gas was used as the batch reactor for the studies.
TCE (3 μL) is added in 173 mL DI water purged with hydrogen gas for 15 mins, together with 0.2 μL pentane as internal standard.
Dynamic headspace analysis using GC has been applied. The experimental condition is concluded in the table below (Table 5.2.1).
Table 5.2.1 The experimental condition in HDC of TCE.
TCE 3 μL

H2 1.5 ppm

Pentane 0.2 μL

DI water 173 mL

1 wt% Pd/Al2O3 50 mg

Temperature 25 °C

Pressure 1 atm

Reaction time 1h

Reaction Kinetics
First order reaction is assumed in the HDC of TCE, 5.2.1, where Kmeans is defined by 5.2.2, and Ccatis equal to the concentration of
Pd metal within the reactor and kcat is the reaction rate with units of L/gPd/min.
−dCT C E /dt  =  kmeas × CT C E (5.2.1)

kmeas   =  kcat × Ccat (5.2.2)

The GC Method
The GC methods used are listed in Table 5.2.3.
Table 5.2.3 GC method for detection of TCE and other related chlorinated compounds.
GC type Agilent 6890N GC

Column Supelco 1-2382 40/60 Carboxen-1000 packed column

Detector FID

Oven temperature 210 °C

Flow rate 35 mL/min

Injection amount 200 μL

Carrier gas Helium

Detect 5 min

Quantitative Method
Since pentane is introduced as the inert internal standard, the relative concentration of TCE in the system can be expressed as the
ratio of area of TCE to pentane in the GC plot, 5.2.3.
CT C E   =  (peak area of  T C E)/(peak area of  pentane) (5.2.3)

5.2.1 https://chem.libretexts.org/@go/page/55897
Results and Analysis
The major analytes (referenced as TCE, pentane, and ethane) are very well separated from each other, allowing for quantitative
analysis. The peak areas of the peaks associated with these compounds are integrated by the computer automatically, and are listed
in (Table 5.2.4) with respect to time.
Table 5.2.2 Peak area of pentane, TCE as a function of reaction time.
Time/min Peak area of pentane Peak area of TCE

0 5992.93 13464

5.92 6118.5 11591

11.25 5941.2 8891

16.92 5873.5 7055.6

24.13 5808.6 5247.4

32.65 5805.3 3726.3

43.65 5949.8 2432.8

53.53 5567.5 1492.3

64.72 5725.6 990.2

77.38 5624.3 550

94.13 5432.5 225.7

105 5274.4 176.8

Normalize TCE concentration with respect to peak area of pentane and then to the initial TCE concentration, and then calculate the
nature logarithm of this normalized concentration, as shown in Table 5.2.3.
Table 5.2.3 Normalized TCE concentration as a function of reaction time
Time (min) TCE/pentane TCE/pentane/TCEinitial In(TCE/Pentane/TCEinitial)

0 2.2466 1.0000 0.0000

5.92 1.8944 0.8432 -0.1705

11.25 1.4965 0.6661 -0.4063

16.92 1.2013 0.5347 -0.6261

24.13 0.9034 0.4021 -0.9110

32.65 0.6419 0.2857 -1.2528

43.65 0.4089 0.1820 -1.7038

53.53 0.2680 0.1193 -2.1261

64.72 0.1729 0.0770 -2.5642

77.38 0.0978 0.0435 -3.1344

94.13 0.0415 0.0185 -3.9904

105 0.0335 0.0149 -4.2050

From a plot normalized TCE concentration against time shows the concentration profile of TCE during reaction (Figure 5.2.1,
while the slope of the logarithmic plot provides the reaction rate constant (5.2.1).

5.2.2 https://chem.libretexts.org/@go/page/55897
Figure 5.2.1 A plot of the normalized concentration profile of TCE.

Figure 5.2.2 A plot of ln(CTCE/C0) versus time.


From Figure 5.2.1, we can see that the linearity, i.e., the goodness of the assumption of first order reaction, is very much satisfied
throughout the reaction. Thus, the reaction kinetic model is validated. Furthermore, the reaction rate constant can be calculated
from the slope of the fitted line, i.e., kmeas = 0.0414 min-1. From this the kcat can be obtained, ??? .

k_{cat}\ =\ k_{meas}/C_{Pd}\ =\ \frac{0.0414min^{-1}{(5 \times 10^{-4}\ g/0.173L)}\ =\ 14.32L/g_{Pd}\ min \label{4}

5.2: Gas Chromatography Analysis of the Hydrodechlorination Reaction of Trichloroethene is shared under a CC BY 4.0 license and was
authored, remixed, and/or curated by Pavan M. V. Raja & Andrew R. Barron via source content that was edited to conform to the style and
standards of the LibreTexts platform; a detailed edit history is available upon request.

5.2.3 https://chem.libretexts.org/@go/page/55897
5.3: Temperature-Programmed Desorption Mass Spectroscopy Applied in Surface
Chemistry
The temperature-programmed desorption (TPD) technique is often used to monitor surface interactions between adsorbed
molecules and substrate surface. Utilizing the dependence on temperature is able to discriminate between processes with different
activation parameters, such as activation energy, rate constant, reaction order and Arrhenius pre-exponential factorIn order to
provide an example of the set-up and results from a TPD experiment we are going to use an ultra-high vacuum (UHV) chamber
equipped with a quadrupole mass spectrometer to exemplify a typical surface gas-solid interaction and estimate several important
kinetic parameters.

Experimental System
Ultra-high Vacuum (UHV) Chamber

When we start to set up an apparatus for a typical surface TPD experiment, we should first think about how we can generate an
extremely clean environment for the solid substrate and gas adsorbents. Ultra-high vacuum (UHV) is the most basic requirement
for surface chemistry experiments. UHV is defined as a vacuum regime lower than 10-9 Torr. At such a low pressure the mean free
path of a gas molecule is approximately 40 Km, which means gas molecules will collide and react with sample substrate in the
UHV chamber many times before colliding with each other, ensuring all interactions take place on the substrate surface.
Most of time UHV chambers require the use of unusual materials in construction and by heating the entire system to ~180 °C for
several hours baking to remove moisture and other trace adsorbed gases around the wall of the chamber in order to reach the ultra-
high vacuum environment. Also, outgas from the substrate surface and other bulk materials should be minimized by careful
selection of materials with low vapor pressures, such as stainless steel, for everything inside the UHV chamber. Thus bulk metal
crystals are chosen as substrates to study interactions between gas adsorbates and crystal surface itself. Figure 5.3.1 shows a
schematic of a TPD system, while Figure 5.3.2 shows a typical TPD instrument equipped with a quadrupole MS spectrometer and
a reflection absorption infrared spectrometer (RAIRS).

Figure 5.3.1 Schematic diagram of a TPD apparatus.

5.3.1 https://chem.libretexts.org/@go/page/55898
Figure 5.3.2 A typical TPD apparatus composed of a UHV chamber equipped with a serious of pumping systems, cooling system,
sample dosing system as well as surface detection instruments including a quadrupole MS Spectrometer and a reflection absorption
infra red spectrometer (RAIRS).

Pumping System
There is no single pump that can operate all the way from atmospheric pressure to UHV. Instead, a series of different pumps are
used, according to the appropriate pressure range for each pump. Pumps are commonly used to achieve UHV include:
Turbomolecular pumps (turbo pumps).
Ionic pumps.
Titanium sublimation pumps.
Non-evaporate mechanical pumps.
UHV pressures are measured with an ion-gauge, either a hot filament or an inverted magnetron type. Finally, special seals and
gaskets must be used between components in a UHV system to prevent even trace leakage. Nearly all such seals are all metal, with
knife edges on both sides cutting into a soft (e.g., copper) gasket. This all-metal seal can maintain system pressures down to ~10-12
Torr.

Manipulator and Bulk Metal Crystal


A UHV manipulator (or sample holder, see Figure 5.3.2) allows an object that is inside a vacuum chamber and under vacuum to be
mechanically positioned. It may provide rotary motion, linear motion, or a combination of both. The manipulator may include
features allowing additional control and testing of a sample, such as the ability to apply heat, cooling, voltage, or a magnetic field.
Sample heating can be accomplished by thermal radiation. A filament is mounted close to the sample and resistively heated to high
temperature. In order to simplify complexity from the interaction between substrate and adsorbates, surface chemistry labs often
carry out TPD experiments by choosing a substrate with single crystal surface instead of polycrystalline or amorphous substrates
(see Figure 5.3.1).

Pretreatment
Before selected gas molecules are dosed to the chamber for adsorption, substrates (metal crystals) need to be cleaned through argon
plasma sputtering, followed by annealing at high temperature for surface reconstruction. After these pretreatments, the system is
again cooled down to very low temperature (liquid N2temp), which facilitating gas molecules adsorbed on the substrate surface.
Adsorption is a process in which a molecule becomes adsorbed onto a surface of another phase. It is distinguished from absorption,
which is used when describing uptake into the bulk of a solid or liquid phase.

Temperature-programmed Desorption Processes


After gas molecules adsorption, now we are going to release theses adsorbates back into gas phase by programmed-heating the
sample holder. A mass spectrometer is set up for collecting these desorbed gas molecules, and then correlation between desorption

5.3.2 https://chem.libretexts.org/@go/page/55898
temperature and fragmentation of desorbed gas molecules will show us certain important information. Figure 5.3.3 shows a typical
TPD experiment carried out by adsorbing CO onto Pd(111) surface, followed by programmed-heating to desorb the CO adsorbates.

Figure 5.3.3 MS spectrum taken from a TPD experiment that CO (m/z = 28) was first adsorbed on Pd(111) surface, followed by
desorbing at a fixed heating rate. The desorption rate which is proportional to the MS signal reaches its maximum around 500 K.

Theory of TPD Experiment


Langmuir Isotherm
The Langmuir isotherm describes the dependence of the surface coverage of an adsorbed gas on the pressure of the gas above the
surface at a fixed temperature. Langmuir isotherm is the simplest assumption, but it provides a useful insight into the pressure
dependence of the extent of surface adsorption. It was Irving Langmuir who first studied the adsorption process quantitatively. In
his proposed model, he supposed that molecules can adsorb only at specific sites on the surface, and that once a site is occupied by
one molecule, it cannot adsorb a second molecule. The adsorption process can be represented as 5.3.1, where A is the adsorbing
molecule, S is the surface site, and A─S stands for an A molecule bound to the surface site.
A  +  S → A  −  S (5.3.1)

In a similar way, it reverse desorption process can be represented as 5.3.2.

A  −  S → A  +  S (5.3.2)

According to the Langmuir model, we know that the adsorption rate should be proportional to ka[A](1-θ), where θ is the fraction of
the surface sites covered by adsorbate A. The desorption rate is then proportional to kdθ. ka and kd are the rate constants for the
adsorption and desorption. At equilibrium, the rates of these two processes are equal, 5.3.3 - 5.3.4.We can replace [A] by P, where
P means the gas partial pressure, 5.3.6.

ka [A](1 − θ)  =  kd θ (5.3.3)

θ ka
  =  [A] (5.3.4)
1  −  θ kd

ka
K  =   (5.3.5)
kd

K[A]
θ  =   (5.3.6)
1 + K[A]

KP
θ  =   (5.3.7)
1 + KP

We can observe the equation above and know that if [A] or P is low enough so that K[A] or KP << 1, then θ ~ K[A] or KP, which
means that the surface coverage should increase linearly with [A] or P. On the contrary, if [A] or P is large enough so that K[A] or
KP >> 1, then θ ~ 1. This behavior is shown in the plot of θ versus [A] or P in Figure 5.3.4.

5.3.3 https://chem.libretexts.org/@go/page/55898
Figure 5.3.4 Simulated Langmuir isotherms. Value of constant K (ka/kd) increases from blue, red, green and brown.

Derivation of Kinetic Parameters Based on TPD Results


Here we are going to show how to use the TPD technique to estimate desorption energy, reaction energy, as well as Arrhenius pre-
exponential factor. Let us assume that molecules are irreversibly adsorbed on the surface at some low temperature T0. The leak
valve is closed, the valve to the pump is opened, and the “density” of product molecules is monitored with a mass spectrometer as
the crystal is heated under programmed temperature 5.3.8, where β is the heating rate (~10 °C/s). We know the desorption rate
depends strongly on temperature, so when the temperature of the crystal reaches a high enough value so that the desorption rate is
appreciable, the mass spectrometer will begin to record a rise in density. At higher temperatures, the surface will finally become
depleted of desorbing molecules; therefore, the mass spectrometer signal will decrease. According to the shape and position of the
peak in the mass signal, we can learn about the activation energy for desorption and the Arrhenius pre-exponential factor.
T   =  T0   +  βT (5.3.8)

First-Order Process
Consider a first-order desorption process 5.3.9, with a rate constant kd, 5.3.10, where A is Arrhenius pre-exponential factor. If θ is
assumed to be the number of surface adsorbates per unit area, the desorption rate will be given by 5.3.11.

A  −  S →  A  +  S (5.3.9)

(−ΔEα
kd   =  Ae RT ) (5.3.10)

−dθ
(−ΔEα
  =  kd θ =  θAe RT ) (5.3.11)
dt

Since we know the relationship between heat rate β and temperature on the crystal surface T, 5.3.12 and 5.3.13.
T   =  T0   +  βt (5.3.12)

1 β
  =  (5.3.13)
dt dT

Multiplying by -dθ gives 5.3.13, since 5.3.14 and 5.3.15. A plot of the form of –dθ/dT versus T is shown in Figure 5.3.5.
−dθ dθ
  =  −β (5.3.14)
dt dT

−Δ E
−dθ (
D

  =  kd   =  θAe RT
) (5.3.15)
dt

−dθ θA (
−Δ Ea
)
  =  e RT
(5.3.16)
dt β

Figure 5.3.5 A simulated TPD experiment: Consider a first order reaction between adsorbates and surface. Values of Tm keep
constant as the initial coverage θ from 1.0 x 1013 to 6.0 x 1013 cm-2. Ea = 30 KJ/mol; β = 1.5 °C/s; A = 1 x 1013.

5.3.4 https://chem.libretexts.org/@go/page/55898
We notice that the Tm (peak maximum) in Figure 5.3.5
keeps constant with increasing θ, which means the value of Tm does not depend on the initial coverage θ in the first-order
desorption. If we want to use different desorption activation energy Ea and see what happens in the corresponding desorption
temperature T. We are able to see the Tm values will increase with increasing Ea.
At the peak of the mass signal, the increase in the desorption rate is matched by the decrease in surface concentration per unit area
so that the change in dθ/dT with T is zero: 5.3.17 - 5.3.18. Since 5.3.19, then 5.3.20 and 5.3.21.
−dθ θA (
−Δ Ea
)
  =  e RT
(5.3.17)
dT β

d (
−Δ Ea
)
[f racθAβ e RT
]  =  0 (5.3.18)
dT

ΔEa 1 dθ
=− ( ) (5.3.19)
2
RT θ dT
M

−dθ θA (−
−Δ Ea
)
−   =  e RT
(5.3.20)
dT β

ΔEa A (−
−Δ Ea
)
  =  e RT
(5.3.21)
2
RT β
M

ΔEa ΔEa
2lnTM   −  lnβ  = + ln (5.3.22)
2
RT RA
M

This tells us if different heating rates β are used and the left-hand side of the above equation is plotted as a function of 1/TM, we
can see that a straight line should be obtained whose slope is ΔEa/R and intercept is ln(ΔEa/RA). So we are able to obtain the
activation energy to desorption ΔEa and Arrhenius pre-exponential factor A.
Second-Order Process
Now let consider a second-order desorption process 5.3.23, with a rate constant kd. We can deduce the desorption kinetics as
5.3.24. The result is different from the first-order reaction whose Tm value does not depend upon the initial coverage, the

temperature of the peak Tm will decrease with increasing initial surface coverage.
2A  − S → A2   +  2S (5.3.23)

dθ 2
Δ Ea

− =  Aθ e RT
(5.3.24)
dT

Figure 5.3.6 A simulated second-order TPD experiment: A second-order reaction between adsorbates and surface. Values of Tm
decrease as the initial coverage θ increases from 1.0 x 1013 to 6.0 x 1013 cm-2; Ea = 30 KJ/mol; β = 1.5 °C/s; A = 1 x 10-1.
Zero-Order Process
The zero-order desorption kinetics relationship as 5.3.25. Looking at desorption rate for the zero-order reaction (Figure 5.3.7), we
can observe that the desorption rate does not depend on coverage and also implies that desorption rate increases exponentially with
T. Also according to the plot of desorption rate versus T, we figure out the desorption rate rapid drop when all molecules have
desorbed. Plus temperature of peak, Tm, moves to higher T with increasing coverage θ.
dθ (−
Δ Ea
)
−   =  Ae RT
(5.3.25)
dT

5.3.5 https://chem.libretexts.org/@go/page/55898
Figure 5.3.7 A simulated zero-order TPD experiment: A zero-order reaction between adsorbates and surface. Values of Tm increase
apparently as the initial coverage θ increases from 1.0 x 1013 to 6.0 x 1013 cm-2; Ea = 30 KJ/mol; β = 1.5 °C/s; A = 1 x 1028.

A Typical Example
A typical TPD spectra of D2 from Rh(100) for different exposures in Langmuirs (L = 10-6 Torr-sec) shows in Figure 5.3.8. First we
figure out the desorption peaks from g to n show two different desorbing regions. The higher one can undoubtedly be ascribed to
chemisorbed D2 on Rh(100) surface, which means chemisorbed molecules need higher energy used to overcome their activation
energy for desorption. The lower desorption region is then due to physisorbed D2 with much lower desorption activation energy
than chemisorbed D2. According to the TPD theory we learnt, we notice that the peak maximum shifts to lower temperature with
increasing initial coverage, which means it should belong to a second-order reaction. If we have other information about heating
rate β and each Tm under corresponding initial surface coverage θ then we are able to calculate the desorption activation energy Ea
and Arrhenius pre-exponential factor A.

Figure 5.3.8 TPD spectra of D2 from Rh(100) for different exposures in L (1 Langmuir = 10-6 Torr-s)6.

Conclusion
Temperature-programmed desorption is an easy and straightforward technique especially useful to investigate gas-solid interaction.
By changing one of parameters, such as coverage or heating rate, followed by running a serious of typical TPD experiments, it is
possible to to obtain several important kinetic parameters (activation energy to desorption, reaction order, pre-exponential factor,
etc). Based on the information, further mechanism of gas-solid interaction can be deduced.

5.3: Temperature-Programmed Desorption Mass Spectroscopy Applied in Surface Chemistry is shared under a CC BY 4.0 license and was
authored, remixed, and/or curated by Pavan M. V. Raja & Andrew R. Barron via source content that was edited to conform to the style and
standards of the LibreTexts platform; a detailed edit history is available upon request.

5.3.6 https://chem.libretexts.org/@go/page/55898
CHAPTER OVERVIEW
6: Dynamic Processes
6.1: NMR of Dynamic Systems- An Overview
6.2: Determination of Energetics of Fluxional Molecules by NMR
6.3: Rolling Molecules on Surfaces Under STM Imaging

6: Dynamic Processes is shared under a CC BY 4.0 license and was authored, remixed, and/or curated by Pavan M. V. Raja & Andrew R. Barron
via source content that was edited to conform to the style and standards of the LibreTexts platform; a detailed edit history is available upon
request.

1
6.1: NMR of Dynamic Systems- An Overview
The study of conformational and chemical equilibrium is an important part of understanding chemical species in solution. NMR is
one of the most useful and easiest to use tools for such kinds of work. Figure 6.1.1 The study of conformational and chemical
equilibrium is an important part of understanding chemical species in solution. NMR is one of the most useful and easiest to use
tools for such kinds of work.
Chemical equilibrium is defined as the state in which both reactants and products (of a chemical reaction) are present at
concentrations which have no further tendency to change with time. Such a state results when the forward reaction proceeds at the
same rate (i.e., Ka in Figure 6.1.1 b) as the reverse reaction (i.e., Kd in Figure 6.1.1 b). The reaction rates of the forward and
reverse reactions are generally not zero but, being equal, there are no net changes in the concentrations of the reactant and product.
This process is called dynamic equilibrium.
Conformational isomerism is a form of stereoisomerism in which the isomers can be interconverted exclusively by rotations about
formally single bonds. Conformational isomers are distinct from the other classes of stereoisomers for which interconversion
necessarily involves breaking and reforming of chemical bonds. The rotational barrier, or barrier to rotation, is the activation
energy required to interconvert rotamers. The equilibrium population of different conformers follows a Boltzmann distribution.

Figure 6.1.1 The process of (a) conformational equilibrium and (b) chemical equilibrium. Adapted from J. Saad, Dynamic NMR
and Application (2008), www.microbio.uab.edu/mic774/lectures/Saad-lecture8.pdf.
If we consider a simple system (Figure 6.1.2)as an example of how to study conformational equilibrium. In this system, the two
methyl groups (one is in red, the other blue) will exchange with each other through the rotation of the C-N bond. When the speed of
the rotation is fast (faster than the NMR timescale of about 10-5s), NMR can no longer recognize the difference of the two methyl
groups, which results in an average peak in the NMR spectrum (as is shown in the red spectrum in Figure 6.1.3).Conversely, when
the speed of the rotation is slowed by cooling (to -50 °C) the two conformations have lifetimes significantly longer that they are
observable in the NMR spectrum (as is shown by the dark blue spectrum in Figure 6.1.3). The changes that occur to this spectrum
with varying temperature is shown in Figure 6.1.3, where it is clearly seen the change of the NMR spectrum with the decreasing of
temperature.

Figure 6.1.2 An example of a process of a conformational equilibrium.

Figure 6.1.2 as a function of temperature. Adapted from J. Saad, Dynamic NMR and Application (2008),
www.microbio.uab.edu/mic774/lectures/Saad-lecture8.pdf.

6.1.1 https://chem.libretexts.org/@go/page/55900
Based upon the above, it should be clear that the presence of an average or separate peaks can be used as an indicator of the speed
of the rotation. As such this technique is useful in probing systems such as molecular motors. One of the most fundamental
problems is to confirm that the motor is really rotating, while the other is to determine the rotation speed of the motors. In this area,
the dynamic NMR measurements is an ideal technique. For example, we can take a look at the molecular motor shown in Figure
6.1.4. This molecular motor is composed of two rigid conjugated parts, which are not in the same plane. The rotation of the C-N

bond will change the conformation of the molecule, which can be shown by the variation of the peaks of the two methyl groups in
NMR spectrum. For the control of the rotation speed of this particular molecule motor, the researchers added additional
functionality. When the nitrogen in the aromatic ring is not protonated the repulsion between the nitrogen and the oxygen atoms is
larger which prohibits the rotation of the five member ring, which separates the peaks of the two methyl groups from each other.
However, when the nitrogen is protonated, the rotation barrier greatly decreases because of the formation of a more stable coplanar
transition state during the rotation process. Therefore, the speed of the rotation of the rotor dramatically increases to make the two
methyl groups unrecognizable by NMR spectrometry to get an average peak. The result of the NMR spectrum versus the addition
of the acid is shown in Figure 6.1.5, which can visually tell that the rotation speed is changing.

Figure 6.1.4 The design of molecule rotor. Reprinted with permission from B. E. Dial, P. J. Pellechia, M. D. Smith, and K. D.
Shimizu, J. Am. Chem. Soc., 2012, 134, 3675. Copyright (2012) American Chemical Society.

Figure 6.1.5 NMR spectra of the diastereotopic methyl groups of the molecular rotor with the addition of 0.0, 0.5, 2.0, and 3.5
equiv of methanesulfonic acid. Reprinted with permission from B. E. Dial, P. J. Pellechia, M. D. Smith, and K. D. Shimizu, J. Am.
Chem. Soc., 2012, 134, 3675. Copyright (2012) American Chemical Society.

6.1: NMR of Dynamic Systems- An Overview is shared under a CC BY 4.0 license and was authored, remixed, and/or curated by Pavan M. V.
Raja & Andrew R. Barron via source content that was edited to conform to the style and standards of the LibreTexts platform; a detailed edit
history is available upon request.

6.1.2 https://chem.libretexts.org/@go/page/55900
6.2: Determination of Energetics of Fluxional Molecules by NMR
Introduction to Fluxionality
It does not take an extensive knowledge of chemistry to understand that as-drawn chemical structures do not give an entirely
correct picture of molecules. Unlike drawings, molecules are not stationary objects in solution, the gas phase, or even in the solid
state. Bonds can rotate, bend, and stretch, and the molecule can even undergo conformational changes. Rotation, bending, and
stretching do not typically interfere with characterization techniques, but conformational changes occasionally complicate analyses,
especially nuclear magnetic resonance (NMR).
For the present discussion, a fluxional molecule can be defined as one that undergoes an intramolecular reversible interchange
between two or more conformations. Fluxionality is specified as intramolecular to differentiate from ligand exchange and
complexation mechanisms, intermolecular processes. An irreversible interchange is more of a chemical reaction than a form of
fluxionality. Most of the following examples alternate between two conformations, but more complex fluxionality is possible.
Additionally, this module will focus on inorganic compounds. In this module, examples of fluxional molecules, NMR procedures,
calculations of energetics of fluxional molecules, and the limitations of the approach will be covered.

Examples of Fluxionality
Bailar Twist
Octahedral trischelate complexes are susceptible to Bailar twists, in which the complex distorts into a trigonal prismatic
intermediate before reverting to its original octahedral geometry. If the chelates are not symmetric, a Δ enantiomer will be inverted
to a Λ enantiomer. For example not how in Figure 6.2.1 with the GaL3 complex of 2,3-dihydroxy-N,N‘-diisopropylterephthalamide
(Figure 6.2.2 he end product has the chelate ligands spiraling the opposite direction around the metal center.

Figure 6.2.1 Bailar twist of a gallium catchetol tris-chelate complex. Adapted from B. Kersting, J. R. Telford, M. Meyer, and K. N.
Raymond, J. Am. Chem. Soc., 1996, 118, 5712.

Figure 6.2.2 Substituted catchetol ligand 2,3-dihydroxy-N,N‘-diisopropylterephthalamide. Adapted from Kersting, B., Telford,
J.R., Meyer, M., Raymond, K.N.; J. Am. Chem. Soc., 1996, 118, 5712.

Berry Psuedorotation
D3h compounds can also experience fluxionality in the form of a Berry pseudorotation (depicted in Figure 6.2.3), in which the
complex distorts into a C4v intermediate and returns to trigonal bipyrimidal geometry, exchanging two equatorial and axial groups .
Phosphorous pentafluoride is one of the simplest examples of this effect. In its 19FNMR, only one peak representing five fluorines
is present at 266 ppm, even at low temperatures. This is due to interconversion faster than the NMR timescale.

Figure 6.2.3 Berry pseudorotation of phosphorus pentafluoride.

Sandwhich and Half-sandwhich Complexes


Perhaps one of the best examples of fluxional metal complexes is (π5-C5H5)Fe(CO)2(π1-C5H5) (Figure 6.2.4. Not only does it have
a rotating η5 cyclopentadienyl ring, it also has an alternating η1 cyclopentadienyl ring (Cp). This can be seen in its NMR spectra in

6.2.1 https://chem.libretexts.org/@go/page/55901
Figure 6.2.5. The signal for five protons corresponds to the metallocene Cp ring (5.6 ppm). Notice how the peak remains a sharp
singlet despite the large temperature sampling range of the spectra. Another noteworthy aspect is how the multiplets corresponding
to the other Cp ring broaden and eventually condense into one sharp singlet.

Figure 6.2.4 Structure of (π5-C5H5)Fe(CO)2(π1-C5H5). Reprinted with permission from M. J. Bennett Jr., F. A. Cotton, A. Davison,
J. W. Faller, S. J. Lippard, and S. M. Morehouse, J. Am. Chem. Soc., 1966, 88, 4371. Copyright: American Chemical Society
(1966).

Figure 6.2.5 Variable temperature NMR spectra of (π5-C5H5)Fe(CO)2(π1-C5H5). Reprinted with permission from M. J. Bennett Jr.,
F. A. Cotton, A. Davison, J. W. Faller, S. J. Lippard, and S. M. Morehouse, J. Am. Chem. Soc., 1966, 88, 4371. Copyright:
American Chemical Society (1966).

An Example Procedure
ample preparation is essentially the same for routine NMR. The compound of interest will need to be dissolved in an NMR
compatible solvent (CDCl3 is a common example) and transferred into an NMR tube. Approximately 600 μL of solution is needed
with only micrograms of compound. Compounds should be at least 99 % pure in order to ease peak assignments and analysis.
Because each spectrometer has its own protocol for shimming and optimization, having the supervision of a trained specialist is
strongly advised. Additionally, using an NMR with temperature control is essential. The basic goal of this experiment is to find
three temperatures: slow interchange, fast interchange, and coalescence. Thus many spectra will be needed to be obtained at
different temperatures in order to determine the energetics of the fluctuation.
The process will be much swifter if the lower temperature range (in which the fluctuation is much slower than the spectrometer
timescale) is known. A spectra should be taken in this range. Spectra at higher temperatures should be taken, preferably in regular
increments (for instance, 10 K), until the peaks of interest condense into a sharp single at higher temperature. A spectrum at the
coalescence temperature should also be taken in case of publishing a manuscript. This procedure should then be repeated in
reverse; that is, spectra should be taken from high temperature to low temperature. This ensures that no thermal reaction has taken
place and that no hysteresis is observed. With the data (spectra) in hand, the energetics can now be determined.

Calculation of Energetics
For intramolecular processes that exchange two chemically equivalent nuclei, the function of the difference in their resonance
frequencies (Δv) and rate of exchange (k) is the NMR spectrum. Slow interchange occurs when Δv >> k, and two separate peaks
are observed. When Δv << k, fast interchange is said to occur, and one sharp peak is observed. At intermediate temperatures, the
peaks are broadened and overlap one another. When they completely merge into one peak, the coalescence temperature, Tc is said
to be reached. In the case of coalescence of an equal doublet (for instance, one proton exchanging with one proton), coalescences
occurs when Δv0t = 1.4142/(2π), where Δv0 is the difference in chemical shift at low interchange and where t is defined by 6.2.1,
where ta and tb are the respective lifetimes of species a and b. This condition only occurs when ta = tb, and as a result, k = ½ t.

6.2.2 https://chem.libretexts.org/@go/page/55901
1 1 1
  =   +  (6.2.1)
t ta tb

For reference, the exact lineshape function (assuming two equivalent groups being exchanged) is given by the Bloch Equation,
6.2.2, where g is the intensity at frequency v,and where K is a normalization constant.

2
Kt(va + vb )
g(v) = (6.2.2)
2 2 2 2 2
[0.5(va + vb ) − u ] + 4 π t (va − v) (vb − v)

Low Temperatures to Coalescence Temperature


At low temperature (slow exchange), the spectrum has two peaks and Δv >> t. As a result, 6.2.3 reduces to 6.2.4, where T2a is the
spin-spin relaxation time. The linewidth of the peak for species a is defined by 6.2.5.
KT2a
g(v)a = g(v)b = (6.2.3)
2 2
1 +T (va − v)
2a

1 1 1
(Δva )1/2 = ( + ) (6.2.4)
π T2a ta

Because the spin-spin relaxation time is difficult to determine, especially in inhomogeneous environments, rate constants at higher
temperatures but before coalescence are preferable and more reliable.
The rate constant k can then be determined by comparing the linewidth of a peak with no exchange (low temp) with the linewidth
of the peak with little exchange using, 6.2.5, where subscript e refers to the peak in the slightly higher temperature spectrum and
subscript 0 refers to the peak in the no exchange spectrum.
π
k = [(Δve )1/2 − (Δv0 )1/2 ] (6.2.5)

√2

Additionally, k can be determined from the difference in frequency (chemical shift) using 6.2.6, where Δv0is the chemical shift
difference in Hz at the no exchange temperature and Δve is the chemical shift difference at the exchange temperature.
π
2 2
k = (Δv − Δve ) (6.2.6)
– 0
√2

The intensity ratio method, 6.2.7, can be used to determine the rate constant for spectra whose peaks have begun to merge, where r
is the ratio between the maximum intensity and the minimum intensity, of the merging peaks, Imax/Imin.
π 2 1/2 −1/2
k = – (r + (r − r) ) (6.2.7)
√2

Additionally, k can be determined from the difference in frequency (chemical shift) using 6.2.8, where Δv0is the chemical shift
difference in Hz at the no exchange temperature and Δve is the chemical shift difference at the exchange temperature.
π
2 2
k  = – (Δv0 − Δve ) (6.2.8)
√2

The intensity ratio method, 6.2.9 can be used to determine the rate constant for spectra whose peaks have begun to merge, where r
is the ratio between the maximum intensity and the minimum intensity, of the merging peaks, Imax/Imin
π 2 1/2 −1/2
k  = – (r + (r − r) ) (6.2.9)
√2

As mentioned earlier, the coalescence temperature, Tc is the temperature at which the two peaks corresponding to the interchanging
groups merge into one broad peak and 6.2.10 may be used to calculate the rate at coalescence.
πΔv0
k  = (6.2.10)

√2

Higher Temperatures
Beyond the coalescence temperature, interchange is so rapid (k >> t) that the spectrometer registers the two groups as equivalent
and as one peak. At temperatures greater than that of coalescence, the lineshape equation reduces to 6.2.11.

6.2.3 https://chem.libretexts.org/@go/page/55901
KT2
g(v)  = (6.2.11)
2
[1  +  π T2 (va   +  vb   +  2v) ]

As mentioned earlier, determination of T2 is very time consuming and often unreliable due to inhomogeneity of the sample and of
the magnetic field. The following approximation (6.2.12) applies to spectra whose signal has not completely fallen (in their
coalescence).
2
0.5πΔv
k  = (6.2.12)
(Δve )1/2 − (Δv0 )1/2

Now that the rate constants have been extracted from the spectra, energetic parameters may now be calculated. For a rough
measure of the activation parameters, only the spectra at no exchange and coalescence are needed. The coalescence temperature is
determined from the NMR experiment, and the rate of exchange at coalescence is given by 6.2.10. The activation parameters can
then be determined from the Eyring equation (6.2.13 ), where kB is the Boltzmann constant, and where ΔH‡ - TΔS‡ = ΔG‡.
‡ ‡
k ΔH ΔS kB
ln( ) = − + ln( ) (6.2.13)
t RT R h

For more accurate calculations of the energetics, the rates at different temperatures need to be obtained. A plot of ln(k/T) versus 1/T
(where T is the temperature at which the spectrum was taken) will yield ΔH‡, ΔS‡, and ΔG ‡ . For a pictorial representation of these
concepts, see Figure 6.2.6.

Figure 6.2.6 Simulated NMR temperature domains of fluxional molecules. Reprinted with permission from F. P. Gasparro and N.
H. Kolodny, J. Chem. Ed., 1977, 4, 258. Copyright: American Chemical Society (1977).

Diverse Populations
For unequal doublets (for instance, two protons exchanging with one proton), a different treatment is needed. The difference in
population can be defined through 6.2.14, where Pi is the concentration (integration) of species i and X = 2πΔvt (counts per
second). Values for Δvt are given in Figure 6.2.7.
2
X −2 3/2
1
ΔP = Pa − Pb = [ ] ( ) (6.2.14)
3 X

6.2.4 https://chem.libretexts.org/@go/page/55901
Figure 6.2.7 Plot of Δvt versus ΔP. Reprinted with permission from H. Shanan-Atidi and K. H. Bar-Eli, J. Phys. Chem., 1970, 74,
961. Copyright: American Chemical Society (1970).
The rates of conversion for the two species, ka and kb, follow kaPa = kbPb (equilibrium), and because ka = 1/taand kb = 1/tb, the rate
constant follows 6.2.15.
1
ki = (1 − ΔP ) (6.2.15)
2t

From Erying's expressions, the Gibbs free activation energy for each species can be obtained through 6.2.16 and 6.2.17.

‡ kTc X
ΔGa =  RTc  ln( × ) (6.2.16)
hπΔv0 1 − ΔPa

‡ kTc X
ΔG =  RTc  ln( × ) (6.2.17)
b
hπΔv0 1 − ΔPb

Taking the difference of 6.2.16 and 6.2.17 gives the difference in energy between species a and b (6.2.18).


Pa 1 +P
ΔG = RTc ln( = RTc ln( ) (6.2.18)
Pb 1 −P

Converting constants will yield the following activation energies in calories per mole (6.2.19 and 6.2.20).


X
ΔGa = 4.57 Tc [10.62  +  log( ) +  log(Tc /Δv)] (6.2.19)
2p(1 − ΔP )


X
ΔG = 4.57 Tc [10.62  +  log( ) +  log(Tc /Δv)] (6.2.20)
b
2p(1 − ΔP )

To obtain the free energys of activation, values of log (X/(2π(1 + ΔP))) need to be plotted against ΔP (values Tc and Δv0 are
predetermined).
This unequal doublet energetics approximation only gives ΔG ‡ at one temperature, and a more rigorous theoretical treatment is
needed to give information about ΔS‡ and ΔH‡.

Example of Determination of Energetic Parameters


Normally ligands such as dipyrido(2,3-a;3′,2′-j)phenazine (dpop’) are tridentate when complexed to transition metal centers.
However, dpop’ binds to rhenium in a bidentate manner, with the outer nitrogens alternating in being coordinated and
uncoordinated. See Figure 6.2.8for the structure of Re(CO)3(dpop')Cl. This fluxionality results in the exchange of the aromatic
protons on the dpop’ ligand, which can be observed via 1HNMR. Because of the complex nature of the coalescence of doublets, the
rate constants at different temperatures were determined via computer simulation (DNMR3, a plugin of Topspin). These spectra are
shown in Figure 6.2.8.

6.2.5 https://chem.libretexts.org/@go/page/55901
Figure 6.2.8 The structure of Re(CO)3(dpop’)Cl. Reprinted with permission from K. D. Zimmera, R. Shoemakerb, and R. R.
Ruminski, Inorg. Chim. Acta., 2006, 5, 1478. Copyright: Elsevier (2006).

Figure 6.2.9 experimental and simulated 1HNMR spectra for Re(CO)3(dpop’)Cl. Reprinted with permission from K. D. Zimmera,
R. Shoemakerb, and R. R. Ruminski, Inorg. Chim. Acta., 2006, 5, 1478. Copyright: Elsevier (2006).
The activation parameters can then be obtained by plotting ln(k/T) versus 1/T (see Figure 6.2.9 for the Eyring plot). ΔS ‡ can be
extracted from the y-intercept, and ΔH ‡ can be obtained through the slope of the plot. For this example, ΔH ‡ , ΔS ‡ and ΔG ‡ . were
determined to be 64.9 kJ/mol, 7.88 J/mol, and 62.4 kJ/mol.

Figure 6.2.10 Eyring plot of ln(k/T) versus 1/T for Re(CO)3(dpop’)Cl. Adapted from K. D. Zimmera, R. Shoemakerb, and R. R.
Ruminski, Inorg. Chim. Acta, 2006, 5, 1478. Copyright: Elsevier (2006).

Limitations to the Approach


Though NMR is a powerful technique for determining the energetics of fluxional molecules, it does have one major limitation. If
the fluctuation is too rapid for the NMR timescale (< 1 ms) or if the conformational change is too slow meaning the coalescence
temperature is not observed, the energetics cannot be calculated. In other words, spectra at coalescence and at no exchange need to
be observable. One is also limited by the capabilities of the available spectrometer. The energetics of very fast fluxionality
(metallocenes, PF5, etc) and very slow fluxionality may not be determinable. Also note that this method does not prove any
fluxionality or any mechanism thereof; it only gives a value for the activation energy of the process. As a side note, sometimes the
coalescence of NMR peaks is not due to fluxionality, but rather temperature-dependent chemical shifts.

6.2: Determination of Energetics of Fluxional Molecules by NMR is shared under a CC BY 4.0 license and was authored, remixed, and/or curated
by Pavan M. V. Raja & Andrew R. Barron via source content that was edited to conform to the style and standards of the LibreTexts platform; a
detailed edit history is available upon request.

6.2.6 https://chem.libretexts.org/@go/page/55901
6.3: Rolling Molecules on Surfaces Under STM Imaging
Introduction to Surface Motions at the Molecular Level
As single molecule imaging methods such as scanning tunneling microscope (STM), atomic force microscope (AFM), and
transmission electron microscope (TEM) developed in the past decades, scientists have gained powerful tools to explore molecular
structures and behaviors in previously unknown areas. Among these imaging methods, STM is probably the most suitable one to
observe detail at molecular level. STM can operate in a wide range of conditions, provides very high resolution, and able to
manipulate molecular motions with the tip. An interesting early example came from IBM in 1990, in which the STM was used to
position individual atoms for the first time, spelling out "I-B-M" in Xenon atoms. This work revealed that observation and control
of single atoms and molecular motions on surfaces were possible.
The IBM work, and subsequent experiments, relied on the fact that STM tip always exerts a finite force toward an adsorbate atom
that contains both van der Waals and electrostatic forces was utilized for manipulation purpose. By adjusting the position and the
voltage of the tip, the interactions between the tip and the target molecule were changed. Therefore, applying/releasing force to a
single atom and make it move was possible Figure 6.3.1.

Figure 6.3.1 Manipulation of STM tip toward a xenon atom. a) STM tip move onto a target atom then change the voltage and
current of the tip to apply a stronger interaction. b) Move the atom to a desire position. c) After reaching the desire position, the tip
released by switching back to the scanning voltage and current.
The actual positioning experiment was carried out in the following process. The nickel metal substrate was prepared by cycles of
argon-ion sputtering, followed by annealing in a partial pressure of oxygen to remove surface carbon and other impurities. After the
cleaning process, the sample was cooled to 4 K, and imaged with the STM to ensure the quality of surface. The nickel sample was
then doped with xenon. An image of the doped sample was taken at constant-current scanning conditions. Each xenon atom
appears as a located randomly 1.6 Å high bump on the surface (Figure 6.3.2 a). Under the imaging conditions (tip bias = 0.010 V
with tunneling current 10-9 A) the interaction of the xenon with the tip is too weak to cause the position of the xenon atom to be
perturbed. To move an atom, the STM tip was placed on top of the atom performing the procedure depicted in Figure 6.3.1 to move
to its target. Repeating this process again and again led the researcher to build of the structure they desired Figure 6.3.2 b and c.

Figure 6.3.2 Manipulation of STM tip starting with a) randomly dosed xenon sample, b) under construction - move xenon atom to
desire position, and c) accomplishment of the manipulation. Adapted from D. M. Eigler and E. K. Schweizer, Nature, 1990, 344,
524.
All motions on surfaces at the single molecule level can be described as by the following (or combination of the following) modes:

6.3.1 https://chem.libretexts.org/@go/page/55902
Sliding
Hopping
Rolling
Pivoting
Although the power of STM imaging has been demonstrated, imaging of molecules themselves is still often a difficult task. The
successful imaging of the IBM work was attributed to selection of a heavy atom. Other synthetic organic molecules without heavy
atoms are much more difficult to be imaged under STM. Determinations of the mechanism of molecular motion is another. Besides
imaging methods themselves, other auxiliary methods such as DFT calculations and imaging of properly designed molecules are
required to determine the mechanism by which a particular molecule moves across a surface.
Herein, we are particularly interested in surface-rolling molecules, i.e., those that are designed to roll on a surface. It is
straightforward to imagine that if we want to construct (and image) surface-rolling molecules, we must think of making highly
symmetrical structures. In addition, the magnitudes of interactions between the molecules and the surfaces have to be adequate;
otherwise the molecules will be more susceptible to slide/hop or stick on the surfaces, instead of rolling. As a result, only very few
molecules are known can roll and be detected on surfaces.

Surface Rolling of Molecules under the Manipulation of STM Tips


As described above, rolling motions are most likely to be observed on molecules having high degree of symmetry and suitable
interactions between themselves and the surface. C60 is not only a highly symmetrical molecule but also readily imageable under
STM due to its size. These properties together make C60 and its derivatives highly suitable to study with regards to surface-rolling
motion.
The STM imaging of C60 was first carried out at At King College, London. Similar to the atom positioning experiment by IBM,
STM tip manipulation was also utilized to achieve C60 displacement. The tip trajectory suggested that a rolling motion took into
account the displacement on the surface of C60. In order to confirm the hypothesis, the researchers also employed ab initio density
function (DFT) calculations with rolling model boundary condition (Figure 6.3.3). The calculation result has supported their
experimental result.

Figure 6.3.3 Proposed mechanism of C60 translation showing the alteration of C60...surface interactions during rolling. a) 2-point
interaction. The left point interaction was dissociated during the interaction. b) 1-point interaction. C60can pivot on surface. c) 2-
point interaction. A new interaction formed to complete part of the rolling motion. a) - c) The black spot on the C60 is moved
during the manipulation. The light blue Si balls represent the first layer of molecules the silicon surface, and the yellow balls are
the second layer.
The results provided insights into the dynamical response of covalently bound molecules to manipulation. The sequential breaking
and reforming of highly directional covalent bonds resulted in a dynamical molecular response in which bond breaking, rotation,
and translation are intimately coupled in a rolling motion, but not performing sliding or hopping motion.
A triptycene wheeled dimeric molecule Figure 6.3.4 was also synthesized for studying rolling motion under STM. This "tripod-
like" triptycene wheel ulike a ball like C60 molecule also demonstrated a rolling motion on the surface. The two triptycene units
were connected via a dialkynyl axle, for both desired molecule orientation sitting on surface and directional preference of the
rolling motion. STM controlling and imaging was demonstrated, including the mechanism Figure 6.3.4.

6.3.2 https://chem.libretexts.org/@go/page/55902
Figure 6.3.4 Scheme of the rolling mechanism (left to right). Step 1 is the tip approach towards the molecule, step 2 is a 120 degree
rotation of a wheel around its molecular axle and in step 3 the tip reaches the other side of the molecule. It shows that, in principle,
only one rotation of a wheel can be induced (the direction of movement is marked by arrows).

Single Molecule Nanocar Under STM Imaging


Another use of STM imaging at single molecule imaging is the single molecule nanocar by the Tour group at Rice University. The
concept of a nanocar initially employed the free rotation of a C-C single bond between a spherical C60 molecule and an alkyne,
Figure 6.3.5. Based on this concept, an “axle” can be designed into which are mounted C60 “wheels” connected with a “chassis” to
construct the “nanocar”. Nanocars with this design are expected to have a directional movement perpendicular to the axle.
Unfortunately, the first generation nanocar (named “nanotruck” Figure 6.3.6) encountered some difficulties in STM imaging due to
its chemical instability and insolubility. Therefore, a new of design of nanocar based on OPE has been synthesized Figure 6.3.7.

Figure 6.3.5 Structure of C60 wheels connecting to an alkyne. The only possible rolling direction is perpendicular to the C-C single
bond between C60 and the alkyne. The arrow indicates the rotational motion of C60.

Figure 6.3.6 Structure of the nanotruck. No rolling motion was observed under STM imaging due to its instability, insolubility and
inseparable unreacted C60.The double head arrow indicates the expected direction of nanocar movement. Y. Shirai, A. J. Osgood, Y.
Zhao, Y. Yao, L. Saudan, H. Yang, Y.-H. Chiu, L. B. Alemany, T. Sasaki, J.-F. Morin, J. M. Guerrero, K. F. Kelly, and J. M. Tour, J.
Am. Chem. Soc., 2006, 128, 4854. Copyright American Chemical Society (2006).

6.3.3 https://chem.libretexts.org/@go/page/55902
Figure 6.3.7 Nanocar based on OPE structure. The size of the nanocar is 3.3 nm X 2.1 nm (W x L). Alkoxy chains were attached to
improve solubility and stability. OPE moiety is also separable from C60. The bold double head arrow indicates the expected
direction of nanocar movement. The dimension of nanocar was 3.3 nm X 2.1 nm which enable direct observation of the orientation
under STM imaging. Y. Shirai, A. J. Osgood, Y. Zhao, K. F. Kelly, and J. M. Tour, Nano Lett., 2005, 5, 2330. Copyright American
Chemical Society (2005).
The newly designed nanocar was studied with STM. When the nanocar was heated to ~200 °C, noticeable displacements of the
nanocar were observed under selected images from a 10 min STM experiment Figure 6.3.8. The phenomenon that the nanocar
moved only at high temperature was attributed their stability to a relatively strong adhesion force between the fullerene wheels and
the underlying gold. The series of images showed both pivotal and translational motions on the surfaces.

Figure 6.3.8 Pivotal and translational movement of OPE based nanocar. Acquisition time of one image is approximately 1 min with
(a – e) images were selected from a series spanning 10 min. The configuration of the nanocar on surface can be determined by the
distances of four wheels. a) – b) indicated the nanocar had made a 80° pivotal motion. b) – e) indicated translation interrupted by
small-angle pivot perturbations. Y. Shirai, A. J. Osgood, Y. Zhao, K. F. Kelly, and J. M. Tour, Nano Lett., 2005, 5, 2330. Copyright
American Chemical Society (2005).

6.3.4 https://chem.libretexts.org/@go/page/55902
Although literature studies suggested that the C60 molecule rolls on the surface, in the nanocar movement studies it is still not
possible to conclusively conclude that the nanocar moves on surface exclusively via a rolling mechanism. Hopping, sliding and
other moving modes could also be responsible for the movement of the nanocar since the experiment was carried out at high
temperature conditions, making the C60 molecules more energetic to overcome interactions between surfaces.
To tackle the question of the mode of translation, a trimeric “nano-tricycle” has been synthesized. If the movement of fullerene-
wheeled nanocar was based on a hopping or sliding mechanism, the trimer should give observable translational motions like the
four-wheeled nanocar, however, if rolling is the operable motion then the nano-tricycle should rotate on an axis, but not translate
across the surface. The result of the imaging experiment of the trimer at ~200 °C (Figure 6.3.9), yielded very small and
insignificant translational displacements in comparison to 4-wheel nanocar (Figure 6.3.9). The trimeric 3-wheel nanocar showed
some pivoting motions in the images. This motion type can be attributed to the directional preferences of the wheels mounted on
the trimer causing the car to rotate. All the experimental results suggested that a C60-based nanocar moves via a rolling motion
rather than hopping and sliding. In addition, the fact that the thermally driven nanocar only moves in high temperature also
suggests that four C60 have very strong interactions to the surface.

Figure 6.3.9 Pivot motion of the trimer. a) - d) Pivot motions of circled trimered were shown in the series of images. No significant
translation were observed in comparison to the nanocar. Y. Shirai, A. J. Osgood, Y. Zhao, K. F. Kelly, and J. M. Tour, Nano Lett.,
2005, 5, 2330. Copyright American Chemical Society (2005).

6.3: Rolling Molecules on Surfaces Under STM Imaging is shared under a CC BY 4.0 license and was authored, remixed, and/or curated by
Pavan M. V. Raja & Andrew R. Barron via source content that was edited to conform to the style and standards of the LibreTexts platform; a
detailed edit history is available upon request.

6.3.5 https://chem.libretexts.org/@go/page/55902
CHAPTER OVERVIEW
7: Molecular and Solid State Structure
7.1: Crystal Structure
7.2: Structures of Element and Compound Semiconductors
7.3: X-ray Crystallography
7.4: Low Energy Electron Diffraction
7.5: Neutron Diffraction
7.6: XAFS
7.7: Circular Dichroism Spectroscopy and its Application for Determination of Secondary Structure of Optically Active Species
7.8: Protein Analysis using Electrospray Ionization Mass Spectroscopy
7.9: The Analysis of Liquid Crystal Phases using Polarized Optical Microscopy

7: Molecular and Solid State Structure is shared under a CC BY 4.0 license and was authored, remixed, and/or curated by Pavan M. V. Raja &
Andrew R. Barron via source content that was edited to conform to the style and standards of the LibreTexts platform; a detailed edit history is
available upon request.

1
7.1: Crystal Structure
In any sort of discussion of crystalline materials, it is useful to begin with a discussion of crystallography: the study of the
formation, structure, and properties of crystals. A crystal structure is defined as the particular repeating arrangement of atoms
(molecules or ions) throughout a crystal. Structure refers to the internal arrangement of particles and not the external appearance of
the crystal. However, these are not entirely independent since the external appearance of a crystal is often related to the internal
arrangement. For example, crystals of cubic rock salt (NaCl) are physically cubic in appearance. Only a few of the possible crystal
structures are of concern with respect to simple inorganic salts and these will be discussed in detail, however, it is important to
understand the nomenclature of crystallography.

Crystallography
Bravais Lattice
The Bravais lattice is the basic building block from which all crystals can be constructed. The concept originated as a topological
problem of finding the number of different ways to arrange points in space where each point would have an identical “atmosphere”.
That is each point would be surrounded by an identical set of points as any other point, so that all points would be indistinguishable
from each other. Mathematician Auguste Bravais discovered that there were 14 different collections of the groups of points, which
are known as Bravais lattices. These lattices fall into seven different "crystal systems”, as differentiated by the relationship between
the angles between sides of the “unit cell” and the distance between points in the unit cell. The unit cell is the smallest group of
atoms, ions or molecules that, when repeated at regular intervals in three dimensions, will produce the lattice of a crystal system.
The “lattice parameter” is the length between two points on the corners of a unit cell. Each of the various lattice parameters are
designated by the letters a, b, and c. If two sides are equal, such as in a tetragonal lattice, then the lengths of the two lattice
parameters are designated a and c, with b omitted. The angles are designated by the Greek letters α, β, and γsize 12{γ} {}, such that
an angle with a specific Greek letter is not subtended by the axis with its Roman equivalent. For example, α is the included angle
between the b and c axis.
Table 7.1.1 shows the various crystal systems, while Figure 7.1.1 shows the 14 Bravais lattices. It is important to distinguish the
characteristics of each of the individual systems. An example of a material that takes on each of the Bravais lattices is shown in
Table 7.1.2.
Table 7.1.1 Geometrical characteristics of the seven crystal systems
System Axial Lengths and Angles Unit Cell Geometry

cubic a=b=c, α = β = γ = 90°

tetragonal a = b ≠ c, α = β = γ= 90°

orthorhombic a ≠ b ≠ c, α = β = γ= 90°

rhombohedral a = b = c, α = β = γ ≠ 90°

hexagonal a = b ≠ c, α = β = 90°, γ = 120°

7.1.1 https://chem.libretexts.org/@go/page/55904
monoclinic a ≠ b ≠ c, α = γ = 90°, β ≠ 90°

triclinic a ≠ b ≠ c, α ≠ β ≠ γ

Figure 7.1.1 Bravais lattices.


Table 7.1.2 Examples of elements and compounds that adopt each of the crystal systems.
Crystal System Example

triclinic K2S2O8

monoclinic As4S4, KNO2

rhombohedral Hg, Sb

hexagonal Zn, Co, NiAs

orthorhombic Ga, Fe3C

tetragonal In, TiO2

cubic Au, Si, NaCl

The cubic lattice is the most symmetrical of the systems. All the angles are equal to 90°, and all the sides are of the same length (a
= b = c). Only the length of one of the sides (a) is required to describe this system completely. In addition to simple cubic, the cubic
lattice also includes body-centered cubic and face-centered cubic (Figure 7.1.1. Body-centered cubic results from the presence of
an atom (or ion) in the center of a cube, in addition to the atoms (ions) positioned at the vertices of the cube. In a similar manner, a
face-centered cubic requires, in addition to the atoms (ions) positioned at the vertices of the cube, the presence of atoms (ions) in
the center of each of the cubes face.
The tetragonal lattice has all of its angles equal to 90°, and has two out of the three sides of equal length (a = b). The system also
includes body-centered tetragonal (Figure 7.1.1.

7.1.2 https://chem.libretexts.org/@go/page/55904
In an orthorhombic lattice all of the angles are equal to 90°, while all of its sides are of unequal length. The system needs only to be
described by three lattice parameters. This system also includes body-centered orthorhombic, base-centered orthorhombic, and
face-centered orthorhombic (Figure 7.1.1.
A base-centered lattice has, in addition to the atoms (ions) positioned at the vertices of the orthorhombic lattice, atoms (ions)
positioned on just two opposing faces.
The rhombohedral lattice is also known as trigonal, and has no angles equal to 90°, but all sides are of equal length (a = b = c), thus
requiring only by one lattice parameter, and all three angles are equal (α = β = γ).
A hexagonal crystal structure has two angles equal to 90°, with the other angle ( γsize 12{γ} {}) equal to 120°. For this to happen,
the two sides surrounding the 120° angle must be equal (a = b), while the third side (c) is at 90° to the other sides and can be of any
length.
The monoclinic lattice has no sides of equal length, but two of the angles are equal to 90°, with the other angle (usually defined as
β) being something other than 90°. It is a tilted parallelogram prism with rectangular bases. This system also includes base-centered
monoclinic (Figure 7.1.2).
In the triclinic lattice none of the sides of the unit cell are equal, and none of the angles within the unit cell are equal to 90°. The
triclinic lattice is chosen such that all the internal angles are either acute or obtuse. This crystal system has the lowest symmetry
and must be described by 3 lattice parameters (a, b, and c) and the 3 angles (α, β, and γ).

Atom Positions, Crystal Directions and Miller Indices


Atom Positions and Crystal Axes
The structure of a crystal is defined with respect to a unit cell. As the entire crystal consists of repeating unit cells, this definition is
sufficient to represent the entire crystal. Within the unit cell, the atomic arrangement is expressed using coordinates. There are two
systems of coordinates commonly in use, which can cause some confusion. Both use a corner of the unit cell as their origin. The
first, less-commonly seen system is that of Cartesian or orthogonal coordinates (X, Y, Z). These usually have the units of
Angstroms and relate to the distance in each direction between the origin of the cell and the atom. These coordinates may be
manipulated in the same fashion are used with two- or three-dimensional graphs. It is very simple, therefore, to calculate inter-
atomic distances and angles given the Cartesian coordinates of the atoms. Unfortunately, the repeating nature of a crystal cannot be
expressed easily using such coordinates. For example, consider a cubic cell of dimension 3.52 Å. Pretend that this cell contains an
atom that has the coordinates (1.5, 2.1, 2.4). That is, the atom is 1.5 Å away from the origin in the x direction (which coincides with
the a cell axis), 2.1 Å in the y (which coincides with the b cell axis) and 2.4 Å in the z (which coincides with the c cell axis). There
will be an equivalent atom in the next unit cell along the x-direction, which will have the coordinates (1.5 + 3.52, 2.1, 2.4) or (5.02,
2.1, 2.4). This was a rather simple calculation, as the cell has very high symmetry and so the cell axes, a, b and c, coincide with the
Cartesian axes, X, Y and Z. However, consider lower symmetry cells such as triclinic or monoclinic in which the cell axes are not
mutually orthogonal. In such cases, expressing the repeating nature of the crystal is much more difficult to accomplish.
Accordingly, atomic coordinates are usually expressed in terms of fractional coordinates, (x, y, z). This coordinate system is
coincident with the cell axes (a, b, c) and relates to the position of the atom in terms of the fraction along each axis. Consider the
atom in the cubic cell discussion above. The atom was 1.5 Å in the a direction away from the origin. As the a axis is 3.52 Å long,
the atom is (1.5/3.52) or 0.43 of the axis away from the origin. Similarly, it is (2.1/3.52) or 0.60 of the b axis and (2.4/3.5) or 0.68
of the c axis. The fractional coordinates of this atom are, therefore, (0.43, 0.60, 0.68). The coordinates of the equivalent atom in the
next cell over in the a direction, however, are easily calculated as this atom is simply 1 unit cell away in a. Thus, all one has to do is
add 1 to the x coordinate: (1.43, 0.60, 0.68). Such transformations can be performed regardless of the shape of the unit cell.
Fractional coordinates, therefore, are used to retain and manipulate crystal information.

Crystal Directions
The designation of the individual vectors within any given crystal lattice is accomplished by the use of whole number multipliers of
the lattice parameter of the point at which the vector exits the unit cell. The vector is indicated by the notation [hkl], where h, k, and
l are reciprocals of the point at which the vector exits the unit cell. The origination of all vectors is assumed defined as [000]. For
example, the direction along the a-axis according to this scheme would be [100] because this has a component only in the a-
direction and no component along either the b or c axial direction. A vector diagonally along the face defined by the a and baxis
would be [110], while going from one corner of the unit cell to the opposite corner would be in the [111] direction. Figure 7.1.2
shows some examples of the various directions in the unit cell. The crystal direction notation is made up of the lowest combination

7.1.3 https://chem.libretexts.org/@go/page/55904
of integers and represents unit distances rather than actual distances. A [222] direction is identical to a [111], so [111] is used.
Fractions are not used. For example, a vector that intercepts the center of the top face of the unit cell has the coordinates x = 1/2, y
= 1/2, z = 1. All have to be inversed to convert to the lowest combination of integers (whole numbers); i.e., [221] in Figure 7.1.2.
Finally, all parallel vectors have the same crystal direction, e.g., the four vertical edges of the cell shown in Figure 7.1.2 all have
the crystal direction [hkl] = [001].

Figure 7.1.2 Some common directions in a cubic unit cell.


Crystal directions may be grouped in families. To avoid confusion there exists a convention in the choice of brackets surrounding
the three numbers to differentiate a crystal direction from a family of direction. For a direction, square brackets [hkl] are used to
indicate an individual direction. Angle brackets <hkl> indicate a family of directions. A family of directions includes any directions
that are equivalent in length and types of atoms encountered. For example, in a cubic lattice, the [100], [010], and [001] directions
all belong to the <100> family of planes because they are equivalent. If the cubic lattice were rotated 90°, the a, b, and cdirections
would remain indistinguishable, and there would be no way of telling on which crystallographic positions the atoms are situated, so
the family of directions is the same. In a hexagonal crystal, however, this is not the case, so the [100] and [010] would both be
<100> directions, but the [001] direction would be distinct. Finally, negative directions are identified with a bar over the negative
number instead of a minus sign.

Crystal Planes
Planes in a crystal can be specified using a notation called Miller indices. The Miller index is indicated by the notation [hkl] where
h, k, and l are reciprocals of the plane with the x, y, and z axes. To obtain the Miller indices of a given plane requires the following
steps:
1. The plane in question is placed on a unit cell.
2. Its intercepts with each of the crystal axes are then found.
3. The reciprocal of the intercepts are taken.
4. These are multiplied by a scalar to insure that is in the simple ratio of whole numbers.
For example, the face of a lattice that does not intersect the y or z axis would be (100), while a plane along the body diagonal
would be the (111) plane. An illustration of this along with the (111) and (110) planes is given in Figure 7.1.3.

Figure 7.1.3 Examples of Miller indices notation for crystal planes.


As with crystal directions, Miller indices directions may be grouped in families. Individual Miller indices are given in parentheses
(hkl), while braces {hkl} are placed around the indices of a family of planes. For example, (001), (100), and (010) are all in the

7.1.4 https://chem.libretexts.org/@go/page/55904
{100} family of planes, for a cubic lattice.

Description of Crystal Structures


Crystal structures may be described in a number of ways. The most common manner is to refer to the size and shape of the unit cell
and the positions of the atoms (or ions) within the cell. However, this information is sometimes insufficient to allow for an
understanding of the true structure in three dimensions. Consideration of several unit cells, the arrangement of the atoms with
respect to each other, the number of other atoms they in contact with, and the distances to neighboring atoms, often will provide a
better understanding. A number of methods are available to describe extended solid-state structures. The most applicable with
regard to elemental and compound semiconductor, metals and the majority of insulators is the close packing approach.

Close Packed Structures: Hexagonal Close Packing and Cubic Close Packing
Many crystal structures can be described using the concept of close packing. This concept requires that the atoms (ions) are
arranged so as to have the maximum density. In order to understand close packing in three dimensions, the most efficient way for
equal sized spheres to be packed in two dimensions must be considered.
The most efficient way for equal sized spheres to be packed in two dimensions is shown in Figure 7.1.4, in which it can be seen
that each sphere (the dark gray shaded sphere) is surrounded by, and is in contact with, six other spheres (the light gray spheres in
Figure 7.1.4. It should be noted that contact with six other spheres the maximum possible is the spheres are the same size, although
lower density packing is possible. Close packed layers are formed by repetition to an infinite sheet. Within these close packed
layers, three close packed rows are present, shown by the dashed lines in Figure 7.1.4.

Figure 7.1.4 Schematic representation of a close packed layer of equal sized spheres. The close packed rows (directions) are shown
by the dashed lines.
The most efficient way for equal sized spheres to be packed in three dimensions is to stack close packed layers on top of each other
to give a close packed structure. There are two simple ways in which this can be done, resulting in either a hexagonal or cubic close
packed structures.

Hexagonal Close Packed


If two close packed layers A and B are placed in contact with each other so as to maximize the density, then the spheres of layer B
will rest in the hollow (vacancy) between three of the spheres in layer A. This is demonstrated in Figure 7.1.5. Atoms in the second
layer, B (shaded light gray), may occupy one of two possible positions (Figure 7.1.5 a or b) but not both together or a mixture of
each. If a third layer is placed on top of layer B such that it exactly covers layer A, subsequent placement of layers will result in the
following sequence ...ABABAB.... This is known as hexagonal close packing or hcp.

Figure 7.1.5 Schematic representation of two close packed layers arranged in A (dark grey) and B (light grey) positions. The
alternative stacking of the B layer is shown in (a) and (b).
The hexagonal close packed cell is a derivative of the hexagonal Bravais lattice system (Figure 7.1.6 with the addition of an atom
inside the unit cell at the coordinates (1/3,2/3,1/2). The basal plane of the unit cell coincides with the close packed layers (Figure
7.1.6. In other words the close packed layer makes-up the {001} family of crystal planes.

7.1.5 https://chem.libretexts.org/@go/page/55904
Figure 7.1.6 A schematic projection of the basal plane of the hcp unit cell on the close packed layers.
The “packing fraction” in a hexagonal close packed cell is 74.05%; that is 74.05% of the total volume is occupied. The packing
fraction or density is derived by assuming that each atom is a hard sphere in contact with its nearest neighbors. Determination of
the packing fraction is accomplished by calculating the number of whole spheres per unit cell (2 in hcp), the volume occupied by
these spheres, and a comparison with the total volume of a unit cell. The number gives an idea of how “open” or filled a structure
is. By comparison, the packing fraction for body-centered cubic (Figure 7.1.5) is 68% and for diamond cubic (an important
semiconductor structure to be described later) is it 34%.

Cubic Close Packed: Face-centered Cubic


In a similar manner to the generation of the hexagonal close packed structure, two close packed layers are stacked (Figure 7.1.7
however, the third layer (C) is placed such that it does not exactly cover layer A, while sitting in a set of troughs in layer B (Figure
7.1.7), then upon repetition the packing sequence will be ...ABCABCABC.... This is known as cubic close packing or ccp.

Figure 7.1.7 Schematic representation of the three close packed layers in a cubic close packed arrangement: A (dark grey), B
(medium grey), and C (light grey).
The unit cell of cubic close packed structure is actually that of a face-centered cubic (fcc) Bravais lattice. In the fcc lattice the close
packed layers constitute the {111} planes. As with the hcp lattice packing fraction in a cubic close packed (fcc) cell is 74.05%.
Since face centered cubic or fcc is more commonly used in preference to cubic close packed (ccp) in describing the structures, the
former will be used throughout this text.

Coordination Number
The coordination number of an atom or ion within an extended structure is defined as the number of nearest neighbor atoms (ions
of opposite charge) that are in contact with it. A slightly different definition is often used for atoms within individual molecules: the
number of donor atoms associated with the central atom or ion. However, this distinction is rather artificial, and both can be
employed.
The coordination numbers for metal atoms in a molecule or complex are commonly 4, 5, and 6, but all values from 2 to 9 are
known and a few examples of higher coordination numbers have been reported. In contrast, common coordination numbers in the
solid state are 3, 4, 6, 8, and 12. For example, the atom in the center of body-centered cubic lattice has a coordination number of 8,
because it touches the eight atoms at the corners of the unit cell, while an atom in a simple cubic structure would have a
coordination number of 6. In both fcc and hcp lattices each of the atoms have a coordination number of 12.

Octahedral and Tetrahedral Vacancies


As was mentioned above, the packing fraction in both fcc and hcp cells is 74.05%, leaving 25.95% of the volume unfilled. The
unfilled lattice sites (interstices) between the atoms in a cell are called interstitial sites or vacancies. The shape and relative size of
these sites is important in controlling the position of additional atoms. In both fcc and hcp cells most of the space within these
atoms lies within two different sites known as octahedral sites and tetrahedral sites. The difference between the two lies in their
“coordination number”, or the number of atoms surrounding each site. Tetrahedral sites (vacancies) are surrounded by four atoms
arranged at the corners of a tetrahedron. Similarly, octahedral sites are surrounded by six atoms which make-up the apices of an
octahedron. For a given close packed lattice an octahedral vacancy will be larger than a tetrahedral vacancy.
Within a face centered cubic lattice, the eight tetrahedral sites are positioned within the cell, at the general fractional coordinate of
(n/4,n/4,n/4) where n = 1 or 3, e.g., (1/4,1/4,1/4), (1/4,1/4,3/4), etc. The octahedral sites are located at the center of the unit cell

7.1.6 https://chem.libretexts.org/@go/page/55904
(1/2,1/2,1/2), as well as at each of the edges of the cell, e.g., (1/2,0,0). In the hexagonal close packed system, the tetrahedral sites
are at (0,0,3/8) and (1/3,2/3,7/8), and the octahedral sites are at (1/3,1/3,1/4) and all symmetry equivalent positions.

Important Structure Types


The majority of crystalline materials do not have a structure that fits into the one atom per site simple Bravais lattice. A number of
other important crystal structures are found, however, only a few of these crystal structures are those of which occur for the
elemental and compound semiconductors and the majority of these are derived from fcc or hcp lattices. Each structural type is
generally defined by an archetype, a material (often a naturally occurring mineral) which has the structure in question and to which
all the similar materials are related. With regard to commonly used elemental and compound semiconductors the important
structures are diamond, zinc blende, Wurtzite, and to a lesser extent chalcopyrite. However, rock salt, β-tin, cinnabar and cesium
chloride are observed as high pressure or high temperature phases and are therefore also discussed. The following provides a
summary of these structures. Details of the full range of solid-state structures are given elsewhere.

Diamond Cubic
The diamond cubic structure consists of two interpenetrating face-centered cubic lattices, with one offset 1/4 of a cube along the
cube diagonal. It may also be described as face centered cubic lattice in which half of the tetrahedral sites are filled while all the
octahedral sites remain vacant. The diamond cubic unit cell is shown in Figure 7.1.8. Each of the atoms (e.g., C) is four coordinate,
and the shortest interatomic distance (C-C) may be determined from the unit cell parameter (a).

√3
C − C   =  a ≈  0.422a (7.1.1)
4

Figure 7.1.8 Unit cell structure of a diamond cubic lattice showing the two interpenetrating face-centered cubic lattices.

Zinc Blende
This is a binary phase (ME) and is named after its archetype, a common mineral form of zinc sulfide (ZnS). As with the diamond
lattice, zinc blende consists of the two interpenetrating fcc lattices. However, in zinc blende one lattice consists of one of the types
of atoms (Zn in ZnS), and the other lattice is of the second type of atom (S in ZnS). It may also be described as face centered cubic
lattice of S atoms in which half of the tetrahedral sites are filled with Zn atoms. All the atoms in a zinc blende structure are 4-
coordinate. The zinc blende unit cell is shown in Figure 7.1.9. A number of inter-atomic distances may be calculated for any
material with a zinc blende unit cell using the lattice parameter (a).

√3
Zn − S  =  a ≈  0.422a (7.1.2)
4

a
Zn − Zn  =  S − S  = – ≈ 0.707 a (7.1.3)
√2

Figure 7.1.9 Unit cell structure of a zinc blende (ZnS) lattice. Zinc atoms are shown in green (small), sulfur atoms shown in red
(large), and the dashed lines show the unit cell.

7.1.7 https://chem.libretexts.org/@go/page/55904
Chalcopyrite
The mineral chalcopyrite CuFeS2 is the archetype of this structure. The structure is tetragonal (a = b ≠ c, α = β = γ = 90°, and is
essentially a superlattice on that of zinc blende. Thus, is easiest to imagine that the chalcopyrite lattice is made-up of a lattice of
sulfur atoms in which the tetrahedral sites are filled in layers, ...FeCuCuFe..., etc. (Figure 7.1.10. In such an idealized structure c =
2a, however, this is not true of all materials with chalcopyrite structures.

Figure 7.1.10 Unit cell structure of a chalcopyrite lattice. Copper atoms are shown in blue, iron atoms are shown in green and
sulfur atoms are shown in yellow. The dashed lines show the unit cell.

Rock Salt
As its name implies the archetypal rock salt structure is NaCl (table salt). In common with the zinc blende structure, rock salt
consists of two interpenetrating face-centered cubic lattices. However, the second lattice is offset 1/2a along the unit cell axis. It
may also be described as face centered cubic lattice in which all of the octahedral sites are filled, while all the tetrahedral sites
remain vacant, and thus each of the atoms in the rock salt structure are 6-coordinate. The rock salt unit cell is shown in Figure
7.1.11. A number of inter-atomic distances may be calculated for any material with a rock salt structure using the lattice parameter

(a).
a
N a − C l  =   ≈ 0.5a (7.1.4)
2

a
N a − N a  =  C l − C l  =   – ≈ 0.707 a (7.1.5)
√2

Figure 7.1.11 Unit cell structure of a rock salt lattice. Sodium ions are shown in purple (small spheres) and chloride ions are shown
in red (large spheres).

Cinnabar
Cinnabar, named after the archetype mercury sulfide, HgS, is a distorted rock salt structure in which the resulting cell is
rhombohedral (trigonal) with each atom having a coordination number of six.

Wurtzite
This is a hexagonal form of the zinc sulfide. It is identical in the number of and types of atoms, but it is built from two
interpenetrating hcp lattices as opposed to the fcc lattices in zinc blende. As with zinc blende all the atoms in a wurtzite structure
are 4-coordinate. The wurtzite unit cell is shown in Figure 7.1.12. A number of inter atomic distances may be calculated for any
material with a wurtzite cell using the lattice parameter (a).

7.1.8 https://chem.libretexts.org/@go/page/55904
−−− 3c
Zn − S  =  a√3/8  =  0.612 a  =   =  0.375 c (7.1.6)
8

Zn − Zn  =  S − S  =  a  =  1.632 c (7.1.7)

However, it should be noted that these formulae do not necessarily apply when the ratio a/c is different from the ideal value of
1.632.

Figure 7.1.12 Unit cell structure of a wurtzite lattice. Zinc atoms are shown in green (small spheres), sulfur atoms shown in red
(large spheres), and the dashed lines show the unit cell.

Cesium Chloride
The cesium chloride structure is found in materials with large cations and relatively small anions. It has a simple (primitive) cubic
cell (Figure 7.1.13) with a chloride ion at the corners of the cube and the cesium ion at the body center. The coordination numbers
of both Cs+ and Cl-, with the inner atomic distances determined from the cell lattice constant (a).

a√3
C s − C l  =   ≈ 0.866a (7.1.8)
2

C s − C s  =  C l − C l  = a (7.1.9)

β-Tin
The room temperature allotrope of tin is β-tin or white tin. It has a tetragonal structure, in which each tin atom has four nearest
neighbors (Sn-Sn = 3.016 Å) arranged in a very flattened tetrahedron, and two next nearest neighbors (Sn-Sn = 3.175 Å). The
overall structure of β-tin consists of fused hexagons, each being linked to its neighbor via a four-membered Sn4 ring.

Defects in Crystalline Solids


Up to this point we have only been concerned with ideal structures for crystalline solids in which each atom occupies a designated
point in the crystal lattice. Unfortunately, defects ordinarily exist in equilibrium between the crystal lattice and its environment.
These defects are of two general types: point defects and extended defects. As their names imply, point defects are associated with
a single crystal lattice site, while extended defects occur over a greater range.

Point Defects: "Too Many or Too Few" or "Just Plain Wrong"


Point defects have a significant effect on the properties of a semiconductor, so it is important to understand the classes of point
defects and the characteristics of each type. Figure 7.1.13 summarizes various classes of native point defects, however, they may
be divided into two general classes; defects with the wrong number of atoms (deficiency or surplus) and defects where the identity
of the atoms is incorrect.

7.1.9 https://chem.libretexts.org/@go/page/55904
Figure 7.1.13 Point defects in a crystal lattice.

Interstitial Impurity
An interstitial impurity occurs when an extra atom is positioned in a lattice site that should be vacant in an ideal structure (Figure
7.1.13 b).Since all the adjacent lattice sites are filled the additional atom will have to squeeze itself into the interstitial site,

resulting in distortion of the lattice and alteration in the local electronic behavior of the structure. Small atoms, such as carbon, will
prefer to occupy these interstitial sites. Interstitial impurities readily diffuse through the lattice via interstitial diffusion, which can
result in a change of the properties of a material as a function of time. Oxygen impurities in silicon generally are located as
interstitials.

Vacancies
The converse of an interstitial impurity is when there are not enough atoms in a particular area of the lattice. These are called
vacancies. Vacancies exist in any material above absolute zero and increase in concentration with temperature. In the case of
compound semiconductors, vacancies can be either cation vacancies (Figure 7.1.13 c) or anion vacancies (Figure 7.1.13 d),
depending on what type of atom are “missing”.

Substitution
Substitution of various atoms into the normal lattice structure is common, and used to change the electronic properties of both
compound and elemental semiconductors. Any impurity element that is incorporated during crystal growth can occupy a lattice
site. Depending on the impurity, substitution defects can greatly distort the lattice and/or alter the electronic structure. In general,
cations will try to occupy cation lattice sites (Figure 7.1.13 e), and anion will occupy the anion site (Figure 7.1.13 f). For example,
a zinc impurity in GaAs will occupy a gallium site, if possible, while a sulfur, selenium and tellurium atoms would all try to
substitute for an arsenic. Some impurities will occupy either site indiscriminately, e.g., Si and Sn occupy both Ga and As sites in
GaAs.

Antisite Defects
Antisite defects are a particular form of substitution defect, and are unique to compound semiconductors. An antisite defect occurs
when a cation is misplaced on an anion lattice site or vice versa ( Figure 7.1.13 g and h).Dependant on the arrangement these are
designated as either AB antisite defects or BA antisite defects. For example, if an arsenic atom is on a gallium lattice site the defect
would be an AsGa defect. Antisite defects involve fitting into a lattice site atoms of a different size than the rest of the lattice, and
therefore this often results in a localized distortion of the lattice. In addition, cations and anions will have a different number of
electrons in their valence shells, so this substitution will alter the local electron concentration and the electronic properties of this
area of the semiconductor.

7.1.10 https://chem.libretexts.org/@go/page/55904
Extended Defects: Dislocations in a Crystal Lattice
Extended defects may be created either during crystal growth or as a consequence of stress in the crystal lattice. The plastic
deformation of crystalline solids does not occur such that all bonds along a plane are broken and reformed simultaneously. Instead,
the deformation occurs through a dislocation in the crystal lattice. Figure shows a schematic representation of a dislocation in a
crystal lattice. Two features of this type of dislocation are the presence of an extra crystal plane, and a large void at the dislocation
core. Impurities tend to segregate to the dislocation core in order to relieve strain from their presence.

Figure 7.1.14 Dislocation in a crystal lattice.

Epitaxy
Epitaxy, is a transliteration of two Greek words epi, meaning "upon", and taxis, meaning "ordered". With respect to crystal growth
it applies to the process of growing thin crystalline layers on a crystal substrate. In epitaxial growth, there is a precise crystal
orientation of the film in relation to the substrate. The growth of epitaxial films can be done by a number of methods including
molecular beam epitaxy, atomic layer epitaxy, and chemical vapor deposition, all of which will be described later.
Epitaxy of the same material, such as a gallium arsenide film on a gallium arsenide substrate, is called homoepitaxy, while epitaxy
where the film and substrate material are different is called heteroepitaxy. Clearly, in homoepitaxy, the substrate and film will have
the identical structure, however, in heteroepitaxy, it is important to employ where possible a substrate with the same structure and
similar lattice parameters. For example, zinc selenide (zinc blende, a = 5.668 Å) is readily grown on gallium arsenide (zinc blende,
a = 5.653 Å). Alternatively, epitaxial crystal growth can occur where there exists a simple relationship between the structures of the
substrate and crystal layer, such as is observed between Al2O3 (100) on Si (100). Whichever route is chosen a close match in the
lattice parameters is required, otherwise, the strains induced by the lattice mismatch results in distortion of the film and formation
of dislocations. If the mismatch is significant epitaxial growth is not energetically favorable, causing a textured film or
polycrystalline untextured film to be grown. As a general rule of thumb, epitaxy can be achieved if the lattice parameters of the two
materials are within about 5% of each other. For good quality epitaxy, this should be less than 1%. The larger the mismatch, the
larger the strain in the film. As the film gets thicker and thicker, it will try to relieve the strain in the film, which could include the
loss of epitaxy of the growth of dislocations. It is important to note that the <100> directions of a film must be parallel to the <100>
direction of the substrate. In some cases, such as Fe on MgO, the [111] direction is parallel to the substrate [100]. The epitaxial
relationship is specified by giving first the plane in the film that is parallel to the substrate [100].

7.1: Crystal Structure is shared under a CC BY 4.0 license and was authored, remixed, and/or curated by Pavan M. V. Raja & Andrew R. Barron
via source content that was edited to conform to the style and standards of the LibreTexts platform; a detailed edit history is available upon
request.

7.1.11 https://chem.libretexts.org/@go/page/55904
7.2: Structures of Element and Compound Semiconductors
A single crystal of either an elemental (e.g., silicon) or compound (e.g., gallium arsenide) semiconductor forms the basis of almost
all semiconductor devices. The ability to control the electronic and opto-electronic properties of these materials is based on an
understanding of their structure. In addition, the metals and many of the insulators employed within a microelectronic device are
also crystalline.

Group IV (14) Elements


Each of the semiconducting phases of the group IV (14) elements, C (diamond), Si, Ge, and α-Sn, adopt the diamond cubic
structure (Figure 7.2.1). Their lattice constants (a, Å) and densities (ρ, g/cm3) are given in Table 7.2.1.

Figure 7.2.1 Unit cell structure of a diamond cubic lattice showing the two interpenetrating face-centered cubic lattices.
Table 7.2.1 : Lattice parameters and densities (measured at 298 K) for the diamond cubic forms of the group IV (14) elements.
Element Lattice Parameter, a (Å) Density (g/cm3)

carbon (diamond) 3.56683(1) 3.51525

silicon 5.4310201(3) 2.319002

germanium 5.657906(1) 5.3234

tin (α-Sn) 6.4892(1)

As would be expected the lattice parameter increase in the order C < Si < Ge < α-Sn. Silicon and germanium form a continuous
series of solid solutions with gradually varying parameters. It is worth noting the high degree of accuracy that the lattice parameters
are known for high purity crystals of these elements. In addition, it is important to note the temperature at which structural
measurements are made, since the lattice parameters are temperature dependent (Figure 7.2.1). The lattice constant (a), in Å, for
high purity silicon may be calculated for any temperature (T) over the temperature range 293 - 1073 K by the formula shown
below.
−5 −9
aT   =  5.4304  +  1.8138 × 10  (T − 298.15 K)  +  1.542 × 10  (T − 298.15 K) (7.2.1)

7.2.1 https://chem.libretexts.org/@go/page/55905
Figure 7.2.2 Temperature dependence of the lattice parameter for (a) Si and (b) Ge.
Even though the diamond cubic forms of Si and Ge are the only forms of direct interest to semiconductor devices, each exists in
numerous crystalline high pressure and meta-stable forms. These are described along with their interconversions, in Table 7.2.2.
Table 7.2.2 : High pressure and metastable phases of silicon and germanium.
Phase Structure Remarks

Si I diamond cubic stable at normal pressure

Si II grey tin structure formed from Si I or Si V above 14 GPa

Si III cubic metastable, formed from Si II above 10 GPa

Si IV hexagonal

stable above 34 GPa, formed from Si II above


Si V unidentified
16 GPa

Si VI hexagonal close packed stable above 45 GPa

Ge I diamond cubic low-pressure phase

Ge II β-tin structure formed from Ge I above 10 GPa

Ge III tetragonal formed by quenching Ge II at low pressure

Ge IV body centered formed by quenching Ge II to 1 atm at 200 K

Group III-V (13-15) Compounds


The stable phases for the arsenides, phosphides and antimonides of aluminum, gallium and indium all exhibit zinc blende structures
(Figure 7.2.3). In contrast, the nitrides are found as wurtzite structures (e.g., Figure 7.2.4). The structure, lattice parameters, and
densities of the III-V compounds are given in Table 7.2.3. It is worth noting that contrary to expectation the lattice parameter of the
gallium compounds is smaller than their aluminum homolog; for GaAs a = 5.653 Å; AlAs a = 5.660 Å. As with the group IV
elements the lattice parameters are highly temperature dependent; however, additional variation arises from any deviation from
absolute stoichiometry. These effects are shown in Figure 7.2.4.

7.2.2 https://chem.libretexts.org/@go/page/55905
Figure 7.2.3 Unit cell structure of a zinc blende (ZnS) lattice. Zinc atoms are shown in green (small), sulfur atoms shown in red
(large), and the dashed lines show the unit cell.

Figure 7.2.4 Unit cell structure of a wurtzite lattice. Zinc atoms are shown in green (small), sulfur atoms shown in red (large), and
the dashed lines show the unit cell.
Table 7.2.4 Lattice parameters and densities (measured at 298 K) for the III-V (13-15) compound semiconductors. Estimated standard
deviations given in parentheses.
Compound Structure Lattice Parameter (Å) Density (g/cm3)

AIN wurtzite a = 3.11(1), c = 4.98(1) 3.255

AIP zinc blende a = 5.4635(4) 2.40(1)

AIAs zinc blende a= 5.660 3.760

AISb zinc blende a = 6.1355(1) 4.26

GaN wurtzite a = 3.190, c=5.187

GaP zinc blende a= 5.4505(2) 4.138

GaAs zinc blende a= 5.56325(2) 5.3176(3)

InN wurtzite a= 3.5446, c= 5.7034 6.81

InP zinc blende a= 5.868(1) 4.81

InAs zinc blende a= 6.0583 5.667

InSb zinc blende a= 6.47937 5.7747(4)

Figure 7.2.5 Temperature dependence of the lattice parameter for stoichiometric GaAs and crystals with either Ga or As excess.
The homogeneity of structures of alloys for a wide range of solid solutions to be formed between III-V compounds in almost any
combination. Two classes of ternary alloys are formed: IIIx-III1-x-V (e.g., Alx-Ga1-x-As) and III-V1-x-Vx (e.g., Ga-As1-x-Px) . While
quaternary alloys of the type IIIx-III1-x-Vy-V1-y allow for the growth of materials with similar lattice parameters, but a broad range

7.2.3 https://chem.libretexts.org/@go/page/55905
of band gaps. A very important ternary alloy, especially in optoelectronic applications, is Alx-Ga1-x-As and its lattice parameter (a)
is directly related to the composition (x).
a  =  5.6533  +  0.0078 x (7.2.2)

Not all of the III-V compounds have well characterized high-pressure phases. however, in each case where a high-pressure phase is
observed the coordination number of both the group III and group V element increases from four to six. Thus, AlP undergoes a zinc
blende to rock salt transformation at high pressure above 170 kbar, while AlSb and GaAs form orthorhombic distorted rock salt
structures above 77 and 172 kbar, respectively. An orthorhombic structure is proposed for the high-pressure form of InP (>133
kbar). Indium arsenide (InAs) undergoes two-phase transformations. The zinc blende structure is converted to a rock salt structure
above 77 kbar, which in turn forms a β-tin structure above 170 kbar.

Group II-VI (12-16) Compounds


The structures of the II-VI compound semiconductors are less predictable than those of the III-V compounds (above), and while
zinc blende structure exists for almost all of the compounds there is a stronger tendency towards the hexagonal wurtzite form. In
several cases the zinc blende structure is observed under ambient conditions, but may be converted to the wurtzite form upon
heating. In general the wurtzite form predominates with the smaller anions (e.g., oxides), while the zinc blende becomes the more
stable phase for the larger anions (e.g., tellurides). One exception is mercury sulfide (HgS) that is the archetype for the trigonal
cinnabar phase.Table 7.2.5 lists the stable phase of the chalcogenides of zinc, cadmium and mercury, along with their high
temperature phases where applicable. Solid solutions of the II-VI compounds are not as easily formed as for the III-V compounds;
however, two important examples are ZnSxSe1-x and CdxHg1-xTe.
Table 7.2.5 Lattice parameters and densities (measured at 298 K) for the II-VI (12-16) compound semiconductors.
Compound Structure Lattice Parameter (Å) Density (g/cm3)

ZnS zinc blende a= 5.410 4.075

wurtzite a = 3.822, c= 6.260 4.087

ZnSe zinc blende a = 5.668 5.27

ZnTe zinc blende a = 6.10 5.636

CdS wurtzite a = 4.136, c = 6.714 4.82

CdSe wurtzite a = 4.300, c = 7.011 5.81

CdTe zinc blende a = 6.482 5.87

HgS cinnabar a = 4.149, c = 9.495

zinc blende a = 5.851 7.73

HgSe zinc blende a = 6.085 8.25

HgTe zinc blende a = 6.46 8.07

The zinc chalcogenides all transform to a cesium chloride structure under high pressures, while the cadmium compounds all form
rock salt high-pressure phases (Figure 7.2.6). Mercury selenide (HgSe) and mercury telluride (HgTe) convert to the mercury
sulfide archetype structure, cinnabar, at high pressure.

Figure 7.2.6 Unit cell structure of a rock salt lattice. Sodium ions are shown in purple and chloride ions are shown in red.

7.2.4 https://chem.libretexts.org/@go/page/55905
I-III-VI2 (11-13-16) Compounds
Nearly all I-III-VI2 compounds at room temperature adopt the chalcopyrite structure (Figure 7.2.7). The cell constants and
densities are given in Table 7.2.6. Although there are few reports of high temperature or high-pressure phases, AgInS2 has been
shown to exist as a high temperature orthorhombic polymorph (a = 6.954, b = 8.264, and c = 6.683 Å), and AgInTe2 forms a cubic
phase at high pressures.

Figure 7.2.7 Unit cell structure of a chalcopyrite lattice. Copper atoms are shown in blue, iron atoms are shown in green and sulfur
atoms are shown in yellow. The dashed lines show the unit cell.
Table 7.2.6 Chalcopyrite lattice parameters and densities (measured at 298 K) for the I-III-VI compound semiconductors. Lattice parameters
for tetragonal cell.
Compound Lattice Parameter a (Å) Lattice parameter c (Å) Density (g cm3)

CuAlS2 5.32 10.430 3.45

CuAlSe2 5.61 10.92 4.69

CuAlTe2 5.96 11.77 5.47

CuGaS2 5.35 10.46 4.38

CuGaSe2 5.61 11.00 5.57

CuGaTe2 6.00 11.93 5.95

CuInS2 5.52 11.08 4.74

CuInSe2 5.78 11.55 5.77

CuInTe2 6.17 12.34 6.10

AgAlS2 6.30 11.84 6.15

AgGaS2 5.75 10.29 4.70

AgGaSe2 5.98 10.88 5.70

AgGaTe2 6.29 11.95 6.08

AgInS2 5.82 11.17 4.97

AgInSe2 6.095 11.69 5.82

AgInTe2 6.43 12.59 6.96

Of the I-III-VI2 compounds, the copper indium chalcogenides (CuInE2) are certainly the most studied for their application in solar
cells. One of the advantages of the copper indium chalcogenide compounds is the formation of solid solutions (alloys) of the
formula CuInE2-xE'x, where the composition variable (x) varies from 0 to 2. The CuInS2-xSex and CuInSe2-xTex systems have
also been examined, as has the CuGayIn1-yS2-xSex quaternary system. As would be expected from a consideration of the relative
ionic radii of the chalcogenides the lattice parameters of the CuInS2-xSex alloy should increase with increased selenium content.
Vergard's law requires the lattice constant for a linear solution of two semiconductors to vary linearly with composition (e.g., as is

7.2.5 https://chem.libretexts.org/@go/page/55905
observed for AlxGa1-xAs), however, the variation of the tetragonal lattice constants (a and c) with composition for CuInS2-xSx are
best described by the parabolic relationships.
2
a  =  5.532  +  0.0801x  +  0.026x (7.2.3)

2
c  =  11.156  +  0.1204x  +  0.0611x (7.2.4)

A similar relationship is observed for the CuInSe2-xTex alloys.


2
a  =  5.783  +  0.1560x  +  0.0212x (7.2.5)

2
c  =  11.628  +  0.3340x  +  0.0277x (7.2.6)

The large difference in ionic radii between S and Te (0.37 Å) prevents formation of solid solutions in the CuInS2-xTex system,
however, the single alloy CuInS1.5Te0.5 has been reported.

Orientation Effects
Once single crystals of high purity silicon or gallium arsenide are produced they are cut into wafers such that the exposed face of
these wafers is either the crystallographic {100} or {111} planes. The relative structure of these surfaces are important with respect
to oxidation, etching and thin film growth. These processes are orientation-sensitive; that is, they depend on the direction in which
the crystal slice is cut.

Atom Density and Dangling Bonds


The principle planes in a crystal may be differentiated in a number of ways, however, the atom and/or bond density are useful in
predicting much of the chemistry of semiconductor surfaces. Since both silicon and gallium arsenide are fcc structures and the
{100} and {111} are the only technologically relevant surfaces, discussions will be limited to fcc {100} and {111}.
The atom density of a surface may be defined as the number of atoms per unit area. Figure shows a schematic view of the {111}
and {100} planes in a fcc lattice. The {111} plane consists of a hexagonal close packed array in which the crystal directions within
the plane are oriented at 60° to each other. The hexagonal packing and the orientation of the crystal directions are indicated in
Figure 7.2.8 b as an overlaid hexagon. Given the intra-planar inter-atomic distance may be defined as a function of the lattice
parameter, the area of this hexagon may be readily calculated. For example in the case of silicon, the hexagon has an area of 38.30
Å2. The number of atoms within the hexagon is three: the atom in the center plus 1/3 of each of the six atoms at the vertices of the
hexagon (each of the atoms at the hexagons vertices is shared by three other adjacent hexagons). Thus, the atom density of the
{111} plane is calculated to be 0.0783 Å-2. Similarly, the atom density of the {100} plane may be calculated. The {100} plane
consists of a square array in which the crystal directions within the plane are oriented at 90° to each other. Since the square is
coincident with one of the faces of the unit cell the area of the square may be readily calculated. For example in the case of silicon,
the square has an area of 29.49 Å2. The number of atoms within the square is 2: the atom in the center plus 1/4 of each of the four
atoms at the vertices of the square (each of the atoms at the corners of the square are shared by four other adjacent squares). Thus,
the atom density of the {100} plane is calculated to be 0.0678 Å-2. While these values for the atom density are specific for silicon,
their ratio is constant for all diamond cubic and zinc blende structures: {100}:{111} = 1:1.155. In general, the fewer dangling
bonds the more stable a surface structure.

Figure 7.2.8 Schematic representation of the (111) and (100) faces of a face centered cubic (fcc) lattice showing the relationship
between the close packed rows.
An atom inside a crystal of any material will have a coordination number (n) determined by the structure of the material. For
example, all atoms within the bulk of a silicon crystal will be in a tetrahedral four-coordinate environment (n = 4). However, at the
surface of a crystal the atoms will not make their full compliment of bonds. Each atom will therefore have less nearest neighbors
than an atom within the bulk of the material. The missing bonds are commonly called dangling bonds. While this description is not
particularly accurate it is, however, widely employed and as such will be used herein. The number of dangling bonds may be

7.2.6 https://chem.libretexts.org/@go/page/55905
defined as the difference between the ideal coordination number (determined by the bulk crystal structure) and the actual
coordination number as observed at the surface.
Figure 7.2.9 shows a section of the {111} surfaces of a diamond cubic lattice viewed perpendicular to the {111} plane. The atoms
within the bulk have a coordination number of four. In contrast, the atoms at the surface (e.g., the atom shown in blue in Figure
7.2.10 are each bonded to just three other atoms (the atoms shown in red in Figure), thus each surface atom has one dangling bond.

As can be seen from Figure 7.2.10, which shows the atoms at the {100} surface viewed perpendicular to the {100} plane, each
atom at the surface (e.g., the atom shown in blue in Figure 7.2.9 is only coordinated to two other atoms (the atoms shown in red in
Figure 7.2.10, leaving two dangling bonds per atom. It should be noted that the same number of dangling bonds are found for the
{111} and {100} planes of a zinc blende lattice. The ratio of dangling bonds for the {100} and {111} planes of all diamond cubic
and zinc blende structures is {100}:{111} = 2:1. Furthermore, since the atom densities of each plane are known then the ratio of the
dangling bond densities is determined to be: {100}:{111} = 1:0.577.

Figure 7.2.9 A section of the {111} surfaces of a diamond cubic lattice viewed perpendicular to the {111} plane.

Figure 7.2.10 A section of the {100} surface of a diamond cubic lattice viewed perpendicular to the {100} plane.

Silicon
For silicon, the {111} planes are closer packed than the {100} planes. As a result, growth of a silicon crystal is therefore slowest in
the <111> direction, since it requires laying down a close packed atomic layer upon another layer in its closest packed form. As a
consequence <111> Si is the easiest to grow, and therefore the least expensive.
The dissolution or etching of a crystal is related to the number of broken bonds already present at the surface: the fewer bonds to be
broken in order to remove an individual atom from a crystal, the easier it will be to dissolve the crystal. As a consequence of having
only one dangling bond (requiring three bonds to be broken) etching silicon is slowest in the <111> direction. The electronic
properties of a silicon wafer are also related to the number of dangling bonds.
Silicon microcircuits are generally formed on a single crystal wafer that is diced after fabrication by either sawing part way through
the wafer thickness or scoring (scribing) the surface, and then physically breaking. The physical breakage of the wafer occurs along
the natural cleavage planes, which in the case of silicon are the {111} planes.

Gallium Arsenide
The zinc blende lattice observed for gallium arsenide results in additional considerations over that of silicon. Although the {100}
plane of GaAs is structurally similar to that of silicon, two possibilities exist: a face consisting of either all gallium atoms or all
arsenic atoms. In either case the surface atoms have two dangling bonds, and the properties of the face are independent of whether
the face is gallium or arsenic.
The {111} plane also has the possibility of consisting of all gallium or all arsenic. However, unlike the {100} planes there is a
significant difference between the two possibilities. Figure 7.2.11 shows the gallium arsenide structure represented by two
interpenetrating fcc lattices. The [111] axis is vertical within the plane of the page. Although the structure consists of alternate
layers of gallium and arsenic stacked along the [111] axis, the distance between the successive layers alternates between large and
small. Assigning arsenic as the parent lattice the order of the layers in the [111] direction is As Ga-As Ga-As Ga, while in the

7.2.7 https://chem.libretexts.org/@go/page/55905
[111] direction the layers are ordered, Ga-As-Ga As-Ga As (Figure 7.2.11).In silicon these two directions are of course identical.
The surface of a crystal would be either arsenic, with three dangling bonds, or gallium, with one dangling bond. Clearly, the latter
is energetically more favorable. Thus, the (111) plane shown in Figure 7.2.11 is called the (111) Ga face. Conversely, the [111]
plane would be either gallium, with three dangling bonds, or arsenic, with one dangling bond. Again, the latter is energetically
more favorable and the [111] plane is therefore called the (111) As face.

Figure 7.2.11 The (111) Ga face of GaAs showing a surface layer containing gallium atoms (green) with one dangling bond per
gallium and three bonds to the arsenic atoms (red) in the lower layer.
The (111) As is distinct from that of (111) Ga due to the difference in the number of electrons at the surface. As a consequence, the
(111) As face etches more rapidly than the (111) Ga face. In addition, surface evaporation below 770 °C occurs more rapidly at the
(111) As face.

7.2: Structures of Element and Compound Semiconductors is shared under a CC BY 4.0 license and was authored, remixed, and/or curated by
Pavan M. V. Raja & Andrew R. Barron via source content that was edited to conform to the style and standards of the LibreTexts platform; a
detailed edit history is available upon request.

7.2.8 https://chem.libretexts.org/@go/page/55905
7.3: X-ray Crystallography
An Introduction to X-ray Diffraction
History of X-ray Crystallography
The birth of X-ray crystallography is considered by many to be marked by the formulation of the law of constant angles by
Nicolaus Steno in 1669 (Figure 7.3.1).
Although Steno is well known for his numerous principles regarding all areas of life, this particular law dealing with geometric
shapes and crystal lattices is familiar ground to all chemists. It simply states that the angles between corresponding faces on
crystals are the same for all specimens of the same mineral. The significance of this for chemistry is that given this fact, crystalline
solids will be easily identifiable once a database has been established. Much like solving a puzzle, crystal structures of
heterogeneous compounds could be solved very methodically by comparison of chemical composition and their interactions.

Figure 7.3.1 Danish pioneer in both anatomy and geology Nicolas Steno (1638 – 1686).
Although Steno was given credit for the notion of crystallography, the man that provided the tools necessary to bring
crystallography into the scientific arena was Wilhelm Roentgen (Figure 7.3.2), who in 1895 successfully pioneered a new form of
photography, one that could allegedly penetrate through paper, wood, and human flesh; due to a lack of knowledge of the specific
workings of this new discovery, the scientific community conveniently labeled the new particles X-rays. This event set off a chain
reaction of experiments and studies, not all performed by physicists. Within one single month, medical doctors were using X-rays
to pinpoint foreign objects such in the human body such as bullets and kidney stones (Figure 7.3.3).

Figure 7.3.2 German physicist Wilhelm Conrad Röentgen (1845 – 1923).

7.3.1 https://chem.libretexts.org/@go/page/55906
Figure 7.3.3 First public X-ray image ever produced. Pictured is the left hand of Anna Berthe Röentgen. The uncharacteristic bulge
is her ring.
The credit for the actual discovery of X-ray diffraction goes to Max von Laue (Figure 7.3.4, to whom the Nobel Prize in physics in
1914 was awarded for the discovery of the diffraction of X-rays. Legend has it that the notion that eventually led to a Nobel prize
was born in a garden in Munich, while von Laue was pondering the problem of passing waves of electromagnetic radiation through
a specific crystalline arrangement of atoms. Because of the relatively large wavelength of visible light, von Laue was forced to turn
his attention to another part of the electromagnetic spectrum, to where shorter wavelengths resided. Only a few decades earlier,
Röentgen had publicly announced the discovery of X-rays, which supposedly had a wavelength shorter than that of visible light.
Having this information, von Laue entrusted the task of performing the experimental work to two technicians, Walter Friedrich and
Paul Knipping. The setup consisted of an X-ray source, which beamed radiation directly into a copper sulfate crystal housed in a
lead box. Film was lined against the sides and back of the box, so as to capture the X-ray beam and its diffraction pattern.
Development of the film showed a dark circle in the center of the film, surrounded by several extremely well defined circles, which
had formed as a result of the diffraction of the X-ray beam by the ordered geometric arrangement of copper sulfate. Max von Laue
then proceeded to work out the mathematical formulas involved in the observed diffraction pattern, for which he was awarded the
Nobel Prize in physics in 1914.

Figure 7.3.4 German physicist Max Theodor Felix von Laue (1879 – 1960) won the Nobel Prize for discovery of the diffraction of
X-rays by crystals.

Principles of X-Ray Diffraction (XRD)


The simplest definition of diffraction is the irregularities caused when waves encounter an object. Diffraction is a phenomenon that
exists commonly in everyday activities, but is often disregarded and taken for granted. For example, when looking at the

7.3.2 https://chem.libretexts.org/@go/page/55906
information side of a compact disc, a rainbow pattern will often appear when it catches light at a certain angle. This is caused by
visible light striking the grooves of the disc, thus producing a rainbow effect (Figure 7.3.5), as interpreted by the observers' eyes.
Another example is the formation of seemingly concentric rings around an astronomical object of significant luminosity when
observed through clouds. The particles that make up the clouds diffract light from the astronomical object around its edges, causing
the illusion of rings of light around the source. It is easy to forget that diffraction is a phenomenon that applies to all forms of
waves, not just electromagnetic radiation. Due to the large variety of possible types of diffractions, many terms have been coined to
differentiate between specific types. The most prevalent type of diffraction to X-ray crystallography is known as Bragg diffraction,
which is defined as the scattering of waves from a crystalline structure.

Figure 7.3.5 The rainbow effects caused by visible light striking the grooves of a compact disc (CD).
Formulated by William Lawrence Bragg (Figure 7.3.6), the equation of Bragg's law relates wavelength to angle of incidence and
lattice spacing, 7.3.1, where n is a numeric constant known as the order of the diffracted beam, λ is the wavelength of the beam, d
denotes the distance between lattice planes, and θ represents the angle of the diffracted wave. The conditions given by this equation
must be fulfilled if diffraction is to occur.
nλ  =  2d sin(θ) (7.3.1)

Figure 7.3.6 Australian-born British physicist Sir William Lawrence Bragg (1890 – 1971).
Because of the nature of diffraction, waves will experience either constructive (Figure 7.3.7) or destructive (Figure 7.3.8)
interference with other waves. In the same way, when an X-ray beam is diffracted off a crystal, the different parts of the diffracted
beam will have seemingly stronger energy, while other parts will have seemed to lost energy. This is dependent mostly on the
wavelength of the incident beam, and the spacing between crystal lattices of the sample. Information about the lattice structure is
obtained by varying beam wavelengths, incident angles, and crystal orientation. Much like solving a puzzle, a three dimensional
structure of the crystalline solid can be constructed by observing changes in data with variation of the aforementioned variables.

Figure 7.3.7 Schematic representation of constructive interference.

7.3.3 https://chem.libretexts.org/@go/page/55906
Figure 7.3.8 Schematic representation of destructive interference.

The X-ray Diffractometer


At the heart of any XRD machine is the X-ray source. Modern day machines generally rely on copper metal as the element of
choice for producing X-rays, although there are variations among different manufacturers. Because diffraction patterns are recorded
over an extended period of time during sample analysis, it is very important that beam intensity remain constant throughout the
entire analysis, or else faulty data will be procured. In light of this, even before an X-ray beam is generated, current must pass
through a voltage regular, which will guarantee a steady stream of voltage to the X-ray source.
Another crucial component to the analysis of crystalline via X-rays is the detector. When XRD was first developed, film was the
most commonly used method for recognizing diffraction patterns. The most obvious disadvantage to using film is the fact that it
has to replaced every time a new specimen is introduced, making data collection a time consuming process. Furthermore, film can
only be used once, leading to an increase in cost of operating diffraction analysis.
Since the origins of XRD, detection methods have progressed to the point where modern XRD machines are equipped with
semiconductor detectors, which produce pulses proportional to the energy absorbed. With these modern detectors, there are two
general ways in which a diffraction pattern may be obtained. The first is called continuous scan, and it is exactly what the name
implies. The detector is set in a circular motion around the sample, while a beam of X-ray is constantly shot into the sample. Pulses
of energy are plotted with respect to diffraction angle, which ensure all diffracted X-rays are recorded. The second and more widely
used method is known as step scan. Step scanning bears similarity to continuous scan, except it is highly computerized and much
more efficient. Instead of moving the detector in a circle around the entire sample, step scanning involves collecting data at one
fixed angle at a time, thus the name. Within these detection parameters, the types of detectors can themselves be varied. A more
common type of detector, known as the charge-coupled device (CCD) detector (Figure 7.3.9, can be found in many XRD
machines, due to its fast data collection capability. A CCD detector is comprised of numerous radiation sensitive grids, each linked
to sensors that measure changes in electromagnetic radiation. Another commonly seen type of detector is a simple scintillation
counter (Figure 7.3.10), which counts the intensity of X-rays that it encounters as it moves along a rotation axis. A comparable
analogy to the differences between the two detectors mentioned would be that the CCD detector is able to see in two dimensions,
while scintillation counters are only able to see in one dimension.

Figure 7.3.9 Single crystal X-ray diffractometer with a CCD detector. The incident beam is generated and delivered through the
silver apparatus on the right side of the sample, and the detector is the large black camera to the left of the sample.

7.3.4 https://chem.libretexts.org/@go/page/55906
Figure 7.3.10 Image of a powder X-ray diffractometer. The incident beam enters from the tube on the left, and the detector is
housed in the black box on the right side of the machine. This particular XRD machine is capable of handling six samples at once,
and is fully automated from sample to sample.
Aside from the above two components, there are many other variables involved in sample analysis by an XRD machine. As
mentioned earlier, a steady incident beam is extremely important for good data collection. To further ensure this, there will often be
what is known as a Söller slit or collimator found in many XRD machines. A Söller slit collimates the direction of the X-ray beam.
In the collimated X-ray beam the rays are parallel, and therefore will spread minimally as they propagates (Figure 7.3.11. Without
a collimator X-rays from all directions will be recorded; for example, a ray that has passed through the top of the specimen (see the
red arrow in Figure 7.3.11a) but happens to be traveling in a downwards direction may be recorded at the bottom of the plate. The
resultant image will be so blurred and indistinct as to be useless. Some machines have a Söller slit between the sample and the
detector, which drastically reduces the amount of background noise, especially when analyzing iron samples with a copper X-ray
source.

Figure 7.3.11 How a Söller collimator filters a stream of rays. (a) without a collimator and (b) with a collimator.
This single crystal XRD machine (Figure 7.3.12) features a cooling gas line, which allows the user to bring down the temperature
of a sample considerably below room temperature. Doing so allows for the opportunities for studies performed where the sample is
kept in a state of extremely low energy, negating a lot of vibrational motion that might interfere with consistent data collection of
diffraction patterns. Furthermore, information can be collected on the effects of temperature on a crystal structure. Also seen in
Figure 7.3.13 is the hook-shaped object located between the beam emitter and detector. It serves the purpose of blocking X-rays
that were not diffracted from being seen by the detector, drastically reducing the amount of unnecessary noise that would otherwise
obscure data analysis.

Evolution of Powder XRD


Over time, XRD analysis has evolved from a very narrow and specific field to something that encompasses a much wider branch of
the scientific arena. In its early stages, XRD was (with the exception of the simplest structures) confined to single crystal analysis,

7.3.5 https://chem.libretexts.org/@go/page/55906
as detection methods had not advanced to a point where more complicated procedures was able to be performed. After many years
of discovery and refining, however, technology has progressed to where crystalline properties (structure) of solids can be gleaned
directly from a powder sample, thus offering information for samples that cannot be obtained as a single crystal. One area in which
this is particularly useful is pharmaceuticals, since many of the compounds studied are not available in single crystal form, only in
a powder.
Even though single crystal diffraction and powder diffraction essentially generate the same data, due to the powdered nature of the
latter sample, diffraction lines will often overlap and interfere with data collection. This is apparently especially when the
diffraction angle 2θ is high; patterns that emerge will be almost to the point of unidentifiable, because of disruption of individual
diffraction patterns. For this particular reason, a new approach to interpreting powder diffraction data has been created.
There are two main methods for interpreting diffraction data:
The first is known as the traditional method, which is very straightforward, and bears resemblance to single crystal data
analysis. This method involves a two step process: 1) the intensities and diffraction patterns from the sample is collected, and 2)
the data is analyzed to produce a crystalline structure. As mentioned before, however, data from a powdered sample is often
obscured by multiple diffraction patterns, which decreases the chance that the generated structure is correct.
The second method is called the direct-space approach. This method takes advantage of the fact that with current technology,
diffraction data can be calculated for any molecule, whether or not it is the molecule in question. Even before the actual
diffraction data is collected, a large number of theoretical patterns of suspect molecules are generated by computer, and
compared to experimental data. Based on correlation and how well the theoretical pattern fits the experimental data best, a
guess is formulated to which compound is under question. This method has been taken a step further to mimic social
interactions in a community. For example, first generation theoretical trial molecules, after comparison with the experimental
data, are allowed to evolve within parameters set by researchers. Furthermore, if appropriate, molecules are produce offspring
with other molecules, giving rise to a second generation of molecules, which fit the experimental data even better. Just like a
natural environment, genetic mutations and natural selection are all introduced into the picture, ultimately giving rise a
molecular structure that represents data collected from XRD analysis.
Another important aspect of being able to study compounds in powder form for the pharmaceutical researcher is the ability to
identify structures in their natural state. A vast majority of drugs in this day and age are delivered through powdered form, either in
the form of a pill or a capsule. Crystallization processes may often alter the chemical composition of the molecule (e.g., by the
inclusion of solvent molecules), and thus marring the data if confined to single crystal analysis. Furthermore, when the sample is in
powdered form, there are other variables that can be adjusted to see real-time effects on the molecule. Temperature, pressure, and
humidity are all factors that can be changed in-situ to glean data on how a drug might respond to changes in those particular
variables.

Powder X-Ray Diffraction


Introduction
Powder X-Ray diffraction (XRD) was developed in 1916 by Debye (Figure 7.3.12) and Scherrer (Figure 7.3.13) as a technique
that could be applied where traditional single-crystal diffraction cannot be performed. This includes cases where the sample cannot
be prepared as a single crystal of sufficient size and quality. Powder samples are easier to prepare, and is especially useful for
pharmaceuticals research.

7.3.6 https://chem.libretexts.org/@go/page/55906
Figure 7.3.12 Dutch physicist and physical chemist Peter Joseph William Debye (1884-1966) recipient of the Nobel Prize in
Chemistry.

Figure 7.3.13 Swiss physicist Paul Scherrer (1890-1969).


Diffraction occurs when a wave meets a set of regularly spaced scattering objects, and its wavelength of the distance between the
scattering objects are of the same order of magnitude. This makes X-rays suitable for crystallography, as its wavelength and crystal
lattice parameters are both in the scale of angstroms (Å). Crystal diffraction can be described by Bragg diffraction, 7.3.2, where λ
is the wavelength of the incident monochromatic X-ray, d is the distance between parallel crystal planes, and θ the angle between
the beam and the plane.
λ  =  2d sinθ (7.3.2)

For constructive interference to occur between two waves, the path length difference between the waves must be an integral
multiple of their wavelength. This path length difference is represented by 2d sinθ Figure 7.3.14. Because sinθ cannot be greater
than 1, the wavelength of the X-ray limits the number of diffraction peaks that can appear.

Figure 7.3.14 Bragg diffraction in a crystal. The angles at which diffraction occurs is a function of the distance between planes and
the X-ray wavelength.

Production and Detection of X-rays


Most diffractometers use Cu or Mo as an X-ray source, and specifically the Kα radiation of wavelengths of 1.54059 Å and 0.70932
Å, respectively. A stream of electrons is accelerated towards the metal target anode from a tungsten cathode, with a potential

7.3.7 https://chem.libretexts.org/@go/page/55906
difference of about 30-50 kV. As this generates a lot of heat, the target anode must be cooled to prevent melting.
Detection of the diffracted beam can be done in many ways, and one common system is the gas proportional counter (GPC). The
detector is filled with an inert gas such as argon, and electron-ion pairs are created when X-rays pass through it. An applied
potential difference separates the pairs and generates secondary ionizations through an avalanche effect. The amplification of the
signal is necessary as the intensity of the diffracted beam is very low compared to the incident beam. The current detected is then
proportional to the intensity of the diffracted beam. A GPC has a very low noise background, which makes it widely used in labs.

Performing X-ray Diffraction


Exposure to X-rays may have health consequences, follow safety procedures when using the diffractometer.
The particle size distribution should be even to ensure that the diffraction pattern is not dominated by a few large particles near the
surface. This can be done by grinding the sample to reduce the average particle size to <10µm. However, if particle sizes are too
small, this can lead to broadening of peaks. This is due to both lattice damage and the reduction of the number of planes that cause
destructive interference.
The diffraction pattern is actually made up of angles that did not suffer from destructive interference due to their special
relationship described by Bragg Law (Figure 7.3.15). If destructive interference is reduced close to these special angles, the peak is
broadened and becomes less distinct. Some crystals such as calcite (CaCO3, Figure 7.3.15 have preferred orientations and will
change their orientation when pressure is applied. This leads to differences in the diffraction pattern of ‘loose’ and pressed samples.
Thus, it is important to avoid even touching ‘loose’ powders to prevent errors when collecting data.

Figure 7.3.15 Calcite crystal structure. Under compression, the c axis orientates subparallel to the direction of pressure.
The sample powder is loaded onto a sample dish for mounting in the diffractometer (Figure 7.3.16), where rotating arms containing
the X-ray source and detector scan the sample at different incident angles. The sample dish is rotated horizontally during scanning
to ensure that the powder is exposed evenly to the X-rays.

Figure 7.3.16 A powder X-ray diffractometer. Two arms containing the X-ray source and detector are positioned around sample
dishes, where the angle between each arm and the plane of the sample dishes is θ.
A sample X-ray diffraction spectrum of germanium is shown in Figure 7.3.17, with peaks identified by the planes that caused that
diffraction. Germanium has a diamond cubic crystal lattice (Figure 7.3.18), named after the crystal structure of prototypical
example. The crystal structure determines what crystal planes cause diffraction and the angles at which they occur. The angles are

7.3.8 https://chem.libretexts.org/@go/page/55906
shown in 2θ as that is the angle measured between the two arms of the diffractometer, i.e., the angle between the incident and the
diffracted beam (Figure 7.3.14).

Figure 7.3.17 Powder XRD spectrum of germanium. Reprinted with permission from H. W. Chiu, C. N. Chervin, and S. M.
Kauzlarich, Chem. Mater., 2005, 17, 4858. Copyright 2013 American Chemical Society.

Figure 7.3.18 Model of diamond cubic crystal lattice.

Determining Crystal Structure for Cubic Lattices


There are three basic cubic crystal lattices, and they are the simple cubic (SC), body-centered cubic (BCC), and the face-centered
cubic (FCC) Figure 7.3.19. These structures are simple enough to have their diffraction spectra analyzed without the aid of
software.

Figure 7.3.19 Models of cubic crystal structures.


Each of these structures has specific rules on which of their planes can produce diffraction, based on their Miller indices (hkl).
SC lattices show diffraction for all values of (hkl), e.g., (100), (110), (111), etc.
BCC lattices show diffraction when the sum of h+k+l is even, e.g., (110), (200), (211), etc.
FCC lattices show diffraction when the values of (hkl) are either all even or all odd, e.g., (111), (200), (220), etc.
Diamond cubic lattices like that of germanium are FCC structures with four additional atoms in the opposite corners of the
tetrahedral interstices. They show diffraction when the values of (hkl) are all odd or all even and the sum h+k+l is a multiple of
4, e.g., (111), (220), (311), etc.
The order in which these peaks appear depends on the sum of h2+k2+l2. These are shown in Table 7.3.1.
Table 7.3.1 Diffraction planes and their corresponding h2+k2+l2 values. The planes which result in diffraction for BCC and FCC structures
are marked with a “Y”.
(hkl) h2+k2+l2 BCC FCC

100 1

110 2 Y

111 3 Y

200 4 Y Y

7.3.9 https://chem.libretexts.org/@go/page/55906
210 5

211 6 Y

220 8 Y Y

300, 221 9

310 10 Y

311 11 Y

222 12 Y Y

320 13

321 14 Y

400 16 Y Y

410, 322 17

411, 330 18 Y

331 19 Y

420 20 Y Y

421 21

The value of d for each of these planes can be calculated using 7.3.3, where a is the lattice parameter of the crystal.
The lattice constant, or lattice parameter, refers to the constant distance between unit cells in a crystal lattice.
2 2 2
1 h +k +l
  =  (7.3.3)
2 2
d a

As the diamond cubic structure of Ge can be complicated, a simpler worked example for sample diffraction of NaCl with Cu-Kα
radiation is shown below. Given the values of 2θ that result in diffraction, Table 7.3.2 can be constructed.
Table 7.3.2 Ratio of diffraction for germanium.
2θ θ Sinθ Sin2θ

27.36 13.68 0.24 0.0559

31.69 15.85 0.27 0.0746

45.43 22.72 0.39 0.1491

53.85 26.92 0.45 0.2050

56.45 28.23 0.47 0.2237

66.20 33.10 0.55 0.2982

73.04 36.52 0.60 0.3541

75.26 37.63 0.61 0.3728

The values of these ratios can then be inspected to see if they corresponding to an expected series of hkl values. In this case, the last
column gives a list of integers, which corresponds to the h2+k2+l2 values of the FCC lattice diffraction. Hence, NaCl has a FCC
structure, shown in angles Figure 7.3.20.

7.3.10 https://chem.libretexts.org/@go/page/55906
Figure 7.3.20 Model of NaCl FCC lattice.
The lattice parameter of NaCl can now be calculated from this data. The first peak occurs at θ = 13.68°. Given that the wavelength
of the Cu-Kα radiation is 1.54059 Å, Bragg's Equation 7.3.4 can be applied as follows:
1.54059  =  2d sin13.68 (7.3.4)

d  =  3.2571  Å (7.3.5)

Since the first peak corresponds to the (111) plane, the distance between two parallel (111) planes is 3.2571 Å. The lattice
parameter can now be worked out using 7.3.6.
2 2 2 2 2
1/ 3.2561   =  (1 +1 +I )/ a (7.3.6)

a  =  5.6414  Å (7.3.7)

The powder XRD spectrum of Ag nanoparticles is given in Figure 7.3.21 as collected using Cu-Kα radiation of 1.54059 Å.
Determine its crystal structure and lattice parameter using the labeled peaks.

Figure 7.3.21 Powder XRD spectra of silver nanoparticles. Adapted from E. C. Njagi, H. Huang, L. Stafford, H. Genuino, H. M.
Galindo, J. B. Collins, G. E. Hoag, and S. L. Suib, Langmuir, 2011, 27, 264. Copyright 2013 American Chemical Society.
Table 7.3.3 Ratio of diffraction angles for Ag.
2θ θ Sinθ Sin2θ Sin2θ/Sin2θ 2 x Sin2θ/Sin2θ 3 x Sin2θ/Sin2θ

38.06 19.03 0.33 0.1063 1.00 2.00 3.00

44.24 22.12 0.38 0.1418 1.33 2.67 4.00

64.35 32.17 0.53 0.2835 2.67 5.33 8

77.28 38.64 0.62 0.3899 3.67 7.34 11

81.41 40.71 0.65 0.4253 4 8 12

97.71 48.86 0.75 0.5671 5.33 10.67 16

110.29 55.15 0.82 0.6734 6.34 12.67 19.01

114.69 57.35 0.84 0.7089 6.67 13.34 20.01

Applying the Bragg Equation 7.3.8,


1.54059  =  2d sin 19.03 (7.3.8)

d  =  2.3624  Å (7.3.9)

7.3.11 https://chem.libretexts.org/@go/page/55906
Calculate the lattice parameter using 7.3.10,
2 2 2 2 2
1/ 2.3624   =  (1 +1 +I )/ a (7.3.10)

a  =  4.0918  Å (7.3.11)

The last column gives a list of integers, which corresponds to the h2+k2+l2 values of the FCC lattice diffraction. Hence, the Ag
nanoparticles have a FCC structure.

Determining Composition
As seen above, each crystal will give a pattern of diffraction peaks based on its lattice type and parameter. These fingerprint
patterns are compiled into databases such as the one by the Joint Committee on Powder Diffraction Standard (JCPDS). Thus, the
XRD spectra of samples can be matched with those stored in the database to determine its composition easily and rapidly.

Solid State Reaction Monitoring


Powder XRD is also able to perform analysis on solid state reactions such as the titanium dioxide (TiO2) anatase to rutile transition.
A diffractometer equipped with a sample chamber that can be heated can take diffractograms at different temperatures to see how
the reaction progresses. Spectra of the change in diffraction peaks during this transition is shown in Figure 7.3.22, Figure 7.3.23,
and Figure 7.3.24.

Figure 7.3.22 Powder XRD spectra of anatase TiO2 at 25 °C. Courtesy of Jeremy Lee.

Figure 7.3.23 Powder XRD spectra of anatase and rutile TiO2 at 750 °C, with labelled peaks for each phase. Courtesy of Jeremy
Lee.

7.3.12 https://chem.libretexts.org/@go/page/55906
Figure 7.3.24 Powder XRD spectra of rutile TiO2 at 1000 °C. Courtesy of Jeremy Lee.

Summary
XRD allows for quick composition determination of unknown samples and gives information on crystal structure. Powder XRD is
a useful application of X-ray diffraction, due to the ease of sample preparation compared to single-crystal diffraction. Its
application to solid state reaction monitoring can also provide information on phase stability and transformation.

An Introduction to Single-Crystal X-Ray Crystallography


Described simply, single-crystal X-ray diffraction (XRD) is a technique in which a crystal of a sample under study is bombarded
with an X-ray beam from many different angles, and the resulting diffraction patterns are measured and recorded. By aggregating
the diffraction patterns and converting them via Fourier transform to an electron density map, a unit cell can be constructed which
indicates the average atomic positions, bond lengths, and relative orientations of the molecules within the crystal.

Fundamental Principles
As an analogy to describe the underlying principles of diffraction, imagine shining a laser onto a wall through a fine sieve. Instead
of observing a single dot of light on the wall, a diffraction pattern will be observed, consisting of regularly arranged spots of light,
each with a definite position and intensity. The spacing of these spots is inversely related to the grating in the sieve— the finer the
sieve, the farther apart the spots are, and the coarser the sieve, the closer together the spots are. Individual objects can also diffract
radiation if it is of the appropriate wavelength, but a diffraction pattern is usually not seen because its intensity is too weak. The
difference with a sieve is that it consists of a grid made of regularly spaced, repeating wires. This periodicity greatly magnifies the
diffraction effect because of constructive interference. As the light rays combine amplitudes, the resulting intensity of light seen on
the wall is much greater because intensity is proportional to the square of the light’s amplitude.
To apply this analogy to single-crystal XRD, we must simply scale it down. Now the sieve is replaced by a crystal and the laser
(visible light) is replaced by an X-ray beam. Although the crystal appears solid and not grid-like, the molecules or atoms contained
within the crystal are arranged periodically, thus producing the same intensity-magnifying effect as with the sieve. Because X-rays
have wavelengths that are on the same scale as the distance between atoms, they can be diffracted by their interactions with the
crystal lattice.
These interactions are dictated by Bragg's law, which says that constructive interference occurs only when 7.3.12 is satisfied;
where n is an integer, λ is the wavelength of light, d is the distance between parallel planes in the crystal lattice, and θ is the angle
of incidence between the X-ray beam and the diffracting planes (see Figure 7.3.25). A complication arises, however, because
crystals are periodic in all three dimensions, while the sieve repeats in only two dimensions. As a result, crystals have many
different diffraction planes extending in certain orientations based on the crystal’s symmetry group. For this reason, it is necessary
to observe diffraction patterns from many different angles and orientations of the crystal to obtain a complete picture of the
reciprocal lattice.
The reciprocal lattice of a lattice (Bravais lattice) is the lattice in which the Fourier transform of the spatial wavefunction of the
original lattice (or direct lattice) is represented. The reciprocal lattice of a reciprocal lattice is the original lattice.

nλ  =  2d sinθ (7.3.12)

7.3.13 https://chem.libretexts.org/@go/page/55906
Figure 7.3.25 Illustration of the requirements of Bragg’s law, 7.3.12 .
The reciprocal lattice is related to the crystal lattice just as the sieve is related to the diffraction pattern: they are inverses of each
other. Each point in real space has a corresponding point in reciprocal space and they are related by 1/d; that is, any vector in real
space multiplied by its corresponding vector in reciprocal space gives a product of unity. The angles between corresponding pairs
of vectors remains unchanged.
Real space is the domain of the physical crystal, i.e. it includes the crystal lattice formed by the physical atoms within the crystal.
Reciprocal space is, simply put, the Fourier transform of real space; practically, we see that diffraction patterns resulting from
different orientations of the sample crystal in the X-ray beam are actually two-dimensional projections of the reciprocal lattice.
Thus by collecting diffraction patterns from all orientations of the crystal, it is possible to construct a three-dimensional version of
the reciprocal lattice and then perform a Fourier transform to model the real crystal lattice.

Technique
Single-crystal Versus Powder Diffraction
Two common types of X-ray diffraction are powder XRD and single-crystal XRD, both of which have particular benefits and
limitations. While powder XRD has a much simpler sample preparation, it can be difficult to obtain structural data from a powder
because the sample molecules are randomly oriented in space; without the periodicity of a crystal lattice, the signal-to-noise ratio is
greatly decreased and it becomes difficult to separate reflections coming from the different orientations of the molecule. The
advantage of powder XRD is that it can be used to quickly and accurately identify a known substance, or to verify that two
unknown samples are the same material.
Single-crystal XRD is much more time and data intensive, but in many fields it is essential for structural determination of small
molecules and macromolecules in the solid state. Because of the periodicity inherent in crystals, small signals from individual
reflections are magnified via constructive interference. This can be used to determine exact spatial positions of atoms in molecules
and can yield bond distances and conformational information. The difficulty of single-crystal XRD is that single crystals may be
hard to obtain, and the instrument itself may be cost-prohibitive.
An example of typical diffraction patterns for single-crystal and powder XRD follows ((Figure 7.3.27 and Figure 7.3.28,
respectively). The dots in the first image correspond to Bragg reflections and together form a single view of the molecule’s
reciprocal space. In powder XRD, random orientation of the crystals means reflections from all of them are seen at once, producing
the observed diffraction rings that correspond to particular vectors in the material’s reciprocal lattice.

7.3.14 https://chem.libretexts.org/@go/page/55906
Figure 7.3.26 Single-crystal diffraction pattern of an enzyme. The white rod protruding from the top is the beamstop. Copyright
Jeff Dahl (2006); used under a creative license.

Figure 7.3.27 Powder X-ray diffraction spectrum of silicon. Taken by XanaG; used under PD license.

Technique
In a single-crystal X-ray diffraction experiment, the reciprocal space of a crystal is constructed by measuring the angles and
intensities of reflections in observed diffraction patterns. These data are then used to create an electron density map of the molecule
which can be refined to determine the average bond lengths and positions of atoms in the crystal.

Instrumentation
The basic setup for single-crystal XRD consist of an X-ray source, a collimator to focus the beam, a goniometer to hold and rotate
the crystal, and a detector to measure and record the reflections. Instruments typically contain a beamstop to halt the primary X-ray
beam from hitting the detector, and a camera to help with positioning the crystal. Many also contain an outlet connected to a cold
gas supply (such as liquid nitrogen) in order to cool the sample crystal and reduce its vibrational motion as data is being collected.
A typical instrument is shown in Figure 7.3.28 and Figure 7.3.31.

7.3.15 https://chem.libretexts.org/@go/page/55906
Figure 7.3.28 Modern single-crystal X-ray diffraction machine; the X-ray source can be seen at the right edge as the gray box that
extends into the background. Note that the goniometer that holds the crystal in place is not shown.

Figure 7.3.29 Close-up view of a single-crystal X-ray diffraction instrument. The large black circle at the left is the detector, and
the X-ray beam comes out of the pointed horizontal nozzle. The beam stop can be seen across from this nozzle, as well as the gas
cooling tube hanging vertically. The mounted crystal rests below the cooling gas supply, directly in the path of the beam. It extends
from a glass fiber on a base (not shown) that attaches to the goniometer. The camera can also be seen as the black tube on the right
side of the photograph.

Obtaining Single Crystals


Despite advances in instrumentation and computer programs that make data collection and solving crystal structures significantly
faster and easier, it can still be a challenge to obtain crystals suitable for analysis. Ideal crystals are single, not twinned, clear, and
of sufficient size to be mounted within the the X-ray beam (usually 0.1-0.3 mm in each direction). They also have clean faces and
smooth edges. Following are images of some ideal crystals (Figure 7.3.30 and Figure 7.3.31), as well as an example of twinned
crystals (Figure 7.3.32).

7.3.16 https://chem.libretexts.org/@go/page/55906
Crystal twinning occurs when two or more crystals share lattice points in a symmetrical manner. This usually results in complex
diffraction patterns which are difficult to analyze and construct a reciprocal lattice.

Figure 7.3.30 Single crystals of insulin, grown in space; taken by NASA. Released under PD license.

Figure 7.3.31 An octahedral-shaped single crystal of synthetic chrome alum. Copyright Ra’ike (2008); used under a creative
license.

Figure 7.3.32 Twinned quartz crystal. Image used under fair use license from the Geology Guide of the Smithsonian National
Museum of Natural History.
Crystal formation can be affected by temperature, pressure, solvent choice, saturation, nucleation, and substrate. Slow crystal
growth tends to be best, as rapid growth creates more imperfections in the crystal lattice and may even lead to a precipitate or gel.
Similarly, too many nucleation sites (points at which crystal growth begins) can lead to many small crystals instead of a few, well-
defined ones.
There are a number of basic methods for growing crystals suitable for single-crystal XRD:
The most basic method is to slowly evaporate a saturated solution until it becomes supersaturated and then forms crystals. This
often works well for growing small-molecule crystals; macroscopic molecules (such as proteins) tend to be more difficult.
A solution of the compound to be crystallized is dissolved in one solvent, then a ‘non-solvent’ which is miscible with the first
but in which the compound itself is insoluble, is carefully layered on top of the solution. As the non-solvent mixes with the
solvent by diffusion, the solute molecules are forced out of solution and may form crystals.
A crystal solution is placed in a small open container which is then set in a larger closed container holding a volatile non-
solvent. As the volatile non-solvent mixes slowly with the solution by vapor diffusion, the solute is again forced to come out of
solution, often leading to crystal growth.
All three of the previous techniques can be combined with seeding, where a crystal of the desired type to be grown is placed in
the saturated solution and acts as a nucleation site and starting place for the crystal growth to begin. In some cases, this can even
cause crystals to grow in a form that they would not normally assume, as the seed can act as a template that might not otherwise
be followed.

7.3.17 https://chem.libretexts.org/@go/page/55906
The hanging drop technique is typically used for growing protein crystals. In this technique, a drop of concentrated protein
solution is suspended (usually by dotting it on a silicon-coated microscope slide) over a larger volume of the solution. The
whole system is then sealed and slow evaporation of the suspended drop causes it to become supersaturated and form crystals.
(A variation of this is to have the drop of protein solution resting on a platform inside the closed system instead of being
suspended from the top of the container.)
These are only the most common ways that crystals are grown. Particularly for macromolecules, it may be necessary to test
hundreds of crystallization conditions before a suitable crystal is obtained. There now exist automated techniques utilizing robots to
grow crystals, both for obtaining large numbers of single crystals and for performing specialized techniques (such as drawing a
crystal out of solution) that would otherwise be too time-consuming to be of practical use.

Wide Angle X-ray Diffraction Studies of Liquid Crystals


Some organic molecules display a series of intermediate transition states between solid and isotropic liquid states (Figure 7.3.33) as
their temperature is raised. These intermediate phases have properties in between the crystalline solid and the corresponding
isotropic liquid state, and hence they are called liquid crystalline phases. Other name is mesomorphic phases where mesomorphic
means of intermediate form. According to the physicist de Gennes (Figure 7.3.34), liquid crystal is ‘an intermediate phase, which
has liquid like order in at least one direction and possesses a degree of anisotropy’. It should be noted that all liquid crystalline
phases are formed by anisotropic molecules (either elongated or disk-like) but not all the anisotropic molecules form liquid
crystalline phases.

Figure 7.3.33 Schematic phase behavior for a molecule that displays an liquid crystal (LC) phase. TCN and TNI represents phase
transition temperatures from crystalline solid to LC phase and LC to isotropic liquid phase, respectively.

Figure 7.3.34 French physicist and the Nobel Prize laureate Pierre-Gilles de Gennes (1932-2007).
Anisotropic objects can possess different types of ordering giving rise to different types of liquid crystalline phases (Figure 7.3.35).

Figure 7.3.35 Schematic illustration of the different types of liquid crystal phases.

7.3.18 https://chem.libretexts.org/@go/page/55906
Nematic Phases
The word nematic comes from the Greek for thread, and refers to the thread-like defects commonly observed in the polarizing
optical microscopy of these molecules. They have no positional order only orientational order, i.e., the molecules all pint in the
same direction. The direction of molecules denoted by the symbol n commonly referred as the ‘director’ (Figure 7.3.36). The
director n is bidirectional that means the states n and -n are indistinguishable.

Smetic Phases
All the smectic phases are layered structures that usually occur at slightly lower temperatures than nematic phases. There are many
variations of smectic phases, and some of the distinct ones are as follows:
Each layer in smectic A is like a two dimensional liquid, and the long axis of the molecules is typically orthogonal to the layers
(Figure 7.3.35.
Just like nematics, the state n and -n are equivalent. They are made up of achiral and non polar molecules.
As with smectic A, the smectic C phase is layered, but the long axis of the molecules is not along the layers normal. Instead it
makes an angle (θ, Figure 7.3.35). The tilt angle is an order parameter of this phase and can vary from 0° to 45-50°.
Smectic C* phases are smectic phases formed by chiral molecules. This added constraint of chirality causes a slight distortion
of the Smectic C structure. Now the tilt direction precesses around the layer normal and forms a helical configuration.

Cholesterics Phases
Sometimes cholesteric phases (Figure 7.3.35) are also referred to as chiral nematic phases because they are similar to nematic
phases in many regards. Many derivatives of cholesterol exhibit this type of phase. They are generally formed by chiral molecules
or by doping the nematic host matrix with chiral molecules. Adding chirality causes helical distortion in the system, which makes
the director, n, rotate continuously in space in the shape of a helix with specific pitch. The magnitude of pitch in a cholesteric phase
is a strong function of temperature.

Columnar Phases
In columnar phases liquid crystals molecules are shaped like disks as opposed to rod-like in nematic and smectics liquid crystal
phases. These disk shaped molecules stack themselves in columns and form a 2D crystalline array structures (Figure 7.3.35). This
type of two dimensional ordering leads to new mesophases.

Introduction to 2D X-ray Diffraction


X-ray diffraction (XRD) is one of the fundamental experimental techniques used to analyze the atomic arrangement of materials.
The basic principle behind X-ray diffraction is Bragg’s Law (Figure 7.3.36). According to this law, X-rays that are reflected from
the adjacent crystal planes will undergo constructive interference only when the path difference between them is an integer multiple
of the X-ray's wavelength, 7.3.13, where n is an integer, d is the spacing between the adjacent crystal planes, θ is the angle between
incident X-ray beam and scattering plane, and λ is the wavelength of incident X-ray.
2dsinθ  =  nλ  (7.3.13)

Figure 7.3.36 Schematic description of Bragg’s Diffraction Law.


Now the atomic arrangement of molecules can go from being extremely ordered (single crystals) to random (liquids).
Correspondingly, the scattered X-rays form specific diffraction patterns particular to that sample. Figure 7.3.37 shows the
difference between X-rays scattered from a single crystal and a polycrystalline (powder) sample. In case of a single crystal the

7.3.19 https://chem.libretexts.org/@go/page/55906
diffracted rays point to discrete directions (Figure 7.3.37a ), while for polycrystalline sample diffracted rays form a series of
diffraction cones (Figure 7.3.37b).

Figure 7.3.37 Diffraction pattern from (a) single crystal and (b) polycrystalline sample
A two dimensional (2D) XRD system is a diffraction system with the capability of simultaneously collecting and analyzing the X-
ray diffraction pattern in two dimensions. A typical 2D XRD setup consists of five major components (Figure 7.3.38):
X-ray source.
X-ray optics.
Goniometer.
Sample alignment and monitoring device.
2D area detector.

Figure 7.3.38 Schematic illustration of basic 2D WAXD setup. Adapted from B. B. He, U. Preckwinkel, and K. L. Smith,
Advances in X-ray Analysis, 2000, 43, 273..
For laboratory scale X-ray generators, X-rays are emitted by bombarding metal targets with high velocity electrons accelerated by
strong electric field in the range 20-60 kV. Different metal targets that can be used are chromium (Cr), cobalt (Co), copper (Cu),
molybdenum (Mo) and iron (Fe). The most commonly used ones are Cu and Mo. Synchrotrons are even higher energy radiation
sources. They can be tuned to generate a specific wavelength and they have much brighter luminosity for better resolution.
Available synchrotron facilities in US are:
Stanford Synchrotron Radiation Lightsource (SSRL), Stanford, CA.
Synchrotron Radiation Center (SRC), University of Wisconsin-Madison, Madison, WI.
Advanced Light Source (ALS), Lawrence Berkeley National, Berkeley, CA.
National Synchrotron Light Source (NSLS), Brookhaven National Laboratory, Upton, NY.
Advanced Photon Source (APS), Argonne National Laboratory, Argonne, IL.
Center for Advanced Microstructures & Devices, Louisiana State University, Baton Rouge, LA.
Cornell High Energy Synchrotron Source (CHESS), Cornell, Ithaca, NY.
The X-ray optics are comprised of the X-ray tube, monochromator, pinhole collimator and beam stop. A monochromator is used to
get rid of unwanted X-ray radiation from the X-ray tube. A diffraction from a single crystal can be used to select a specific
wavelength of radiation. Typical materials used are pyrolytic graphite and silicon. Monochromatic X-ray beams have three
components: parallel, convergent and divergent X-rays. The function of a pinhole collimator is to filter the incident X-ray beam and
allow passage of parallel X-rays. A 2D X-ray detector can either be a film or a digital detector, and its function is to measure the
intensity of X-rays diffracted from a sample as a function of position, time, and energy.

Advantages of 2D XRD as Compared to 1D XRD


2D diffracton data has much more information in comparison diffraction pattern, which is acquired using a 1D detector. Figure
7.3.39 shows the diffraction pattern from a polycrystalline sample. For illustration purposes only, two diffraction cones are shown

7.3.20 https://chem.libretexts.org/@go/page/55906
in the schematic. In the case of 1D X-ray diffraction, measurement area is confined within a plane labeled as diffractometer plane.
The 1D detector is mounted along the detection circle and variation of diffraction pattern in the z direction are not considered. The
diffraction pattern collected is an average over a range defined by a beam size in the z direction. The diffraction pattern measured is
a plot of X-ray intensity at different 2θ angles. For 2D X-ray diffraction, the measurement area is not limited to the diffractometer
plane. Instead, a large portion of the diffraction rings are measured simultaneously depending on the detector size and position from
the sample.

Figure 7.3.39 Diffraction patterns from a powder sample. Adapted from B. B. He, U. Preckwinkel, and K. L. Smith, Advances in
X-ray Analysis, 2000, 43, 273.
One such advantage is the measurement of percent crystallinity of a material. Determination of material crystallinity is required
both for research and quality control. Scattering from amorphous materials produces a diffuse intensity ring while polycrystalline
samples produce sharp and well-defined rings or spots are seen. The ability to distinguish between amorphous and crystalline is the
key in determining percent of crystallinity accurately. Since most crystalline samples have preferred orientation, depending on the
sample is oriented it is possible to measure different peak or no peak using conventional diffraction system. On the other hand,
sample orientation has no effect on the full circle integrated diffraction measuring done using 2D detector. A 2D XRD can therefore
measure percent crystallinity more accurately.

2D Wide Angle X-ray Diffraction Patterns of LCs


As mentioned in the introduction section, liquid crystal is an intermediate state between solid and liquid phases. At temperatures
above the liquid crystal phase transition temperature (Figure 7.3.40), they become isotropic liquid, i.e., absence of long-range
positional or orientational order within molecules. Since an isotropic state cannot be aligned, its diffraction pattern consists of
weak, diffuse rings Figure 7.3.40a. The reason we see any diffraction pattern in the isotropic state is because in classical liquids
there exists a short range positional order. The ring has of radius of 4.5 Å and it mostly appears at 20.5°. It represents the distance
between the molecules along their widths.

Figure 7.3.40 Schematic of 2D X-ray diffraction of different types of liquid crystal phases: (a) isotopic, (b) nematic, (c) smectic A,
and (d) smectic C.
Nematic liquid crystalline phases have long range orientational order but no positional order. An unaligned sample of nematic
liquid crystal has similar diffraction pattern as an isotropic state. But instead of a diffuse ring, it has a sharper intensity distribution.
For an aligned sample of nematic liquid crystal, X-ray diffraction patterns exhibit two sets of diffuse arcs (Figure 7.3.40 b). The
diffuse arc at the larger radius (P1, 4.5 Å) represents the distance between molecules along their widths. Under the presence of an
external magnetic field, samples with positive diamagnetic anisotropy align parallel to the field and P1 is oriented perpendicularly

7.3.21 https://chem.libretexts.org/@go/page/55906
to the field. While samples with negative diamagnetic anisotropy align perpendicularly to the field with P1 being parallel to the
field. The intensity distribution within these arcs represents the extent of alignment within the sample; generally denoted by S.
The diamagnetic anistropy of all liquid crystals with an aromatic ring is positive, and on order of 10-7. The value decreases with the
substitution of each aromatic ring by a cyclohexane or other aliphatic group. A negative diamagnetic anistropy is observed for
purely cycloaliphatic LCs.
When a smectic phase is cooled down slowly under the presence the external field, two sets of diffuse peaks are seen in diffraction
pattern (Figure 7.3.40 c). The diffuse peak at small angles condense into sharp quasi-Bragg peaks. The peak intensity distribution
at large angles is not very sharp because molecules within the smectic planes are randomly arranged. In case of smectic C phases,
the angle between the smectic layers normal and the director (θ) is no longer collinear (Figure 7.3.40 d). This tilt can easily be seen
in the diffraction pattern as the diffuse peaks at smaller and larger angles are no longer orthogonal to each other.

Sample Preparation
In general, X-ray scattering measurements of liquid crystal samples are considered more difficult to perform than those of
crystalline samples. The following steps should be performed for diffraction measurement of liquid crystal samples:
1. The sample should be free of any solvents and absorbed oxygen, because their presence affects the liquid crystalline character
of the sample and its thermal response. This can be achieved by performing multiple melting and freezing cycles in a vacuum to
get rid of unwanted solvents and gases.
2. For performing low resolution measurements, liquid crystal sample can be placed inside a thin-walled glass capillary. The ends
of the capillary can be sealed by epoxy in case of volatile samples. The filling process tends to align the liquid crystal molecules
along the flow direction.
3. For high resolution measurements, the sample is generally confined between two rubbed polymer coated glass coverslips
separated by an o-ring as a spacer. The rubbing causes formation of grooves in the polymer film which tends to the align the
liquid crystal molecules.
4. Aligned samples are necessary for identifying the liquid crystalline phase of the sample. Liquid crystal samples can be aligned
by heating above the phase transition temperature and cooling them slowly in the presence of an external electric or magnetic
field. A magnetic field is effective for samples with aromatic cores as they have high diamagnetic anisotropy. A common
problem in using electric field is internal heating which can interfere with the measurement.
5. Sample size should be sufficient to avoid any obstruction to the passage of the incident X-ray beam.
6. The sample thickness should be around one absorption length of the X-rays. This allows about 63% of the incident light to pass
through and get optimum scattering intensity. For most hydrocarbons absorption length is approximately 1.5 mm with a copper
metal target (λ = 1.5418 Å). Molybdenum target can be used for getting an even higher energy radiation (λ = 0.71069 Å ).

Data Analysis
Identification of the phase of a liquid crystal sample is critical in predicting its physical properties. A simple 2D X-ray diffraction
pattern can tell a lot in this regard (Figure 7.3.40). It is also critical to determine the orientational order of a liquid crystal. This is
important to characterize the extent of sample alignment.
For simplicity, the rest of the discussion focuses on nematic liquid crystal phases. In an unaligned sample, there isn't any specific
macroscopic order in the system. In the micrometer size domains, molecules are all oriented in a specific direction, called a local
director. Because there is no positional order in nematic liquid crystals, this local director varies in space and assumes all possible
orientations. For example, in a perfectly aligned sample of nematic liquid crystals, all the local directors will be oriented in the
same direction. The specific alignment of molecules in one preferred direction in liquid crystals makes their physical properties
such as refractive index, viscosity, diamagnetic susceptibility, directionally dependent.
When a liquid crystal sample is oriented using external fields, local directors preferentially align globally along the field director.
This globally preferred direction is referred to as the director and is denoted by unit vector n. The extent of alignment within a
liquid crystal sample is typically denoted by the order parameter, S, as defined by 7.3.14, where θ is the angle between long axis of
molecule and the preferred direction, n.
2
3cos θ  −  1
S  =  ( ) (7.3.14)
2

For isotropic samples, the value of S is zero, and for perfectly aligned samples it is 1. Figure 7.3.41 shows the structure of a most
extensively studied nematic liquid crystal molecule, 4-cyano-4'-pentylbiphenyl, commonly known as 5CB. For preparing a

7.3.22 https://chem.libretexts.org/@go/page/55906
polydomain sample 5CB was placed inside a glass capillary via capillary forces (Figure 7.3.41). Figure 7.3.42 shows the 2D X-ray
diffraction of the as prepared polydomain sample. For preparing monodomain sample, a glass capillary filled with 5CB was heated
to 40 °C (i.e., above the nematic-isotropic transition temperature of 5CB, ~35 °C) and then cooled slowly in the presence of
magnetic field (1 Testla, Figure 7.3.43. This gives a uniformly aligned sample with the nematic director n oriented along the
magnetic field. Figure 7.3.44 shows the collected 2D X-ray diffraction measurement of a monodomain 5CB liquid crystal sample
using Rigaku Raxis-IV++, and it consists of two diffuse arcs (as mentioned before). Figure 7.3.45 shows the intensity distribution
of a diffuse arc as a function of Θ, and the calculated order parameter value, S, is -0.48.

Figure 7.3.41 Chemical structure of a nematic liquid crystal molecule 4-cyano-4'-pentylbiphenyl (also known as 5CB).

Figure 7.3.42 Schematic representation of a polydomain liquid crystal samples (5CB) inside a glass capillary.

Figure 7.3.43 2D X-ray diffraction of polydomain nematic liquid crystal sample of 5CB. Data was acquired using a Rigaku Raxis-
IV++ equipped with an incident beam monochromator, pinhole collimation (0.3 mm) and Cu X-ray tube (λ = 1.54 Å). The sample
to detector distance was 100 mm.

Figure 7.3.44 Magnetic field setup used to prepare a monodomain sample of 5CB. The glass capillary can just be seen between the
sides of the holder.

Figure 7.3.45 2D X-ray diffraction of polydomain nematic liquid crystal sample of 5CB. Data was acquired using a Rigaku Raxis-
IV++ equipped with an incident beam monochromator, pinhole collimation (0.3 mm) and Cu X-ray tube (λ = 1.54 Å). The sample
to detector distance was 100 mm.

7.3.23 https://chem.libretexts.org/@go/page/55906
Figure 7.3.46 Plot of intensity versus 2θ (°) for a 2D X-ray diffraction measurement of the monodomain sample of 5CB.

Refinement of Crystallographic Disorder in the Tetrafluoroborate Anion


Through the course of our structural characterization of various tetrafluoroborate salts, the complex cation has nominally been the
primary subject of interest; however, we observed that the tetrafluoroborate anion (BF4-) anions were commonly disordered (13 out
of 23 structures investigated). Furthermore, a consideration of the Cambridge Structural Database as of 14th December 2010
yielded 8,370 structures in which the tetrafluoroborate anion is present; of these, 1044 (12.5%) were refined as having some kind of
disorder associated with the BF4- anion. Several different methods have been reported for the treatment of these disorders, but the
majority was refined as a non-crystallographic rotation along the axis of one of the B-F bonds.
Unfortunately, the very property that makes fluoro-anions such good candidates for non-coordinating counter-ions (i.e., weak
intermolecular forces) also facilitates the presence of disorder in crystal structures. In other words, the appearance of disorder is
intensified with the presence of a weakly coordinating spherical anion (e.g., BF4- or PF6-) which lack the strong intermolecular
interactions needed to keep a regular, repeating anion orientation throughout the crystal lattice. Essentially, these weakly
coordinating anions are loosely defined electron-rich spheres. All considered it seems that fluoro-anions, in general, have a
propensity to exhibit apparently large atomic displacement parameters (ADP's), and thus, are appropriately refined as having
fractional site-occupancies.

Refining Disorder
In crystallography the observed atomic displacement parameters are an average of millions of unit cells throughout entire volume
of the crystal, and thermally induced motion over the time used for data collection. A disorder of atoms/molecules in a given
structure can manifest as flat or non-spherical atomic displacement parameters in the crystal structure. Such cases of disorder are
usually the result of either thermally induced motion during data collection (i.e., dynamic disorder), or the static disorder of the
atoms/molecules throughout the lattice. The latter is defined as the situation in which certain atoms, or groups of atoms, occupy
slightly different orientations from molecule to molecule over the large volume (relatively speaking) covered by the crystal lattice.
This static displacement of atoms can simulate the effect of thermal vibration on the scattering power of the "average" atom.
Consequently, differentiation between thermal motion and static disorder can be ambiguous, unless data collection is performed at
low temperature (which would negate much of the thermal motion observed at room temperature).
In most cases, this disorder is easily resolved as some non-crystallographic symmetry elements acting locally on the weakly
coordinating anion. The atomic site occupancies can be refined using the FVAR instruction on the different parts (see PART 1 and
PART 2 in Figure 7.3.47) of the disorder, having a site occupancy factor (s.o.f.) of x and 1-x, respectively. This is accomplished by
replacing 11.000 (on the F-atom lines in the “NAME.INS” file) with 21.000 or -21.000 for each of the different parts of the
disorder. For instance, the "NAME.INS" file would look something like that shown in Figure 7.3.47. Note that for more heavily
disordered structures, i.e., those with more than two disordered parts, the SUMP command can be used to determine the s.o.f. of
parts 2, 3, 4, etc. the combined sum of which is set at s.o.f. = 1.0. These are designated in FVAR as the second, third, and fourth
terms.

7.3.24 https://chem.libretexts.org/@go/page/55906
Figure 7.3.47 General layout of the SHELXTL "NAME.INS" file for treatment of disordered tetrafluoroborate.a For more than two
site occupancies “SUMP = 1.0 0.01 1.0 2 1.0 3 1.0 4” is added in addition to the FVAR instruction.
In small molecule refinement, the case will inevitably arise in which some kind of restraints or constraints must be used to achieve
convergence of the data. A restraint is any additional information concerning a given structural feature, i.e., limits on the possible
values of parameters, may be added into the refinement, thereby increasing the number of refined parameters. For example,
aromatic systems are essentially flat, so for refinement purposes, a troublesome ring system could be restrained to lie in one plane.
Restraints are not exact, i.e., they are tied to a probability distribution, whereas constraints are exact mathematical conditions.
Restraints can be regarded as falling into one of several general types:
Geometric restraints, which relates distances that should be similar.
Rigid group restraints.
Anti-bumping restraints.
Linked parameter restraints.
Similarity restraints.
ADP restraints (Figure 7.3.48
Sum and average restraints.
Origin fixing and shift limiting restraints.
Those imposed upon atomic displacement parameters.

Figure 7.3.48 Consequence of the anisotropic displacement parameter (ADP) restraints DELU, SIMU, and ISOR on the shape and
directionality of atomic displacement parameters. Adapted from P. Müller, Crystal Structure Refinement, A Crystallographer's
Guide to SHELXL, Oxford University Press, UK (2006).
Geometric Restraints
SADI - similar distance restraints for named pairs of atoms.
DFIX - defined distance restraint between covalently bonded atoms.
DANG - defined non-bonding distance restraints, e.g., between F atoms belonging to the same PART of a disordered BF4-.
FLAT - restrains group of atoms to lie in a plane.
Anisotropic Displacement Parameter Restraints
DELU - rigid bond restraints (Figure 7.3.48)
SIMU - similar ADP restraints on corresponding Uij components to be approximately equal for atoms in close proximity
(Figure 7.3.48)
ISOR - treat named anisotropic atoms to have approximately isotropic behavior (Figure 7.3.48)

7.3.25 https://chem.libretexts.org/@go/page/55906
Constraints (different than "restraints")
EADP - equivalent atomic displacement parameters.
AFIX - fitted group; e.g., AFIX 66 would fit the next six atoms into a regular hexagon.
HFIX - places H atoms in geometrically ideal positions, e.g., HFIX 123 would place two sets of methyl H atoms disordered
over two sites, 180° from each other.

Classess of Disorder for the Tetrafluoroborate Anion


Rotating about a non-crystallographic axis along a B-F bond
The most common case of disorder is a rotation about an axis, the simplest of which involves a non-crystallographic symmetry
related rotation axis about the vector made by one of the B-F bonds; this operation leads to three of the four F-atoms having two
site occupancies (Figure 7.3.49). This disorder is also seen for tBu and CF3 groups, and due to the C3 symmetry of the C(CH3)3,
CF3 and BF3 moieties actually results in a near C2rotation.

Figure 7.3.49 Schematic representation of the rotational relationship between two disordered orientations of the BF4-anion.
In a typical example, the BF4- anion present in the crystal structure of [H(Mes-dpa)]BF4 (Figure 7.3.50) was found to have a 75:25
site occupancy disorder for three of the four fluorine atoms (Figure 7.3.51). The disorder is a rotation about the axis of the B(1)-
F(1) bond. For initial refinement cycles, similar distance restraints (SADI) were placed on all B-F and F-F distances, in addition to
similar ADP restraints (SIMU) and rigid bond restraints (DELU) for all F atoms. Restraints were lifted for final refinement cycles.
A similar disorder refinement was required for [H(2-iPrPh-dpa)]BF4 (45:55), while refinement of the disorder in [Cu(2-iPrPh-dpa)
(styrene)]BF4(65:35) was performed with only SADI and DELU restraints were lifted in final refinement cycles.

Figure 7.3.50 Structures of (a) substituted bis(2-pyridyl)amines (R-dpa) and (b) substituted bis(2-quinolyl)amines [R-N(quin)2]
ligands.

Figure 7.3.51 Structure for the BF4- anion in compound [H(Mes-dpa)]BF4 with both parts of the disorder present. Thermal
ellipsoids are shown at the 20% level. Adapted from J. J. Allen, C. E. Hamilton, and A. R. Barron, Dalton Trans., 2010,11451.
In the complex [Ag(H-dpa)(styrene)]BF4 use of the free variable (FVAR) led to refinement of disordered fluorine atoms F(2A)-
F(4A) and F(2B)-F(4B) as having a 75:25 site-occupancy disorder (Figure 7.3.52). For initial refinement cycles, all B-F bond
lengths were given similar distance restraints (SADI). Similar distance restraints (SADI) were also placed on F…F distances for
each part, i.e., F(2A)…F(3A) = F(2B)…F(3B), etc. Additionally, similar ADP restraints (SIMU) and rigid bond restraints (DELU)
were placed on all F atoms. All restraints, with the exception of SIMU, were lifted for final refinement cycles.

7.3.26 https://chem.libretexts.org/@go/page/55906
Figure 7.3.52 Structure of the disordered BF4- anion in [Ag(H-dpa)(styrene)]BF4 viewed down the axis of disorder. Thermal
ellipsoids are shown at the 30% probability level. Adapted from J. J. Allen and A. R. Barron, J. Chem. Cryst., 2009, 39, 935.

Rotation About a Non-Crystallographic Axis not Along a B-F Bond


The second type of disorder is closely related to the first, with the only difference being that the rotational axis is tilted slightly off
the B-F bond vector, resulting in all four F-atoms having two site occupancies (Figure 7.3.53). Tilt angles range from 6.5° to 42°.

Figure 7.3.53 Molecular structure for the anion in [Cu(H-dpa)(cis-3-octene)]BF4 with both parts of the disordered BF4-present.
Thermal ellipsoids are shown at the 20% level. Adapted from J. J. Allen and A. R. Barron, Dalton Trans., 2009, 878.
The disordered BF4- anion present in the crystal structure of [Cu(Ph-dpa)(styrene)]BF4 was refined having fractional site
occupancies for all four fluorine atoms about a rotation slightly tilted off the B(1)-F(2A) bond. However, it should be noted that
while the U(eq) values determined for the data collected at low temperature data is roughly half that of that found at room
temperature, as is evident by the sizes and shapes of fluorine atoms in Figure 7.3.54, the site occupancies were refined to 50:50 in
each case, and there was no resolution in the disorder.

Figure 7.3.54 Comparison of the atomic displacement parameters observed in the disordered BF4- anion from [Cu(Ph-dpa)
(styrene)]BF4 at data collection temperature (a) T = 213 K and (b) T = 298 K. Thermal ellipsoids are set at the 25% level.
An extreme example of rotation off-axis is observed where refinement of more that two site occupancies (Figure 7.3.55 ) with as
many as thirteen different fluorine atom locations on only one boron atom.

7.3.27 https://chem.libretexts.org/@go/page/55906
Figure 7.3.55 Structure for the tetrafluoroborate anion with twelve fluorine atom locations. Adapted from S. Martinez-Vargas, R.
Toscano, and J. Valdez-Martinez, Acta Cryst., 2007, E63, m1975.

Constrained Rotation About a Non-Crystallographic Axis not Along a B-F Bond


Although a wide range of tilt angles are possible, in some systems the angle is constrained by the presence of hydrogen bonding.
For example, the BF4- anion present in [Cu(Mes-dpa)(μ-OH)(H2O)]2[BF4]2 was found to have a 60:40 site occupancy disorder of
the four fluorine atoms, and while the disorder is a C2-rotation slightly tilted off the axis of the B(1)-F(1A) bond, the angle is
restricted by the presence of two B-F…O interactions for one of the isomers (Figure 7.3.56).

Figure 7.3.56 Structure of the disordered BF4- in [Cu(Mes-dpa)(μ-OH)(H2O)]2[BF4]2 showing interaction with bridging hydroxide
and terminal water ligands. Thermal ellipsoids are shown at the 20% level. Adapted from J. J. Allen, C. E. Hamilton, and A. R.
Barron, Dalton Trans., 2010, 11451.
An example that does adhere to global symmetry elements is seen in the BF4- anion of [Cu{2,6-iPr2C6H3N(quin)2}2]BF4.MeOH
(Figure 7.3.57), which exhibits a hydrogen-bonding interaction with a disordered methanol solvent molecule. The structure of R-
N(quin)2 is shown in Figure 7.3.54 b. By crystallographic symmetry, the carbon atom from methanol and the boron atom from the
BF4- anion lie on a C2-axis. Fluorine atoms [F(1)-F(4)], the methanol oxygen atom, and the hydrogen atoms attached to methanol
O(1S) and C(1S) atoms were refined as having 50:50 site occupancy disorder (Figure 7.3.57).

Figure 7.3.57 H-bonding interaction in [Cu{2,6-iPr2C6H3N(quin)2}2]BF4.MeOH between anion and solvent of crystallization, both
disordered about a crystallographic C2-rotation axis running through the B(1)…C(1S) vector. Adapted from J. J. Allen, C. E.
Hamilton, and A. R. Barron, Dalton Trans., 2010, 11451.

Non Crystallographic Inversion Center at the Boron Atom


Multiple disorders can be observed with a single crystal unit cell. For example, the two BF4- anions in [Cu(Mes-dpa)(styrene)]BF4
both exhibited 50:50 site occupancy disorders, the first is a C2-rotation tilted off one of the B-F bonds, while the second is
disordered about an inversion centered on the boron atom. Refinement of the latter was carried out similarly to the aforementioned

7.3.28 https://chem.libretexts.org/@go/page/55906
cases, with the exception that fixed distance restraints for non-bonded atoms (DANG) were left in place for the disordered fluorine
atoms attached to B(2) (Figure 7.3.58).

Figure 7.3.58 Structure for the disordered BF4- anion due to a NCS-inversion center, in compound [Cu(Mes-dpa)(styrene)]BF4
with both parts of the disorders present. Thermal ellipsoids are shown at the 20% level. Adapted from J. J. Allen, C. E. Hamilton,
and A. R. Barron, Dalton Trans., 2010, 11451.

Disorder on a Crystallographic Mirror Plane


Another instance in which the BF4- anion is disordered about a crystallographic symmetry element is that of [Cu(H-dpa)(1,5-
cyclooctadiene)]BF4. In this instance fluorine atoms F(1) through F(4) are present in the asymmetric unit of the complex.
Disordered atoms F(1A)-F(4A) were refined with 50% site occupancies, as B(1) lies on a mirror plane (Figure 7.3.59). For initial
refinement cycles, similar distance restraints (SADI) were placed on all B-F and F-F distances, in addition to similar ADP restraints
(SIMU) and rigid bond restraints (DELU) for all F atoms. Restraints were lifted for final refinement cycles, in which the boron
atom lies on a crystallographic mirror plane, and all four fluorine atoms are reflected across.

Figure 7.3.59 Molecular structure for the anion in [Cu(H-dpa)(1,5-cyclooctadiene)]BF4 with both parts of the disordered BF4-
present. For clarity, thermal ellipsoids are shown at the 20% level. Adapted from J. J. Allen and A. R. Barron, Dalton Trans., 2009,
878.

Disorder on a Non-Crystallographic Mirror Plane


It has been observed that the BF4- anion can exhibit site occupancy disorder of the boron atom and one of the fluorine atoms across
an NCS mirror plane defined by the plane of the other three fluorine atoms (Figure 7.3.60) modeling the entire anion as disordered
(including the boron atom).

Figure 7.3.60 Disordered anion across the plane of three fluorine atoms. Adapted from J. T. Mague and S. W. Hawbaker, J. Chem.
Cryst., 1997, 27, 603.

7.3.29 https://chem.libretexts.org/@go/page/55906
Disorder of the Boron Atom Core
The extreme case of a disorder involves refinement of the entire anion, with all boron and all fluorine atoms occupying more than
two sites (Figure 7.3.61). In fact, some disorders of the latter types must be refined isotropically, or as a last-resort, not at all, to
prevent one or more atoms from turning non-positive definite.

Figure 7.3.61 An example of a structure of a highly disordered BF4- anion refined with four site occupancies for all boron and
fluorine atoms. Adapted from P. Szklarz, M. Owczarek, G. Bator, T. Lis, K. Gatner, and R. Jakubas, J. Mol. Struct., 2009, 929, 48.

7.3: X-ray Crystallography is shared under a CC BY 4.0 license and was authored, remixed, and/or curated by Pavan M. V. Raja & Andrew R.
Barron via source content that was edited to conform to the style and standards of the LibreTexts platform; a detailed edit history is available upon
request.

7.3.30 https://chem.libretexts.org/@go/page/55906
7.4: Low Energy Electron Diffraction
Low energy electron diffraction (LEED) is a very powerful technique that allows for the characterization of the surface of
materials. Its high surface sensitivity is due to the use of electrons with energies between 20-200 eV, which have wavelengths equal
to 2.7 – 0.87 Å (comparable to the atomic spacing). Therefore, the electrons can be elastically scattered easily by the atoms in the
first few layers of the sample. Its features, such as little penetration of low–energy electrons have positioned it as one of the most
common techniques in surface science for the determination of the symmetry of the unit cell (qualitative analysis) and the position
of the atoms in the crystal surface (quantitative analysis).

History: Davisson and Germer Experiment


In 1924 Louis de Brogile postulated that all forms of matter, such as electrons, have a wave-particle nature. Three years later after
this postulate, the American physicists Clinton J. Davisson and Lester H. Germer (Figure 7.4.1) proved experimentally the wave
nature of electrons at Bell Labs in New York, see Figure 1. At that time, they were investigating the distribution-in-angle of the
elastically scattered electrons (electrons that have suffered no loss of kinetic energy) from the (111) face of a polycrystalline nickel,
material composed of many randomly oriented crystals.

Figure 7.4.1 Clinton Davisson (right) and Lester Germer (left) in their laboratory, where they proved that electrons could act like
waves in 1927. Author unknown, public domain.
The experiment consisted of a beam of electrons from a heated tungsten filament directed against the polycrystalline nickel and an
electron detector, which was mounted on an arc to observe the electrons at different angles. During the experiment, air entered in
the vacuum chamber where the nickel was, producing an oxide layer on its surface. Davisson and Clinton reduced the nickel by
heating it at high temperature. They did not realize that the thermal treatment changed the polycrystalline nickel to a nearly
monocrystalline nickel, material composed of many oriented crystals. When they repeated the experiment, it was a great surprise
that the distribution-in-angle of the scattered electrons manifested sharp peaks at certain angles. They soon realized that these peaks
were interference patterns, and, in analogy to X-ray diffraction, the arrangement of atoms and not the structure of the atoms was
responsible for the pattern of the scattered electrons.
The results of Davisson and Germer were soon corroborated by George Paget Thomson, J. J. Thomson’s son. In 1937, both
Davisson and Thomson were awarded with the Nobel Prize in Physics for their experimental discovery of the electron diffraction
by crystals. It is noteworthy that 31 years after J. J. Thomson showed that the electron is a particle, his son showed that it is also a
wave.
Although the discovery of low-energy electron diffraction was in 1927, it became popular in the early 1960’s, when the advances in
electronics and ultra-high vacuum technology made possible the commercial availability of LEED instruments. At the beginning,
this technique was only used for qualitative characterization of surface ordering. Years later, the impact of computational
technologies allowed the use of LEED for quantitative analysis of the position of atoms within a surface. This information is
hidden in the energetic dependence of the diffraction spot intensities, which can be used to construct a LEED I-V curve.

Principles and Diffraction Patterns


Electrons can be considered as a stream of waves that hit a surface and are diffracted by regions with high electron density (the
atoms). The electrons in the range of 20 to 200 eV can penetrate the sample for about 10 Å without loosing energy. Because of this
reason, LEED is especially sensitive to surfaces, unlike X-ray diffraction, which gives information about the bulk-structure of a
crystal due to its larger mean free path (around micrometers). Table 7.4.1 compares general aspects of both techniques.

7.4.1 https://chem.libretexts.org/@go/page/55907
Table 7.4.1 Comparison between low energy electron diffraction and X-ray diffraction.
Low Energy Electron Diffraction X-ray Diffraction

Surface structure determination (high surface sensitivity) Bulk structures determination

Sample single crystal Sample single-crystal or polycrystalline

Sample must be have an oriented surface, sensitive to impurities Surface impurities not important

Experiment in ultra-high vacuum Experiment usually at atmospheric pressure

Experiment done mostly at constant incidence angle and variable


Constant wavelength and variable incidence angle
wavelength (electron energy)

Diffraction pattern consists of beams flashing out at specific


Diffraction pattern consists of beams visible at almost all energies
wavelengths and angles

Like X-ray diffraction, electron diffraction also follows the Bragg’s law, see Figure 7.4.2, where λ is the wavelength, a is the
atomic spacing, d is the spacing of the crystal layers, θ is the angle between the incident beam and the reflected beam, and n is an
integer. For constructive interference between two waves, the path length difference (2a sinθ / 2d sinθ) must be an integral multiple
of the wavelength.

Figure 7.4.2 Representation of the electron and X-ray diffraction.


In LEED, the diffracted beams impact on a fluorescent screen and form a pattern of light spots (Figure 7.4.3 a), which is a to-scale
version of the reciprocal lattice of the unit cell. The reciprocal lattice is a set of imaginary points, where the direction of a vector
from one point to another point is equal to the direction of a normal to one plane of atoms in the unit cell (real space). For example,
an electron beam penetrates a few 2D-atomic layers, Figure 7.4.3 b), so the reciprocal lattice seen by LEED consists of continues
rods and discrete points per atomic layer, see Figure 7.4.3 c. In this way, LEED patterns can give information about the size and
shape of the real space unit cell, but nothing about the positions of the atoms. To gain this information about atomic positions,
analysis of the spot intensities is required. For further information about reciprocal lattice and crystals refer to Crystal Structure and
An Introduction to Single-Crystal X-Ray Crystallography.

Figure 7.4.3 (a) LEED pattern of Cu (100) surface, (b) 2D atomic layer (real space), and its (c) reciprocal lattice. (a) adapted from
Z. Robinson, E. Ong, T. Mowll, P. Tyagi, D. Gaskill, H. Geisler, C. Ventrice, J. Phys. Chem. C, 2013, 117, 23919. Copyright:
American Chemical Society 2013.
Thanks to the hemispheric geometry of the green screen of LEED, we can observe the reciprocal lattice without distortion. It is
important to take into account that the separation of the points in the reciprocal lattice and the real interplanar distance are inversely
proportional, which means that if the atoms are more widely spaced, the spots in the pattern get closer and vice versa. In the case of
superlattices, a periodic structure composed of layers of two materials, new points arise in addition to the original diffraction
pattern.

LEED Experimental Equipment


The typical diagram of a LEED system is shown in Figure 7.4.4. This system sends an electron beam to the surface of the sample,
which comes from an electron gun behind a transparent hemispherical fluorescent screen. The electron gun consists of a heated

7.4.2 https://chem.libretexts.org/@go/page/55907
cathode and a set of focusing lenses which send electrons at low energies. The electrons collide with the sample and diffract in
different directions depending on the surface. Once diffracted, they are directed to the fluorescent screen. Before colliding with the
screen, they must pass through four different grids (known as retarding grids), which contain a central hole through which the
electron gun is inserted. The first grid is the nearest one to the sample and is connected to earth ground. A negative potential is
applied to the second and third grids, which act as suppressor grids, given that they repel all electrons coming from non–elastic
diffractions. These grids perform as filters, which only allow the highest–energy electrons to pass through; the electrons with the
lowest energies are blocked in order to prevent a bad resolution image. The fourth grid protects the phosphor screen, which
possesses positive charge from the negative grids. The remaining electrons collide with the luminescent screen, creating a phosphor
glow (left side of Figure 7.4.4), where the light intensity depends on the electron intensity.

Figure 7.4.4 Schematic diagram of a typical LEED instrument and an example of the LEED pattern view by the CCD camera.
Adapted from L. Meng, Y. Wang, L. Zhang, S. Du, R. Wu, L. Li, Y. Zhang, G. Li, H. Zhou, W. Hofer, H. Gao, Nano Letters, 2013,
13, 685. Copyright: American Chemical Society 2013.
For conventional systems of LEED, it is necessary a method of data acquisition. In the past, the general method for analyzing the
diffraction pattern was to manually take several dozen pictures. After the development of computers, the photographs were scanned
and digitalized for further analysis through computational software. Years later, the use of the charge–coupled device (CCD)
camera was incorporated, allowing rapid acquisition, the possibility to average frames during the acquisition in order to improve
the signal, the immediate digitalization and channeling of LEED pattern. In the case of the IV curves, the intensities of the points
are extracted making use of special algorithms. Figure 7.4.5 shows a commercial LEED spectrometer with the CCD camera, which
has to be in an ultra-high vacuum vessel.

Figure 7.4.5 Commercial LEED Spectrometer (OCI Vacuum Micro engineering Inc).

LEED Applications
We have previously talked about the discovery of LEED and its principles, along with the experimental setup of a LEED system. It
was also mentioned that LEED provides qualitative and quantitative surface analysis. In the following section, we will discuss the
most common applications of LEED and the information that one can obtain with this technique.

Study of Adsorbates on the Surface and Disorder Layers


ne of the principal applications of LEED is the study of adsorbates on catalysts, due to its high surface sensitivity. In order to
illustrate the application of LEED in the study of adsorbates. As an example, Figure 7.4.6 a shows the surface of Cu (100) single

7.4.3 https://chem.libretexts.org/@go/page/55907
crystal, the pristine material. This surface was cleaned carefully by various cycles of sputtering with ions of argon, followed by
annealing. The LEED patter of Cu (100) presents four well-defined spots corresponding to its cubic unit cell.

Figure 7.4.6 LEED patterns of (a) the clean Cu(100) surface, (b) the Cu(100) surface following graphene growth at 800 °C, and (c)
the Cu(100) surface following graphene growth at 900 °C. Adapted from Z. Robinson, E. Ong, T. Mowll, P. Tyagi, D. Gaskill, H.
Geisler, C. Ventrice, J. Phys. Chem. C, 2013, 117, 23919. Copyright: American Chemical Society 2013.
Figure 7.4.6 b shows the LEED pattern after the growth of graphene on the surface of Cu (100) at 800 °C, we can observe the four
spots that correspond to the surface of Cu (100) and a ring just outside these spots, which correspond to the domains of graphene
with four different primary rotational alignments with respect to the Cu (100) substrate lattice, see Figure 7.4.7. When increasing
the temperature of growth of graphene to 900 °C, we can observe a ring of twelve spots (as seen in Figure 7.4.6 c), which indicates
that the graphene has a much higher degree of rotational order. Only two domains are observed with an alignment of one of the
lattice vectors to one of the Cu (100) surface lattice vectors, given that graphene has a hexagonal geometry, so that only one vector
can coincide with the cubic lattice of Cu (100).

Figure 7.4.7 Simulated LEED image for graphene domains with four different rotational orientations with respect to the Cu(100)
surface. Adapted from Z. Robinson, E. Ong, T. Mowll, P. Tyagi, D. Gaskill, H. Geisler, C. Ventrice, J. Phys. Chem. C, 2013, 117,
23919. Copyright: American Chemical Society 2013.
One possible explanation for the twelve spots observed at 900 ˚C is that when the temperature of all domains is increased the four
different domains observed at 800 ˚C, may possess enough energy to adopt the two orientations in which the vectors align with the
surface lattice vector of Cu (100). In addition, at 900 ˚C, a decrease in the size and intensity of the Cu (100) spots is observed,
indicating a larger coverage of the copper surface by the domains of graphene.
When the oxygen is chemisorbed on the surface of Cu (100), the new spots correspond to oxygen, Figure 7.4.8 a. Once graphene is
allowed to grow on the surface with oxygen at 900 ˚C, the LEED pattern turns out different: the twelve spots corresponding to
graphene domains are not observed due to nucleation of graphene domains in the presence of oxygen in multiple orientations,
Figure 7.4.8 b.

Figure 7.4.8 LEED patterns of (a) the clean Cu(100) surface dosed with oxygen, (b) the oxygen predosed Cu(100) surface
following graphene growth at 900 °C. Adapted from Z. Robinson, E. Ong, T. Mowll, P. Tyagi, D. Gaskill, H. Geisler, C. Ventrice,
J. Phys. Chem. C, 2013, 117, 23919. Copyright: American Chemical Society 2013.

7.4.4 https://chem.libretexts.org/@go/page/55907
A way to study the disorder of the adsorbed layers is through the LEED–IV curves, see Figure 7.4.9. In this case, the intensities are
in relation to the angle of the electron beam. The spectrum of Cu (100) with only four sharp peaks shows a very organized surface.
In the case of the graphene sample growth over the copper surface, twelve peaks are shown, which correspond to the main twelve
spots of the LEED pattern. These peaks are sharp, which indicate an high level of order. For the case of the sample of graphene
growth over copper with oxygen, the twelve peaks widen, which is an effect of the increase of disorder in the layers.

Figure 7.4.9 LEED-IV using angles for the clean Cu(100) surface (top), graphene grown on the oxygen reconstructed surface
(middle), and graphene grown on the clean Cu(100) surface (bottom). Adapted from Z. Robinson, E. Ong, T. Mowll, P. Tyagi, D.
Gaskill, H. Geisler, C. Ventrice, J. Phys. Chem. C, 2013, 117, 23919. Copyright: American Chemical Society 2013.

Structure Determination
As previously mentioned, LEED–IV curves may give us exact information about the position of the atoms in a crystal. These
curves are related to a variation of intensities of the diffracted electron (spots) with the energy of the electron beam. The process of
determination of the structure by this technique works as ‘proof and error’ and consists of three main parts: the measurement of the
intensity spectra, the calculations for various models of atomic positions and the search for the best-fit structure which is
determined by an R-factor.
The first step consists of obtaining the experimental LEED pattern and all the electron beam intensities for every spot of the
reciprocal lattice in the pattern. Theoretical LEED–IV curves are calculated for a large number of geometrical models and these are
compared with the experimental curves. The agreement is quantified by means of a reliability factor or R–factor. The closest this
value to zero is, the more perfect the agreement between experimental and theoretical curves. In this way, the level of precision of
the crystalline structure will depend on the smallest R–factor that can be achieved.
Pure metals with pure surfaces allow R–factor values of around 0.1. When moving to more complex structures, these values
increase. The main reason for this gradually worse agreement between theoretical and experimental LEED-IV curves lies in the
approximations in conventional LEED theory, which treats the atoms as perfect spheres with constant scattering potential in
between. This description results in inaccurate scattering potential for more open surfaces and organic molecules. In consequence, a
precision of 1-2 pm can be achieved for atoms in metal surfaces, whereas the positions of atoms within organic molecules are
typically determined within ±10-20 pm. The values of the R-factor are usually between 0.2 and 0.5, where 0.2 represents a good
agreement, 0.35 a mediocre agreement and 0.5 a poor agreement.
Figure 7.4.10 shows an example of a typical LEED–IV curve for Ir (100), which has a quasi-hexagonal unit cell. One can observe
the parameters used to calculate the theoretical LEED–IV curve and the best-fitted curve obtained experimentally, which has an R–
factor value of 0.144. The model used is also shown.

7.4.5 https://chem.libretexts.org/@go/page/55907
Figure 10. Experimental and theoretical LEED-IV curves for Ir (100) using two different electron beams (left), and the structural
parameters using for the LEED-IV theoretical curve (right). Adapted from K. Heinz and L. Hammer, J. Phys. Chem. B, 2004, 108,
14579. Copyright: American Chemical Society 2004.

7.4: Low Energy Electron Diffraction is shared under a CC BY 4.0 license and was authored, remixed, and/or curated by Pavan M. V. Raja &
Andrew R. Barron via source content that was edited to conform to the style and standards of the LibreTexts platform; a detailed edit history is
available upon request.

7.4.6 https://chem.libretexts.org/@go/page/55907
7.5: Neutron Diffraction
The first neutron diffraction experiment was in 1945 by Ernest O. Wollan (Figure 7.5.1) using the Graphite Reactor at Oak Ridge.
Along with Clifford Shull (Figure 7.5.1) they outlined the principles of the technique. However, the concept that neutrons would
diffract like X-rays was first proposed by Dana Mitchell and Philip Powers. They proposed that neutrons have a wave like
structure, which is explained by the de Broglie equation, 7.5.1, where λ is the wavelength of the source usually measured in Å, h is
Planck’s constant, v is the velocity of the neutron, and finally m represents the mass of the neutron.
λ  =  h/mv (7.5.1)

Figure 7.5.1 American physicists Ernest Wollan (1902 - 1984) and (standing) Clifford Shull (1915 – 2001).
The great majority of materials that are studied by diffraction methods are composed of crystals. X-rays where the first type of
source tested with crystals in order to determine their structural characteristics. Crystals are said to be perfect structures although
some of them show defects on their structure. Crystals are composed of atoms, ions or molecules, which are arranged, in a uniform
repeating pattern. The basic concept to understand about crystals is that they are composed of an array of points, which are called
lattice points, and the motif, which represents the body part of the crystal. Crystals are composed of a series of unit cells. A unit cell
is the repeating portion of the crystal. Usually there are another eight unit cells surrounding each unit cell. Unit cells can be
categorized as primitive, which have only one lattice point. This means that the unit cell will only have lattice points on the corners
of the cell. This point is going to be shared with eight other unit cells. Whereas in a non primitive cell there will also be point in the
corners of the cell but in addition there will be lattice points in the faces or the interior of the cell, which similarly will be shared by
other cells. The only primitive cell known is the simple crystal system and for nonprimitive cells there are known face-centered
cubic, base centered cubic and body centered cubic.
Crystals can be categorized depending on the arrangement of lattice points; this will generate different types of shapes. There are
known seven crystal systems, which are cubic, tetragonal, orthorhombic, rhombohedral, hexagonal, monoclinic and triclinic. All of
these have different angles and the axes are equally the same or different in others. Each of these type of systems have different
bravais lattice.

Braggs Law
Braggs Law was first derived by physicist Sir W.H. Bragg (Figure 7.5.2) and his son W. L Bragg (Figure 7.5.3) in 1913.

7.5.1 https://chem.libretexts.org/@go/page/55910
Figure 7.5.2 British physicist, chemist, mathematician and active sportsman Sir William H. Bragg (1862 - 1942).

Figure 7.5.3 Australian-born British physicist William L. Bragg (1890 - 1971).


It has been used to determine the spacing of planes and angles formed between these planes and the incident beam that had been
applied to the crystal examined. Intense scattered X-rays are produced when X-rays with a set wavelength are executed to a crystal.
These scattered X-rays will interfere constructively due the equality in the differences between the travel path and the integral
number of the wavelength. Since crystals have repeating units patterns, diffraction can be seen in terms of reflection from the
planes of the crystals. The incident beam, the diffracted beam and normal plane to diffraction need to lie in the same geometric
plane. The angle, which the incident beam forms when it hits the plane of the crystal, is called 2θ. Figure 7.5.4 shows a schematic
representation of how the incident beam hits the plane of the crystal and is reflected at the same angle 2θ, which the incident beam
hits. Bragg’s Law is mathematically expressed, 7.5.2, where,n= integer order of reflection, λ= wavelength, d= plane spacing.

Figure 7.5.4 Bragg’s Law construction

nλ  =  2d sinθ (7.5.2)

7.5.2 https://chem.libretexts.org/@go/page/55910
Bragg’s Law is essential in determining the structure of an unknown crystal. Usually the wavelength is known and the angle of the
incident beam can be measured. Having these two known values, the plane spacing of the layer of atoms or ions can be obtained.
All reflections collected can be used to determine the structure of the unknown crystal material.
Bragg’s Law applies similarly to neutron diffraction. The same relationship is used the only difference being is that instead of using
X-rays as the source, neutrons that are ejected and hit the crystal are being examined.

Neutron Diffraction
Neutrons have been studied for the determination of crystalline structures. The study of materials by neutron radiation has many
advantages against the normally used such as X-rays and electrons. Neutrons are scattered by the nucleus of the atoms rather than
X-rays, which are scattered by the electrons of the atoms. These generates several differences between them such as that scattering
of X-rays highly depend on the atomic number of the atoms whereas neutrons depend on the properties of the nucleus. These lead
to a greater and accurately identification of the unknown sample examined if neutron source is being used. The nucleus of every
atom and even from isotopes of the same element is completely different. They all have different characteristics, which make
neutron diffraction a great technique for identification of materials, which have similar elemental composition. In contrast, X-rays
will not give an exact solution if similar characteristics are known between materials. Since the diffraction will be similar for
adjacent atoms further analysis needs to be done in order to determine the structure of the unknown. Also, if the sample contains
light elements such as hydrogen, it is almost impossible to determine the exact location of each of them just by X-ray diffraction or
any other technique. Neutron diffraction can tell the number of light elements and the exact position of them present in the
structure.

Neutron Inventors
Neutrons were first discovered by James Chadwick in 1932 Figure 7.5.5 when he showed that there were uncharged particles in the
radiation he was using. These particles had a similar mass of the protons but did not have the same characteristics as them.
Chadwick followed some of the predictions of Rutherford who first worked in this unknown field. Later, Elsasser designed the first
neutron diffraction in 1936 and the ones responsible for the actual constructing were Halban and Preiswerk. This was first
constructed for powders but later Mitchell and Powers developed and demonstrated the single crystal system. All experiments
realized in early years were developed using radium and beryllium sources. The neutron flux from these was not sufficient for the
characterization of materials. Then, years passed and neutron reactors had to be constructed in order to increase the flux of neutrons
to be able to realize a complete characterization the material being examined.
Between mid and late 40s neutron sources began to appear in countries such as Canada, UK and some other of Europe. Later in
1951 Shull and Wollan presented a paper that discussed the scattering lengths of 60 elements and isotopes, which generated a broad
opening of neutron diffraction for the structural information that can be obtained from neutron diffraction.

Figure 7.5.5 English Nobel laureate in physics James Chadwick (1891-1974)

7.5.3 https://chem.libretexts.org/@go/page/55910
Neutron Sources
The first source of neutrons for early experiments was gathered from radium and beryllium sources. The problem with this, as
already mentioned, was that the flux was not enough to perform huge experiments such as the determination of the structure of an
unknown material. Nuclear reactors started to emerge in early 50s and these had a great impact in the scientific field. In the 1960s
neutron reactors were constructed depending on the desired flux required for the production of neutron beams. In USA the first one
constructed was the High Flux Beam Reactor (HFBR). Later, this was followed by one at Oak Ridge Laboratory (HFIR) (Figure
7.5.6), which also was intended for isotope production and a couple of years later the ILL was built. This last one is the most

powerful so far and it was built by collaboration between Germany and France. These nuclear reactors greatly increased the flux
and so far there has not been constructed any other better reactor. It has been discussed that probably the best solution to look for
greater flux is to look for other approaches for the production of neutrons such as accelerator driven sources. These could greatly
increase the flux of neutrons and in addition other possible experiments could be executed. The key point in these devices is
spallation, which increases the number of neutrons executed from a single proton and the energy released is minimal. Currently,
there are several of these around the world but investigations continue searching for the best approach of the ejection of neutrons.

Figure 7.5.6 Schematic representation of HIFR. Courtesy of Oak Ridge National Laboratory, US Dept. of Energy

Neutron Detectors
Although neutrons are great particles for determining complete structures of materials they have some disadvantages. These
particles experiment a reasonably weak scattering when looking especially to soft materials. This is a huge concern because there
can be problems associated with the scattering of the particles which can lead to a misunderstanding in the analysis of the structure
of the material.
Neutrons are particles that have the ability to penetrate through the surface of the material being examined. This is primarily due to
the nuclear interaction produced from the particles and the nucleus from the material. This interaction is much greater that the one
performed from the electrons, which it is only an electrostatic interaction. Also, it cannot be omitted the interaction that occurs
between the electrons and the magnetic moment of the neutrons. All of these interactions discussed are of great advantage for the
determination of the structure since neutrons interacts with every single nucleus in the material. The problem comes when the
material is being analyzed because neutrons being uncharged materials make them difficult to detect them. For this reason,
neutrons need to be reacted in order to generate charged particles, ions. Some of the reactions uusually used for the detection of
neutrons are:
3 3 1
n  +   H e →   H   +   H   +  0.764M eV (7.5.3)

10 7 4
n  +   B →   Li  +   H e  +  γ  +  2.3M eV (7.5.4)

6 4 3
n  +   Li →   H e  +   H   +  4.79M eV (7.5.5)

The first two reactions apply when the detection is performed in a gas environment whereas the third one is carried out in a solid.
In each of these reaction there is a large cross section, which makes them ideal for neutron capture. The neutron detection hugely
depends on the velocity of the particles. As velocity increases, shorter wavelengths are produced and the less efficient the detection
becomes. The particles that are executed to the material need to be as close as possible in order to have an accurate signal from the
detector. These signal needs to be quickly transduced and the detector should be ready to take the next measurement.
In gas detectors the cylinder is filled up with either 3He or BF3. The electrons produced by the secondary ionization interact with
the positively charged anode wire. One disadvantage of this detector is that it cannot be attained a desired thickness since it is very

7.5.4 https://chem.libretexts.org/@go/page/55910
difficult to have a fixed thickness with a gas. In contrast, in scintillator detectors since detection is developed in a solid, any
thickness can be obtained. The thinner the thickness of the solid the more efficient the results obtained become. Usually the
absorber is 6Li and the substrate, which detects the products, is phosphor, which exhibits luminescence. This emission of light
produced from the phosphor results from the excitation of this when the ions pass thorough the scintillator. Then the signal
produced is collected and transduced to an electrical signal in order to tell that a neutron has been detected.

Neutron Scattering
One of the greatest features of neutron scattering is that neutrons are scattered by every single atomic nucleus in the material
whereas in X-ray studies, these are scattered by the electron density. In addition, neutron can be scattered by the magnetic moment
of the atoms. The intensity of the scattered neutrons will be due to the wavelength at which it is executed from the source. Figure
7.5.7 shows how a neutron is scattered by the target when the incident beam hits it.

Figure 7.5.7 Schematic representation of scattering of neutrons when it hits the target. Adapted from W. Marshall and S. W.
Lovesey, Theory of thermal neutron scattering: the use of neutrons for the investigation of condensed matter, Clarendon Press,
Oxford (1971).
The incident beam encounters the target and the scattered wave produced from the collision is detected by a detector at a defined
position given by the angles θ, ϕ which are joined by the dΩ. In this scenario there is assumed that there is no transferred energy
between the nucleus of the atoms and the neutron ejected, leads to an elastic scattering.
When there is an interest in calculating the diffracted intensities the cross sectional area needs to be separated into scattering and
absorption respectively. In relation to the energies of these there is moderately large range for constant scattering cross section.
Also, there is a wide range cross sections close to the nuclear resonance. When the energies applied are less than the resonance the
scattering length and scattering cross section are moved to the negative side depending on the structure being examined. This
means that there is a shift on the scattering, therefore the scattering will not be in a 180° phase. When the energies are higher that
resonance it means that the cross section will be asymptotic to the nucleus area. This will be expected for spherical structures.
There is also resonance scattering when there are different isotopes because each produce different nuclear energy levels.

Coherent and Incoherent Scattering


Usually in every material, atoms will be arranged differently. Therefore, neutrons when scattered will be either coherently or
incoherently. It is convenient to determine the differential scattering cross section, which is given by 7.5.6, where b represents the
mean scattering length of the atoms, k is the scattering vector, r nis the position of the vector of the analyzed atom and lastly N is
the total number of atoms in the structure.This equation can be separated in two parts, which one corresponds to the coherent
scattering and the incoherent scattering as labeled below. Usually the particles scattered will be coherent which facilitates the
solution of the cross section but when there is a difference in the mean scattering length, there will be a complete arrangement of
the formula and these new changes (incoherent scattering) should be considered. Incoherent scattering is usually due to the isotopes
and nuclear spins of the atoms in the structure.
2 (ik. rn ) 2 2
dσ/dΩ  =  |b |  |Σ e  |   +  N |b − b | (7.5.6)

Coherent Exp:
2 (ik. rn ) 2
|b |  |Σ e  | (7.5.7)

Incoherent Exp:
2
N  |b − b| (7.5.8)

The ability to distinguish atoms with similar atomic number or isotopes is proportional to the square of their corresponding
scattering lengths. There are already known several coherent scattering lengths of some atoms which are very similar to each other.
Therefore, it makes even easier to identify by neutrons the structure of a sample. Also neutrons can find ions of light elements

7.5.5 https://chem.libretexts.org/@go/page/55910
because they can locate very low atomic number elements such as hydrogen. Due to the negative scattering that hydrogen develops
it increases the contrast leading to a better identification of it, although it has a very large incoherent scattering which causes
electrons to be removed from the incident beam applied.

Magnetic Scattering
As previously mentioned one of the greatest features about neutron diffraction is that neutrons because of their magnetic moment
can interact with either the orbital or the spin magnetic moment of the material examined. Not all every single element in the
periodic table can exhibit a magnetic moment. The only elements that show a magnetic moment are those, which have unpaired
electrons spins. When neutrons hit the solid this produces a scattering from the magnetic moment vector as well as the scattering
vector from the neutron itself. Below Figure 7.5.8 shows the different vectors produced when the incident beam hits the solid.

Figure 7.5.8 Diagram of magnetic Scattering of neutrons. Adapted from G. E. Bacon, Neutron Diffraction, Clarendon Press,
Oxford (1975).
When looking at magnetic scattering it needs to be considered the coherent magnetic diffraction peaks where the magnetic
contribution to the differential cross section is p2q2 for an unpolarized incident beam. Therefore the magnetic structure amplitude
will be given by ??? , where qn is the magnetic interaction vector, pn is the magnetic scattering length and the rest of the terms are
used to know the position of the atoms in the unit cell. When this term Fmag is squared, the result is the intensity of magnetic
contribution from the peak analyzed. This equation only applies to those elements which have atoms that develop a magnetic
moment.
\[ F_{\text{mag}}\ =\ \Sigma p_{n}q_{n} e^
\label{9} \]
Magnetic diffraction becomes very important due to its d-spacing dependence. Due to the greater effect produced from the
electrons in magnetic scattering the forward scattering has a greater strength than the backward scattering. There can also be
developed similar as in X-ray, interference between the atoms which makes structure factor also be considered. These interference
effects could be produced by the wide range in difference between the electron distribution and the wavelength of the thermal
neutrons. This factor quickly decreases as compared to X-rays because the beam only interacts with the outer electrons of the
atoms.

Sample Preparation and Environment


In neutron diffraction there is not a unique protocol of factors that should be considered such as temperature, electric field and
pressure to name a few. Depending on the type of material and data that has been looked the parameters are assigned. There can be
reached very high temperatures such as 1800K or it can go as low as 4K. Usually to get to these extreme temperatures a special
furnace capable of reaching these temperatures needs to be used. For example, one of the most common used is the He refrigerator
when working with very low temperatures. For high temperatures, there are used furnaces with a heating element cylinder such as
vanadium (V), niobium (Nb), tantalum (Ta) or tungsten (W) that is attached to copper bars which hold the sample. Figure 7.5.9
shows the design for the vacuum furnaces used for the analysis. The metal that works best at the desired temperature range will be
the one chosen as the heating element. The metal that is commonly used is vanadium because it prevents the contribution of other
factors such as coherent scattering. Although with this metal this type of scattering is almost completely reduced. Other important
factor about this furnaces is that the material been examined should not decompose under vacuum conditions. The crystal needs to
be as stable as possible when it is being analyzed. When samples are not able to persist at a vacuum environment, they are heated in
the presence of several gases such as nitrogen or argon.

7.5.6 https://chem.libretexts.org/@go/page/55910
Figure 7.5.9 Metallic chamber which holds the sample. Courtesy of Nuclear Physics Institute.
Usually in order to prepare the samples that are being examined in neutron diffraction it is needed large crystals rather small ones
as the one needed for X-ray studies. This one of the main disadvantages of this instrument. Most experiments are carried out using
a four-circle diffractometer. The main reason being is because several experiment are carried out using very low temperatures and
in order to achieve a good spectra it is needed the He refrigerator. First, the crystal being analyzed is mounted on a quartz slide,
which needs to be a couple millimeters in size. Then, it is inserted into the sample holder, which is chosen depending on the
temperatures wanted to be reached. In addition, neutrons can also analyze powder samples and in order to prepare the sample for
these they need to be completely rendered into very fine powders and then inserted into the quartz slide similarly to the crystal
structures. The main concern with this method is that when samples are grounded into powders the structure of the sample being
examined can be altered.

Summary
Neutron diffraction is a great technique used for complete characterization of molecules involving light elements and also very
useful for the ones that have different isotopes in the structure. Due to the fact that neutrons interact with the nucleus of the atoms
rather than with the outer electrons of the atoms such as X-rays, it leads to a more reliable data. In addition, due to the magnetic
properties of the neutrons there can be characterized magnetic compounds due to the magnetic moment that neutrons develop.
There are several disadvantages as well, one of the most critical is that there needs to be a good amount of sample in order to be
analyzed by this technique. Also, great amounts of energy are needed to produce large amounts of neutrons. There are several
powerful neutron sources that have been developed in order to conduct studies of largest molecules and a smaller quantity of
sample. However, there is still the need of devices which can produce a great amount of flux to analyze more sophisticated
samples. Neutron diffraction has been widely studied due to the fact that it works together with X-rays studies for the
characterization of crystalline samples. The properties and advantages of this technique can greatly increased if some of the
disadvantages are solved. For example, the study of molecules which exhibit some type of molecular force can be characterized.
This will be because neutrons can precisely locate hydrogen atoms in a sample. Neutrons have gives a better answer to the
chemical interactions that are present in every single molecule, whereas X-rays help to give an idea of the macromolecular
structure of the samples being examined.

7.5: Neutron Diffraction is shared under a CC BY 4.0 license and was authored, remixed, and/or curated by Pavan M. V. Raja & Andrew R.
Barron via source content that was edited to conform to the style and standards of the LibreTexts platform; a detailed edit history is available upon
request.

7.5.7 https://chem.libretexts.org/@go/page/55910
7.6: XAFS
X-ray absorption fine structure (XAFS) spectroscopy includes both X-ray absorption near edge structure (XANES) and extended
X-ray absorption fine structure (EXAFS) spectroscopies. The difference between both techniques is the area to analyze, as shown
in Figure 7.6.1 and the information each technique provides. The complete XAFS spectrum is collected across an energy range of
around 200 eV before the absorption edge of interest and until 1000 eV after it (Figure 7.6.2). The absorption edge is defined as the
X-ray energy when the absorption coefficient has a pronounced increasing. This energy is equal to the energy required to excite an
electron to an unoccupied orbital.

Figure 7.6.1 Characteristic spectra areas for X-ray absorption near edge structure (XANES) and extended X-ray absorption fine
structure (EXAFS) spectroscopies. Adapted from S. D. Kelly, D. Hesterberg, and B. Ravel in Methods of Soil Analysis: Part 5,
Mineralogical Methods, Ed. A. L. Urely and R. Drees, Soil Science Society of America Book Series, Madison (2008).
X-ray absorption near edge structure (XANES) is used to determine the valence state and coordination geometry, whereas extended
X-ray absorption fine structure (EXAFS) is used to determine the local molecular structure of a particular element in a sample.

X-Ray Absorption Near Edge Structure (XANES) spectra


XANES is the part of the absorption spectrum closer an absorption edge. It covers from approximately -50 eV to +200 eV relative
to the edge energy (Figure 7.6.2).
Because the shape of the absorption edge is related to the density of states available for the excitation of the photoelectron, the
binding geometry and the oxidation state of the atom affect the XANES part of the absorption spectrum.
Before the absorption edge, there is a linear and smooth area. Then, the edge appears as a step, which can have other extra shapes
as isolated peaks, shoulders or a white line, which is a strong peak onto the edge. Those shapes give some information about the
atom. For example, the presence of a white line indicates that after the electron releasing, the atomic states of the element is
confined by the potential it feels. This peak sharp would be smoothed if the atom could enter to any kind of resonance. Important
information is given because of the absorption edge position. Atoms with higher oxidation state have fewer electrons than protons,
so, the energy states of the remaining electrons are lowered slightly, which causes a shift of the absorption edge energy up to
several eV to a higher X-ray energy.

Extended X-ray absorption fine structure (EXAFS) spectra


The EXAFS part of the spectrum is the oscillatory part of the absorption coefficient above around 1000 eV of the absorption edge.
This region is used to determine the molecular bonding environments of the elements. EXAFS gives information about the types
and numbers of atoms in coordination a specific atom and their inter-atomic distances. The atoms at the same radial distance from a
determinate atom form a shell. The number of the atoms in the shell is the coordination number (e.g., Figure 7.6.2).

7.6.1 https://chem.libretexts.org/@go/page/55911
Figure 7.6.2 A schematic representation of coordination number in different layers in which there are two shells around the center
atom. Both shells, green (x) and red (+), have coordination numbers of 4, but the radial distance of the red one (+) is bigger than the
green one (x). Based on S. D. Kelly, D. Hesterberg, and B. Ravel in Methods of Soil Analysis: Part 5, Mineralogical Methods, Ed.
A. L. Urely and R. Drees, Soil Science Society of America Book Series, Madison (2008).
An EXAFS signal is given by the photoelectron scattering generated for the center atom. The phase of the signal is determinate by
the distance and the path the photoelectrons travel. A simple scheme of the different paths is shown by Figure 7.6.3. In the case of
two shells around the centered atom, there is a degeneracy of four for the path between the main atom to the first shell, a
degeneracy of four for the path between the main atom to the second shell, and a degeneracy of eight for the path between the main
atom to the first shell, to the second one and to the center atom.

Figure 7.6.3 A two shell diagram in which there are three kinds of paths. From the center atom to the green one (x) and then going
back (1); from the center atom to the red one (+) and the going back (2); and from the center atom to the first shell to the second
one, and the returning to the center atom (3). Based on S. D. Kelly, D. Hesterberg, and B. Ravel in Methods of Soil Analysis: Part
5, Mineralogical Methods, Ed. A. L. Urely and R. Drees, Soil Science Society of America Book Series, Madison (2008).
The analysis of EXAFS spectra is accomplished using Fourier transformation to fit the data to the EXAFS equation. The EXAFS
equation is a sum of the contribution from all scattering paths of the photoelectrons 7.6.1, where each path is given by 7.6.2.

χ(k)  =   ∑ χi (k) (7.6.1)

2
(Ni S )Fef f (k) 2 2
−2R
i
0 i −2σi k
χi (k) ≡ sin[2kRi   +  ϕi (k)] e e λ( k)
(7.6.2)
2
kR
i

The terms Feffi(k), φi(k), and λi(k) are the effective scattering amplitude of the photoelectron, the phase shift of the photoelectron,
and the mean free path of the photoelectron, respectively. The term Ri is the half path length of the photoelectron (the distance
between the centered atom and a coordinating atom for a single-scattering event). And the k2 is given by 7.6.3. The remaining
variable are frequently determined by modeling the EXAFS spectrum.
2 me (E − E0   +  ΔE0 )
2
k   = (7.6.3)

XAFS Analysis for Arsenic Adsorption onto Iron Oxides


The absorption of arsenic species onto iron oxide offers n example of the information that can be obtained by EXAFS. Because the
huge impact that the presence of arsenic in water can produce in societies there is a lot of research in the adsorption of arsenic in
several kinds of materials, in particular nano materials. Some of the materials more promising for this kind of applications are iron
oxides. The elucidation of the mechanism of arsenic coordination onto the surfaces of those materials has been studied lately using
X-ray absorption spectroscopy.

7.6.2 https://chem.libretexts.org/@go/page/55911
There are several ways how arsenate (AsO43−, Figure 7.6.4) can be adsorbed onto the surfaces. Figure 7.6.5 shows the three ways
that Sherman proposes arsenate can be adsorbed onto goethite (α-FeOOH): bidentate cornersharing (2C), bidentate edge sharing
(2E) and monodentate corner-sharing (1V) shapes. Figure 7.6.6 shows that the bidentate corner sharing (2C) is the configuration
that corresponds with the calculated parameters not only for goethite, but for several iron oxides.

Figure 7.6.4 Structure of the arsenate anion.

Figure 7.6.5 Possible configurations of arsenate onto goethite. The tetrahedral with the small spheres represents the arsenate ions.
Adapted from D. M. Sherman and S. R. Randal. Geochim. Cosmochim. Ac. 2003, 67, 4223.

Figure 7.6.6 Fourier transforms of the EXAFS for arsenate sorbed onto goethite, lepidocrocite, hematite and ferrihydrite. Adapted
from D. M. Sherman and S. R. Randal. Geochim. Cosmochim. Ac. 2003, 67, 4223.
Several studies have confirmed that the bidentate corner sharing (2C) is the one present in the arsenate adsorption but also one
similar, a tridentate corner sharing complex (3C), for the arsenite adsorption onto most of iron oxides as shows Figure 7.6.7. Table
7.6.1 shows the coordination numbers and distances reported in the literature for the As(III) and As(V) onto goethite.

Figure 7.6.7 Proposed structural model for arsenic(III) tridante. Adapted from G. Morin, Y. Wang, G. Ona-Nguema, F. Juillot, G.
Calas, N. Menguy, E. Aubry, J. R. Bargar, and G. E. Brown. Langmuir 2009, 25, 9119.
Table 7.6.1 Coordination numbers (CN) and inter-atomic distances (R) reported in the literature for the As(III) and As(V) adsorption onto
goethite.
As CN As-O R As-O (Å) CN As-Fe R As-Fe(Å)

III 3.06±0.03 1.79±0.8 2.57±0.01 3.34±3

3.19 1.77±1 1.4 3.34±5

3 1.78 2 3.55±5

V 1.03 1.631 2 3.30

7.6.3 https://chem.libretexts.org/@go/page/55911
As CN As-O R As-O (Å) CN As-Fe R As-Fe(Å)

4.6 1.68 -- 3.55±5

7.6: XAFS is shared under a CC BY 4.0 license and was authored, remixed, and/or curated by Pavan M. V. Raja & Andrew R. Barron via source
content that was edited to conform to the style and standards of the LibreTexts platform; a detailed edit history is available upon request.

7.6.4 https://chem.libretexts.org/@go/page/55911
7.7: Circular Dichroism Spectroscopy and its Application for Determination of
Secondary Structure of Optically Active Species
Circular dichroism (CD) spectroscopy is one of few structure assessmet methods that can be utilized as an alternative and
amplification to many conventional analysis techniques with advatages such as rapid data collection and ease of use. Since most of
the efforts and time spent in advancement of chemical sciences are devoted to elucidation and analysis of structure and composition
of synthesized molecules or isolated natural products rather than their preparation, one should be aware of all the relevant
techniques available and know which instrument can be employed as an alternative to any other technique.
The aim of this module is to introduce CD technique and discuss what kind of information one can collect using CD. Additionally,
the advantages of CD compared to other analysis techniques and its limitations will be shown.

Optical Activity
As CD spectroscopy can analyze only optically active species, it is convenient to start the module with a brief introduction of
optical activity. In nature almost every life form is handed, meaning that there is certain degree of asymmetry, just like in our
hands. One cannot superimpose right hand on the left because they are non-identical mirror images of one another. So are the chiral
(handed) molecules, they exist as enantiomers, which mirror images of each other (Figure 7.7.1). One interesting phenomena
related to chiral molecules is their ability to rotate plane of polarized light. Optical activity property is used to determine specific
rotation, [ α ]Tλ, of pure enantiomer. This feature is used in polarimetery to find the enantiomeric excess, (ee), present in sample.

Figure 7.7.1 Schematic depiction of chirality/handedness of an amino acid.

Circular Dichroism
Circular dichroism (CD) spectroscopy is a powerful yet straightforward technique for examining different aspects of optically
active organic and inorganic molecules. Circular dichroism has applications in variety of modern research fields ranging from
biochemistry to inorganic chemistry. Such widespread use of the technique arises from its essential property of providing structural
information that cannot be acquired by other means. One other laudable feature of CD is its being a quick, easy technique that
makes analysis a matter of minutes. Nevertheless, just like all methods, CD has a number of limitations, which will be discussed
while comparing CD to other analysis techniques.
CD spectroscopy and related techniques were considered as esoteric analysis techniques needed and accessible only to a small
clandestine group of professionals. In order to make the reader more familiar with the technique, first of all, the principle of
operation of CD and its several types, as well as related techniques will be shown. Afterwards, sample preparation and instrument
use will be covered for protein secondary structure study case.
Depending on the light source used for generation of circularly polarized light, there are:
Far UV CD, used to study secondary structure proteins.
Near UV CD, used to investigate tertiary structure of proteins.
Visible CD, used for monitoring metal ion protein interactions.

Principle of Operation
In the CD spectrometer the sample is places in a cuvette and a beam of light is passed through the sample. The light (in the present
context all electromagnetic waves will be refer to as light) coming from source is subjected to circular polarization, meaning that its

7.7.1 https://chem.libretexts.org/@go/page/55912
plane of polarization is made to rotate either clockwise (right circular polarization) or anti-clockwise (left circular polarization)
with time while propagating, see Figure 7.7.2.

Figure 7.7.2 Schematic representation of (a) right circularly polarized and (b) left circularly polarized light. Adapted from L. Que,
Physical Methods in Bioinorganic Chemistry – Spectroscopy and Magnetism, University Science Books, Sausalito (2000).
The sample is, firstly irradiated with left rotating polarized light, and the absorption is determined by 7.7.1. A second irradiation is
performed with right polarized light. Now, due to the intrinsic asymmetry of chiral molecules, they will interact with circularly
polarized light differently according to the direction of rotation there is going to be a tendency to absorb more for one of rotation
directions. The difference between absorption of left and right circularly polarized light is the data, which is obtained from 7.7.2,
where εL and εR are the molar extinction coefficients for left and right circularly polarized light, c is the molar concentration, l is
the path length, the cuvette width (in cm). The difference in absorption can be related to difference in extinction, Δε, by 7.7.3.

A  = εcl (7.7.1)

ΔA  =  AL − AR   =  (εL   −  εR )cl (7.7.2)

Δε  =  εL   −  εR (7.7.3)

Usually, due to historical reasons the CD is reported not only as difference in absorption or extinction coefficients but as degree of
ellipticity, [θ]. The relationship between [θ] and Δε is given by 7.7.4.

[θ]  =  3, 298Δε (7.7.4)

Since the absorption is monitored in a range of wavelengths, the output is a plot of [θ] versus wavelength or Δε versus wavelength.
Figure 7.7.3 shows the CD spectrum of Δ–[Co(en)3]Cl3.

Figure 7.7.3 CD spectrum of Δ–[Co(en)3]Cl3.

7.7.2 https://chem.libretexts.org/@go/page/55912
Related Techniques
Magnetic Circular Dichroism
Magnetic circular dichroism (MCD) is a sister technique to CD, but there are several distinctions:
MCD does not require the sample to possess intrinsic asymmetry (i.e., chirality/optical activity), because optical activity is
induced by applying magnetic field parallel to light.
MCD and CD have different selection rules, thus information obtained from these two sister techniques is different. CD is good
for assessing environment of the samples’ absorbing part while MCD is superior for obtaining detailed information about
electronic structure of absorbing part.
MCD is powerful method for studying magnetic properties of materials and has recently been employed for analysis of iron-
nitrogen compound, the strongest magnet known. Moreover, MCD and its variation, variable temperature MCD are complementary
techniques to Mossbauer spectroscopy and electron paramagnetic resonance (EPR) spectroscopy. Hence, these techniques can give
useful amplification to the chapter about Mossbauer and EPR spectroscopy.

Linear Dichroism
Linear dichrosim (LD) is also a very closely related technique to CD in which the difference between absorbance of
perpendicularly and parallel polarized light is measured. In this technique the plane of polarization of light does not rotate. LD is
used to determine the orientation of absorbing parts in space.

Advantages and Limitations of CD


Just like any other instrument CD has its strengths and limits. The comparison between CD and NMR shown in Table 7.7.1 gives a
good sense of capabilities of CD.
Table 7.7.1 A comparison of CD spectroscopy to NMR spectroscopy.
CD NMR

Molecules of any size can be studied There is size limitation

The experiments are quick to perform; single wavelength


This is not the case all of the time.
measurements require milliseconds

Unique sensitivity to asymmetry in sample's structure. Special conditions are required to differentiate between enantiomers.

Can work with very small concentrations, by lengthening the cuvette


There is a limit to sensitivity of instrument.
width until discernable absorption is achieved.

Timescale is much shorter (UV) thus allowing to study dynamic Timescale is long, use of radio waves gives average of all dynamic
systems and kinetics. systems.

Quantitative data analysis can be performed to estimate chemical


Only qualitative analysis of data is possible.
composition.
Very powerful for atomic level analysis, providing essential information
Does not provide atomic level structure analysis
about chemical bonds in system.

The observed spectrum is not enough for claiming one and only
The NMR spectrum is key information for assigning a unique structure.
possible structure

What Kind of Data is Obtained from CD?


One effective way to demonstrate capabilities of CD spectroscopy is to cover the protein secondary structure study case, since CD
spectroscopy is well-established technique for elucidation of secondary structure of proteins as well as any other macromolecules.
By using CD one can estimate the degree of conformational order (what percent of the sample proteins is in α-helix and/or β-sheet
conformation), see Figure 7.7.4.

7.7.3 https://chem.libretexts.org/@go/page/55912
Figure 7.7.4 CD spectra of samples with representative conformaitons. Adapted by permission from N. Greenfield, Nat. Proto.,
2006, 1, 6.
Key points for visual estimation of secondary structure by looking at a CD spectrum:
α-helical proteins have negative bands at 222 nm and 208 nm and a positive band at 193 nm.
β-helices have negative bands at 218 nm and positive bands at 195 nm.
Proteins lacking any ordered secondary structure will not have any peaks above 210 nm.
Since the CD spectra of proteins uniquely represent their conformation, CD can be used to monitor structural changes (due to
complex formation, folding/unfolding, denaturation because of rise in temperature, denaturants, change in amino acid
sequence/mutation, etc. ) in dynamic systems and to study kinetics of protein. In other words CD can be used to perform stability
investigations and interaction modeling.

CD Instrument
Figure 7.7.5 shows a typical CD instrument.

Figure 7.7.5 A CD instrument.

Protocol for Collecting a CD Spectrum


Most of proteins and peptides will require using buffers in order to prevent denaturation. Caution should be shown to avoid using
any optically active buffers. Clear solutions are required. CD is taken in high transparency quartz cuvettes to ensure least
interference. There are cuvettes available that have path-length ranging from 0.01 cm to 1 cm. Depending on UV activity of buffers
used one should choose a cuvette with path-length (distance the beam of light passes through the sample) that compensates for UV
absorbance of buffer. Solutions should be prepared according to cuvette that will be used, see Table 7.7.2.
Table 7.7.2 Choosing the appropriate cuvette based upon the sample concentration.
Cuvette Path (cm) Concentration of Sample (mg/mL)

0.01-0.02 0.2-1.0

7.7.4 https://chem.libretexts.org/@go/page/55912
0.1 0.05-0.2

1 0.005-0.01

Besides, just like salts used to prepare pallets in FT-IR, the buffers in CD will show cutoffs at a certain point in low wavelength
region, meaning that buffers start to absorb after certain wavelengh. The cutoff values for most of common buffers are known and
can be found from manufacturer. Oxygen absorbs light below 200 nm. Therefore, in order to remove interference buffers should be
prepared from distilled water or the water should be degassed before use. Another important point is to accurately determine
concentration of sample, because concentration should be known for CD data analysis. Concentration of sample can be determined
from extinction coefficients, if such are reported in literature also for protein samples quantitative amino acid analysis can be used.
Many CD instrument come bundled with a sample compartment temperature control unit. This is very handy when doing stability
and unfolding/denaturation studies of proteins. Check to make sure the heat sink is filled with water. Turn the temperature control
unit on and set to chosen temperature.
UV source in CD is very powerful lamp and can generates large amounts of Ozone in its chamber. Ozone significantly reduces the
life of the lamp. Therefore, oxygen should be removed before turning on the main lamp (otherwise it will be converted to ozone
near lamp). For this purpose nitrogen gas is constantly flushed into lamp compartment. Let Nitrogen flush at least for 15 min.
before turning on the lamp.

Collecting Spectra for Blank, Water, Buffer Background, and Sample


1. Collect spectrum of air blank (Figure 7.7.6). This will be essentially a line lying on x-axis of spectrum, zero absorbance.
2. Fill the cuvette with water and take a spectrum.
3. Water droplets left in cuvette may change concentration of your sample, especially when working with dilute samples. Hence, it
is important to thoroughly dry the cuvette. After drying the cuvette, collect spectrum of buffer of exactly same concentration as
used for sample (Figure 7.7.6). This is the step where buffer is confirmed to be suitable spectrum of the buffer and water should
overlap within experimental error, except for low wavelength region where signal-to-noise ratio is low.
4. Clean the cuvette as described above and fill with sample solution. Collect the CD spectrum for three times for better accuracy
(Figure 7.7.6). For proteins multiple scans should overlap and not drift with time.

Figure 7.7.6 CD spectra of blank and water (left), buffer (center), and sample (right). Lysozyme in 10 mM sodium phosphate pH 7.
Adapted by permission from N. Greenfield, Nat. Protoc., 2006, 1, 6.

Data Handling and Analysis


After saving the data for both the spectra of the sample and blank is smoothed using built-in commands of controller software. The
smoothed baseline is subtracted from the smoothed spectrum of the sample. The next step is to use software bundles which have
algorithms for estimating secondary structure of proteins. Input the data into the software package of choice and process it. The
output from algorithms will be the percentage of a particular secondary structure conformation in sample. The data shown in Figure
7.7.7 lists commonly used methods and comparers them for several proteins. The estimated secondary structure is compared to X-

ray data, and one can see that it is best to use several methods for best accuracy.

7.7.5 https://chem.libretexts.org/@go/page/55912
Figure 7.7.7 Comparison of secondary structure estimation methods. Adapted by permission from N. Greenfield, Nat. Protoc.,
2006, 1, 6.

Conclusion
What advantages CD has over other analysis methods? CD spectroscopy is an excellent, rapid method for assessing the secondary
structure of proteins and performing studies of dynamic systems like folding and binding of proteins. It worth noting that CD does
not provide information about the position of those subunits with specific conformation. However, CD outrivals other techniques in
rapid assessing of the structure of unknown protein samples and in monitoring structural changes of known proteins caused by
ligation and complex formation, temperature change, mutations, denaturants. CD is also widely used to juxtapose fused proteins
with wild type counterparts, because CD spectra can tell whether the fused protein retained the structure of wild type or underwent
changes.

7.7: Circular Dichroism Spectroscopy and its Application for Determination of Secondary Structure of Optically Active Species is shared under a
CC BY 4.0 license and was authored, remixed, and/or curated by Pavan M. V. Raja & Andrew R. Barron via source content that was edited to
conform to the style and standards of the LibreTexts platform; a detailed edit history is available upon request.

7.7.6 https://chem.libretexts.org/@go/page/55912
7.8: Protein Analysis using Electrospray Ionization Mass Spectroscopy
Electrospray ionization-mass spectrometry (ESI-MS) is an analytical method that focuses on macromolecular structural
determination. The unique component of ESI-MS is the electrospray ionization. The development of electrospraying, the process of
charging a liquid into a fine aerosol, was completed in the 1960’s when Malcolm Dole (Figure 7.8.1) demonstrated the ability of
chemical species to be separated through electrospray techniques. With this important turn of events, the combination of ESI and
MS was feasible and was later developed by John B. Fenn (Figure 7.8.2), as a functional analytical method that could provide
beneficial information about the structure and size of a protein. Fenn shared the Nobel Prize in 2002, with Koichi Tanaka (Figure
7.8.3 and Kurt Wuthrich (Figure 7.8.4) for the development of ESI-MS.

Figure 7.8.1 American chemist Malcolm Dole (on right) (1903 – 1990).

Figure 7.8.2 American chemist John Bennett Fenn (1917 - 2010) shared the Nobel Prize for his work in ESI-MS and other
identification and structural analyses of biological molecules.

Figure 7.8.3 Japanese chemist and Nobel laureate Tanaka (1959 – ).

7.8.1 https://chem.libretexts.org/@go/page/55914
Figure 7.8.4 Swiss chemist and Nobel laureate Kurt Wüthrich (1938 – ).
ESI-MS is the process through which proteins, or macromolecules, in the liquid phase are charged and fragmented into smaller
aerosol droplets. These aerosol droplets lose their solvent and propel the charged fragments into the gas phase in several
components that vary by charge. These components can then be detected by a mass spectrometer. The recent boom and
development of ESI-MS is attributed to its benefits in characterizing and analyzing macromolecules, specifically biologically
important macromolecules such as proteins.
How does ESI-MS Function?
ESI-MS is a process that requires the sample to be in liquid solution, so that tiny droplets may be ionized and analyzed individually
by a mass spectrometer. The following delineates the processes that occur as relevant to Figure 7.8.5:
Spray needle/capillary- The liquid solution of the desired macromolecule is introduced into the system through this needle. The
needle is highly charged via an outside voltage source that maintains the charge constant across the needle. The normal charge
for a needle is approximately 2.5 to 4 kV. The voltage causes the large droplets to fragment into small droplets based on charge
that is accumulated from the protein constituent parts, and the liquid is now in the gas phase.
Droplet formation- The droplets that are expelled from the needle are smaller than initially, and as a result the solvent will
evaporate. The smaller droplets then start increasing their charge density on the surface as the volume decreases. As the droplets
near the Rayleigh limit, Coulombic interactions of the droplet equal the surface tension of the droplet, a Coulombic explosion
occurs that further breaks the droplet into minute fractions, including the isolated analyte with charge.
Vacuum interface/cone - This portion of the device allows for the droplets to align in a small trail and pass through to the mass
spectrometer. Alignment occurs because of the similarity and differences in charges amongst all the droplets. All the droplets
are ionized to positive charges through addition of protons to varying basic sites on the droplets, yet all the charges vary in
magnitude dependent upon the number of basic sites available for protonation. The receiving end or the cone has the opposite
charge of the spray needle, causing an attraction between the cone and the droplets.
Mass spectrometer- The charged particles then reach the mass spectrometer and are deflected based on the charge of each
particle. Deflection occurs by the quadrupole magnet of the mass spectrometer. The different deflection paths of the ions occur
due to the strength of the interaction with the magnetic field. This leads to various paths based on a mass/charge (m/z) ratio. The
particles are then read by the ion detector, as they arrive, providing a spectrum based on m/z ratio.

Figure 7.8.5 The process of ESI-MS. A focus on the capillary spray needle and the generation of aerosol particles.

7.8.2 https://chem.libretexts.org/@go/page/55914
What Data is Provided by ESI-MS?
As implied by the name, the data produced from this technique is a mass spectrometry spectrum. Without delving too deeply into
the topic of mass spectrometry, which is out of the true scope of this module, a slight explanation will be provided here. The mass
spectrometer separates particles based on a magnetic field created by a quadrupole magnet. The strength of the interaction varies on
the charge the particles carry. The amount of deflection or strength of interaction is determined by the ion detector and quantified
into a mass/charge (m/z) ratio. Because of this information, determination of chemical composition or peptide structure can easily
be managed as is explained in greater detail in the following section.
Interpretation of a Typical MS Spectrum
Interpreting the mass spectrometry data involves understanding the m/z ratio. The knowledge necessary to understanding the
interpretation of the spectrum is that the peaks correspond to portions of the whole molecule. That is to say, hypothetically, if you
put a human body in the mass spectrometer, one peak would coincide with one arm, another peak would coincide with the arm and
the abdomen, etc. The general idea behind these peaks, is that an overlay would paint the entire picture, or in the case of the
hypothetical example, provide the image of the human body. The m/z ratio defines these portions based on the charges carried by
them; thus the terminology of the mass/charge ratio. The more charges a portion of the macromolecule or protein holds, the smaller
the m/z ratio will be and the farther left it will appear on the spectrum. The fundamental concept behind interpretation involves
understanding that the peaks are interrelated, and thus the math calculations may be carried out to provide relevant information of
the protein or macromolecule being analyzed.
Calculations of m/z of the MS Spectrum Peaks
As mentioned above, the pertinent information to be obtained from the ESI-MS data is extrapolated from the understanding that the
peaks are interrelated. The steps for calculating the data are as follow:
Determine which two neighboring peaks will be analyzed.
Establish the first peak (the one farthest left) as the peak with the greatest m/z ratio. This is mathematically defined as our z+1
peak.
Establish the adjacent peak to the right of our first peak as the peak with the lower m/z ratio. This is mathematically our z peak.
Our z+1 peak will also be our m+1 peak as the difference between the two peaks is the charge of one proton. Consequently, our
z peak will be defined as our m peak.
Solve both equations for m to allow for substitution. Both sides of the equation should be in terms of zand can be solved.
Determine the charge of the z peak and subsequently, the charge of the z+1 peak.
Subtract one from the m/z ratio and multiply the m/z ratio of each peak by the previous charges determined to obtain the mass of
the protein or macromolecule.
Average the results to determine the average mass of the macromolecule or protein.
1. Determine which two neighboring peaks will be analyzed from the MS (Figure 7.8.6) as the m/z = 5 and m/z = 10 peaks.

Figure 7.8.6 Hypothetical mass spectrometry data; not based off of any particular compound. The example steps are based off of
this spectrum.
2. Establish the first peak (the one farthest left in Figure 7.8.1 as the z + 1 peak (i.e., z + 1 = 5).
3. Establish the adjacent peak to the right of the first peak as the z peak (i.e., z = 10).
4. Establish the peak ratios, 7.8.1 and 7.8.2.
m +1
=  5 (7.8.1)
z+1

m
= 10 (7.8.2)
z

5. Solve the ratios for m: 7.8.3 and 7.8.4.

7.8.3 https://chem.libretexts.org/@go/page/55914
m  =  5z  +  4 (7.8.3)

m  =  10z (7.8.4)

6. Substitute one equation for m: 7.8.5.


5z  +  4  =  10z (7.8.5)

7. Solve for z: 7.8.6.


z  = 4/5 (7.8.6)

8. Find z+1: 7.8.7.


z  +  1  =  9/5 (7.8.7)

Find average molecular mass by subtracting the mass by 1 and multiplying by the charge: 7.8.8 and 7.8.9. Hence, the average mass
= 7.2
(10  −  1)(4/5)  =  7.2 (7.8.8)

(5  −  1)(9/5)  =  7.2 (7.8.9)

Sample Preparation
Samples for ESI-MS must be in a liquid state. This requirement provides the necessary medium to easily charge the
macromolecules or proteins into a fine aerosol state that can be easily fragmented to provide the desired outcomes. The benefit to
this technique is that solid proteins that were once difficult to analyze, like metallothionein, can dissolved in an appropriate solvent
that will allow analysis through ESI-MS. Because the sample is being delivered into the system as a liquid, the capillary can easily
charge the solution to begin fragmentation of the protein into smaller fractions Maximum charge of the capillary is approximately 4
kV. However, this amount of charge is not necessary for every macromolecule. The appropriate charge is dependent on the size and
characteristic of the solvent and each individual macromolecule. This has allowed for the removal of the molecular weight limit
that was once held true for simple mass spectrometry analysis of proteins. Large proteins and macromolecules can now easily be
detected and analyzed through ESI-MS due to the facility with which the molecules can fragment.
Related Techniques
A related technique that was developed at approximately the same time as ESI-MS is matrix assisted laser desorption/ionization
mass spectrometry (MALDI-MS). This technique that was developed in the late 1980’s as wells, serves the same fundamental
purpose; allowing analysis of large macromolecules via mass spectrometry through an alternative route of generating the necessary
gas phase for analysis. In MALDI-MS, a matrix, usually comprised of crystallized 3,5-dimethoxy-4-hydroxycinnamic acid (Figure
7.8.7, water, and an organix solvent, is used to mix the analyte, and a laser is used to charge the matrix. The matrix then co-

crystallizes the analyte and pulses of the laser are then used to cause desorption of the matrix and some of the analyte crystals with
it, leading to ionization of the crystals and the phase change into the gaseous state. The analytes are then read by the tandem mass
spectrometer. Table 7.8.1 directly compares some attributes between ESI-MS and MALDI-MS. It should be noted that there are
several variations of both ESI-MS and MALDI-MS, with the methods of data collection varying and the piggy-backing of several
other methods (liquid chromatography, capillary electrophoresis, inductively coupled plasma mass spectrometry, etc.), yet all of
them have the same fundamental principles as these basic two methods.

Figure 7.8.7 Structure of 3,5-dimethoxy-4-hydroxycinnamic acid.

Experimental Details ESI-MS MALDI-MS

Starting analyte state Liquid Liquid/solid

Method of ionization Charged capillary needle Matrix laser desorption

Final analyte state Gas Gas

7.8.4 https://chem.libretexts.org/@go/page/55914
Quantity of protein needed 1 μL 1 μL

Spectrum method Mass spectrometry Mass spectrometry

Table 7.8.1 Comparison of the general experimental details of ESI-MS and MALDI-MS.

Problems with ESI-MS


ESI-MS has proven to be useful in determination of tertiary structure and molecular weight calculations of large macromolecules.
However, there are still several problems incorporated with the technique and macromolecule analysis. One problem is the isolation
of the desired protein for analysis. If the protein is unable to be extracted from the cell, this is usually done through gel
electrophoresis, there is a limiting factor in what proteins can be analyzed. Cytochrome c (Figure 7.8.7) is an example of a protein
that can be isolated and analyzed, but provides an interesting limitation on how the analytical technique does not function for a
completely effective protein analysis. The problem with cytochrome c is that even if the protein is in its native confirmation, it can
still show different charge distribution. This occurs due to the availability of basic sites for protonation that are consistently
exposed to the solvent. Any slight change to the native conformation may cause basic sites, such as in cytochrome c to be blocked
causing different m/z ratios to be seen. Another interesting limitation is seen when inorganic elements, such as in metallothioneins
proteins that contain zinc, are analyzed using ESI-MS. Metallothioneins have several isoforms that show no consistent trend in
ESI-MS data between the varied isoforms. The marked differences occur due to the metallation of each isoform being different,
which causes the electrospraying and as a result protonation of the protein to be different. Thus, incorporation of metal atoms in
proteins can have various effects on ESI-MS data due to the unexpected interactions between the metal center and the protein itself.

Figure 7.8.8 The 3-D structure of human cytochrome P450 2A13, a sub class of human cytochrome c.

7.8: Protein Analysis using Electrospray Ionization Mass Spectroscopy is shared under a CC BY 4.0 license and was authored, remixed, and/or
curated by Pavan M. V. Raja & Andrew R. Barron via source content that was edited to conform to the style and standards of the LibreTexts
platform; a detailed edit history is available upon request.

7.8.5 https://chem.libretexts.org/@go/page/55914
7.9: The Analysis of Liquid Crystal Phases using Polarized Optical Microscopy
Liquid Crystal Phases
Liquid crystals are a state of matter that has the properties between solid crystal and common liquid. There are basically three
different types of liquid crystal phases:
Thermotropic liquid crystal phases are dependent on temperature.
Lyotropic liquid crystal phases are dependent on temperature and the concentration of LCs in the solvent.
Metallotropic LCs are composed of organic and inorganic molecules, and the phase transition not only depend on temperature
and concentration, but also depend on the ratio between organic and inorganic molecules.
Thermotropic LCs are the most widely used one, which can be divided into five categories:
Nematic phase in which rod-shaped molecules have no positional order, but they self-align to have long-range directional order
with their long axes roughly parallel (Figure 7.9.1a).
Smactic phase where the molecules are positionally ordered along one direction in well-defined layers oriented either along the
layer normal (smectic A) or tilted away from the layer normal (smectic C), see Figure 7.9.1b.
Chiral phase which exhibits a twisting of the molecules perpendicular to the director, with the molecular axis parallel to the
director Figure 7.9.1 c.
Blue phase having a regular three-dimensional cubic structure of defects with lattice periods of several hundred nanometers,
and thus they exhibit selective Bragg reflections in the wavelength range of light Figure 7.9.2.
Discotic phase in which disk-shaped LC molecules can orient themselves in a layer-like fashion Figure 7.9.3.

Figure 7.9.1 Schematic representations of (a) a nematic LC phase, (b) smactic LC phases oriented along (left) and away (right)
from the normal of the layer, and (c) a chiral LC phase.

Figure 7.9.2 A schematic representation of the ordered structure of a blue LC phase.

Figure 7.9.3 Schematic representations of (a) a discotic nematic LC phase and (b) a discotic columnar LC phase.
Thermotropic LCs are very sensitive to temperature. If the temperature is too high, thermal motion will destroy the ordering of
LCs, and push it into a liquid phase. If the temperature is too low, thermal motion is hard to perform, so the material will become
crystal phase.
The existence of liquid crystal phase can be detected by using polarized optical microscopy, since liquid crystal phase exhibits its
unique texture under microscopy. The contrasting areas in the texture correspond to domains where LCs are oriented towards
different directions.

7.9.1 https://chem.libretexts.org/@go/page/55915
Polarized Optical Microscopy
Polarized optical microscopy is typically used to detect the existence of liquid crystal phases in a solution.The principle of this is
corresponding to the polarization of light. A polarizer is a filter that only permits the light oriented in a specific direction with its
polarizing direction to pass through. There are two polarizers in a polarizing optical microscope (POM) (Figure 7.9.4) and they are
designed to be oriented at right angle to each other, which is termed as cross polar. The fundamental of cross polar is illustrated in
Figure 7.9.5, the polarizing direction of the first polarizer is oriented vertically to the incident beam, so only the waves with
vertical direction can pass through it. The passed wave is subsequently blocked by the second polarizer, since this polarizer is
oriented horizontally to the incident wave.

Figure 7.9.4 The basic configuration of polarized optical microscope. Copyright: Nikon Corporation.

Figure 7.9.5 A schematic representation of the polarization of light waves. Copyright: Nikon Corporation.
Theory of Birefringence
Birefringent or doubly-refracting sample has a unique property that it can produce two individual wave components while one
wave passes through it, those two components are termed as ordinary and extraordinary waves. Figure 7.9.6 is an illustration of a
typical construction of Nicol polarizing prism, as we can see, the non-plarized white light are splitted into two ray as it passes
through the prism. The one travels out of the prism is called ordinary ray, and the other one is called extraordinary ray. So if we
have a birefringent specimen located between the polarizer and analyzer, the initial light will be separated into two waves when it
passes though the specimen. After exiting the specimen, the light components become out of phase, but are recombined with
constructive and destructive interference when they pass through the analyzer. Now the combined wave will have elliptically or
circularly polarized light wave, see Figure 7.9.7, image contrast arises from the interaction of plane-polarized light with a
birefringent specimen so some amount of wave will pass through the analyzer and give a bright domain on the specimen.

Figure 7.9.6 A schematic representation of a Nicol polarzing prism. Copyright: Nikon Corporation.

7.9.2 https://chem.libretexts.org/@go/page/55915
Figure 7.9.7 A schematic representation of elliptically and circularly polarized light waves. Copyright: Nikon Corporation.
Liquid Crystal Display
The most common application of LCs are in liquid crystals displays (LCD). Figure 7.9.8 is a simple demonstration of how LCD
works in digit calculators. There are two crossed polarizers in this system, and liquid crystal (cholesteric spiral pattern) sandwich
with positive and negative charging is located between these two polarizers. When the liquid crystal is charged, waves can pass
through without changing orientations. When the liquid crystal is out of charge, waves will be rotated 90° as it passes through LCs
so it can pass through the second polarizer. There are seven separately charged electrodes in the LCD, so the LCD can exhibit
different numbers from 0 to 9 by adjusting the electrodes. For example, when the upper right and lower left electrodes are charged,
we can get 2 on the display.

Figure 7.9.8 Demonstration of a seven-segment liquid crystal display. Copyright: Nikon Corporation.
Microscope Images of Liquid Crystal Phase
The first order retardation plate is frequently utilized to determine the optical sign of a birefringent specimen in polarized light
microscopy. The optical sign includes positive and negative. If the ordinary wavefront is faster than the extraordinary wavefront
(see Figure 7.9.9 a). When a first order retardation plate is added, the structure of the cell become all apparent compared with the
one without retardation plate, Figure 7.9.9 b).

Figure 7.9.9 Microscope images of thin section of human tongue, (a) without first order retardation plate and (b) with first order
retardation plate. Copyright: Olympus.
Images of Liquid Crystal Phases
Figure 7.9.10 shows the images of liquid crystal phases from different specimens. First order retardation plates are utilized in all of
these images. Apparent contrasts are detected here in the image which corresponds to the existence of liquid crystal phase within
the specimen.

7.9.3 https://chem.libretexts.org/@go/page/55915
Figure 7.9.10 Microscope images in polarized light with a first-order retardation plate inserted between the specimen and analyzer:
(a) polyethylene glycol, (b) polycarbonate, and (c) liquid crystalline DNA. Copyright from Nikon.
The Effect of Rotation of the Polarizer
The effect of the angle between horizontal direction and polarizer transmission axis on the appearance of liquid crystal phase may
be analyzed. In Figure 7.9.11 is show images of an ascorbic acid (Figure 7.9.12) sample under cross polar mode. When the
polarizer rotates from 0° to 90°, big variations on the shape of bright domains and domain colors appear due to the change of wave
vibrating directions. By rotating the polarizer, we can have a comprehensive understanding of the overall texture.

Figure 7.9.11 Cross polarized Microscope images of ascorbic acid specimen with polarizer rotation of (a) 0°, (b) 45°, and (c) 90°.
Copyright: Nikon Corporation.

Figure 7.9.12 The structure of ascorbic acid.

7.9: The Analysis of Liquid Crystal Phases using Polarized Optical Microscopy is shared under a CC BY 4.0 license and was authored, remixed,
and/or curated by Pavan M. V. Raja & Andrew R. Barron via source content that was edited to conform to the style and standards of the
LibreTexts platform; a detailed edit history is available upon request.

7.9.4 https://chem.libretexts.org/@go/page/55915
CHAPTER OVERVIEW
8: Structure at the Nano Scale
8.1: Microparticle Characterization via Confocal Microscopy
8.2: Transmission Electron Microscopy
8.3: Scanning Tunneling Microscopy
8.4: Spectroscopic Characterization of Nanoparticles
8.5: Using UV-Vis for the detection and characterization of silicon quantum dots
8.6: Characterization of Graphene by Raman Spectroscopy
8.7: Characterization of Graphene by Raman Spectroscopy
8.8: Characterization of Bionanoparticles by Electrospray-Differential Mobility Analysis
8.9: Characterization of Bionanoparticles by Electrospray-Differential Mobility Analysis
Index

8: Structure at the Nano Scale is shared under a CC BY 4.0 license and was authored, remixed, and/or curated by Pavan M. V. Raja & Andrew R.
Barron via source content that was edited to conform to the style and standards of the LibreTexts platform; a detailed edit history is available upon
request.

1
CHAPTER OVERVIEW
Front Matter
TitlePage
InfoPage

1
Rice University
8: Structure at the Nano Scale

Pavan M. V. Raja & Andrew R. Barron


This text is disseminated via the Open Education Resource (OER) LibreTexts Project (https://LibreTexts.org) and like the hundreds
of other texts available within this powerful platform, it is freely available for reading, printing and "consuming." Most, but not all,
pages in the library have licenses that may allow individuals to make changes, save, and print this book. Carefully
consult the applicable license(s) before pursuing such effects.
Instructors can adopt existing LibreTexts texts or Remix them to quickly build course-specific resources to meet the needs of their
students. Unlike traditional textbooks, LibreTexts’ web based origins allow powerful integration of advanced features and new
technologies to support learning.

The LibreTexts mission is to unite students, faculty and scholars in a cooperative effort to develop an easy-to-use online platform
for the construction, customization, and dissemination of OER content to reduce the burdens of unreasonable textbook costs to our
students and society. The LibreTexts project is a multi-institutional collaborative venture to develop the next generation of open-
access texts to improve postsecondary education at all levels of higher learning by developing an Open Access Resource
environment. The project currently consists of 14 independently operating and interconnected libraries that are constantly being
optimized by students, faculty, and outside experts to supplant conventional paper-based books. These free textbook alternatives are
organized within a central environment that is both vertically (from advance to basic level) and horizontally (across different fields)
integrated.
The LibreTexts libraries are Powered by MindTouch® and are supported by the Department of Education Open Textbook Pilot
Project, the UC Davis Office of the Provost, the UC Davis Library, the California State University Affordable Learning Solutions
Program, and Merlot. This material is based upon work supported by the National Science Foundation under Grant No. 1246120,
1525057, and 1413739. Unless otherwise noted, LibreTexts content is licensed by CC BY-NC-SA 3.0.
Any opinions, findings, and conclusions or recommendations expressed in this material are those of the author(s) and do not
necessarily reflect the views of the National Science Foundation nor the US Department of Education.
Have questions or comments? For information about adoptions or adaptions contact [email protected]. More information on our
activities can be found via Facebook (https://facebook.com/Libretexts), Twitter (https://twitter.com/libretexts), or our blog
(http://Blog.Libretexts.org).

This text was compiled on 06/03/2022


8.1: Microparticle Characterization via Confocal Microscopy
A Brief History of Confocal Microscopy
Confocal microscopy was invented by Marvin Minsky (FIGURE) in 1957, and subsequently patented in 1961. Minsky was trying
to study neural networks to understand how brains learn, and needed a way to image these connections in their natural state (in
three dimensions). He invented the confocal microscope in 1955, but its utility was not fully realized until technology could catch
up. In 1973 Egger published the first recognizable cells, and the first commercial microscopes were produced in 1987.

Figure 8.1.1 American cognitive scientist in the field of artificial intelligence Marvin Lee Minsky (1927 - ).
In the 1990's confocal microscopy became near routine due to advances in laser technology, fiber optics, photodetectors, thin film
dielectric coatings, computer processors, data storage, displays, and fluorophores. Today, confocal microscopy is widely used in
life sciences to study cells and tissues.

The Basics of Fluorescence


Fluorescence is the emission of a secondary photon upon absorption of a photon of higher wavelength. Most molecules at normal
temperatures are at the lowest energy state, the so-called 'ground state'. Occasionally, a molecule may absorb a photon and increase
its energy to the excited state. From here it can very quickly transfer some of that energy to other molecules through collisions;
however, if it cannot transfer enough energy it spontaneously emits a photon with a lower wavelength Figure 8.1.2. This is
fluorescence.

Figure 8.1.2 An energy diagram shows the principle of fluorescence. A molecule absorbs a high energy photon (blue) which
excites the molecule to a higher energy state. The molecule then dissipates some of the extra energy via molecular collisions (red),
and emits the remaining energy by emitting a photon (green) to return to the ground state.
In fluorescence microscopy, fluorescent molecules are designed to attach to specific parts of a sample, thus identifying them when
imaged. Multiple fluorophores can be used to simultaneously identify different parts of a sample. There are two options when using
multiple fluorophores:
Fluorophores can be chosen that respond to different wavelengths of a multi-line laser.
Fluorophores can be chosen that respond to the same excitation wavelength but emit at different wavelengths.
In order to increase the signal, more fluorophores can be attached to a sample. However, there is a limit, as high fluorophore
concentrations result in them quenching each other, and too many fluorophores near the surface of the sample may absorb enough
light to limit the light available to the rest of the sample. While the intensity of incident radiation can be increased, fluorophores
may become saturated if the intensity is too high.

8.1.1 https://chem.libretexts.org/@go/page/55917
Photobleaching is another consideration in fluorescent microscopy. Fluorophores irreversibly fade when exposed to excitation light.
This may be due to reaction of the molecules’ excited state with oxygen or oxygen radicals. There has been some success in
limiting photobleaching by reducing the oxygen available or by using free-radical scavengers. Some fluorophores are more robust
than others, so choice of fluorophore is very important. Fluorophores today are available that emit photons with wavelengths
ranging 400 - 750 nm.

How Confocal Microscopy is Different from Optical Microscopy


A microscope’s lenses project the sample plane onto an image plane. An image can be formed at many image planes; however, we
only consider one of these planes to be the ‘focal plane’ (when the sample image is in focus). When a pinhole screen in placed at
the image focal point, it allows in-focus light to pass while effectively blocking light from out-of-focus locations Figure 8.1.3. This
pinhole is placed at the conjugate image plane to the focal plane, thus the name "confocal". The size of this pinhole determines the
depth-of-focus; a bigger pinhole collects light from a larger volume. The pinhole can only practically be made as small as
approximately the radius of the Airy disk, which is the best possible light spot from a circular aperture Figure 8.1.4, because
beyond that more signal is blocked resulting in a decreased signal-to-boise ratio.
In optics, the Airy disk and Airy pattern are descriptions of the best focused spot of light that a perfect lens with a circular aperture
can make, limited by the diffraction of light.

Figure 8.1.3 A schematic of a simplified microscope objective. Red and blue lines represent light rays refracted through the
objective, indicating the focal points and corresponding image points.

Figure 8.1.4 A representation of an Airy disk. An intense peak of light forms at the middle, surrounded by rings of lower intensity
formed due to the diffraction of light. Adapted with permission from Confocal Microscopy, Eric Weeks
To further reduce the effect of scattering due to light from other parts of the sample, the sample is only illuminated at a tiny point
through the use of a pinhole in front of the light source. This greatly reduces the interference of scattered light from other parts of
the sample. The combination of a pinhole in front of both the light source and detector is what makes confocal unique.

Parts of a Confocal Microscope


A simple confocal microscope generally consists of a laser, pinhole aperture, dichromatic mirror, scanning mirrors, microscope
objectives, a photomultiplier tube, and computing software used to reconstruct the image Figure 8.1.5. Because a relatively small
volume of the sample is being illuminated at any given time, a very bright light source must be used to produce a detectable signal.
Early confocal microscopes used zirconium arc lamps, but recent advances in laser technology have made lasers in the UV-visible
and infrared more stable and affordable. A laser allows for a monochromatic (narrow wavelength range) light source that can be
used to selectively excite fluorophores to emit photons of a different wavelength. Sometimes filters are used to further screen for
single wavelengths.

8.1.2 https://chem.libretexts.org/@go/page/55917
Figure 8.1.5 A schematic of a confocal microscope. Rays represent the path of light from source to detector.
The light passes through a dichromatic (or "dichroic") mirror Figure 8.1.6 which allows light with a higher wavelength (from the
laser) to pass but reflects light of a lower wavelength (from the sample) to the detector. This allows the light to travel the same path
through the majority of the instrument, and eliminates signal due to reflection of the incident light.
The light is then reflects across a pair of mirrors or crystals, one each for the x and y directions, which enable the beam to scan
across the sample (Figure 8.1.6). The speed of the scan is usually the limiting factor in the speed of image acquisition. Most
confocal microscopes can create an image in 0.1 - 1 second. Usually the sample is raster scanned quickly in the x-direction and
slowly in the y direction (like reading a paragraph left to right, Figure 8.1.6).

Figure 8.1.6 Raster scanning is usually performed quickly in the x direction, line-by-line. Other scanning patterns are also used, but
this is most common.
The rastering is controlled by galvanometers that move the mirrors back and forth in a sawtooth motion. The disadvantage to
scanning with the light beam is that the angle of light hitting the sample changes. Fortunately, this change is small. Interestingly,
Minsky's original design moved the stage instead of the beam, as it was difficult to maintain alignment of the sensitive optics.
Despite the obvious disadvantages of moving a bulky specimen, there are some advantages of moving the stage and keeping the
optics stationary:
The light illuminates the specimen axially everywhere circumventing optical aberrations, and
The field of view can be made much larger by controlling the amplitude of the stage movements.
An alternative to light-reflecting mirrors is the acousto-optic deflector (AOD). The AOD allows for fast x-direction scans by
creating a diffraction grating from high-frequency standing sound (pressure) waves which locally change the refractive index of a
crystal. The disadvantage to AODs is that the amount of deflection depends on the wavelength, so the emission light cannot be
descanned (travel back through the same path as the excitation light). The solution to this is to descan only in the y direction
controlled by the slow galvanometer and collect the light in a slit instead of a pinhole. This results in reduced optical sectioning and
slight distortion due to the loss of radial symmetry, but good images can still be formed. Keep in mind this is not a problem for
reflected light microscopy which has the same wavelength for incident and reflected light!
Another alternative is the Nipkow disk, which has a spiral array of pinholes that create the simultaneous sampling of many points
in the sample. A single rotation covers the entire specimen several times over (at 40 revolutions per second, that's over 600 frames
per second). This allows descanning, but only about 1% of the excitation light passes through. This is okay for reflected light
microscopy, but the signal is relatively weak and signal-to-noise ratio is low. The pinholes could be made bigger to increase light
transmission but then the optical sectioning is less effective (remember depth of field is dependent on the diameter of the pinhole)
and xy resolution is poorer. Highly responsive, efficient fluorophores are needed with this method.
Returning to the confocal microscope (Figure 8.1.5), light then passes through the objective which acts as a well-corrected
condenser and objective combination. The illuminated fluorophores fluoresce and emitted light travels up the objective back to the
dichromatic mirror. This is known as epifluorescence when the incident light has the same path as detected light. Since the emitted
light now has a lower wavelength than the incident, it cannot pass through the dichromatic mirror and is reflected to the detector.
When using reflected light, a beamsplitter is used instead of a dichromatic mirror. Fluorescence microscopy when used properly
can be more sensitive than reflected light microscopy.

8.1.3 https://chem.libretexts.org/@go/page/55917
Though the signal’s position is well-defined according to the position of the xy mirrors, the signal from fluorescence is relatively
weak after passing through the pinhole, so a photomultiplier tube is used to detect emitted photons. Detecting all photons without
regard to spatial position increases the signal, and the photomultiplier tube further increases the detection signal by propagating an
electron cascade resulting from the photoelectric effect (incident photons kicking off electrons). The resulting signal is an analog
electrical signal with continuously varying voltage that corresponds to the emission intensity. This is periodically sampled by an
analog-to-digital converter.
It is important to understand that the image is a reconstruction of many points sampled across the specimen. At any given time the
microscope is only looking at a tiny point, and no complete image exists that can be viewed at an instantaneous point in time.
Software is used to recombine these points to form an image plane, and combine image planes to form a 3-D representation of the
sample volume.

Two-photon Microscopy
Two-photon microscopy is a technique whereby two beams of lower intensity are directed to intersect at the focal point. Two
photons can excite a fluorophore if they hit it at the same time, but alone they do not have enough energy to excite any molecules.
The probability of two photons hitting a fluorophore at nearly the exact same time (less than 10-16) is very low, but more likely at
the focal point. This creates a bright point of light in the sample without the usual cone of light above and below the focal plane,
since there are almost no excitations away from the focal point.

Figure 8.1.7 Schematic representation of the difference between single photon and two photon microscopy. Copyright: J. Mertz,
Boston University.
To increase the chance of absorption, an ultra-fast pulsed laser is used to create quick, intense light pulses. Since the hourglass
shape is replaced by a point source, the pinhole near the detector (used to reduce the signal from light originating from outside the
focal plane) can be eliminated. This also increases the signal-to-noise ratio (here is very little noise now that the light source is so
focused, but the signal is also small). These lasers have lower average incident power than normal lasers, which helps reduce
damage to the surrounding specimen. This technique can image deeper into the specimen (~400 μm), but these lasers are still very
expensive, difficult to set up, require a stronger power supply, intensive cooling, and must be aligned in the same optical table
because pulses can be distorted in optical fibers.

Microparticle Characterization
Confocal microscopy is very useful for determining the relative positions of particles in three dimensions Figure 8.1.8. Software
allows measurement of distances in the 3D reconstructions so that information about spacing can be ascertained (such as packing
density, porosity, long range order or alignment, etc.).

FIgure 8.1.8 A reconstruction of a colloidal suspension of poly(methyl methacrylate) (PMMA) microparticles approximately 2
microns in diameter. Adapted from Confocal Microscopy of Colloids, Eric Weeks.

8.1.4 https://chem.libretexts.org/@go/page/55917
If imaging in fluorescence mode, remember that the signal will only represent the locations of the individual fluorophores. There is
no guarantee fluorophores will completely attach to the structures of interest or that there will not be stray fluorophores away from
those structures. For microparticles it is often possible to attach the fluorophores to the shell of the particle, creating hollow spheres
of fluorophores. It is possible to tell if a sample sphere is hollow or solid but it would depend on the transparency of the material.
Dispersions of microparticles have been used to study nucleation and crystal growth, since colloids are much larger than atoms and
can be imaged in real-time. Crystalline regions are determined from the order of spheres arranged in a lattice, and regions can be
distinguished from one another by noting lattice defects.
Self-assembly is another application where time-dependent, 3-D studies can help elucidate the assembly process and determine the
position of various structures or materials. Because confocal is popular for biological specimens, the position of nanoparticles such
as quantum dots in a cell or tissue can be observed. This can be useful for determining toxicity, drug-delivery effectiveness,
diffusion limitations, etc.

A Summary of Confocal Microscopy's Strengths and Weaknesses


Strengths
Less haze, better contrast than ordinary microscopes.
3-D capability.
Illuminates a small volume.
Excludes most of the light from the sample not in the focal plane.
Depth of field may be adjusted with pinhole size.
Has both reflected light and fluorescence modes.
Can image living cells and tissues.
Fluorescence microscopy can identify several different structures simultaneously.
Accommodates samples with thickness up to 100 μm.
Can use with two-photon microscopy.
Allows for optical sectioning (no artifacts from physical sectioning) 0.5 - 1.5 μm.

Weaknesses
Images are scanned slowly (one complete image every 0.1-1 second).
Must raster scan sample, no complete image exists at any given time.
There is an inherent resolution limit because of diffraction (based on numerical aperture, ~200 nm).
Sample should be relatively transparent for good signal.
High fluorescence concentrations can quench the fluorescent signal.
Fluorophores irreversibly photobleach.
Lasers are expensive.
Angle of incident light changes slightly, introducing slight distortion.

8.1: Microparticle Characterization via Confocal Microscopy is shared under a CC BY 4.0 license and was authored, remixed, and/or curated by
Pavan M. V. Raja & Andrew R. Barron via source content that was edited to conform to the style and standards of the LibreTexts platform; a
detailed edit history is available upon request.

8.1.5 https://chem.libretexts.org/@go/page/55917
8.2: Transmission Electron Microscopy
TEM: An Overview
Transmission electron microscopy (TEM) is a form of microscopy which in which a beam of electrons transmits through an
extremely thin specimen, and then interacts with the specimen when passing through it. The formation of images in a TEM can be
explained by an optical electron beam diagram in Figure 8.2.1. TEMs provide images with significantly higher resolution than
visible-light microscopes (VLMs) do because of the smaller de Broglie wavelength of electrons. These electrons allow for the
examination of finer details, which are several thousand times higher than the highest resolution in a VLM. Nevertheless, the
magnification provide in a TEM image is in contrast to the absorption of the electrons in the material, which is primarily due to the
thickness or composition of the material.

Figure 8.2.1 The optical electron beam diagram of TEM.


When a crystal lattice spacing (d) is investigated with electrons with wavelength λ, diffracted waves will be formed at specific
angles 2θ, satisfying the Bragg condition, 8.2.1.
2dsinθ  =  λ (8.2.1)

The regular arrangement of the diffraction spots, the so-called diffraction pattern (DP), can be observed. While the transmitted and
the diffracted beams interfere on the image plane, a magnified image (electron microscope image) appears. The plane where the DP
forms is called the reciprocal space, which the image plane is called the real space. A Fourier transform can mathematically
transform the real space to reciprocal space.
By adjusting the lenses (changing their focal lengths), both electron microscope images and DP can be observed. Thus, both
observation modes can be successfully combined in the analysis of the microstructures of materials. For instance, during
investigation of DPs, an electron microscope image is observed. Then, by inserting an aperture (selected area aperture), adjusting
the lenses, and focusing on a specific area that we are interested in, we will get a DP of the area. This kind of observation mode is
called a selected area diffraction. In order to investigate an electron microscope image, we first observe the DP. Then by passing the
transmitted beam or one of the diffracted beams through a selected aperture and changing to the imaging mode, we can get the
image with enhanced contrast, and precipitates and lattice defects can easily be identified.
Describing the resolution of a TEM in terms of the classic Rayleigh criterion for VLMs, which states that the smallest distance that
can be investigated, δ, is given approximately by 8.2.2, where λ is the wavelength of the electrons, µ is the refractive index of the
viewing medium, and β is the semi-angle of collection of the magnifying lens.
0.61λ
δ  = (8.2.2)
μ sinβ

ccording to de Broglie’s ideas of the wave-particle duality, the particle momentum p is related to its wavelength λ through Planck’s
constant h, 8.2.3.
h
λ = (8.2.3)
p

Momentum is given to the electron by accelerating it through a potential drop, V, giving it a kinetic energy, eV. This potential
energy is equal to the kinetic energy of the electron, 8.2.4.
2
mo u
eV   =   (8.2.4)
2

8.2.1 https://chem.libretexts.org/@go/page/55918
Based upon the foregoing, we can equate the momentum (p) to the electron mass (mo), multiplied by the velocity (v) and
substituting for v from 8.2.5 i.e., 8.2.6.
1

p  =  mo u  =  (2 mo eV ) 2 (8.2.5)

These equations define the relationship between the electron wavelength, λ, and the accelerating voltage of the electron microscope
(V), Eq. However, we have to consider about the relative effects when the energy of electron more than 100 keV. So in order to be
exact we must modify 8.2.6 to give 8.2.7.
h
λ  = (8.2.6)
1

(2 mo eV ) 2

h
λ  = (8.2.7)
1
eV
[2 mo eV (1  +   2
)] 2

2mo e

From 8.2.2 and 8.2.3, if a higher resolution is desired a decrease in the electron wavelength is accomplished by increasing the
accelerating voltage of the electron microscope. In other words, the higher accelerating rating used, the better resolution obtained.
Why the Specimen Should be Thin
The scattering of the electron beam through the material under study can form different angular distribution (Figure 8.2.2) and it
can be either forward scattering or back scattering. If an electron is scattered < 90o, then it is forward scattered, otherwise, it is
backscattered. If the specimen is thicker, fewer electrons are forward scattered and more are backscattered. Incoherent,
backscattered electrons are the only remnants of the incident beam for bulk, non-transparent specimens. The reason that electrons
can be scattered through different angles is related to the fact that an electron can be scattered more than once. Generally, the more
times of scattering happen, the greater the angle of scattering.

Figure 8.2.2 Two different kinds of electron scattering form (a) a thin specimen and (b) a bulk specimen.
All scattering in the TEM specimen is often approximated as a single scattering event since it is the simplest process. If the
specimen is very thin, this assumption will be reasonable enough. If the electron is scattered more than once, it is called ‘plural
scattering.’ It is generally safe to assume single scattering occurs, unless the specimen is particularly thick. When the times of
scattering increase, it is difficult to predict what will happen to the electron and to interpret the images and DPs. So, the principle is
‘thinner is better’, i.e., if we make thin enough specimens so that the single-scattering assumption is plausible, and the TEM
research will be much easier.
In fact, forward scattering includes the direct beam, most elastic scattering, refraction, diffraction, particularly Bragg diffraction,
and inelastic scattering. Because of forward scattering through the thin specimen, a DP or an image would be showed on the
viewing screen, and an X-ray spectrum or an electron energy-loss spectrum can be detected outside the TEM column. However,
backscattering still cannot be ignored, it is an important imagine mode in the SEM.
Limitations of TEM

Interpreting Transmission Images

8.2.2 https://chem.libretexts.org/@go/page/55918
One significant problem that might encounter when TEM images are analyzed is that the TEM present us with 2D images of a 3D
specimen, viewed in transmission. This problem can be illustrated by showing a picture of two rhinos side by side such that the
head of one appears attached to the rear of the other (Figure 8.2.3).

Figure 8.2.3 In projection, this photograph of two rhinos appears as one two-headed beast, because sometimes people have
difficulty to translate a 2D image to a 3D image. Adapted from D. B. Williams and C. B. Carter, Transmission Electron
Microscopy: A Textbook for Material Science, 2nd Ed., Springer, New York (2009).
One aspect of this particular drawback is that a single TEM images has no depth sensitivity. There often is information about the
top and bottom surfaces of the specimen, but this is not immediately apparent. There has been progress in overcoming this
limitation, by the development of electron tomography, which uses a sequence of images taken at different angles. In addition, there
has been improvement in specimen-holder design to permit full 360o rotation and, in combination with easy data storage and
manipulation; nanotechnologists have begun to use this technique to look at complex 3D inorganic structures such as porous
materials containing catalyst particles.

Electron Beam Damage


A detrimental effect of ionizing radiation is that it can damage the specimen, particularly polymers (and most organics) or certain
minerals and ceramics. Some aspects of beam damage made worse at higher voltages. Figure 8.2.4 shows an area of a specimen
damaged by high-energy electrons. However, the combination of more intense electron sources with more sensitive electron
detectors, and the use computer enhancement of noisy images, can be used to minimize the total energy received by the sample.

Figure 8.2.4 High-resolution TEM images at the slit edge of the GaAs samples prepared by slit focused ion beam. GaAs samples
prepared at (a) 3 kV, (b) 5 kV, (c) 10 kV, (d) 20 kV, and (e) 30 kV. The thickness of the amorphous layer produced by focused ion
beam is shown in each image. Adapted from Y. Yabuuchi, S. Tametou, T. Okano, S. Inazato, S. Sadayamn, and Y. Tamamoto, J.
Electron Micros., 2004, 53, 5.
Sample Preparation
The specimens under study have to be thin if any information is to be obtained using transmitted electrons in the TEM. For a
sample to be transparent to electrons, the sample must be thin enough to transmit sufficient electrons such that enough intensity
falls on the screen to give an image. This is a function of the electron energy and the average atomic number of the elements in the

8.2.3 https://chem.libretexts.org/@go/page/55918
sample. Typically for 100 keV electrons, a specimen of aluminum alloy up to ~ 1 µm would be thin, while steel would be thin up to
about several hundred nanometers. However, thinner is better and specimens < 100 nm should be used wherever possible.
The method to prepare the specimens for TEM depends on what information is required. In order to observe TEM images with high
resolution, it is necessary to prepare thin films without introducing contamination or defects. For this purpose, it is important to
select an appropriate specimen preparation method for each material, and to find an optimum condition for each method.
Crushing

A specimen can be crushed with an agate mortar and pestle. The flakes obtained are suspended in an organic solvent (e.g., acetone),
and dispersed with a sonic bath or simply by stirring with a glass stick. Finally, the solvent containing the specimen flakes is
dropped onto a grid. This method is limited to materials which tend to cleave (e.g., mica).
Electropolishing
Slicing a bulk specimen into wafer plates of about 0.3 mm thickness by a fine cutter or a multi-wire saw. The wafer is further
thinned mechanically down to about 0.1 mm in thickness. Electropolishing is performed in a specific electrolyte by supplying a
direct current with the positive pole at the thin plate and the negative pole at a stainless steel plate. In order to avoid preferential
polishing at the edge of the specimen, all the edges are cover with insulating paint. This is called the window method. The
electropolishing is finished when there is a small hole in the plate with very thin regions around it (Figure 8.2.5). This method is
mainly used to prepare thin films of metals and alloys.

Figure 8.2.5 Principle of jet electropolishing method. The specimen and the stainless steel plate is electronic positive and negative,
respectively.
Chemical Polishing
Thinning is performed chemically, i.e., by dipping the specimen in a specific solution. As for electropolishing, a thin plate of
0.1~0.2 mm in thickness should be prepared in advance. If a small dimple is made in the center of the plate with a dimple grinder, a
hole can be made by etching around the center while keeping the edge of the specimen relatively thick. This method is frequently
used for thinning semiconductors such as silicon. As with electro-polishing, if the specimen is not washed properly after chemical
etching, contamination such as an oxide layer forms on the surface.
Ultramicrotomy
Specimens of thin films or powders are usually fixed in an acrylic or epoxy resin and trimmed with a glass knife before being
sliced with a diamond knife. This process is necessary so that the specimens in the resin can be sliced easily by a diamond knife.
Acrylic resins are easily sliced and can be removed with chloroform after slicing. When using an acrylic resin, a gelatin capsule is
used as a vessel. Epoxy resin takes less time to solidify than acrylic resins, and they remain strong under electron irradiation. This
method has been used for preparing thin sections of biological specimens and sometimes for thin films of inorganic materials
which are not too hard to cut.
Ion Milling
A thin plate (less than 0.1 mm) is prepared from a bulk specimen by using a diamond cutter and by mechanical thinning. Then, a
disk 3 mm in diameter is made from the plate using a diamond knife or a ultrasonic cutter, and a dimple is formed in the center of
the surface with a dimple grinder. If it is possible to thin the disk directly to 0.03 mm in thickness by mechanical thinning without
using a dimple grinder, the disk should be strengthened by covering the edge with a metal ring. Ar ions are usually used for the
sputtering, and the incidence angle against the disk specimen and the accelerating voltage are set as 10 - 20o and a few kilovolts,
respectively. This method is widely used to obtain thin regions of ceramics and semiconductors in particular, and also for cross
section of various multilayer films.

8.2.4 https://chem.libretexts.org/@go/page/55918
Focused Ion Beam (FIB)
This method was originally developed for the purpose of fixing semiconductor devices. In principle, ion beams are sharply focused
on a small area, and the specimen in thinned very rapidly by sputtering. Usually Ga ions are used, with an accelerating voltage of
about 30 kV and a current of about 10 A/cm2. The probe size is several tens of nanometers. This method is useful for specimens
containing a boundary between different materials, where it may be difficult to homogeneously thin the boundary region by other
methods such as ion milling.
Vacuum Evaporation

The specimen to be studied is set in a tungsten-coil or basket. Resistance heating is applied by an electric current passing through
the coil or basket, and the specimen is melted, then evaporated (or sublimed), and finally deposited onto a substrate. The deposition
process is usually carried under a pressure of 10-3-10-4 Pa, but in order to avoid surface contamination, a very high vacuum is
necessary. A collodion film or cleaved rock salt is used as a substrate. Rock salt is especially useful in forming single crystals with
a special orientation relationship between each crystal and the substrate. Salt is easily dissolved in water, and then the deposited
films can be fixed on a grid. Recently, as an alternative to resistance heating, electron beam heating or an ion beam sputtering
method has been used to prepare thin films of various alloys. This method is used for preparing homogeneous thin films of metals
and alloys, and is also used for coating a specimen with the metal of alloy.
The Characteristics of the Grid

The types of TEM specimens that are prepared depend on what information is needed. For example, a self-supporting specimen is
one where the whole specimen consists of one material (which may be a composite). Other specimens are supported on a grid or on
a Cu washer with a single slot. Some grids are shown in Figure 8.2.6. Usually the specimen or grid will be 3 mm in diameter.

Figure 8.2.6 TEM sample support mesh grids. A diameter of a grid is usually 3.05 mm, however, some grids with diameters of
2.30 mm are also be used for earlier microscopes. Adapted from D. B. Williams and C. B. Carter, Transmission Electron
Microscopy: A Textbook for Material Science, 2nd Ed., Springer, New York (2009).
TEM specimen stage designs include airlocks to allow for insertion of the specimen holder into the vacuum with minimal increase
in pressure in other areas of the microscope. The specimen holders are adapted to hold a standard size of grid upon which the
sample is placed or a standard size of self-supporting specimen. Standard TEM grid sizes is a 3.05 mm diameter ring, with a
thickness and mesh size ranging from a few to 100 µm. The sample is placed onto the inner meshed area having diameter of
approximately 2.5 mm. The grid materials usually are copper, molybdenum, gold or platinum. This grid is placed into the sample
holder which is paired with the specimen stage. A wide variety of designs of stages and holders exist, depending upon the type of
experiment being performed. In addition to 3.05 mm grids, 2.3 mm grids are sometimes, if rarely, used. These grids were
particularly used in the mineral sciences where a large degree of tilt can be required and where specimen material may be
extremely rare. Electron transparent specimens have a thickness around 100 nm, but this value depends on the accelerating voltage.
Once inserted into a TEM, the sample is manipulated to allow study of the region of interest. To accommodate this, the TEM stage
includes mechanisms for the translation of the sample in the XY plane of the sample, for Z height adjustment of the sample holder,
and usually at least one rotation degree of freedom. Most TEMs provide the ability for two orthogonal rotation angles of movement
with specialized holder designs called double-tilt sample holders.
A TEM stage is required to have the ability to hold a specimen and be manipulated to bring the region of interest into the path of
the electron beam. As the TEM can operate over a wide range of magnifications, the stage must simultaneously be highly resistant
to mechanical drift as low as a few nm/minute while being able to move several µm/minute, with repositioning accuracy on the
order of nanometers.

8.2.5 https://chem.libretexts.org/@go/page/55918
Transmission Electron Microscopy Image for Multilayer-Nanomaterials
Although, TEMs can only provide 2D analysis for a 3D specimen; magnifications of 300,000 times can be routinely obtained for
many materials making it an ideal methodfor the study of nanomaterials. Besides from the TEM images, darker areas of the image
show that the sample is thicker or denser in these areas, so we can observe the different components and structures of the specimen
by the difference of color. For investigating multilayer-nanomaterials, a TEM is usually the first choice, because not only does it
provide a high resolution image for nanomaterials but also it can distinguish each layer within a nanostructured material.
Observations of Multilayer-nanomaterials

TEM was been used to analyze the depth-graded W/Si multilayer films. Multilayer films were grown on polished, 100 mm thick Si
wafers by magnetron sputtering in argon gas. The individual tungsten and silicon layer thicknesses in periodic and depth-graded
multilayers are adjusted by varying the computer-controlled rotational velocity of the substrate platen. The deposition times
required to produce specific layer thicknesses were determined from detailed rate calibrations. Samples for TEM were prepared by
focused ion beam milling at liquid N2 temperature to prevent any beam heating which might result in re-crystallization and/or re-
growth of any amorphous or fine grained polycrystalline layers in the film.
TEM measurements were made using a JEOL-4000 high-resolution transmission electron microscope operating at 400 keV; this
instrument has a point-to-point resolution of 0.16 nm. Large area cross-sectional images of a depth-graded multilayer film obtained
under medium magnification (~100 kX) were acquired at high resolution. A cross-sectional TEM image showed 150 layers W/Si
film with the thickness of layers in the range of 3.33 ~ 29.6 nm (Figure 8.2.7 shows a part of layers). The dark layers are tungsten
and the light layers are silicon and they are separated by the thin amorphous W–Si interlayers (gray bands). By the high resolution
of the TEM and the nature characteristics of the material, each layer can be distinguished clearly with their different darkness.

Figure 8.2.7 Cross-sectional transmission electron micrograph of the top portion of a depth-graded W/Si multilayer structure.
Selected bilayer indices and thicknesses are indicated. The tungsten (dark bands) and silicon (light bands) layers are separated by
thin amorphous W–Si interlayers (gray bands). The topmost silicon layer is not completely visible in this image. Adapted from D.
L. Windt, F. E. Christensen, W. W. Craig, C. Hailey, F. A. Harrison, M. Jimenez-Garate, R. Kalyanaraman, and P. H. Mao, J. Appl.
Phys., 2000, 88, 460.
Not all kinds of multilayer nanomaterials can be observed clearly under TEM. A materials consist of pc-Si:H multilayers were
prepared by a photo-assisted chemical vapor deposition (photo-CVD) using a low-pressure mercury lamp as an UV light source to
dissociate the gases. The pc-Si:H multilayer included low H2-diluted a-Si:H sublayers (SL’s) and highly H2-diluted a-Si:H
sublayers (SH’s). Control of the CVD gas flow (H2|SiH4) under continuous UV irradiation resulted in the deposition of multilayer
films layer by layer.
For a TEM measurement, a 20 nm thick undiluted a-Si:H film on a c-Si wafer before the deposition of multilayer to prevent from
any epitaxial growth. Figure 8.2.8 shows a cross-sectional TEM image of a six-cycled pc-Si:H multilayer specimen. The white
dotted lines are used to emphasize the horizontal stripes, which have periodicity in the TEM image. As can be seen, there are no
significant boundaries between SL and SH could be observed because all sublayers are prepared in H2 gas. In order to get the more
accurate thickness of each sublayer, other measurements might be necessary.

8.2.6 https://chem.libretexts.org/@go/page/55918
Figure 8.2.8 Cross-sectional TEM image of a 6-cycled pc-Si:H multilayer. Before the multilayer deposition, a 20 nm thick a-Si:H
was deposited on a c-Si substrate. Adapted from S. W. Kwon, J. Kwak, S. Y. Myong, and K. S. Lim, J. Non-Cryst. Solid, 2006,
352, 1132.

TEM Imaging of Carbon Nanomaterials


Transmission electron microscopy (TEM) is a form of microscopy that uses an high energy electron beam (rather than optical
light). A beam of electrons is transmitted through an ultra thin specimen, interacting with the specimen as it passes through. The
image (formed from the interaction of the electrons with the sample) is magnified and focused onto an imaging device, such as a
photographic film, a fluorescent screen,or detected by a CCD camera. In order to let the electrons pass through the specimen, the
specimen has to be ultra thin, usually thinner than 10 nm.
The resolution of TEM is significantly higher than light microscopes. This is because the electron has a much smaller de Broglie
wavelength than visible light (wavelength of 400~700 nm). Theoretically, the maximum resolution, d, has been limited by λ, the
wavelength of the detecting source (light or electrons) and NA, the numerical aperture of the system.
λ λ
d  = ≈ (8.2.8)
2n sinα 2N A

For high speed electrons (in TEM, electron velocity is close to the speed of light, c, so that the special theory of relativity has to be
considered), the λe:
h
λe =   −−−−−−−−−−−−−−− − (8.2.9)
2
√ 2 m0 E(1 + E/2 m0 c )

According to this formula, if we increase the energy of the detecting source, its wavelength will decrease, and we can get higher
resolution. Today, the energy of electrons used can easily get to 200 keV, sometimes as high as 1 MeV, which means the resolution
is good enough to investigate structure in sub-nanometer scale. Because the electrons is focused by several electrostatic and
electromagnetic lenses, like the problems optical camera usually have, the image resolution is also limited by aberration, especially
the spherical aberration called Cs. Equipped with a new generation of aberration correctors, transmission electron aberration-
corrected microscope (TEAM) can overcome spherical aberration and get to half angstrom resolution.
Although TEAM can easily get to atomic resolution, the first TEM invented by Ruska in April 1932 could hardly compete with
optical microscope, with only 3.6×4.8 = 14.4 magnification. The primary problem was the electron irradiation damage to sample in
poor vacuum system. After World War II, Ruska resumed his work in developing high resolution TEM. Finally, this work brought
him the Nobel Prize in physics 1986. Since then, the general structure of TEM hasn’t changed too much as shown in Figure 8.2.9.
The basic components in TEM are: electron gun, condenser system, objective lens (most important len in TEM which determines
the final resolution), diffraction lens, projective lenses (all lens are inside the equipment column, between apertures), image
recording system (used to be negative films, now is CCD cameras) and vacuum system.

8.2.7 https://chem.libretexts.org/@go/page/55918
Figure 8.2.9 Position of the basic components in a TEM.
The Family of Carbon Allotropes and Carbon Nanomaterials
Common carbon allotropes include diamond, graphite, amorphrous C (a-C), fullerene (also known as buckyball), carbon nanotube
(CNT, including single wall CNT and multi wall CNT), graphene. Most of them are chemically inert and have been found in
nature. We can also define carbon as sp2 carbon (which is graphite), sp3 carbon (which is diamond) or hybrids of sp2 and sp3
carbon. As shown in Figure, (a) is the structure of diamond, (b) is the structure of graphite, (c) graphene is a single sheet of
graphite, (d) is amorphous carbon, (e) is C60, and (f) is single wall nanotube. As for carbon nanomaterials, fullerene, CNT and
graphene are the three most well investigated, due to their unique properties in both mechanics and electronics. Under TEM, these
carbon nanomaterials will display three different projected images.

Figure 8.2.10 Six allotropes of carbon: a) diamond, b) graphite, c) graphene, d) amorphous carbon, e) C60 (Buckminsterfullerene
or buckyball), f) single-wall carbon nanotube or buckytube.
Atomic Structure of Carbon Nanomaterials under TEM
All carbon naomaterials can be investigated under TEM. Howerver, because of their difference in structure and shape, specific
parts should be focused in order to obtain their atomic structure.
For C60, which has a diameter of only 1 nm, it is relatively difficult to suspend a sample over a lacey carbon grid (a common kind
of TEM grid usually used for nanoparticles). Even if the C60 sits on a thin a-C film, it also has some focus problems since the
surface profile variation might be larger than 1 nm. One way to solve this problem is to encapsulate the C60 into single wall CNTs,
which is known as nano peapods. This method has two benefits:
CNT helps focus on C60. Single wall is aligned in a long distance (relative to C60). Once it is suspended on lacey carbon film, it is
much easier to focus on it. Therefore, the C60 inside can also be caught by minor focus changes.

8.2.8 https://chem.libretexts.org/@go/page/55918
The CNT can protect C60 from electron irradiation. Intense high energy electrons can permanently change the structure of the
CNT. For C60, which is more reactive than CNTs, it can not survive after exposing to high dose fast electrons.
In studying CNT cages, C92 is observed as a small circle inside the walls of the CNT. While a majority of electron energy is
absorbed by the CNT, the sample is still not irradiation-proof. Thus, as is seen in Figure 8.2.11, after a 123 s exposure, defects can
be generated and two C92 fused into one new larger fullerene.

Figure 8.2.11 C92 encapsulated in SWNTs under different electron irradiation time. Courtesy of Dr. Kazutomo SUENAGA,
adapted from K. Urita, Y. Sato, K. Suenaga, A. Gloter, A. Hasimoto, M. Ishida, T. Shimada, T. Shinohara, S. Iijima, Nano Lett.,
2004, 4, 2451. Copyright American Chemical Society (2004).
Although, the discovery of C60 was first confirmed by mass spectra rather than TEM. When it came to the discovery of CNTs, mass
spectra was no longer useful because CNTs shows no individual peak in mass spectra since any sample contains a range of CNTs
with different lengths and diameters. On the other hand, HRTEM can provide a clear image evidence of their existence. An
example is shown in Figure 8.2.12.

Figure 8.2.11 TEM images of SWNT and DWCNTs. Parallel dark lines corresponds to (002) lattice image of graphite. (a) and (b)
SWNTs have 1 layer graphene sheet, diameter 3.2 nm. (c) DWCNT, diameter 4.0 nm.
Graphene is a planar fullerene sheet. Until recently, Raman, AFM and optical microscopy (graphene on 300 nm SiO2 wafer) were
the most convenient methods to characterize samples. However, in order to confirm graphene’s atomic structure and determine the
difference between mono-layer and bi-layer, TEM is still a good option. In Figure 8.2.13, a monolayer suspended graphene is
observed with its atomic structure clearly shown. Inset is the FFT of the TEM image, which can be used as a filter to get an

8.2.9 https://chem.libretexts.org/@go/page/55918
optimized structure image. High angle annular dark field (HAADF) image usually gives better contrast for different particles on it.
It is also sensitive with changes of thickness, which allows a determination of the number of graphene layers.

Figure 8.2.13 HRTEM of monolayer graphene. (a) Bright filed. (b) High Angle Annular Dark Field. Courtesy of Dr M. H. Gass,
adapted from M. H. Gass, U. Bangert, A. L. Bleloch, P. Wang, R. R. Nair, and A. K. Geim, Nature Nanotechnol., 2008, 3, 676.
Graphene Stacking and Edges Direction
Like the situation in CNT, TEM image is a projected image. Therefore, even with exact count of edge lines, it is not possible to
conclude that a sample is a single layer graphene or multi-layer. If folding graphene has AA stacking (one layer is superposed on
the other), with a projected direction of [001], one image could not tell the thickness of graphene. In order to distinguish such a
bilayer of graphene from a single layer of graphene, a series of tilting experiment must be done. Different stacking structures of
graphene are shown in Figure 8.2.13 a.
Theoretically, graphene has the potential for interesting edge effects. Based upon its sp2 structure, its edge can be either that of a
zigzag or armchair configuration. Each of these possess different electronic properties similar to that observed for CNTs. For both
research and potential application, it is important to control the growth or cutting of graphene with one specific edge. But before
testing its electronic properties, all the edges have to be identified, either by directly imaging with STM or by TEM. Detailed
information of graphene edges can be obtained with HRTEM, simulated with fast fourier transform (FFT). In Figure 8.2.14 b,
armchair directions are marked with red arrow respectively. A clear model in Figurec shows a 30 degree angle between zigzag edge
and armchair edge.

Figure 8.2.14 (a) Graphene stacking structure; (b) HRTEM image of graphene edges: zigzag and armchain (inset is FFT); (c)
graphene edge model, a 30° angle between zigzag and armchair direction.

Transmission Electron Energy Loss Spectroscopy


Electron energy loss spectroscopy (EELS) is a technique that measures electronic excitations within solid-state materials. When an
electron beam with a narrow range of kinetic energy is directed at a material some electrons will be inelastically scattered, resulting

8.2.10 https://chem.libretexts.org/@go/page/55918
in a kinetic energy loss. Electrons can be inelastically scattered from phonon excitations, plasmon excitations, interband transitions,
or inner shell ionization. EELS measures the energy loss of these inelastically scattered electrons and can yield information on
atomic composition, bonding, electronic properties of valance and conduction bands and surface properties. An example of atomic
level composition mapping is shown in Figure 8.2.15 a. EELS has even been used to measure pressure and temperature within
materials.

Figure 8.2.15 EEL Map showing atomic composition with atomic scale spatial resolution. (a) Shows a high angle annular dark
field (HAADF) image with atomic resolution of a LaxSr1-xMnO3 film grow on TiO2. (b) La (c) Mn (d) Ti elemental data obtain
from a STEM-EELS measurement. (e) Overlayed image of b-d showing relative positions of La, Mn, and Ti compared to atomic
resolution imaging. (f) Model of theoretical packing in LaxSr1-xMnO3 film. Reproduced from J. A. Mundy, et al., Nat. Commun.,
2014, 5, 3464. Copyright Nature Publishing Group 2015.
The EEl Spectrum

An idealized EEL spectrum is show in Figure 8.2.16. The most prominent feature of any EEL spectrum is the zero loss peak
(ZLP). The ZLP is due to those electrons from the electron beam that do not inelastically scatter and reach the detector with their
original kinetic energy; typically 100-300 keV. By definition the ZLP is set to 0 eV for further analysis and all signals arising from
inelastically scatter electrons occur at >0 eV. The second largest feature is often the plasmon resonance - the collective excitation of
conduction band electrons within a material. The plasmon resonance and other peaks attributed to weakly bound, or outer shell
electrons, occur in the “low-loss” region of the spectrum. The low-loss regime is typically thought of as energy loss <50 eV, but
this cut-off from low-loss to high-loss is arbitrary. Shown in the inset of Figure 8.2.16 is an edge from atom core-loss and further
fine structure. Inner shell ionizations, represented by the core-loss peaks, are useful in determining elemental compositions as these
peaks can act as fingerprints for specific elements. For example, if there is a peak at 532 eV in a EEL spectrum, there is a high
probability that the sample contains a considerable amount of oxygen because this is known to be the energy needed to remove an
inner shell electron from oxygen. This idea is further explored by looking at sudden changes in the bulk plasmon for aluminum in
different chemical environments as shown in Figure 8.2.16.

Figure 8.2.16 Idealized electron energy loss spectrum (EELS). Inset shows core loss and fine structure. This image is filed under
the Creative Commons Attribution-Share Alike 3.0 Unported License. Original Author Hat’n’Coat.

8.2.11 https://chem.libretexts.org/@go/page/55918
Figure 8.2.17 Slight shifts in the plasmon peak of pure Al, AlN, and various aluminum oxides. Slight shifts in peak shape and
energy can allow EEL spectroscopists to determine slight variations in chemical and electron enviroments. Adapted from D.B.
William, C.B. Carter. Transmission Electron Microscopy: A Textbook for Materials Science. Springer, New York, NY, 2nd Ed.,
2009. 760p.
Of course, there are several other techniques available for probing atomic compositions many of which are covered in this text.
These include Energy dispersive X-ray spectroscopy, X-ray photoelectron spectroscopy, and Auger electron spectroscopy. Please
reference these chapters thorough introduction to these techniques.
Electron Energy Loss Spectroscopy Versus Energy Dispersive X-ray Spectroscopy

As a technique EELS is most frequently compared to energy dispersive X-ray spectroscopy (EDX) also known as energy dispersive
spectroscopy (EDS). Energy dispersive X-ray detectors are commonly found as analytical probes on both scanning and
transmission electron microscopes. The popularity of EDS can be understood by recognizing the simplicity of compositional
analysis using this technique. However, EELS data can offer complementary compositional analysis while also generally yielding
further insight into the solid-state physics and chemistry in a system at the cost of a steeper learning curve. EDS and EELS spectra
are both derived from the electronic excitations of materials, however, EELS probes the initial excitation while EDS looks at X-ray
emissions from the decay of this excited state. As a result, EEL spectra investigate energy ranges from 0-3 keV while EDS spectra
analyze a wider energy range from 1-40 keV. The difference in ranges makes EDS suited particularly well for heavy elements while
EELS complements measurement elements lighter than Zn.
History and Implementation
In the early 1940s, James Hillier (Figure 8.2.18) and R.F. Baker were looking to develop a method for pairing the size, shape, and
structure available from electron microscopes to a convenient method for “determining the composition of individual particles in a
mixed specimen”. Their instrument, shown in Figure 8.2.19,reported in the Journal of Applied Physics in September 1994 was the
first electron-optical instrument used to measure the velocity distribution in an electron beam transmitting through a sample.

8.2.12 https://chem.libretexts.org/@go/page/55918
Figure 8.2.18 Dr. James Hillier (August 22, 1915 – January 15, 2007) received his Ph.D. in Physics from the University of Toronto.
He is recognized as a pioneer in the field of electron microscopy and as a graduate student in 1938 designed and built the first
successful electron microscope in the western hemisphere. Dr. Hillier authored 150 papers on both technical subjects and
management and held 41 patents for his inventions. He was inducted into the National Inventors Hall of Fame for the electron
microscope in 1980.

Figure 8.2.19 The first instrument capable of performing electron energy loss spectroscopy built by Hillier and Baker in 1944.
Reproduced from J. Hillier and R. F. Baker, J. Appl. Phys., 1944, 15, 663–675. Copyright AIP Publishing LLC
The instrument was built from a repurposed transmission electron microscope (TEM). It consisted of an electron source and three
electromagnetic focusing lenses, standard for TEMs at the time, but also incorporated a magnetic deflecting lenses, which when
turned on, would redirect the electrons 180° into a photographic plate. The electrons with varying kinetic energies dispersed across
the photographic plate and could be correlated to the energy loss of each peak depending on position. In this groundbreaking work,
Hillier and Baker were able to find the discrete energy loss corresponding to the K levels of both carbon and oxygen.
The vast majority of EEL spectrometers are found as secondary analyzers in transmission electron microscopes. It wasn’t until the
1990s when EELS became a widely used research tool because of advances in electron beam aberration correction and vacuum
technologies. Today, EELS is capable of spatial resolutions down to the single atom level, and if the electron beam is
monochromated energy resolution can be as low as 0.01eV. Figure 8.2.20 depicts the typical layout of an EEL spectrometer at the
base of a TEM.

8.2.13 https://chem.libretexts.org/@go/page/55918
Figure 8.2.20 Schematic to show the simplified positions of standard EELS components in a standard TEM.

8.2: Transmission Electron Microscopy is shared under a CC BY 4.0 license and was authored, remixed, and/or curated by Pavan M. V. Raja &
Andrew R. Barron via source content that was edited to conform to the style and standards of the LibreTexts platform; a detailed edit history is
available upon request.

8.2.14 https://chem.libretexts.org/@go/page/55918
8.3: Scanning Tunneling Microscopy
Scanning tunneling microscopy (STM) is a powerful instrument that allows one to image the sample surface at the atomic level. As
the first generation of scanning probe microscopy (SPM), STM paves the way for the study of nano-science and nano-materials.
For the first time, researchers could obtain atom-resolution images of electrically conductive surfaces as well as their local electric
structures. Because of this milestone invention, Gerd Binnig (Figure 8.3.1) and Heinrich Rohrer (Figure 8.3.2) won the Nobel
Prize in Physics in 1986.

Figure 8.3.1 German physicist Gerd Binnig (1947 - ).

Figure 8.3.2 Swiss physicist Heinrich Rohrer (1933 - )


Principles of Scanning Tunneling Microscopy
The key physical principle behind STM is the tunneling effect. In terms of their wave nature, the electrons in the surface atoms
actually are not as tightly bonded to the nucleons as the electrons in the atoms of the bulk. More specifically, the electron density is
not zero in the space outside the surface, though it will decrease exponentially as the distance between the electron and the surface
increases (Figure 8.3.3 a). So, when a metal tip approaches to a conductive surface within a very short distance, normally just a few
Å, their perspective electron clouds will starting to overlap, and generate tunneling current if a small voltage is applied between
them, as shown in Figure \(\PageIndex{3}) b.

8.3.1 https://chem.libretexts.org/@go/page/55919
Figure 8.3.3 Schematic diagram of the principles of AFM showing (a) the interactions between tip and surface and (b) the
tunneling current generated from tip and surface is measured and used as feedback to control the movement of the tip.
When we consider the separation between the tip and the surface as an ideal one-dimensional tunneling barrier, the tunneling
probability, or the tunneling current I, will depend largely on s, the distance between the tip and surface, 8.3.1, where m is the
electron mass, e the electron charge, h the Plank constant, ϕ the averaged work function of the tip and the sample, and V the bias
voltage.
2 1/2
−2s [2m/ h (<ϕ> − e|V |/2) ]
I ∝e (8.3.1)

A simple calculation will show us how strongly the tunneling current is affected by the distance (s). If s is increased by ∆s = 1 Å,
8.3.2 and 8.3.3.

−2 k0 Δs
ΔI   =  e (8.3.2)

2 1/2
k0   =  [2m/ h (< ϕ >   −  e|V |/2)] (8.3.3)

-1 1
Usually (<ϕ> e|V|/2) is about 5 eV, which k0 about 1 Å , then ∆I/I = /8. That means, if s changes by 1 Å, the current will change
by one order of the magnitude. That’s the reason why we can get atom-level image by measuring the tunneling current between the
tip and the sample.
In a typical STM operation process, the tip is scanning across the surface of sample in x-y plain, the instrument records the x-y
position of the tip, measures the tunneling current, and control the height of the tip via a feedback circuit. The movements of the tip
in x, y and z directions are all controlled by piezo ceramics, which can be elongated or shortened according to the voltage applied
on them.
Normally, there are two modes of operation for STM, constant height mode and constant current mode. In constant height mode,
the tip stays at a constant height when it scans through the sample, and the tunneling current is measured at different (x, y) position
(Figure 8.3.4b). This mode can be applied when the surface of sample is very smooth. But, if the sample is rough, or has some
large particles on the surface, the tip may contact with the sample and damage the surface. In this case, the constant current mode is
applied. During this scanning process, the tunneling current, namely the distance between the tip and the sample, is settled to an
unchanged target value. If the tunneling current is higher than that target value, that means the height of the sample surface is
increasing, the distance between the tip and sample is decreasing. In this situation, the feedback control system will respond
quickly and retract the tip. Conversely, if the tunneling current drops below the target value, the feedback control will have the tip
closer to the surface. According to the output signal from feedback control, the surface of the sample can be imaged.
Comparison of Atomic Force Microscopy (AFM) and Scanning Tunneling Microscopy (STM)
Both AFM and STM are widely used in nano-science. According to the different working principles though, they have their own
advantages and disadvantages when measuring specific properties of sample (Table 8.3.1). STM requires an electric circuit
including the tip and sample to let the tunneling current go through. That means, the sample for STM must be conducting. In case
of AFM however, it just measures the deflection of the cantilever caused by the van der Waals forces between the tip and sample.
Thus, in general any kind of sample can be used for AFM. But, because of the exponential relation of the tunneling current and
distance, STM has a better resolution than AFM. In STM image one can actually “see” an individual atom, while in AFM it’s
almost impossible, and the quality of AFM image is largely depended on the shape and contact force of the tip. In some cases, the
measured signal would be rather complicated to interpret into morphology or other properties of sample. On the other side, STM
can give straight forward electric property of the sample surface.
Table 8.3.1 Comparison of AFM and STM

8.3.2 https://chem.libretexts.org/@go/page/55919
AFM STM

Sample Requirement - Conducting

Work environment Air, liquid Vacuum

Lateral resolution ~1 nm ~0.1 nm

Vertical resolution ~0.05 nm ~0.05 nm

Working mode Tapping, contact Constant current, constant height

Applications of Scanning Tunneling Microscopy in Nanoscience


STM provides a powerful method to detect the surface of conducting and semi-conducting materials. Recently STM can also be
applied in the imaging of insulators, superlattice assemblies and even the manipulation of molecules on surface. More importantly,
STM can provide the surface structure and electric property of surface at atomic resolution, a true breakthrough in the development
of nano-science. In this sense, the data collected from STM could reflect the local properties even of single molecule and atom.
With these valuable measurement data, one could give a deeper understanding of structure-property relations in nanomaterials.
An excellent example is the STM imaging of graphene on Ru(0001), as shown in Figure 8.3.4. Clearly seen is the superstructure
with a periodicity of ~30 Å , coming from the lattice mismatch of 12 unit cells of the graphene and 11 unit cells of the underneath
Ru(0001) substrate. This so-called moiré structure can also be seen in other systems when the adsorbed layers have strong chemical
bonds within the layer and weak interaction with the underlying surface. In this case, the periodic superstructure seen in graphene
tells us that the formed graphene is well crystallized and expected to have high quality.

Figure 8.3.4 Atomically resolved image of the graphene overlayer. The scanning area is 40 x 40 Å, the operation mode is constant
current mode, It is l nA, VBias is -0.05 V. Adapted with permission from S. Marchini, S. Gunther, and J. Wintterlin, Phys. Rev. B,
2007, 76, 075429. Copyrighted by the American Physical Society.
Another good example is shown to see that the measurement from STM could tell us the bonding information in single-molecular
level. In thiol- and thiophene-functionalization of single-wall carbon nanotubes (SWNTs), the use of Au nanoparticles as chemical
markers for AFM gives misleading results, while STM imaging could give correct information of substituent location. From AFM
image, Au-thiol-SWNT (Figure 8.3.6a) shows that most of the sidewalls are unfunctionalized, while Au-thiophene-SWNT (Figure
8.3.6 c)shows long bands of continuous functionalized regions on SWNT. This could lead to the estimation that thiophene is better

functionalized to SWNT than thiol. Yet, if we look up to the STM image (Figure 8.3.6b and d), in thiol-SWNTs the multiple
functional groups are tightly bonded in about 5 - 25 nm, while in thiophene-SWNTs the functionalization is spread out uniformly
along the whole length of SWNT. This information indicates that actually the functionalization levels of thiol- and thiophene-
SWNTs are comparable. The difference is that, in thiol-SWNTs, functional groups are grouped together and each group is bonded
to a single gold nanoparticle, while in thiophene-SWNTs, every individual functional group is bonded to a nanoparticle.

Figure 8.3.5 Structure of (a) thiol-functionalized SWNTs and thiophene-functionalized SWNTs.

8.3.3 https://chem.libretexts.org/@go/page/55919
Figure 8.3.6 Difference between AFM and STM images of functionalized SWNTs. (a) and (c) are tapping mode AFM images and
height profiles of thiol- and thiophene-SWNTs. (b) and (d) are STM images of thiol-SWNTs (scanning area is 4560 x 4000 Å, the
operation mode is constant current mode, It is 3.25 pA, VBias is -0.5 V) and thiophene-SWNTs (scanning area is 4560 x 4000 Å, the
operation mode is constant current mode, It is 5.66 pA, VBias is -0.8 V). Inset in (d) is a higher resolution image of the local defects
on thiophene-SWNT (500 x 140 Å, the operation mode is constant current mode, It is 25.5 pA, VBias is -0.8 V). Adapted from L.
Zhang, J. Zhang, N. Schmandt, J. Cratty, V. N. Khabashesku, K. F. Kelly, and A. R. Barron, Chem. Commun., 2005, 5429
(http://dx.doi.org/10.1039/b509257d). Reproduced by permission of The Royal Society of Chemistry.

Adaptations to Scanning Tunneling Microscopy


Scanning tunneling microscopy (STM) is a relatively recent imaging technology that has proven very useful for determining the
topography of conducting and semiconducting samples with angstrom (Å) level precision. STM was invented by Gerd Binnig
(Figure 8.3.7) and Heinrich Rohrer (Figure 8.3.8), who both won the 1986 Nobel Prize in physics for their technological advances.

Figure 8.3.7 German physicist Gerd Binnig (1947 - ).

8.3.4 https://chem.libretexts.org/@go/page/55919
Figure 8.3.8 Swiss physicist Heinrich Rohrer (1933 - ).
The main component of a scanning tunneling microscope is a rigid metallic probe tip, typically composed of tungsten, connected to
a piezodrive containing three perpendicular piezoelectric transducers (Figure 8.3.9). The tip is brought within a fraction of a
nanometer of an electrically conducting sample. At close distances, the electron clouds of the metal tip overlap with the electron
clouds of the surface atoms (Figure 8.3.9 inset). If a small voltage is applied between the tip and the sample a tunneling current is
generated. The magnitude of this tunneling current is dependent on the bias voltage applied and the distance between the tip and the
surface. A current amplifier can covert the generated tunneling current into a voltage. The magnitude of the resulting voltage as
compared to the initial voltage can then be used to control the piezodrive, which controls the distance between the tip and the
surface (i.e., the z direction). By scanning the tip in the x and y directions, the tunneling current can be measured across the entire
sample. The STM system can operate in either of two modes: Constant height or constant current

Figure 8.3.9 Schematic drawing of a STM apparatus.


In constant height mode, the tip is fixed in the z direction and the change in tunneling current as the tip changes in the x,y direction
is collected and plotted to describe the change in topography of the sample. This method is dangerous for use in samples with
fluctuations in height as the fixed tip might contact and destroy raised areas of the sample. A common method for non-uniformly
smooth samples is constant current mode. In this mode, a target current value, called the set point, is selected and the tunneling
current data gathered from the sample is compared to the target value. If the collected voltage deviates from the set point, the tip is
moved in the z direction and the voltage is measured again until the target voltage is reached. The change in the z direction required
to reach the set point is recorded across the entire sample and plotted as a representation of the topography of the sample. The
height data is typically displayed as a gray scale image of the topography of the sample, where lighter areas typically indicate
raised sample areas and darker spots indicate protrusions. These images are typically colored for better contrast.

8.3.5 https://chem.libretexts.org/@go/page/55919
The standard method of STM, described above, is useful for many substances (including high precision optical components, disk
drive surfaces, and buckyballs) and is typically used under ultrahigh vacuum to avoid contamination of the samples from the
surrounding systems. Other sample types, such as semiconductor interfaces or biological samples, need some enhancements to the
traditional STM apparatus to yield more detailed sample information. Three such modifications, spin-polarized STM (SP-STM),
ballistic electron emission microscopy (BEEM) and photon STM (PSTM) are summarized in Table 8.3.2 and in described in detail
below.
Table 8.3.2 Comparison of conventional and altered STM types
Alterations to Conventional
Name Sample Types Limitations
STM

STM None Conducting surface Rigidity of probe

Needs to be overlaid with STM,


SP-STM Magnetized STM tip Magnetic
magnetized tip type

Three-terminal with base Voltage, changes due to barrier


BEEM Interfaces
electrode and current collector height

PSTM Optical fiber tip Biological Optical tip and psrim manufacture

Spin Polarized STM


Spin-polarized scanning tunneling microscopy (SP-STM) can be used to provide detailed information of magnetic phenomena on
the single-atom scale. This imaging technique is particularly important for accurate measurement of superconductivity and high-
density magnetic data storage devices. In addition, SP-STM, while sensitive to the partial magnetic moments of the sample, is not a
field-sensitive technique and so can be applied in a variety of different magnetic fields.
Device setup and sample preparation
In SP-STM, the STM tip is coated with a thin layer of magnetic material. As with STM, voltage is then applied between tip and
sample resulting in tunneling current. Atoms with partial magnetic moments that are aligned in the same direction as the partial
magnetic moment of the atom at the very tip of the STM tip show a higher magnitude of tunneling current due to the interactions
between the magnetic moments. Likewise, atoms with partial magnetic moments opposite that of the atom at the tip of the STM tip
demonstrate a reduced tunneling current (Figure 8.3.10). A computer program can then translate the change in tunneling current to
a topographical map, showing the spin density on the surface of the sample.

Figure 8.3.10 Schematic illustration of magnetized tip for SP-STM.


The sensitivity to magnetic moments depends greatly upon the direction of the magnetic moment of the tip, which can be
controlled by the magnetic properties of the material used to coat the outermost layer of the tungsten STM probe. A wide variety of
magnetic materials have been studied as possible coatings, including both ferromagnetic materials, such as a thin coat of iron or of
gadolinium, and antiferromagnetic materials such as chromium. Another method that has been used to make a magnetically
sensitive probe tip is irradiation of a semiconducting GaAs tip with high energy circularly polarized light. This irradiation causes a
splitting of electrons in the GaAs valence band and population of the conduction band with spin-polarized electrons. These spin-
polarized electrons then provide partial magnetic moments which in turn influence the tunneling current generated by the sample
surface.

8.3.6 https://chem.libretexts.org/@go/page/55919
Sample preparation for SP-STM is essentially the same as for STM. SP-STM has been used to image samples such as thin films
and nanoparticle constructs as well as determining the magnetic topography of thin metallic sheets such as in Figure 8.3.11. The
upper image is a traditional STM image of a thin layer of cobalt, which shows the topography of the sample. The second image is
an SP-STM image of the same layer of cobalt, which shows the magnetic domain of the sample. The two images, when combined
provide useful information about the exact location of the partial magnetic moments within the sample.

Figure 8.3.11 A thin layer of Co(0001) as imaged by (a) STM, showing the topography, and (b) SP-STM, showing the magnetic
domain structure. Image adapted from W. Wulfhekel and J. Kirschner, Appl. Phys. Lett., 1999, 75, 1944.
Limitations

One of the major limitations with SP-STM is that both distance and partial magnetic moment yield the same contrast in a SP-STM
image. This can be corrected by combination with conventional STM to get multi-domain structures and/or topological information
which can then be overlaid on top of the SP-STM image, correcting for differences in sample height as opposed to magnetization.
The properties of the magnetic tip dictate much of the properties of the technique itself. If the outermost atom of the tip is not
properly magnetized, the technique will yield no more information than a traditional STM. The direction of the magnetization
vector of the tip is also of great importance. If the magnetization vector of the tip is perpendicular to the magnetization vector of the
sample, there will be no spin contrast. It is therefore important to carefully choose the coating applied to the tungsten STM tip in
order to align appropriately with the expected magnetic moments of the sample. Also, the coating makes the magnetic tips more
expensive to produce than standard STM tips. In addition, these tips are often made of mechanically soft materials, causing them to
wear quickly and require a high cost of maintenance.
Ballistic Electron Emission Microscopy
Ballistic electron emission microscopy (BEEM) is a technique commonly used to image semiconductor interfaces. Conventional
surface probe techniques can provide detailed information on the formation of interfaces, but lack the ability to study fully formed
interfaces due to inaccessibility to the surface. BEEM allows for the ability to obtain a quantitative measure of electron transport
across fully formed interfaces, something necessary for many industrial applications.
Device Setup and Sample Preparation
BEEM utilizes STM with a three-electrode configuration, as seen in Figure 8.3.12. In this technique, ballistic electrons are first
injected from a STM tip into the sample, traditionally composed of at least two layers separated by an interface, which rests on
three indium contact pads that provide a connection to a base electrode (Figure 8.3.12). As the voltage is applied to the sample,
electrons tunnel across the vacuum and through the first layer of the sample, reaching the interface, and then scatter. Depending on
the magnitude of the voltage, some percentage of the electrons tunnel through the interface, and can be collected and measured as a
current at a collector attached to the other side of the sample. The voltage from the STM tip is then varied, allowing for
measurement of the barrier height. The barrier height is defined as the threshold at which electrons will cross the interface and are
measurable as a current in the far collector. At a metal/n-type semiconductor interface this is the difference between the conduction
band minimum and the Fermi level. At a metal/p-type semiconductor interface this is the difference between the valence band
maximum of the semiconductor and the metal Fermi level. If the voltage is less than the barrier height, no electrons will cross the
interface and the collector will read zero. If the voltage is greater than the barrier height, useful information can be gathered about
the magnitude of the current at the collector as opposed to the initial voltage.

8.3.7 https://chem.libretexts.org/@go/page/55919
Figure 8.3.12 Diagram of a STM/BEEM system. The tip is maintained at the tunneling voltage, V, and the tunneling current, It =
VI/RF, is held constant by the STM feedback circuit. The sample base layer is grounded and current into the semiconductor is
measured by a virtual ground current amplifier.
Samples are prepared from semiconductor wafers by chemical oxide growth-strip cycles, ending with the growth of a protective
oxide layer. Immediately prior to imaging the sample is spin-etched in an inert environment to remove oxides of oxides and then
transferred directly to the ultra-high vacuum without air exposure. The BEEM apparatus itself is operated in a glove box under
inert atmosphere and shielded from light.
Nearly any type of semiconductor interface can be imaged with BEEM. This includes both simple binary interfaces such as Au/n-
Si(100) and more chemically complex interfaces such as Au/n-GaAs(100), such as seen in Figure 8.3.13.

Figure 8.3.13 Images of Au/n-GaAs(100) layer (image area 510 Å x 390 Å) showing (a) the topography of the Au surface and (b)
the BEEM grey-scale interface image. Image adapted from M. H. Hecht, L. D. Bell, W. J. Kaiser, and F. J. Grunthaner, Appl. Phys.
Lett., 1989, 55, 780.
Limitations
Expected barrier height matters a great deal in the desired setup of the BEEM apparatus. If it is necessary to measure small
collector currents, such as with an interface of high-barrier-height, a high-gain, low-noise current preamplifier can be added to the
system. If the interface is of low-barrier-height, the BEEM apparatus can be operated at very low temperatures, accomplished by
immersion of the STM tip in liquid nitrogen and enclosure of the BEEM apparatus in a nitrogen-purged glove box.
Photon STM
Photon scanning tunneling microscopy (PSTM) measures light to determine more information about characteristic sample
topography. It has primarily been used as a technique to measure the electromagnetic interaction of two metallic objects in close
proximity to one another and biological samples, which are both difficult to measure using many other common surface analysis
techniques.
Device Setup and Sample Preparation
This technique works by measuring the tunneling of photons to an optical tip. The source of these photons is the evanescent field
generated by the total internal reflection (TIR) of a light beam from the surface of the sample (Figure 8.3.14). This field is
characteristic of the sample material on the TIR surface, and can be measured by a sharpened optical fiber probe tip where the light
intensity is converted to an electrical signal (Figure 8.3.15). Much like conventional STM, the force of this electrical signal
modifies the location of the tip in relation to the sample. By mapping these modifications across the entire sample, the topography
can be determined to a very accurate degree as well as allowing for calculations of polarization, emission direction and emission
time.

8.3.8 https://chem.libretexts.org/@go/page/55919
Figure 8.3.14 A schematic of a PSTM system

Figure 8.3.15 A TIR light beam generates an evanescent field which is modulated by the sample. A sharpened fiber optic probe tip
receives light from the evanescent field and spatial variations in evanescent field intensity form the basis for imaging.
In PSTM, the vertical resolution is governed only by the noise, as opposed to conventional STM where the vertical resolution is
limited by the tip dimensions. Therefore, this technique provides advantages over more conventional STM apparatus for samples
where subwavelength resolution in the vertical dimension is a critical measurement, including fractal metal colloid clusters,
nanostructured materials and simple organic molecules.
Samples are prepared by placement on a quartz or glass slide coupled to the TIR face of a triangular prism containing a laser beam,
making the sample surface into the TIR surface (Figure 8.3.16). The optical fiber probe tips are constructed from UV grade quartz
optical fibers by etching in HF acid to have nominal end diameters of 200 nm or less and resemble either a truncated cone or a
paraboloid of revolution (Figure 8.3.16).

Figure 8.3.16 Possible optical fiber tip configurations: (a) truncated cone and (b) paraboloid of rotation.
PSTM shows much promise in the imaging of biological materials due to the increase in vertical resolution and the ability to
measure a sample within a liquid environment with a high index TIR substrate and probe tip. This would provide much more
detailed information about small organisms than is currently available.
Limitations
The majority of the limitations in this technique come from the materials and construction of the optical fibers and the prism used
in the sample collection. The sample needs to be kept at low temperatures, typically around 100K, for the duration of the imaging
and therefore cannot decompose or be otherwise negatively impacted by drastic temperature changes.
Conclusion
Scanning tunneling microscopy can provide a great deal of information into the topography of a sample when used without
adaptations, but with adaptations, the information gained is nearly limitless. Depending on the likely properties of your sample
surface, SP-STM, BEEM and PSTM can provide much more accurate topographical pictures than conventional forms of STM
(Table 8.3.2). All of these adaptations to STM have their limitations and all work within relatively specialized categories and
subsets of substances, but they are very strong tools that are constantly improving to provide more useful information about
materials to the nanometer scale.

8.3.9 https://chem.libretexts.org/@go/page/55919
Scanning Transmission Electron Microscope- Electron Energy Loss Spectroscopy (STEM-EELS)
History
STEM-EELS is a terminology abbreviation for scanning transmission electron microscopy (STEM) coupled with electron energy
loss spectroscopy (EELS). It works by combining two instruments, obtaining an image through STEM and applying EELS to
detect signals on the specific selected area of the image. Therefore, it can be applied for many research, such as characterizing
morphology, detecting different elements, and different valence state. The first STEM was built by Baron Manfred von Arden
(Figure 8.3.17) in around 1983, since it was just the prototype of STEM, it was not as good as transmission electron microscopy
(TEM) by that time. Development of STEM was stagnant until the field emission gun was invented by Albert Crewe (Figure
8.3.18) in 1970s; he also came with the idea of annular dark field detector to detect atoms. In 1997, its resolution increased to 1.9

Å, and further increased to 1.36 Å in 2000. 4D STEM-EELS was developed recently, and this type of 4D STEM-EELS has high
brightness STEM equipped with a high acquisition rate EELS detector, and a rotation holder. The rotation holder plays quite an
important role to achieve this 4D aim, because it makes observation of the sample in 360° possible, the sample could be rotated to
acquire the sample’s thickness. High acquisition rate EELS enables this instrument the acquisition of the pixel spectrum in a few
minutes.

Figure 8.3.17 German physicist and inventor Baron Manfred von Arden (1907–1997).

Figure 8.3.18 British physicist Albert Crewe (1927–2009).


Basics of STEM-EELS
Interaction between Electrons and Sample
When electrons interact with the samples, the interaction between those two can be classified into two types, namely, elastic and
inelastic interactions (Figure 8.3.19). In the elastic interaction, if electrons do not interact with the sample and pass through it, these
electrons will contribute to the direct beam. The direct beam can be applied in STEM. In another case, electrons’ moving direction
in the sample is guided by the Coulombic force; the strength of the force is decided by charge and the distance between electrons
and the core. In both cases, these is no energy transfer from electrons to the samples, that’s the reason why it is called elastic
interaction. In inelastic interaction, energy transfers from incident electrons to the samples, thereby, losing energy. The lost energy
can be measured and how many electrons amounted to this energy can also be measured, and these data yield the electron energy
loss spectrum (EELS).

8.3.10 https://chem.libretexts.org/@go/page/55919
Figure 8.3.19 Demonstration of interaction between sample and electrons. Adapted from
http://www.microscopy.ethz.ch/downlo...teractions.pdf

How do TEM, STEM and STEM-EELS work?


In transmission electron microscopy (TEM), a beam of electrons is emitted from tungsten source and then accelerated by
electromagnetic field. Then with the aid of lens condenser, the beam will focus on and pass through the sample. Finally, the
electrons will be detected by a charge-coupled device (CCD) and produce images, Figure 8.3.20. STEM works differently from
TEM, the electron beam focuses on a specific spot of the sample and then raster scans the sample pixel by pixel, the detector will
collect the transmitted electrons and visualize the sample. Moreover, STEM-EELS allows to analyze these electrons, the
transmitted electrons could be characterized by adding a magnetic prism, the more energy the electrons lose, the more they will be
deflected. Therefore, STEM-EELS can be used to characterize the chemical properties of thin samples.

Figure 8.3.20 Scheme of TEM, STEM and STEM-EELS experiments. Adapted from http://toutestquantique.fr/en/scanning-
electron/.

Principles of STEM-EELS
A brief illustration of STEM-EELS is displayed in Figure 8.3.21. The electron source provides electrons, and it usually comes from
a tungsten source located in a strong electrical field. The electron field will provide electrons with high energy. The condenser and
the object lens also promote electrons forming into a fine probe and then raster scanning the specimen. The diameter of the probe
will influence STEM’s spatial resolution, which is caused by the lens aberrations. Lens aberration results from the refraction
difference between light rays striking the edge and center point of the lens, and it also can happen when the light rays pass through
with different energy. Base on this, an aberration corrector is applied to increase the objective aperture, and the incident probe will
converge and increase the resolution, then promote sensitivity to single atoms. For the annular electron detector, the installment

8.3.11 https://chem.libretexts.org/@go/page/55919
sequence of detectors is a bright field detector, a dark field detector and a high angle annular dark field detector. Bright field
detector detects the direct beam that transmits through the specimen. Annular dark field detector collects the scattered electrons,
which only go through at an aperture. The advantage of this is that it will not influence the EELS to detect signals from direct
beam. High angle annular dark field detector collects electrons which are Rutherford scattering (elastic scattering of charged
electrons), and its signal intensity is related with the square of atomic number (Z). So, it is also named as Z-contrast image. The
unique point about STEM in acquiring image is that the pixels in image are obtained in a point by point mode by scanning the
probe. EELS analysis is based on the energy loss of the transmitted electrons, so the thickness of the specimen will influence the
detecting signal. In other words, if the specimen is too thick, the intensity of plasmon signal will decrease and may cause difficulty
distinguishing these signals from the background.

Figure 8.3.21 Schematic representation of STEM-EELS.


Typical features of EELS Spectra
As shown in Figure 8.3.22, a significant peak appears at energy zero in EELS spectra and is therefore called zero-loss peak. Zero-
loss peak represents the electrons which undergo elastic scattering during the interaction with specimen. Zero-loss peak can be used
to determine the thickness of specimen according to 8.3.4, where t stands for the thickness, λinel is inelastic mean free path, It
stands for the total intensity of the spectrum and IZLP is the intensity of zero loss peak.

Figure 8.3.22 Typical features of EELS spectra. Adapted from http://www.mardre.com/homepage/mic/t...ls/sld001.html.


t  =  λinel  ln[ It / IZLP ] (8.3.4)

The low loss region is also called valence EELS. In this region,valence electrons will be excited to the conduction band. Valence
EELS can provide the information about band structure, bandgap, and optical properties. In the low loss region, plasmon peak is
the most important. Plasmon is a phenomenon originates from the collective oscillation of weakly bound electrons. Thickness of
the sample will influence the plasmon peak. The incident electrons will go through inelastic scattering several times when they
interact with a very thick sample, and then result in convoluted plasmon peaks. It is also the reason why STEM-EELS favors
sample with low thickness (usually less than 100 nm).
The high loss region is characterized by the rapidly increasing intensity with a gradually falling, which called ionization edge. The
onset of ionization edges equals to the energy that inner shell electron needs to be excited from the ground state to the lowest
unoccupied state. The amount of energy is unique for different shells and elements. Thus, this information will help to understand
the bonding, valence state, composition and coordination information.
Energy resolution affects the signal to background ratio in the low loss region and is used to evaluate EELS spectrum. Energy
resolution is based on the full width at half maximum of zero-loss peak.
Background signal in the core-loss region is caused by plasmon peaks and core-loss edges, and can be described by the following
power law, 8.3.5, where IBG stands for the background signal, E is the energy loss, A is the scaling constant and r is the slope

8.3.12 https://chem.libretexts.org/@go/page/55919
exponent:
−r
IBG   =  AE (8.3.5)

Therefore, when quantification the spectra data, the background signal can be removed by fitting pre-edge region with the above-
mentioned equation and extrapolating it to the post-edge region.
Advantages and Disadvantages of STEM-EELS
STEM-EELS has advantages over other instruments, such as the acquisition of high resolution of images. For example, the
operation of TEM on samples sometimes result in blurring image and low contrast because of chromatic aberration. STEM-EELS
equipped with aberration corrector, will help to reduce the chromatic aberration and obtain high quality image even at atomic
resolution. It is very direct and convenient to understand the electron distributions on surface and bonding information. STEM-
EELS also has the advantages in controlling the spread of energy. So, it becomes much easier to study the ionization edge of
different material.
Even though STEM-EELS does bring a lot of convenience for research in atomic level, it still has limitations to overcome. One of
the main limitation of STEM-EELS is controlling the thickness of the sample. As discussed above, EELS detects the energy loss of
electrons when they interact with samples and the specimen, then the thickness of samples will impact on the energy lost detection.
Simplify, if the sample is too thick, then most of the electrons will interact with the sample, signal to background ratio and edge
visibility will decrease. Thus, it will be hard to tell the chemical state of the element. Another limitation is due to EELS needs to
characterize low-loss energy electrons, which high vacuum condition is essential for characterization. To achieve such a high
vacuum environment, high voltage is necessary. STEM-EELS also requires the sample substrates to be conductive and flat.
Application of STEM-EELS

STEM-EELS can be used to detect the size and distribution of nanoparticles on a surface. For example, CoO on MgO catalyst
nanoparticles may be prepared by hydrothermal methods. The size and distribution of nanoparticles will greatly influence the
catalytic properties, and the distribution and morphology change of CoO nanoparticles on MgO is important to understand. Co
L3/L2 ratios display uniformly around 2.9, suggesting that Co2+ dominates the electron state of Co. The results show that the ratios
of O:(Co+Mg) and Mg:(Co+Mg) are not consistence, indicating that these three elements are in a random distribution. STEM-
EELS mapping images results further confirm the non-uniformity of the elemental distribution, consistent with a random
distribution of CoO on the MgO surface (Figure 8.3.23).

Figure 8.3.23 EELS data for a CoO/MgO sample. (a) EELS signal ratio of Co L3/L2, and O and Mg EELS signals relative to
combined Co + Mg signals. (b) STEM image and EELS maps acquired at O K, Co L and Mg K edges. Reproduced from S.
Alayoglu, D. J. Rosenberg, and M. Ahmed, Dalton Trans., 2016, 45, 9932 with permission of The Royal Society of Chemistry.
Figure 8.3.24 shows the K-edge absorption of carbon and transition state information could be concluded. Typical carbon based
materials have the features of the transition state, such that 1s transits to π* state and 1s to σ* states locate at 285 and 292 eV,
respectively. The two-transition state correspond to the electrons in the valence band electrons being excited to conduction state.

8.3.13 https://chem.libretexts.org/@go/page/55919
Epoxy exhibits a sharp peak around 285.3 eV compared to GO and GNPs. Meanwhile, GNPs have the sharpest peak around 292
eV, suggesting the most C atoms in GNPs are in 1s to σ* state. Even though GO is in oxidation state, part of its carbon still behaves
1s transits to π*.

Figure 8.3.24 EELS spectrum of graphene nanoplatelets (GNPs), graphene oxide (GO) in comparison with an epoxide resin.
Reprinted with permission from Y. Liu, A. L. Hamon, P. Haghi-Ashtiani, T. Reiss, B. Fan, D. He, and J. Bai, ACS Appl. Mater.
Inter., 2016, 8, 34151). Copyright (2017) American Chemical Society.
The annular dark filed (ADF) mode of STEM provides information about atomic number of the elements in a sample. For example,
the ADF image of La1.2Sr1.8Mn2O7 (Figure 8.3.25 a and b) along [010] direction shows bright spots and dark spots, and even for
bright spots (p and r), they display different levels of brightness. This phenomenon is caused by the difference in atomic numbers.
Bright spots are La and Sr, respectively. Dark spots are Mn elements. O is too light to show on the image. EELS result shows the
core-loss edge of La, Mn and O (Figure 8.3.25 c), but the researchers did not give information on core-loss edge of Sr, Sr has N2,3
edge at 29 eV and L3 edge at 1930 eV and L2 edge at 2010 eV.

Figure 8.3.25 (a) Crystal structure of La1.2Sr21.8Mn2O7, (b) the ADF image of the specimen observed along the [010] direction ,
and (c) STEM-EELS data of La1.2Sr21.8Mn2O7 obtained from rectangular area of (b) and the blue area equals the core-loss of
each element. Reproduced from K. Kimoto, T. Asaka, T. Nagai, M. Saito, Y. Matsui, K. Ishizuka, Nature, 2007, 450, 702.
Copyright © 2007, Rights Managed by Nature Publishing Group

8.3: Scanning Tunneling Microscopy is shared under a CC BY 4.0 license and was authored, remixed, and/or curated by Pavan M. V. Raja &
Andrew R. Barron via source content that was edited to conform to the style and standards of the LibreTexts platform; a detailed edit history is
available upon request.

8.3.14 https://chem.libretexts.org/@go/page/55919
8.4: Spectroscopic Characterization of Nanoparticles
Magnetic force microscopy (MFM) is a natural extension of scanning tunneling microscopy (STM), whereby both the physical
topology of a sample surface and the magnetic topology may be seen. Scanning tunneling microscopy was developed in 1982 by
Gerd Binnig and Heinrich Rohrer, and the two shared the 1986 Nobel prize for their innovation. Binnig later went on to develop the
first atomic force microscope (AFM) along with Calvin Quate and Christoph Gerber (Figure 8.4.1). Magnetic force microscopy
was not far behind, with the first report of its use in 1987 by Yves Martin and H. Kumar Wickramasinge (Figure 8.4.2). An AFM
with a magnetic tip was used to perform these early experiments, which proved to be useful in imaging both static and dynamic
magnetic fields.

Figure 8.4.1 Photograph of German physicist Gerd Binnig (left) and Swiss physicist Heinrich Rohrer (right). Reproduced with
permission from “The Scanning Tunneling Microscope.” Nobelprize.org. Nobel Media AB, 2017. Copyright Nobel Media AB
2017.

Figure 8.4.2 H. Kumar Wickramasinge, now a professor at the University of California, Irvine. Reproduced from “H. Kumar
Wickramasinge.” UCI Samueli, University of California, Irvine. Copyright The Henry Samueli School of Engineering, 2017.
Nobelprize.org. Nobel Media AB, 2017. Copyright Nobel Media AB 2017.
MFM, AFM, and STM all have similar instrumental setups, all of which are based on the early scanning tunneling microscopes. In
essence, STM uses a very small conductive tip attached to a piezoelectric cylinder to carefully scan across a small sample space.
The electrostatic forces between the conducting sample and tip are measured, and the output is a picture that shows the surface of
the sample. AFM and MFM are essentially derivative types of STM, which explains why a typical MFM device is very similar to
an STM, with a piezoelectric driver and magnetized tip as seen in Figure 8.4.3 and Figure 8.4.4.

8.4.1 https://chem.libretexts.org/@go/page/55920
Figure 8.4.3 This image shows a typical MFM setup. Reproduced with permission from A. Méndez-Vilas, Modern research and
educational topics in microscopy. Vol. 2: Applications in physical/chemical sciences, techniques. Formatex, Badajoz (2007)..
Copyright: FORMATEX 2007.

Figure 8.4.4 Illustration of an MFM tip on the instrument cantilever.


One may notice that this MFM instrument very closely resembles an atomic force microscope, and this is for good reason. The
simplest MFM instruments are no more than AFM instruments with a magnetic tip. The differences between AFM and MFM lie in
the data collected and its processing. Where AFM gives topological data through tapping, noncontact, or contact mode, MFM gives
both topological (tapping) and magnetic topological (non-contact) data through a two-scan process known as interleave scanning.
The relationships between basic STM, AFM, and MFM are summarized in Table 8.4.1.
Table 8.4.1 A summary of the capabilities of MFM, SPM, and AFM instrumentation.
Techniques Samples Qualities Observed Modes Benefits Limitations

Resolution depends on
Electrostatic
tip size; different tips
interactions; magnetic
Magnetic and for various
Any film or powder forces/domains; van
MFM Tapping; non-contact physical properties; applications;
surface; magnetic der Waals'
high resolution complicated data
interactions;
processing and
topology; morphology
analysis

Simplest instrumental Resolution depends on


Only conductive Topology; Constant height;
STM setup; many tip size; tips wear out
surfaces morphology constant current
variations easily; rare technique

Common,
Resolution depends on
Any film or powder Particle size; Tapping;contact; non- standardized; often do
AFM tip size; easy to break
surface topology; morphology contact not need special tip;
tips; slow process
ease of data analysis

Data Collection
Interleave scanning, also known as two-pass scanning, is a process typically used in an MFM experiment. The magnetized tip is
first passed across the sample in tapping mode, similar to an AFM experiment, and this gives the surface topology of the sample.
Then, a second scan is taken in non-contact mode, where the magnetic force exerted on the tip by the sample is measured. These
two types of scans are shown in Figure 8.4.5.

8.4.2 https://chem.libretexts.org/@go/page/55920
Figure 8.4.5 Interleave (two-pass) scanning across a sample surface

In non-contact mode (also called dynamic or AC mode), the magnetic force gradient from the sample affects the resonance
frequency of the MFM cantilever, and can be measured in three different ways.
Phase detection: the phase difference between the oscillation of the cantilever and piezoelectric source is measured
Amplitude detection: the changes in the cantilever’s oscillations are measured
Frequency modulation: the piezoelectric source’s oscillation frequency is changed to maintain a 90° phase lag between the
cantilever and the piezoelectric actuator. The frequency change needed for the lag is measured.
Regardless of the method used in determining the magnetic force gradient from the sample, a MFM interleave scan will always
give the user information about both the surface and magnetic topology of the sample. A typical sample size is 100x100 μm, and
the entire sample is scanned by rastering from one line to another. In this way, the MFM data processor can compose an image of
the surface by combining lines of data from either the surface or magnetic scan. The output of an MFM scan is two images, one
showing the surface and the other showing magnetic qualities of the sample. An idealized example is shown in Figure 8.4.6.

Figure 8.4.6 Idealized images of a mixture of ferromagnetic and non-ferromagnetic nanoparticles from MFM.

Types of MFM Tips


Any suitable magnetic material or coating can be used to make an MFM tip. Some of the most commonly used standard tips are
coated with FeNi, CoCr, and NiP, while many research applications call for individualized tips such as carbon nanotubes. The
resolution of the end image in MFM is dependent directly on the size of the tip, therefore MFM tips must come to a sharp point on
the angstrom scale in order to function at high resolution. This leads to tips being costly, an issue exacerbated by the fact that
coatings are often soft or brittle, leading to wear and tear. The best materials for MFM tips, therefore, depend on the desired
resolution and application. For example, a high coercivity coating such as CoCr may be favored for analyzing bulk or strongly
magnetic samples, whereas a low coercivity material such as FeNi might be preferred for more fine and sensitive applications.

Data Output and Applications


From an MFM scan, the product is a 2D scan of the sample surface, whether this be the physical or magnetic topographical image.
Importantly, the resolution depends on the size of the tip of the probe; the smaller the probe, the higher the number of data points
per square micrometer and therefore the resolution of the resulting image. MFM can be extremely useful in determining the
properties of new materials, as in Figure 8.4.7, or in analyzing already known materials’ magnetic landscapes. This makes MFM
particularly useful for the analysis of hard drives. As people store more and more information on magnetic storage devices, higher
storage capacities need to be developed and emergency backup procedures for this data must be developed. MFM is an ideal
procedure for characterizing the fine magnetic surfaces of hard drives for use in research and development, and also can show the
magnetic surfaces of already-used hard drives for data recovery in the event of a hard drive malfunction. This is useful both in
forensics and in researching new magnetic storage materials.

8.4.3 https://chem.libretexts.org/@go/page/55920
Figure 8.4.7 Images of Fe40Ni38Mo4B18 ribbons from MFM. Left images: surface topography. Right images: magnetic
topography. Reproduced with permission from I. García, N. Iturriza, J. José del Val, H. Grande, J. A. Pomposo, and J. González, J.
Magn. Magn. Mater., 2010, 13, 1822. Copyright: Elsevier (2010).

MFM has also found applications on the frontiers of research, most notably in the field of Spintronics. In general, Spintronics is the
study of the spin and magnetic moment of solid-state materials, and the manipulation of these properties to create novel electronic
devices. One example of this is quantum computing, which is promising as a fast and efficient alternative to traditional transistor-
based computing. With regards to Spintronics, MFM can be used to characterize non-homogenous magnetic materials and unique
samples such as dilute magnetic semiconductors (DMS). This is useful for research in magnetic storage such as MRAM,
semiconductors , and magnetoresistive materials.

MFM for Characterization of Magnetic Storage Devices


In device manufacturing, the smoothness and/or roughness of the magnetic coatings of hard drive disks is significant in their ability
to operate. Smoother coatings provide a low magnetic noise level, but stick to read/write heads, whereas rough surfaces have the
opposite qualities. Therefore, fine tuning not only of the magnetic properties but the surface qualities of a given magnetic film is
extremely important in the development of new hard drive technology. Magnetic force microscopy allows the manufacturers of
hard drives to analyze disks for magnetic and surface topology, making it easier to control the quality of drives and determine
which materials are suitable for further research. Industrial competition for higher bit density (bits per square millimeter), which
means faster processing and increased storage capability, means that MFM is very important for characterizing films to very high
resolution.

Conclusion
Magnetic force microscopy is a powerful surface technique used to deduce both the magnetic and surface topology of a given
sample. In general, MFM offers high resolution, which depends on the size of the tip, and straightforward data once processed. The
images outputted by the MFM raster scan are clear and show structural and magnetic features of a 100x100 μm square of the given
sample. This information can be used not only to examine surface properties, morphology, and particle size, but also to determine
the bit density of hard drives, features of magnetic computing materials, and identify exotic magnetic phenomena at the atomic
level. As MFM evolves, thinner and thinner magnetic tips are being fabricated to finer applications, such as in the use of carbon
nanotubes as tips to give high atomic resolution in MFM images. The customizability of magnetic coatings and tips, as well as the
use of AFM equipment for MFM, make MFM an important technique in the electronics industry, making it possible to see
magnetic domains and structures that otherwise would remain hidden.

8.4.4 https://chem.libretexts.org/@go/page/55920
8.4: Spectroscopic Characterization of Nanoparticles is shared under a CC BY 4.0 license and was authored, remixed, and/or curated by Pavan M.
V. Raja & Andrew R. Barron via source content that was edited to conform to the style and standards of the LibreTexts platform; a detailed edit
history is available upon request.

8.4.5 https://chem.libretexts.org/@go/page/55920
8.5: Using UV-Vis for the detection and characterization of silicon quantum dots
Using UV-vis for the Detection and Characterization of Silicon Quantum Dots
What are Quantum Dots?
Quantum dots (QDs) are small semiconductor nanoparticles generally composed of two elements that have extremely high
quantum efficiencies when light is shined on them. The most common quantum dots are CdSe, PbS, and ZnSe, but there are many
many other varieties of these particles that contain other elements as well. QDs can also be made of just three elements or just one
element such as silicon.
Synthesis of Silicon Quantum Dots
Silicon quantum dots are synthesized in inverse micelles. SiCl4 is reduced using a two fold excess of LiAlH4 (Figure 8.5.1). After
the silicon has been fully reduced and the excess reducing agent quenched, the particles are capped with hydrogens and are
hydrophobic. A platinum catalyzed ligand exchange of hydrogen for allylamine will produce hydrophilic particles (Figure 8.5.2).
All reactions in making these particles are extremely air sensitive, and silica is formed readily, so the reactions should be performed
in a highly controlled atmosphere, such as a glove box. The particles are then washed in DMF, and finally filtered and stored in
deionized water. This will allow the Si QDs to be pure in water, and the particles are ready for analysis. This technique yields Si
QDs of 1 - 2 nm in size.

Figure 8.5.1 A schematic representation of the inverse micelle used for the synthesis of Si QDs.

Figure 8.5.2 Conversion of hydrophobic Si QDs to hydrophillic Si QDs. Adapted from J. H. Warner, A. Hoshino, K. Yamamoto,
and R. D. Tilley, Angew. Chem., Int. Ed., 2005, 44, 4550. Copyright: American Chemical Society (2005).
Sample Preparation of Silicon Quantum Dots
The reported absorbtion wavelength for 1 - 2 nm Si QDs absorb is 300 nm. With the hydrophobic Si QDs, UV-vis absorbance
analysis in toluene does not yield an acceptable spectrum because the UV-vis absorbance cutoff is 287 nm, which is very close to
300 nm for the peaks to be resolvable. A better hydrophobic solvent would be hexanes. All measurements of these particles would
require a quartz cuvette since the glass aborbance cutoff (300 nm) is exactly where the particles would be observed. Hydrophilic
substituted particles do not need to be transferred to another solvent because water’s absorbance cutoff is much lower. There is
usually a slight impurity of DMF in the water due to residue on the particles after drying. If there is a DMF peak in the spectrum
with the Si QDs the wavelengths are far enough apart to be resolved.
What Information can be Obtained from UV-Visible Spectra?
Quantum dots are especially interesting when it comes to UV-vis spectroscopy because the size of the quantum dot can be
determined from the position of the absorbtion peak in the UV-vis spectrum. Quantum dots absorb different wavelengths depending
on the size of the particles (e.g., Figure 8.5.3). Many calibration curves would need to be done to determine the exact size and
concentration of the quantum dots, but it is entirely possible and very useful to be able to determine size and concentration of
quantum dots in this way since other ways of determining size are much more expensive and extensive (electron microscopy is
most widely used for this data).

8.5.1 https://chem.libretexts.org/@go/page/55921
Figure 8.5.3 Absorbance of different sized CdSe QDs. Reprinted with permission from C. B. Murray, D. J. Norris, and M. G.
Bawendi, J. Am. Chem. Soc., 1993, 115, 8706. Copyright: American Chemical Society (1993).
An example of silicon quantum dot data can be seen in Figure 8.5.4. The wider the absorbance peak is, the less monodispersed the
sample is.

Figure 8.5.4 UV-vis absorbance spectrum of 1 - 2 nm Si QDs with a DMF reference spectrum.
Why is Knowing the Size of Quantum Dots Important?
Different size (different excitation) quantum dots can be used for different applications. The absorbance of the QDs can also reveal
how monodispersed the sample is; more monodispersity in a sample is better and more useful in future applications. Silicon
quantum dots in particular are currently being researched for making more efficient solar cells. The monodispersity of these
quantum dots is particularly important for getting optimal absorbance of photons from the sun or other light source. Different sized
quantum dots will absorb light differently, and a more exact energy absorption is important in the efficiency of solar cells. UV-vis
absorbance is a quick, easy, and cheap way to determine the monodispersity of the silicon quantum dot sample. The peak width of
the absorbance data can give that information. The other important information for future applications is to get an idea about the
size of the quantum dots. Different size QDs absorb at different wavelengths; therefore, specific size Si QDs will be required for
different cells in tandem solar cells.

UV-Visible Spectrocopy of Noble Metal Nanoparticles


Noble metal nanoparticles have been used for centuries to color stained glass windows and provide many opportunities for novel
sensing and optical technologies due to their intense scattering (deflection) and absorption of light. One of the most interesting and
important properties of noble metal nanoparticles is their localized surface plasmon resonance (LSPR). The LSPR of noble metal
nanoparticles arises when photons of a certain frequency induce the collective oscillation of conduction electrons on the
nanoparticles’ surface. This causes selective photon absorption, efficient scattering, and enhanced electromagnetic field strength
around the nanoparticles. More information about the properties and potential applications of noble metal nanoparticles can be
found in Silver Nanoparticles: A Case Study in Cutting Edge Research
Synthesis of Noble Metal Nanoparticles
Noble metal nanoparticles can be synthesized via the reduction of metal salts. Spherical metal nanoparticle “seeds” are first
synthesized by reducing metal salts in water with a strong reducing agent such as sodium borohydride (Figure 8.5.5). The seeds are
then "capped" to prevent aggregation with a surface group such as citrate (Figure 8.5.5).

8.5.2 https://chem.libretexts.org/@go/page/55921
Figure 8.5.5 Synthesis reaction of citrate-capped silver nanoparticle seeds.
Adjusting the Geometry of Metal Nanoparticles
After small nanoparticle seeds have been synthesized, the seeds can be grown into nanoparticles of various sizes and shapes. Seeds
are added to a solution of additional metal salt and a structure-directing agent, and are then reduced with a weak reducing agent
such as ascorbic acid (see Figure 8.5.6). The structure-directing agent will determine the geometry of the nanoparticles produced.
For example, cetyltrimethylammonium bromide (CTAB) is often used to produce nanorods (Figure 8.5.6).

Figure 8.5.6 Synthesis reaction of cetyltrimethylammonium bromide (CTAB)-capped silver nanorods.


Assemblies of Metal Nanoparticles
Once synthesized, noble metal nanoparticles can be assembled into various higher-order nanostructures. Nanoparticle dimers,
linear chains of two nanoparticles, can be assembled using a linker molecule that binds the two nanoparticles together (Figure
8.5.7). Less-organized nanoparticle assemblies can be formed through the addition of counterions. Counterions react with the

surface groups on nanoparticles, causing the nanoparticles to be stripped of their protective surface coating and inducing their
aggregation.

Figure 8.5.7 TEM images of a gold nanosphere (A) a gold nanorod (B) and a gold nanosphere dimer (C).
UV-Visible Spectroscopy of Noble Metal Nanoparticles
UV-visible absorbance spectroscopy is a powerful tool for detecting noble metal nanoparticles, because the LSPR of metal
nanoparticles allows for highly selective absorption of photons. UV-visible absorbance spectroscopy can also be used to detect
various factors that affect the LSPR of noble metal nanoparticles. More information about the theory and instrumentation of UV-
visible absorbance spectroscopy can be found in the section related to UV-Vis Spectroscopy.
Mie Theory
Mie theory, a theory that describes the interaction of light with a homogenous sphere, can be used to predict the UV-visible
absorbance spectrum of spherical metallic nanoparticles. One equation that can be obtained using Mie theory is 8.5.1, which
describes the extinction, the sum of absorption and scattering of light, of spherical nanoparticles. In 8.5.1, E(λ) is the extinction,
NA is the areal density of the nanoparticles, a is the radius of the nanoparticles, εm is the dielectric constant of the environment
surrounding the nanoparticles, λ is the wavelength of the incident light, and εr and εi are the real and imaginary parts of the
nanoparticles’ dielectric function. From this relation, we can see that the UV-visible absorbance spectrum of a solution of
nanoparticles is dependent on the radius of the nanoparticles, the composition of the nanoparticles, and the environment
surrounding the nanoparticles.
3 3/2
24πNA a εm εi
E(λ)  = [ ] (8.5.1)
2 2
λ ln(10) (εr   +  2 εm )   +  ε
i

8.5.3 https://chem.libretexts.org/@go/page/55921
More Advanced Theoretical Techniques
Mie theory is limited to spherical nanoparticles, but there are other theoretical techniques that can be used to predict the UV-visible
spectrum of more complex noble metal nanostructures. These techniques include surface-based methods such as the generalized
multipole technique and T-matrix method, as well as volume-based techniques such as the discrete dipole approximation and the
finite different time domain method.
Using UV-Vis Spectroscopy to Predict Nanoparticle Geometry

Just as the theoretical techniques described above can use nanoparticle geometry to predict the UV-visible absorbance spectrum of
noble metal nanoparticles, nanoparticles’ UV-visible absorbance spectrum can be used to predict their geometry. As shown in
Figure 8.5.8 below, the UV-visible absorbance spectrum is highly dependent on nanoparticle geometry. The shapes of the two
spectra are quite different despite the two types of nanoparticles having similar dimensions and being composed of the same
material (Figure 8.5.8).

Figure 8.5.8 UV-visible absorbance spectra of 50 nm diameter gold nanospheres (A) and 25 nm diameter, 60 nm length gold
nanorods (B).
Using UV-Visible Spectroscopy to Determine Nanoparticle Aggregation States

The UV-visible absorbance spectrum is also dependent on the aggregation state of the nanoparticles. When nanoparticles are in
close proximity to each other, their plasmons couple, which affects their LSPR and thus their absorption of light. Dimerization of
nanospheres causes a “red shift,” a shift to longer wavelengths, in the UV-visible absorbance spectrum as well as a slight increase
in absorption at higher wavelengths (see Figure 8.5.9). Unlike dimerization, aggregation of nanoparticles causes a decrease in the
intensity of the peak absorbance without shifting the wavelength at which the peak occurs (λmax). Information about the
calculation of λmax can be found in the earlier section about silver nanoparticles. Figure 8.5.9 illustrates the increase in
nanoparticle aggregation with increased salt concentrations based on the decreased absorbance peak intensity.

Figure 8.5.9 UV-visible absorbance spectrum of 50 nm gold nanosphere dimers with a reference spectrum of single gold
nanospheres (A) and UV-visible absorbance spectrum of 50 nm gold nanospheres exposed to various concentrations of NaCl (B).
Using UV-Visible Spectroscopy to Determine Nanoparticle Surface Composition
The λmax of the UV-visible absorbance spectrum of noble metal nanoparticles is highly dependent on the environment surrounding
the nanoparticles. Because of this, shifts in λmax can be used to detect changes in the surface composition of the nanoparticles. One
potential application of this phenomenon is using UV-visible absorbance spectroscopy to detect the binding of biomolecules to the
surface of noble metal nanoparticles. The red shift in the λmax of the UV-visible absorbance spectrum in Figure 8.5.10 below with
the addition of human serum albumin protein indicates that the protein is binding to the surface of the nanoparticles.

8.5.4 https://chem.libretexts.org/@go/page/55921
Figure 8.5.10 UV-visible absorbance spectrum of 50 nm gold nanospheres exposed to human serum albumin protein with a
reference spectrum of nanospheres exposed to deionized water.

Optical Properties of Group 12-16 (II-VI) Semiconductor Nanoparticles


What are Group 12-16 semiconductors?
Semiconductor materials are generally classified on the basis of the periodic table group that their constituent elements belong to.
Thus, Group 12-16 semiconductors, formerly called II-VI semiconductors, are materials whose cations are from the Group 12 and
anions are from Group 16 in the periodic table (Figure 8.5.11). Some examples of Group 12-16 semiconductor materials are
cadmium selenide (CdSe), zinc sulfide (ZnS), cadmium teluride (CdTe), zinc oxide (ZnO), and mercuric selenide (HgSe) among
others.
The new IUPAC (International Union of Pure and Applied Chemistry) convention is being followed in this document, to avoid any
confusion with regard to conventions used earlier. In the old IUPAC convention, Group 12 was known as Group IIB with the roman
numeral ‘II’ referring to the number of electrons in the outer electronic shells and B referring to being on the right part of the table.
However, in the CAS (Chemical Abstracts Service), the alphabet B refers to transition elements as compared to main group
elements, though the roman numeral has the same meaning. Similarly, Group 16 was earlier known as Group VI because all the
elements in this group have 6 valence shell electrons.

Figure 8.5.11 The red box indicates the Group 12 and Group 16 elements in the periodic table.
What are Group 12-16 (II-VI) Semiconductor Nanoparticles?
From the Greek word nanos - meaning "dwarf" this prefix is used in the metric system to mean 10-9 or one billionth
(1/1,000,000,000). Thus a nanometer is 10-9 or one billionth of a meter, and a nanojoule is 10-9 or one billionth of a Joule, etc. A
nanoparticle is ordinarily defined as any particle with at least one of its dimensions in the 1 - 100 nm range.
Nanoscale materials often show behavior which is intermediate between that of a bulk solid and that of an individual molecule or
atom. An inorganic nanocrystal can be imagined to be comprised of a few atoms or molecules. It thus will behave differently from
a single atom; however, it is still smaller than a macroscopic solid, and hence will show different properties. For example, if one
would compare the chemical reactivity of a bulk solid and a nanoparticle, the latter would have a higher reactivity due to a
significant fraction of the total number of atoms being on the surface of the particle. Properties such as boiling point, melting point,
optical properties, chemical stability, electronic properties, etc. are all different in a nanoparticle as compared to its bulk
counterpart. In the case of Group 12-16 semiconductors, this reduction in size from bulk to the nanoscale results in many size
dependent properties such as varying band gap energy, optical and electronic properties.
Optical Properties of Semiconductor Quantum Nanoparticles
In the case of semiconductor nanocrystals, the effect of the size on the optical properties of the particles is very interesting.
Consider a Group 12-16 semiconductor, cadmium selenide (CdSe). A 2 nm sized CdSe crystal has a blue color fluorescence

8.5.5 https://chem.libretexts.org/@go/page/55921
whereas a larger nanocrystal of CdSe of about 6 nm has a dark red fluorescence (Figure 8.5.12). In order to understand the size
dependent optical properties of semiconductor nanoparticles, it is important to know the physics behind what is happening at the
nano level.

Figure 8.5.12 Fluorescing CdSe quantum dots synthesized in a heat transfer liquid of different sizes (M. S. Wong, Rice University).
Energy Levels in a Semiconductor
The electronic structure of any material is given by a solution of Schrödinger equations with boundary conditions, depending on the
physical situation. The electronic structure of a semiconductor (Figure 8.5.13 can be described by the following terms:

Figure 8.5.13 Simplified representation of the energy levels in a bulk semiconductor.

Energy Level
By the solution of Schrödinger’s equations, the electrons in a semiconductor can have only certain allowable energies, which are
associated with energy levels. No electrons can exist in between these levels, or in other words can have energies in between the
allowed energies. In addition, from Pauli’s Exclusion Principle, only 2 electrons with opposite spin can exist at any one energy
level. Thus, the electrons start filling from the lowest energy levels. Greater the number of atoms in a crystal, the difference in
allowable energies become very small, thus the distance between energy levels decreases. However, this distance can never be zero.
For a bulk semiconductor, due to the large number of atoms, the distance between energy levels is very small and for all practical
purpose the energy levels can be described as continuous (Figure 8.5.13).

Band Gap
From the solution of Schrödinger’s equations, there are a set of energies which is not allowable, and thus no energy levels can exist
in this region. This region is called the band gap and is a quantum mechanical phenomenon (Figure 8.5.13). In a bulk
semiconductor the bandgap is fixed; whereas in a quantum dot nanoparticle the bandgap varies with the size of the nanoparticle.

Conduction Band
The conduction band consists of energy levels from the upper edge of the bandgap and higher (Figure 8.5.13). To reach the
conduction band, the electrons in the valence band should have enough energy to cross the band gap. Once the electrons are
excited, they subsequently relax back to the valence band (either radiatively or non-radiatively) followed by a subsequent emission
of radiation. This property is responsible for most of the applications of quantum dots.

8.5.6 https://chem.libretexts.org/@go/page/55921
Exciton and Exciton Bohr Radius
When an electron is excited from the valence band to the conduction band, corresponding to the electron in the conduction band a
hole (absence of electron) is formed in the valence band. This electron pair is called an exciton. Excitons have a natural separation
distance between the electron and hole, which is characteristic of the material. This average distance is called exciton Bohr radius.
In a bulk semiconductor, the size of the crystal is much larger than the exciton Bohr radius and hence the exciton is free to move
throughout the crystal.
Energy Levels in a Quantum Dot Semiconductor
Before understanding the electronic structure of a quantum dot semiconductor, it is important to understand what a quantum dot
nanoparticle is. We earlier studied that a nanoparticle is any particle with one of its dimensions in the 1 - 100 nm. A quantum dot is
a nanoparticle with its diameter on the order of the materials exciton Bohr radius. Quantum dots are typically 2 - 10 nm wide and
approximately consist of 10 to 50 atoms. With this understanding of a quantum dot semiconductor, the electronic structure of a
quantum dot semiconductor can be described by the following terms.

Figure 8.5.14 Energy levels in quantum dot. Allowed optical transitions are shown. Adapted from T. Pradeep, Nano: The
Essentials. Understanding Nanoscience and Nanotechnology, Tata McGraw-Hill, New Delhi (2007).

Quantum Con nement


When the size of the semiconductor crystal becomes comparable or smaller than the exciton Bohr radius, the quantum dots are in a
state of quantum confinement. As a result of quantum confinement, the energy levels in a quantum dot are discrete (Figure 8.5.14
as opposed to being continuous in a bulk crystal (Figure 8.5.13).

Discrete Energy Levels


In materials that have small number of atoms and are considered as quantum confined, the energy levels are separated by an
appreciable amount of energy such that they are not continuous, but are discrete (see Figure 8.5.13). The energy associated with an
electron (equivalent to conduction band energy level) is given by is given by 8.5.2, where h is the Planck’s constant, me is the
effective mass of electron and n is the quantum number for the conduction band states, and n can take the values 1, 2, 3 and so on.
Similarly, the energy associated with the hole (equivalent to valence band energy level) is given by 8.5.2, where n' is the quantum
number for the valence states, and n' can take the values 1, 2, 3, and so on. The energy increases as one goes higher in the quantum
number. Since the electron mass is much smaller than that of the hole, the electron levels are separated more widely than the hole
levels.
2 2
e
h n
E   =  (8.5.2)
2 2
8π me d

2 ′2
h
h n
E   =  (8.5.3)
2 2
8π mh d

Tunable Band Gap


As seen from 8.5.2 and 8.5.3, the energy levels are affected by the diameter of the semiconductor particles. If the diameter is very
small, since the energy is dependent on inverse of diameter squared, the energy levels of the upper edge of the band gap (lowest

8.5.7 https://chem.libretexts.org/@go/page/55921
conduction band level) and lower edge of the band gap (highest valence band level) change significantly with the diameter of the
particle and the effective mass of the electron and the hole, resulting in a size dependent tunable band gap. This also results in the
discretization of the energy levels.
Qualitatively, this can be understood in the following way. In a bulk semiconductor, the addition or removal of an atom is
insignificant compared to the size of the bulk semiconductor, which consists of a large number of atoms. The large size of bulk
semiconductors makes the changes in band gap so negligible on the addition of an atom, that it is considered as a fixed band gap. In
a quantum dot, addition of an atom does make a difference, resulting in the tunability of band gap.

UV-Visible Absorbance
Due to the presence of discrete energy levels in a QD, there is a widening of the energy gap between the highest occupied
electronic states and the lowest unoccupied states as compared to the bulk material. As a consequence, the optical properties of the
semiconductor nanoparticles also become size dependent.
The minimum energy required to create an exciton is the defined by the band gap of the material, i.e., the energy required to excite
an electron from the highest level of valence energy states to the lowest level of the conduction energy states. For a quantum dot,
the bandgap varies with the size of the particle. From 8.5.2 and 8.5.3, it can be inferred that the band gap becomes higher as the
particle becomes smaller. This means that for a smaller particle, the energy required for an electron to get excited is higher. The
relation between energy and wavelength is given by 8.5.4, where h is the Planck’s constant, c is the speed of light, λ is the
wavelength of light. Therefore, from 8.5.4 to cross a bandgap of greater energy, shorter wavelengths are absorbed, i.e., a blue shift
is seen.
E  =  hc (8.5.4)

For Group 12-16 semiconductors, the bandgap energy falls in the UV-visible range. That is ultraviolet light or visible light can be
used to excite an electron from the ground valence states to the excited conduction states. In a bulk semiconductor the band gap is
fixed, and the energy states are continuous. This results in a rather uniform absorption spectrum (Figure 8.5.15 a).

Figure 8.5.15 UV-vis spectra of (a) bulk CdS and (b) 4 nm CdS. Adapted from G. Kickelbick, Hybrid Materials: Synthesis,
Characterization and Applications, Wiley-VCH, Weinheim (2007).
In the case of Group 12-16 quantum dots, since the bandgap can be changed with the size, these materials can absorb over a range
of wavelengths. The peaks seen in the absorption spectrum (Figure 8.5.15 b) orrespond to the optical transitions between the
electron and hole levels. The minimum energy and thus the maximum wavelength peak corresponds to the first exciton peak or the
energy for an electron to get excited from the highest valence state to the lowest conduction state. The quantum dot will not absorb
wavelengths of energy longer than this wavelength. This is known as the absorption onset.
Fluorescence
Fluorescence is the emission of electromagnetic radiation in the form of light by a material that has absorbed a photon. When a
semiconductor quantum dot (QD) absorbs a photon/energy equal to or greater than its band gap, the electrons in the QD’s get
excited to the conduction state. This excited state is however not stable. The electron can relax back to its ground state by either
emitting a photon or lose energy via heat losses. These processes can be divided into two categories – radiative decay and non-
radiative decay. Radiative decay is the loss of energy through the emission of a photon or radiation. Non-radiative decay involves
the loss of heat through lattice vibrations and this usually occurs when the energy difference between the levels is small. Non-
radiative decay occurs much faster than radiative decay.
Usually the electron relaxes to the ground state through a combination of both radiative and non-radiative decays. The electron
moves quickly through the conduction energy levels through small non-radiative decays and the final transition across the band gap
is via a radiative decay. Large nonradiative decays don’t occur across the band gap because the crystal structure can’t withstand
large vibrations without breaking the bonds of the crystal. Since some of the energy is lost through the non-radiative decay, the

8.5.8 https://chem.libretexts.org/@go/page/55921
energy of the emitted photon, through the radiative decay, is much lesser than the absorbed energy. As a result the wavelength of
the emitted photon or fluorescence is longer than the wavelength of absorbed light. This energy difference is called the Stokes shift.
Due this Stokes shift, the emission peak corresponding to the absorption band edge peak is shifted towards a higher wavelength
(lower energy), i.e., Figure 8.5.16.

Figure 8.5.16 Absorption spectra (a) and emission spectra (b) of CdSe tetrapod.
Intensity of emission versus wavelength is a bell-shaped Gaussian curve. As long as the excitation wavelength is shorter than the
absorption onset, the maximum emission wavelength is independent of the excitation wavelength. Figure 8.5.16 shows a combined
absorption and emission spectrum for a typical CdSe tetrapod.
Factors Affecting the Optical Properties of NPs
There are various factors that affect the absorption and emission spectra for Group 12-16 semiconductor quantum crystals.
Fluorescence is much more sensitive to the background, environment, presence of traps and the surface of the QDs than UV-visible
absorption. Some of the major factors influencing the optical properties of quantum nanoparticles include:
Surface defects, imperfection of lattice, surface charges- The surface defects and imperfections in the lattice structure of
semiconductor quantum dots occur in the form of unsatisfied valencies. Similar to surface charges, unsatisfied valencies provide
a sink for the charge carriers, resulting in unwanted recombinations.
Surface ligands- The presence of surface ligands is another factor that affects the optical properties. If the surface ligand
coverage is a 100%, there is a smaller chance of surface recombinations to occur.
Solvent polarity- The polarity of solvents is very important for the optical properties of the nanoparticles. If the quantum dots
are prepared in organic solvent and have an organic surface ligand, the more non-polar the solvent, the particles are more
dispersed. This reduces the loss of electrons through recombinations again, since when particles come in close proximity to
each other, increases the non-radiative decay events.
Applications of the Optical Properties of Group 12-16 Semiconductor NPs
The size dependent optical properties of NP’s have many applications from biomedical applications to solar cell technology, from
photocatalysis to chemical sensing. Most of these applications use the following unique properties.
For applications in the field of nanoelectronics, the sizes of the quantum dots can be tuned to be comparable to the scattering
lengths, reducing the scattering rate and hence, the signal to noise ratio. For Group 12-16 QDs to be used in the field of solar cells,
the bandgap of the particles can be tuned so as to form absorb energy over a large range of the solar spectrum, resulting in more
number of excitons and hence more electricity. Since the nanoparticles are so small, most of the atoms are on the surface. Thus, the
surface to volume ratio is very large for the quantum dots. In addition to a high surface to volume ratio, the Group 12-16 QDs
respond to light energy. Thus quantum dots have very good photocatalytic properties. Quantum dots show fluorescence properties,
and emit visible light when excited. This property can be used for applications as biomarkers. These quantum dots can be tagged to
drugs to monitor the path of the drugs. Specially shaped Group 12-16 nanoparticles such as hollow shells can be used as drug
delivery agents. Another use for the fluorescence properties of Group 12-16 semiconductor QDs is in color-changing paints, which
can change colors according to the light source used.

Characterization of Group 12-16 (II-VI) Semiconductor Nanoparticles by UV-Visible Spectroscopy


Quantum dots (QDs) as a general term refer to nanocrystals of semiconductor materials, in which the size of the particles are
comparable to the natural characteristic separation of an electron-hole pair, otherwise known as the exciton Bohr radius of the
material. When the size of the semiconductor nanocrystal becomes this small, the electronic structure of the crystal is governed by
the laws of quantum physics. Very small Group 12-16 (II-VI) semiconductor nanoparticle quantum dots, in the order of 2 - 10 nm,
exhibit significantly different optical and electronic properties from their bulk counterparts. The characterization of size dependent
optical properties of Group 12-16 semiconductor particles provide a lot of qualitative and quantitative information about them –

8.5.9 https://chem.libretexts.org/@go/page/55921
size, quantum yield, monodispersity, shape and presence of surface defects. A combination of information from both the UV-visible
absorption and fluorescence, complete the analysis of the optical properties.
UV-Visible Absorbance Spectroscopy
Absorption spectroscopy, in general, refers to characterization techniques that measure the absorption of radiation by a material, as
a function of the wavelength. Depending on the source of light used, absorption spectroscopy can be broadly divided into infrared
and UV-visible spectroscopy. The band gap of Group 12-16 semiconductors is in the UV-visible region. This means the minimum
energy required to excite an electron from the valence states of the Group 12-16 semiconductor QDs to its conduction states, lies in
the UV-visible region. This is also a reason why most of the Group 12-16 semiconductor quantum dot solutions are colored.
This technique is complementary to fluorescence spectroscopy, in that UV-visible spectroscopy measures electronic transitions
from the ground state to the excited state, whereas fluorescence deals with the transitions from the excited state to the ground state.
In order to characterize the optical properties of a quantum dot, it is important to characterize the sample with both these techniques
In quantum dots, due to the very small number of atoms, the addition or removal of one atom to the molecule changes the
electronic structure of the quantum dot dramatically. Taking advantage of this property in Group 12-16 semiconductor quantum
dots, it is possible to change the band gap of the material by just changing the size of the quantum dot. A quantum dot can absorb
energy in the form of light over a range of wavelengths, to excite an electron from the ground state to its excited state. The
minimum energy that is required to excite an electron, is dependent on the band gap of the quantum dot. Thus, by making accurate
measurements of light absorption at different wavelengths in the ultraviolet and visible spectrum, a correlation can be made
between the band gap and size of the quantum dot. Group 12-16 semiconductor quantum dots are of particular interest, since their
band gap lies in the visible region of the solar spectrum.
The UV-visible absorbance spectroscopy is a characterization technique in which the absorbance of the material is studied as a
function of wavelength. The visible region of the spectrum is in the wavelength range of 380 nm (violet) to 740 nm (red) and the
near ultraviolet region extends to wavelengths of about 200 nm. The UV-visible spectrophotometer analyzes over the wavelength
range 200 – 900 nm.
When the Group 12-16 semiconductor nanocrystals are exposed to light having an energy that matches a possible electronic
transition as dictated by laws of quantum physics, the light is absorbed and an exciton pair is formed. The UV-visible
spectrophotometer records the wavelength at which the absorption occurs along with the intensity of the absorption at each
wavelength. This is recorded in a graph of absorbance of the nanocrystal versus wavelength.
Instrumentation
A working schematic of the UV-visible spectrophotometer is show in Figure 8.5.17.

Figure 8.5.17 Schematic of UV-visible spectrophotometer.


The Light Source
Since it is a UV-vis spectrophotometer, the light source (Figure 8.5.17) needs to cover the entire visible and the near ultra-violet
region (200 - 900 nm). Since it is not possible to get this range of wavelengths from a single lamp, a combination of a deuterium
lamp for the UV region of the spectrum and tungsten or halogen lamp for the visible region is used. This output is then sent through
a diffraction grating as shown in the schematic.
The Diffraction Grating and the Slit
The beam of light from the visible and/or UV light source is then separated into its component wavelengths (like a very efficient
prism) by a diffraction grating (Figure 8.5.17). Following the slit is a slit that sends a monochromatic beam into the next section of

8.5.10 https://chem.libretexts.org/@go/page/55921
the spectrophotometer.
Rotating Discs
Light from the slit then falls onto a rotating disc (Figure 8.5.17). Each disc consists of different segments – an opaque black
section, a transparent section and a mirrored section. If the light hits the transparent section, it will go straight through the sample
cell, get reflected by a mirror, hits the mirrored section of a second rotating disc, and then collected by the detector. Else if the light
hits the mirrored section, gets reflected by a mirror, passes through the reference cell, hits the transparent section of a second
rotating disc and then collected by the detector. Finally if the light hits the black opaque section, it is blocked and no light passes
through the instrument, thus enabling the system to make corrections for any current generated by the detector in the absence of
light.
Sample Cell, Reference Cell and Sample Preparation
For liquid samples, a square cross section tube sealed at one end is used. The choice of cuvette depends on the following factors:
Type of solvent - For aqueous samples, specially designed rectangular quartz, glass or plastic cuvettes are used. For organic
samples glass and quartz cuvettes are used.
Excitation wavelength – Depending on the size and thus, bandgap of the 12-16 semiconductor nanoparticles, different
excitation wavelengths of light are used. Depending on the excitation wavelength, different materials are used
Table 8.5.1 Cuvette materials and their wavelengths.
Cuvette Wavelength (nm)

Visible only glass 380-780

Visible only plastic 380-780

UV plastic 220-780

Quartz 200-900

Cost - Plastic cuvettes are the least expensive and can be discarded after use. Though quartz cuvettes have the maximum utility,
they are the most expensive, and need to reused. Generally, disposable plastic cuvettes are used when speed is more important
than high accuracy.
The best cuvettes need to be very clear and have no impurities that might affect the spectroscopic reading. Defects on the cuvette
such as scratches, can scatter light and hence should be avoided. Some cuvettes are clear only on two sides, and can be used in the
UV-Visible spectrophotometer, but cannot be used for fluorescence spectroscopy measurements. For Group 12-16 semiconductor
nanoparticles prepared in organic solvents, the quartz cuvette is chosen.
In the sample cell the quantum dots are dispersed in a solvent, whereas in the reference cell the pure solvent is taken. It is important
that the sample be very dilute (maximum first exciton absorbance should not exceed 1 au) and the solvent is not UV-visible active.
For these measurements, it is required that the solvent does not have characteristic absorption or emission in the region of interest.
Solution phase experiments are preferred, though it is possible to measure the spectra in the solid state also using thin films,
powders, etc. The instrumentation for solid state UV-visible absorption spectroscopy is slightly different from the solution phase
experiments and is beyond the scope of discussion.
Detector
Detector converts the light into a current signal that is read by a computer. Higher the current signal, greater is the intensity of the
light. The computer then calculates the absorbance using the in 8.5.5, here A denotes absorbance, I is sample cell intensity and Io is
the reference cell intensity.

A  =  log10 (I0 /I ) (8.5.5)

The following cases are possible:


Where I < I0 and A < 0. This usually occurs when the solvent absorbs in the wavelength range. Preferably the solvent should be
changed, to get an accurate reading for actual reference cell intensity.
Where I = I0 and A= 0. This occurs when pure solvent is put in both reference and sample cells. This test should always be done
before testing the sample, to check for the cleanliness of the cuvettes.
When A = 1. This occurs when 90% or the light at a particular wavelength has been absorbed, which means that only 10% is seen

8.5.11 https://chem.libretexts.org/@go/page/55921
at the detector. So I0/I becomes 100/10 = 10. Log10 of 10 is 1.
When A > 1. This occurs in extreme case where more than 90% of the light is absorbed.
Output
The output is the form of a plot of absorbance against wavelength, e.g., Figure 8.5.18.

Figure 8.5.18 Representative UV-visble absorption spectrum for CdSe tetrapods.


Beer-Lambert Law
In order to make comparisons between different samples, it is important that all the factors affecting absorbance should be constant
except the sample itself.
Effect of Concentration on Absorbance
The extent of absorption depends on the number of absorbing nanoparticles or in other words the concentration of the sample. If it
is a reasonably concentrated solution, it will have a high absorbance since there are lots of nanoparticles to interact with the light.
Similarly in an extremely dilute solution, the absorbance is very low. In order to compare two solutions, it is important that we
should make some allowance for the concentration.
Effect of Container Shape
Even if we had the same concentration of solutions, if we compare two solutions – one in a rectagular shaped container (e.g.,
Figure 8.5.19) so that light travelled 1 cm through it and the other in which the light travelled 100 cm through it, the absorbance
would be different. This is because if the length the light travelled is greater, it means that the light interacted with more number of
nanocrystals, and thus has a higher absorbance. Again, in order to compare two solutions, it is important that we should make some
allowance for the concentration.

Figure 8.5.19 A typical rectangular cuvette for UV-visible spectroscopy.


The Law
The Beer-Lambert law addresses the effect of concentration and container shape as shown in 8.5.5, 8.5.6 and 8.5.7, where A
denotes absorbance; ε is the molar absorptivity or molar absorption coefficient; l is the path length of light (in cm); and c is the
concentration of the solution (mol/dm3).

log10 (I0 /I )  =  εlc (8.5.6)

A  =  εlc (8.5.7)

Molar Absorptivity
From the Beer-Lambert law, the molar absorptivity 'ε' can be expressed as shown in 8.5.8.
c  =  A/lε (8.5.8)

Molar absorptivity corrects for the variation in concentration and length of the solution that the light passes through. It is the value
of absorbance when light passes through 1 cm of a 1 mol/dm3 solution.
Limitations of Beer-Lambert Law
The linearity of the Beer-Lambert law is limited by chemical and instrumental factors.

8.5.12 https://chem.libretexts.org/@go/page/55921
At high concentrations (> 0.01 M), the relation between absorptivity coefficient and absorbance is no longer linear. This is due
to the electrostatic interactions between the quantum dots in close proximity.
If the concentration of the solution is high, another effect that is seen is the scattering of light from the large number of quantum
dots.
The spectrophotometer performs calculations assuming that the refractive index of the solvent does not change significantly
with the presence of the quantum dots. This assumption only works at low concentrations of the analyte (quantum dots).
Presence of stray light.
Analysis of Data
The data obtained from the spectrophotometer is a plot of absorbance as a function of wavelength. Quantitative and qualitative data
can be obtained by analysing this information.
Quantitative Information
The band gap of the semiconductor quantum dots can be tuned with the size of the particles. The minimum energy for an electron
to get excited from the ground state is the energy to cross the band gap. In an absorption spectra, this is given by the first exciton
peak at the maximum wavelength (λmax).

Size of the Quantum Dots


The size of quantum dots can be approximated corresponding to the first exciton peak wavelength. Emperical relationships have
been determined relating the diameter of the quantum dot to the wavelength of the first exciton peak. The Group 12-16
semiconductor quantum dots that they studied were cadmium selenide (CdSe), cadmium telluride (CdTe) and cadmium sulfide
(CdS). The empirical relationships are determined by fitting experimental data of absorbance versus wavelength of known sizes of
particles. The empirical equations determined are given for CdTe, CdSe, and CdS in 8.5.9, 8.5.10 and 8.5.11 respectively, where D
is the diameter and λ is the wavelength corresponding to the first exciton peak. For example, if the first exciton peak of a CdSe
quantum dot is 500 nm, the corresponding diameter of the quantum dot is 2.345 nm and for a wavelength of 609 nm, the
corresponding diameter is 5.008 nm.
−7 3 −3 2
D  =  (9.8127 x 10 )λ   −  (1.7147 x 10 )λ   +  (1.0064)λ  −  194.84 (8.5.9)

−7 3 −3 2
D  =  (1.6122 x 10 )λ   −  (2.6575 x 10 )λ   +  (1.6242)λ  −  41.57 (8.5.10)

−7 3 −3 2
D  =  (−6.6521 x 10 )λ   −  (1.9577 x 10 )λ   +  (9.2352)λ  −  13.29 (8.5.11)

Concentration of Sample
Using the Beer-Lambert law, it is possible to calculate the concentration of the sample if the molar absorptivity for the sample is
known. The molar absorptivity can be calculated by recording the absorbance of a standard solution of 1 mol/dm3 concentration in
a standard cuvette where the light travels a constant distance of 1 cm. Once the molar absorptivity and the absorbance of the sample
are known, with the length the light travels being fixed, it is possible to determine the concentration of the sample solution.
Empirical equations can be determined by fitting experimental data of extinction coefficient per mole of Group 12-16
semiconductor quantum dots, at 250 °C, to the diameter of the quantum dot, 8.5.12, 8.5.13, and 8.5.14.
2.12
ε  =  10043xD (8.5.12)

2.65
ε  =  5857 x D (8.5.13)

2.3
ε  =  21536 x D (8.5.14)

The concentration of the quantum dots can then be then be determined by using the Beer Lambert law as given by 8.5.8.
Qualitative Information

Apart from quantitative data such as the size of the quantum dots and concentration of the quantum dots, a lot of qualitative
information can be derived from the absorption spectra.

8.5.13 https://chem.libretexts.org/@go/page/55921
Size Distribution
If there is a very narrow size distribution, the first exciton peak will be very sharp (Figure 8.5.20). his is because due to the narrow
size distribution, the differences in band gap between different sized particles will be very small and hence most of the electrons
will get excited over a smaller range of wavelengths. In addition, if there is a narrow size distribution, the higher exciton peaks are
also seen clearly.

Figure 8.5.20 Narrow emission spectra (a) and broad emission spectra (b) of CdSe QDs.

Shapd Particles
In the case of a spherical quantum dot, in all dimensions, the particle is quantum confined (Figure 8.5.21). In the case of a nanorod,
whose length is not in the quantum regime, the quantum effects are determined by the width of the nanorod. Similar is the case in
tetrapods or four legged structures. The quantum effects are determined by the thickness of the arms. During the synthesis of the
shaped particles, the thickness of the rod or the arm of the tetrapod does not vary among the different particles, as much as the
length of the rods or arms changes. Since the thickness of the rod or tetrapod is responsible for the quantum effects, the absorption
spectrum of rods and tetrapods has sharper features as compared to a quantum dot. Hence, qualitatively it is possible to differentiate
between quantum dots and other shaped particles.

Figure 8.5.21 Different shaped nanoparticles with the arrows indicating the dimension where quantum confinement effects are
observed.

Crystal Lattice Information


In the case of CdSe semiconductor quantum dots it has been shown that it is possible to estimate the crystal lattice of the quantum
dot from the adsorption spectrum (Figure 8.5.22), and hence determine if the structure is zinc blend or wurtzite.

Figure 8.5.22 Zinc blende and wurtzite CdSe absorption spectra. Adapted from J. Jasieniak, C. Bullen, J. van Embden, and P.
Mulvaney, J. Phys. Chem. B, 2005, 109, 20665.
UV-Vis Absorption Spectra of Group 12-16 Semiconductor Nanoparticles

Cadmium Selenide (CdSe)


Cadmium selenide (CdSe) is one of the most popular Group 12-16 semiconductors. This is mainly because the band gap (712 nm
or 1.74 eV) energy of CdSe. Thus, the nanoparticles of CdSe can be engineered to have a range of band gaps throughout the visible

8.5.14 https://chem.libretexts.org/@go/page/55921
range, corresponding to the major part of the energy that comes from the solar spectrum. This property of CdSe along with its
fluorescing properties is used in a variety of applications such as solar cells and light emitting diodes. Though cadmium and
selenium are known carcinogens, the harmful biological effects of CdSe can be overcome by coating the CdSe with a layer of zinc
sulfide. Thus CdSe, can also be used as bio-markers, drug-delivery agents, paints and other applications.
A typical absorption spectrum of narrow size distribution wurtzite CdSe quantum dot is shown in Figure 8.5.23. A size evolving
absorption spectra is shown in Figure 8.5.24. However, a complete analysis of the sample is possible only by also studying the
fluorescence properties of CdSe.

Figure 8.5.23 Wurtzite CdSe quantum dot. Adapted from X. Zhong, Y. Feng, and Y. Zhang, J. Phys. Chem. C, 2007, 111, 526.

Figure 8.5.24 Size evolving absorption spectra of CdSe quantum dots.

Cadmium Telluride (CdTe)


Cadmium telluride has a band gap of 1.44 eV (860 nm) and as such it absorbs in the infrared region. Like CdSe, the sizes of CdTe
can be engineered to have different band edges and thus, different absorption spectra as a function of wavelength. A typical CdTe
spectra is shown in Figure 8.5.25. Due to the small bandgap energy of CdTe, it can be used in tandem with CdSe to absorb in a
greater part of the solar spectrum.

Figure 8.5.25 Size evolving absorption spectra of CdTe quantum dots from 3 nm to 7 nm. Adapted from C. Qi-Fan, W. Wen-Xing,
G. Ying-Xin, L. Meng-Ying, X. Shu-Kun and Z. Xiu-Juan, Chin. J. Anal. Chem., 2007, 35, 135.

Other Group 12-16 Semiconductor Systems


Table 8.5.1 shows the bulk band gap of other Group 12-16 semiconductor systems. The band gap of ZnS falls in the UV region,
while those of ZnSe, CdS, and ZnTe fall in the visible region.
Table 8.5.1 Bulk band gaps of different Group 12-16 semiconductors.
Material Band Gap (eV) Wavelength (nm)

ZnS 3.61 343.2

ZnSe 2.69 460.5

8.5.15 https://chem.libretexts.org/@go/page/55921
ZnTe 2.39 518.4

CdS 2.49 497.5

CdSe 1.74 712.1

CsTe 1.44 860.3

Heterostructures of Group 12-16 Semiconductor Systems


It is often desirable to have a combination of two Group 12-16 semiconductor system quantum heterostructures of different shapes
like dots and tetrapods, for applications in solar cells, bio-markers, etc. Some of the most interesting systems are ZnS shell-CdSe
core systems, such as the CdSe/CdS rods and tetrapods.
Figure 8.5.26 shows a typical absorption spectra of CdSe-ZnS core-shell system. This system is important because of the
drastically improved fluorescence properties because of the addition of a wide band gap ZnS shell than the core CdSe. In addition
with a ZnS shell, CdSe becomes bio-compatible.

Figure 8.5.26 Absorption spectra of CdSe core, ZnS shell. Adapted from C. Qing-Zhu, P. Wang, X. Wang and Y. Li, Nanoscale
Res. Lett., 2008, 3, 213.
A CdSe seed, CdS arm nanorods system is also interesting. Combining CdSe and CdS in a single nanostructure creates a material
with a mixed dimensionality where holes are confined to CdSe while electrons can move freely between CdSe and CdS phases.

Optical Characterization of Group 12-16 (II-VI) Semiconductor Nanoparticles by Fluorescence


Spectroscopy
Group 12-16 semiconductor nanocrystals when exposed to light of a particular energy absorb light to excite electrons from the
ground state to the excited state, resulting in the formation of an electron-hole pair (also known as excitons). The excited electrons
relax back to the ground state, mainly through radiative emission of energy in the form of photons.
Quantum dots (QD) refer to nanocrystals of semiconductor materials where the size of the particles is comparable to the natural
characteristic separation of an electron-hole pair, otherwise known as the exciton Bohr radius of the material. In quantum dots, the
phenomenon of emission of photons associated with the transition of electrons from the excited state to the ground state is called
fluorescence.
Fluorescence Spectroscopy
Emission spectroscopy, in general, refers to a characterization technique that measures the emission of radiation by a material that
has been excited. Fluorescence spectroscopy is one type of emission spectroscopy which records the intensity of light radiated from
the material as a function of wavelength. It is a nondestructive characterization technique.
After an electron is excited from the ground state, it needs to relax back to the ground state. This relaxation or loss of energy to
return to the ground state, can be achieved by a combination of non-radiative decay (loss of energy through heat) and radiative
decay (loss of energy through light). Non-radiative decay by vibrational modes typically occurs between energy levels that are
close to each other. Radiative decay by the emission of light occurs when the energy levels are far apart like in the case of the band
gap. This is because loss of energy through vibrational modes across the band gap can result in breaking the bonds of the crystal.
This phenomenon is shown in Figure 8.5.27.

8.5.16 https://chem.libretexts.org/@go/page/55921
Figure 8.5.27 Emission of luminescence photon for Group 12-16 semiconductor quantum dot.
The band gap of Group 12-16 semiconductors is in the UV-visible region. Thus, the wavelength of the emitted light as a result of
radiative decay is also in the visible region, resulting in fascinating fluorescence properties.
A fluorimeter is a device that records the fluorescence intensity as a function of wavelength. The fluorescence quantum yield can
then be calculated by the ratio of photons absorbed to photons emitted by the system. The quantum yield gives the probability of
the excited state getting relaxed via fluorescence rather than by any other non-radiative decay.
Difference between Fluorescence and Phosphorescence

Photoluminescence is the emission of light from any material due to the loss of energy from excited state to ground state. There are
two main types of luminescence – fluorescence and phosphorescence. Fluorescence is a fast decay process, where the emission rate
is around 108 s-1 and the lifetime is around 10-9 - 10-7 s. Fluorescence occurs when the excited state electron has an opposite spin
compared to the ground state electrons. From the laws of quantum mechanics, this is an allowed transition, and occurs rapidly by
emission of a photon. Fluorescence disappears as soon as the exciting light source is removed.
Phosphorescence is the emission of light, in which the excited state electron has the same spin orientation as the ground state
electron. This transition is a forbidden one and hence the emission rates are slow (103 - 100 s-1). So the phosphorescence lifetimes
are longer, typically seconds to several minutes, while the excited phosphors slowly returned to the ground state. Phosphorescence
is still seen, even after the exciting light source is removed. Group 12-16 semiconductor quantum dots exhibit fluorescence
properties when excited with ultraviolet light.
Instrumentation
The working schematic for the fluorometer is shown in Figure 8.5.28.

Figure 8.5.28 Schematic of fluorometer.


The Light Source
The excitation energy is provided by a light source that can emit wavelengths of light over the ultraviolet and the visible range.
Different light sources can be used as excitation sources such as lasers, xenon arcs and mercury-vapor lamps. The choice of the
light source depends on the sample. A laser source emits light of a high irradiance at a very narrow wavelength interval. This
makes the need for the filter unnecessary, but the wavelength of the laser cannot be altered significantly. The mercury vapor lamp is
a discrete line source. The xenon arc has a continuous emission spectrum between the ranges of 300 - 800 nm.
The Diffraction Grating and Primary Filter
The diffraction grating splits the incoming light source into its component wavelengths (Figure 8.5.29). The monochromator can
then be adjusted to choose with wavelengths to pass through. Following the primary filter, specific wavelengths of light are
irradiated onto the sample.

8.5.17 https://chem.libretexts.org/@go/page/55921
Sample Cell and Sample Preparation
A proportion of the light from the primary filter is absorbed by the sample. After the sample gets excited, the fluorescent substance
returns to the ground state, by emitting a longer wavelength of light in all directions (Figure 8.5.28). Some of this light passes
through a secondary filter. For liquid samples, a square cross section tube sealed at one end and all four sides clear, is used as a
sample cell. The choice of cuvette depends on three factors:
1. Type of Solvent - For aqueous samples, specially designed rectangular quartz, glass or plastic cuvettes are used. For organic
samples glass and quartz cuvettes are used.
2. Excitation Wavelength - Depending on the size and thus, bandgap of the Group 12-16 semiconductor nanoparticles, different
excitation wavelengths of light are used. Depending on the excitation wavelength, different materials are used (Table 8.5.2).
Table 8.5.2 Cuvette Materials and their wavelengths.
Cuvette Wavelength (nm)

Visible only glass 380-780

Visible only plastic 380-780

UV plastic 220-780

Quartz 200-900

3. Cost - Plastic cuvettes are the least expensive and can be discarded after use. Though quartz cuvettes have the maximum utility,
they are the most expensive, and need to reused. Generally, disposable plastic cuvettes are used when speed is more important than
high accuracy.

Figure 8.5.29 A typical cuvette for fluorescence spectroscopy.


The cuvettes have a 1 cm path length for the light (Figure 8.5.29). The best cuvettes need to be very clear and have no impurities
that might affect the spectroscopic reading. Defects on the cuvette, such as scratches, can scatter light and hence should be avoided.
Since the specifications of a cuvette are the same for both, the UV-visible spectrophotometer and fluorimeter, the same cuvette that
is used to measure absorbance can be used to measure the fluorescence. For Group 12-16 semiconductor nanoparticles preparted in
organic solvents, the clear four sided quartz cuvette is used. The sample solution should be dilute (absorbance <1 au), to avoid very
high signal from the sample to burn out the detector. The solvent used to disperse the nanoparticles should not absorb at the
excitation wavelength.
Secondary Filter
The secondary filter is placed at a 90° angle (Figure 8.5.28) to the original light path to minimize the risk of transmitted or
reflected incident light reaching the detector. Also this minimizes the amount of stray light, and results in a better signal-to-noise
ratio. From the secondary filter, wavelengths specific to the sample are passed onto the detector.
Detector
The detector can either be single-channeled or multichanneled (Figure 8.5.28). The single-channeled detector can only detect the
intensity of one wavelength at a time, while the multichanneled detects the intensity at all wavelengths simultaneously, making the
emission monochromator or filter unnecessary. The different types of detectors have both advantages and disadvantages.
Output
The output is the form of a plot of intensity of emitted light as a function of wavelength as shown in Figure 8.5.30).

8.5.18 https://chem.libretexts.org/@go/page/55921
Figure 8.5.30 Emission spectra of CdSe quantum dot.
Analysis of Data
The data obtained from fluorimeter is a plot of fluorescence intensity as a function of wavelength. Quantitative and qualitative data
can be obtained by analysing this information.
Quantitative Information
From the fluorescence intensity versus wavelength data, the quantum yield (ΦF) of the sample can be determined. Quantum yield is
a measure of the ratio of the photons absorbed with respect to the photons emitted. It is important for the application of Group 12-
16 semiconductor quantum dots using their fluorescence properties, for e.g., bio-markers.
The most well-known method for recording quantum yield is the comparative method which involves the use of well characterized
standard solutions. If a test sample and a standard sample have similar absorbance values at the same excitation wavelength, it can
be assumed that the number of photons being absorbed by both the samples is the same. This means that a ratio of the integrated
fluorescence intensities of the test and standard sample measured at the same excitation wavelength will give a ratio of quantum
yields. Since the quantum yield of the standard solution is known, the quantum yield for the unknown sample can be calculated.
A plot of integrated fluorescence intensity versus absorbance at the excitation wavelength is shown in Figure 8.5.31. The slope of
the graphs shown in Figure 8.5.31 are proportional to the quantum yield of the different examples. Quantum yield is then
calculated using 8.5.15, where subscripts ST denotes standard sample and X denotes the test sample; QY is the quantum yield; RI
is the refractive index of the solvent.

Figure 8.5.31 Integrated fluoresncene intensity as a function of absorbance.


2
QYX slop eX (RIX )
  =  (8.5.15)
2
QYST slop eST (RIST )

Take the example of Figure 8.5.32. If the same solvent is used in both the sample and the standard solution, the ratio of quantum
yields of the sample to the standard is given by 8.5.16. If the quantum yield of the standard is known to 0.95, then the quantum
yield of the test sample is 0.523 or 52.3%.
QYX 1.41
  =  (8.5.16)
QYST 2.56

The assumption used in the comparative method is valid only in the Beer-Lambert law linear regime. Beer-Lambert law states that
absorbance is directly proportional to the path length of light travelled within the sample, and concentration of the sample. The
factors that affect the quantum yield measurements are the following:
Concentration - Low concentrations should be used (absorbance < 0.2 a.u.) to avoid effects such as self quenching.
Solvent - It is important to take into account the solvents used for the test and standard solutions. If the solvents used for both
are the same then the comparison is trivial. However, if the solvents in the test and standard solutions are different, this

8.5.19 https://chem.libretexts.org/@go/page/55921
difference needs to be accounted for. This is done by incorporating the solvent refractive indices in the ratio calculation.
Standard Samples - The standard samples should be characterized thoroughly. In addition, the standard sample used should
absorb at the excitation wavelength of the test sample.
Sample Preparation - It is important that the cuvettes used are clean, scratch free and clear on all four sides. The solvents used
must be of spectroscopic grade and should not absorb in the wavelength range.
Slit Width - The slit widths for all measurements must be kept constant.
The quantum yield of the Group 12-16 semiconductor nanoparticles are affected by many factors such as the following.
Surface Defects - The surface defects of semiconductor quantum dots occur in the form of unsatisfied valencies. Thus resulting
in unwanted recombinations. These unwanted recombinations reduce the loss of energy through radiative decay, and thus
reducing the fluorescence.
Surface Ligands - If the surface ligand coverage is a 100%, there is a smaller chance of surface recombinations to occur.
Solvent Polarity - If the solvent and the ligand have similar solvent polarities, the nanoparticles are more dispersed, reducing
the loss of electrons through recombinations.
Qualitative Information
Apart from quantum yield information, the relationship between intensity of fluorescence emission and wavelength, other useful
qualitative information such as size distribution, shape of the particle and presence of surface defects can be obtained.
As shown in Figure 8.5.32, the shape of the plot of intensity versus wavelength is a Gaussian distribution. In Figure 8.5.32, the full
width at half maximum (FWHM) is given by the difference between the two extreme values of the wavelength at which the
photoluminescence intensity is equal to half its maximum value. From the full width half max (FWHM) of the fluorescence
intensity Gaussian distribution, it is possible to determine qualitatively the size distribution of the sample. For a Group 12-16
quantum dot sample if the FWHM is greater than 30, the system is very polydisperse and has a large size distribution. It is desirable
for all practical applications for the FWHM to be lesser than 30.

Figure 8.5.32 Emission spectra of CdSe QDs showing the full width half maximum (FWHM).
From the FWHM of the emission spectra, it is also possible to qualitatively get an idea if the particles are spherical or shaped.
During the synthesis of the shaped particles, the thickness of the rod or the arm of the tetrapod does not vary among the different
particles, as much as the length of the rods or arms changes. The thickness of the arm or rod is responsible for the quantum effects
in shaped particles. In the case of quantum dots, the particle is quantum confined in all dimensions. Thus, any size distribution
during the synthesis of quantum dots greatly affects the emission spectra. As a result the FWHM of rods and tetrapods is much
smaller as compared to a quantum dot. Hence, qualitatively it is possible to differentiate between quantum dots and other shaped
particles.
Another indication of branched structures is the decrease in the intensity of fluorescence peaks. Quantum dots have very high
fluorescence values as compared to branched particles, since they are quantum confined in all dimensions as compared to just 1 or
2 dimensions in the case of branched particles.
Fluorescence Spectra of Different Group 12-16 Semiconductor Nanoparticles

The emission spectra of all Group 12-16 semiconductor nanoparticles are Gaussian curves as shown in Figure 8.5.30 and Figure
8.5.32. The only difference between them is the band gap energy, and hence each of the Group 12-16 semiconductor nanoparticles

fluoresce over different ranges of wavelengths.

Cadmium Selenide

8.5.20 https://chem.libretexts.org/@go/page/55921
Since its bulk band gap (1.74 eV, 712 nm) falls in the visible region cadmium Selenide (CdSe) is used in various applications such
as solar cells, light emitting diodes, etc. Size evolving emission spectra of cadmium selenide is shown in Figure 8.5.33. Different
sized CdSe particles have different colored fluorescence spectra. Since cadmium and selenide are known carcinogens and being
nanoparticles are easily absorbed into the human body, there is some concern regarding these particles. However, CdSe coated with
ZnS can overcome all the harmful biological effects, making cadmium selenide nanoparticles one of the most popular 12-16
semiconductor nanoparticle.

Figure 8.5.33 Size evolving CdSe emission spectra. Adapted from http://www.physics.mq.edu.au.
A combination of the absorbance and emission spectra is shown in Figure 8.5.34 for four different sized particles emitting green,
yellow, orange, and red fluorescence.

Figure 8.5.34 Absorption and emission spectra of CdSe quantum dots. Adapted from G. Schmid, Nanoparticles: From Theory to
Application, Wiley-VCH, Weinham (2004).

Cadmium Telluride
Cadmium Telluride (CdTe) has a band gap of 1.44 eV and thus absorbs in the infra red region. The size evolving CdTe emission
spectra is shown in Figure 8.5.35.

Figure 8.5.35 Size evolution spectra of CdTe quantum dots.

Adding Shells to QDs


Capping a core quantum dot with a semiconductor material with a wider bandgap than the core, reduces the nonradiative
recombination and results in brighter fluorescence emission. Quantum yields are affected by the presences of free surface charges,
surface defects and crystal defects, which results in unwanted recombinations. The addition of a shell reduces the nonradiative
transitions and majority of the electrons relax radiatively to the valence band. In addition, the shell also overcomes some of the
surface defects.

8.5.21 https://chem.libretexts.org/@go/page/55921
For the CdSe-core/ZnS-shell systems exhibit much higher quantum yield as compared to core CdSe quantum dots as seen in Figure
8.5.36.

Figure 8.5.36 Emission spectra of core CdSe only and CdSe-core/ZnS-shell.

Band Gap Measurement


In solid state physics a band gap also called an energy gap, is an energy range in an ideal solid where no electron states can exist.
As shown in Figure 8.5.37 for an insulator or semiconductor the band gap generally refers to the energy difference between the top
of the valence band and the bottom of the conduction band. This is equivalent to the energy required to free an outer shell electron
from its orbit about the nucleus to become a mobile charge carrier, able to move freely within the solid material.

Figure 8.5.37 Schematic explanation of band gap.


The band gap is a major factor determining the electrical conductivity of a solid. Substances with large band gaps are generally
insulators (i.e., dielectric), those with smaller band gaps are semiconductors, while conductors either have very small band gaps or
no band gap (because the valence and conduction bands overlap as shown in Figure 8.5.38.

Figure 8.5.38 Schematic representation of the band gap difference in a metal, a semiconductor and an insulator.
The theory of bands in solids is one of the most important steps in the comprehension of the properties of solid matter. The
existence of a forbidden energy gap in semiconductors is an essential concept in order to be able to explain the physics of
semiconductor devices. For example, the magnitude of the bad gap of solid determines the frequency or wavelength of the light,
which will be adsorbed. Such a value is useful for photocatalysts and for the performance of a dye sensitized solar cell.
Nanocomposites materials are of interest to researchers the world over for various reasons. One driver for such research is the
potential application in next-generation electronic and photonic devices. Particles of a nanometer size exhibit unique properties

8.5.22 https://chem.libretexts.org/@go/page/55921
such as quantum effects, short interface migration distances (and times) for photoinduced holes and electrons in photochemical and
photocatalytic systems, and increased sensitivity in thin film sensors.
Measurement Methods
Electrical measurement method
For a p-n junction, the essential electrical characteristic is that it constitutes a rectifier, which allows the easy flow of a charge in
one direction but restrains the flow in the opposite direction. The voltage-current characteristic of such a device can be described
by the Shockley equation, 8.5.17, in which, I0 is the reverse bias saturation current, q the charge of the electron, k is Boltzmann’s
constant, and T is the temperature in Kelvin.
qV /kT
I   =  I0 (e − 1) (8.5.17)

When the reverse bias is very large, the current I is saturated and equal to I0. This saturation current is the sum of several different
contributions. They are diffusion current, generation current inside the depletion zone, surface leakage effects and tunneling of
carriers between states in the band gap. In a first approximation at a certain condition, I0 can be interpreted as being solely due to
minority carriers accelerated by the depletion zone field plus the applied potential difference. Therefore it can be shown that,
8.5.18, where A is a constant, Eg the energy gap (slightly temperature dependent), and γ an integer depending on the temperature

dependence of the carrier mobility µ.


(3 + γ/2) −Eg (T )/KT
I0   =  AT e (8.5.18)

We can show that γ is defined by the relation by a more advanced treatment, 8.5.19.
2 γ
T μ   =  T (8.5.19)

After substituting the value of I0 given by 8.5.17 into 8.5.18, we take the napierian logarithm of the two sides and multiply them
by kT for large forward bias (qV > 3kT); thus, rearranging, we have 8.5.20.
qV   =  Eg (T )  +  T [k ln(1/A)]  −  (3 + γ/2)klnT (8.5.20)

As InT can be considered as a slowly varying function in the 200 - 400 K interval, therefore for a constant current, I, flowing
through the junction a plot of qV versus the temperature should approximate a straight line, and the intercept of this line with the
qV axis is the required value of the band gap Eg extrapolated to 0 K. Through 8.5.21 instead of qV, we can get a more precise
value of Eg.
q Vc   =  qV   +  (3 + γ/2)klnT (8.5.21)

8.5.20 shows that the value of γ depends on the temperature and µ that is a very complex function of the particular materials,
doping and processing. In the 200 - 400 K range, one can estimate that the variation ΔEg produced by a change of Δγ in the value
of γ is 8.5.22. So a rough value of γ is sufficient for evaluating the correction. By taking the experimental data for the temperature
dependence of the mobility µ, a mean value for γ can be found. Then the band gap energy qV can be determined.
−2
ΔEg   =  10 eV Δγ (8.5.22)

The electrical circuit required for the measurement is very simple and the constant current can be provided by a voltage regulator
mounted as a constant current source (see Figure 8.5.39). The potential difference across the junction can be measured with a
voltmeter. Five temperature baths were used: around 90 °C with hot water, room temperature water, water-ice mixture, ice-salt-
water mixture and mixture of dry ice and acetone. The result for GaAs is shown in Figure 8.5.40. The plot qV corrected (qVc)
versus temperature gives E1 = 1.56±0.02 eV for GaAs. This may be compared with literature value of 1.53 eV.

Figure 8.5.39 Schematic of the constant current source. (Ic = 5V/R). Adapted from Y. Canivez, Eur. J. Phys., 1983, 4, 42.

8.5.23 https://chem.libretexts.org/@go/page/55921
Figure 8.5.40 Plot of corrected voltage versus temperature for GaAs. Adapted from Y. Canivez, Eur. J. Phys., 1983, 4, 42.
Optical Measurement Method
Optical method can be described by using the measurement of a specific example, e.g., hexagonal boron nitride (h-BN, Figure
8.5.41. The UV-visible absorption spectrum was carried out for investigating the optical energy gap of the h-BN film based on its

optically induced transition.

Figure 8.5.41 The structure of hexagonal boron nitride (h-BN).


For this study, a sample of h-BN was first transferred onto an optical quartz plate, and a blank quartz plate was used for the
background as the reference substrate. The following Tauc’s equation was used to determine the optical band gap Eg, 8.5.23, where
ε is the optical absorbance, λ is the wavelength and ω = 2π/λ is the angular frequency of the incident radiation.
2 2
ω ε  =  (hω  −  Eg ) (8.5.23)

As Figure 8.5.42a shows, the absorption spectrum has one sharp absorption peak at 201 - 204 nm. On the basis of Tauc’s
formulation, it is speculated that the plot of ε1/2/λ versus 1/λ should be a straight line at the absorption range. Therefore, the
intersection point with the x axis is 1/λg (λg is defined as the gap wavelength). The optical band gap can be calculated based on Eg)
hc/λg. The plot in Figure 8.5.42b shows ε1/2/λ versus 1/λ curve acquired from the thin h-BN film. For more than 10 layers sample,
he calculated gap wavelength λg is about 223 nm, which corresponds to an optical band gap of 5.56 eV.

8.5.24 https://chem.libretexts.org/@go/page/55921
Figure 8.5.42 Ultraviolet-visible adsorption spectra of h-BN films of various thicknesses taken at room temperature. (a) UV
adsorption spectra of 1L, 5L and thick (>10L) h-BN films. (b) Corresponding plots of ε 1/2/λ versus 1/λ. (c) Calculated optical
band gap for each h-BN films.
Previous theoretical calculations of a single layer of h-BN shows 6 eV band gap as the result. The thickness of h-BN film are 1
layer, 5 layers and thick (>10 layers) h-BN films, the measured gap is about 6.0, 5.8, 5.6 eV, respectively, which is consistent with
the theoretical gap value. For thicker samples, the layer-layer interaction increases the dispersion of the electronic bands and tends
to reduce the gap. From this example, we can see that the band gap is relative to the size of the materials, this is the most important
feature of nano material.

Band Gap Measurements of Quantum Dots


A semiconductor is a material that has unique properties in the way it reacts to electrical current. A semiconductor’s ability to
conduct an electrical current is intermediate between that of an insulator (such as rubber or glass) and a conductor (such as copper).
However, the conductivity of a semiconductor material increases with increasing temperature, a behavior opposite to that of a
metal. Semiconductors may also have a lower resistance to the flow of current in one direction than in the other.
Band Theory

The properties of semiconductors can best be understood by band theory, where the difference between conductors,
semiconductors, and insulators can be understood by increasing separations between a valence band and a conduction band, as
shown in Figure 8.5.43. In semiconductors a small energy gap separates the valence band and the conduction band. This energy
gap is smaller than that of insulators – which is too large for essentially any electrons from the valence band to enter the conduction
band – and larger than that of conductors, where the valence and conduction bands overlap. At 0 K all of the electrons in a
semiconductor lie in the valence band, but at higher temperatures some electrons will have enough energy to be promoted to the
conduction band

Figure 8.5.43 A schematic presentation of band theory, showing the differences in energy separation between valence bands and
conduction bands of insulators, conductors, and semiconductors.

8.5.25 https://chem.libretexts.org/@go/page/55921
Carrier Generation and Recombination
In addition to the band structure of solids, the concept of carrier generation and recombination is very important to the
understanding of semiconducting materials. Carrier generation and recombination is the process by which mobile charge carriers
(electrons and electron holes) are created and eliminated. The valence band in semiconductors is normally very full and its
electrons immobile, resulting in no flow as electrical current. However, if an electron in the valence band acquires enough energy
to reach the conduction band, it can flow freely in the nearly empty conduction band. Furthermore, it will leave behind an electron
hole that can flow as current exactly like a physical charged particle. The energy of an electron-electron hole pair is quantified in
the form of a neutrally-charged quasiparticle called an exciton. For semiconducting materials, there is a characteristic separation
distance between the electron and the hole in an exciton called the exciton Bohr radius. The exciton Bohr radius has large
implications for the properties of quantum dots.
The process by which electrons gain energy and move from the valence to the conduction band is termed carrier generation, while
recombination describes the process by which electrons lose energy and re-occupy the energy state of an electron hole in the
valence band. Carrier generation is accompanied by the absorption of radiation, while recombination is accompanied by the
emission of radiation.
Quantum Dots

In the 1980s, a new nanoscale (~1-10 nm) semiconducting structure was developed that exhibits properties intermediate between
bulk semiconductors and discrete molecules. These semiconducting nanocrystals, called quantum dots, are small enough to be
subjected to quantum effects, which gives them interesting properties and the potential to be useful in a wide-variety of
applications. The most important characteristic of quantum dots (QDs) is that they are highly tunable, meaning that the
optoelectronic properties are dependent on the particles size and shape. As Figure 8.5.44 illustrates, the band gap in a QD is
inversely related to its size, which produces a blue shift in emitted light as the particle size decreases. The highly tunable nature of
QDs result not only from the inverse relationship between band gap size and particle size, but also from the ability to set the size of
QDs and make QDs out of a wide variety of materials. The potential to produce QDs with properties tailored to fulfill a specific
function has produce an enormous amount of interest in quantum dots (see the section on Optical Properties of Group 12-16 (II-VI)
Semiconductor Nanoparticles).

Figure 8.5.44 A picture of different-sized CdSe quantum dots synthesized in a heat transfer liquid (M.S. Wong, Rice University).
Band Gap Measurements of QDs
As previously mentioned, QDs are small enough that quantum effects influence their properties. At sizes under approximately 10
nm, quantum confinement effects dominate the optoelectronic properties of a material. Quantum confinement results from
electrons and electron holes being squeezed into a dimension that approaches a critical quantum measurement, called the exciton
Bohr radius. As explained above, the distance between the electron and the hole within an exciton is called the exciton Bohr radius.
In bulk semiconductors the exciton can move freely in all directions, but when the size of a semiconductor is reduced to only a few
nanometers, quantum confinement effects occur and the band gap properties are changed. Confinement of the exciton in one
dimension produces a quantum well, confinement in two dimensions produces a quantum wire, and confinement in all three
dimensions produces a quantum dot.
Recombination occurs when an electron from a higher energy level relaxes to a lower energy level and recombines with an electron
hole. This process is accompanied by the emission of radiation, which can be measured to give the band gap size of a
semiconductor. The energy of the emitted photon in a recombination process of a QD can be modeled as the sum of the band gap
energy, the confinement energies of the excited electron and the electron hole, and the bound energy of the exciton as show in
8.5.24.

E  =  Ebandgap   +  Econf inement   +  Eexciton (8.5.24)

8.5.26 https://chem.libretexts.org/@go/page/55921
The confinement energy can be modeled as a simple particle in a one-dimensional box problem and the energy levels of the exciton
can be represented as the solutions to the equation at the ground level (n = 1) with the mass replaced by the reduced mass. The
confinement energy is given by 8.5.25, where ħ is the reduced Plank’s constant, µ is the reduced mass, and a is the particle radius.
me and mh are the effective masses of the electron and the hole, respectively.
2 2 2 2
ℏ π 1 1 ℏ π
Econf inement   =   ( + ) =  (8.5.25)
2 2
2a me mh 2μa

The bound exciton energy can be modeled by using the Coulomb interaction between the electron and the positively charged
electron-hole, as shown in 8.5.26. The negative energy is proportional to Rydberg’s energy (Ry) (13.6 eV) and inversely
proportional to the square of the size-dependent dielectric constant, εr. µ and me are the reduced mass and the effective mass of the
electron, respectively.
1 μ

E  =   − Ry +   − Ry (8.5.26)
2
εr m e

Using these models and spectroscopic measurements of the emitted photon energy (E), it is possible to measure the band gap of
QDs.
Photoluminescence Spectroscopy
Photoluminescence (PL) Spectroscopy is perhaps the best way to measure the band gap of QDs. PL spectroscopy is a contactless,
nondestructive method that is extremely useful in measuring the separation between different energy levels. PL spectroscopy works
by directing light onto a sample, where energy is absorbed by electrons in the sample and elevated to a higher energy-state through
a process known as photo-excitation. Photo-excitation produces the electron-electron hole pair. The recombination of the electron-
electron hole pair then occurs with the emission of radiation (light). The energy of the emitted light (photoluminescence) relates to
the difference in energy levels between the lower (ground) electronic state and the higher (excited) electronic state. This amount of
energy is measured by PL spectroscopy to give the band gap size.
PL spectroscopy can be divided in two different categories: fluorescence and phosphorescence. It is fluorescent PL spectroscopy
that is most relevant to QDs. In fluorescent PL spectroscopy, an electron is raised from the ground state to some elevated excited
state. The electron than relaxes (loses energy) to the lowest electronic excited state via a non-radiative process. This non-radiative
relaxation can occur by a variety of mechanisms, but QDs typically dissipate this energy via vibrational relaxation. This form of
relaxation causes vibrations in the material, which effectively heat the QD without emitting light. The electron then decays from
the lowest excited state to the ground state with the emission of light. This means that the energy of light absorbed is greater than
the energy of the light emitted. The process of fluorescence is schematically summarized in the Jablonski diagram in Figure 8.5.45.

Figure 8.5.45 A Jablonski diagram of a fluorescent process.


Instrumentation
A schematic of a basic design for measuring fluorescence is shown in Figure 8.5.46. The requirements for PL spectroscopy are a
source of radiation, a means of selecting a narrow band of radiation, and a detector. Unlike optical absorbance spectroscopy, the
detector must not be placed along the axis of the sample, but rather at 90º to the source. This is done to minimize the intensity of
transmitted source radiation (light scattered by the sample) reaching the detector. Figure 8.5.46 shows two different ways of
selecting the appropriate wavelength for excitation: a monochromator and a filter. In a fluorimeter the excitation and emission
wavelengths are selected using absorbance or interference filters. In a spectrofluorimeterthe excitation and emission wavelengths
are selected by a monochromator.

8.5.27 https://chem.libretexts.org/@go/page/55921
Figure 8.5.46 A schematic representation of a fluorescent spectrometer.
Excitation vs. Emission Spectra
PL spectra can be recorded in two ways: by measuring the intensity of emitted radiation as a function of the excitation wavelength,
or by measuring the emitted radiation as a function of the the emission wavelength. In an excitation spectrum, a fixed wavelength is
used to monitor emission while the excitation wavelength is varied. An excitation spectrum is nearly identical to a sample’s
absorbance spectrum. In an emission spectrum, a fixed wavelength is used to excite the sample and the intensity of the emitted
radiation is monitored as a function of wavelength.
Optical Absorbance Spectroscopy

PL spectroscopy data is frequently combined with optical absorbance spectroscopy data to produce a more detailed description of
the band gap size of QDs. UV-visible spectroscopy is a specific kind of optical absorbance spectroscopy that measures the
transitions from ground state to excited state. This is the opposite of PL spectroscopy, which measures the transitions from excited
states to ground states. UV-visible spectroscopy uses light in the visible or ultraviolet range to excite electrons and measures the
absorbance of radiation verses wavelength. A sharp peak in the UV-visible spectrum indicates the wavelength at which the sample
best absorbs radiation.
As mentioned before, an excitation spectrum is a graph of emission intensity versus excitation wavelength. This spectrum often
looks very similar to the absorbance spectrum and in some instances they are the exact same. However, slight differences in the
theory behind these techniques do exist. Broadly speaking, an absorption spectrum measures wavelengths at which a molecule
absorbs lights, while an excitation spectrum determines the wavelength of light necessary to produce emission or fluorescence from
the sample, as monitored at a particular wavelength. It is quite possible then for peaks to appear in the absorbance spectrum that
would not occur on the PL excitation spectrum.
Instrumentation
A schematic diagram for a UV-vis spectrometer is shown in Figure 8.5.47. Like PL spectroscopy, the instrument requires a source
of radiation, a means of selecting a narrow band of radiation (monochromator), and a detector. Unlike PL spectroscopy, the detector
is placed along the same axis as the sample, rather than being directed 90º away from it.

Figure 8.5.47 A schematic representation of UV-Vis spectrometer.


Sample Spectra
A UV-Vis spectrum, such as the one shown in Figure 8.5.48, can be used not only to determine the band gap of QDs, but to also
determine QD size. Because QDs absorb different wavelengths of light based on the size of the particles, UV-Vis (and PL)
spectroscopy can provide a convenient and inexpensive way to determine the band gap and/or size of the particle by using the
peaks on the spectrum.

8.5.28 https://chem.libretexts.org/@go/page/55921
Figure 8.5.48 A standard absorbance spectrum of different sized CdSe QDs. Reprinted with permission form C.B. Murray, D. J.
Norris, and M.G. Bawendi, J. Am. Chem. Soc., 1993, 115, 8706. Copyright: American Chemical Society (1993).

The highly tunable nature of QDs, as well as their high extinction coefficient, makes QDs well-suited to a large variety of
applications and new technologies. QDs may find use as inorganic fluorophores in biological imaging, as tools to improve
efficiency in photovoltaic devices, and even as a implementations for qubits in quantum computers. Knowing the band gap of QDs
is essential to understanding how QDs may be used in these technologies. PL and optical absorbance spectroscopies provide ideal
ways of obtaining this information.

8.5: Using UV-Vis for the detection and characterization of silicon quantum dots is shared under a CC BY 4.0 license and was authored, remixed,
and/or curated by Pavan M. V. Raja & Andrew R. Barron via source content that was edited to conform to the style and standards of the
LibreTexts platform; a detailed edit history is available upon request.

8.5.29 https://chem.libretexts.org/@go/page/55921
8.6: Characterization of Graphene by Raman Spectroscopy
Surface area is a property of immense importance in the nano-world, especially in the area of heterogeneous catalysis. A solid
catalyst works with its active sites binding to the reactants, and hence for a given active site reactivity, the higher the number of
active sites available, the faster the reaction will occur. In heterogeneous catalysis, if the catalyst is in the form of spherical
nanoparticles, most of the active sites are believed to be present on the outer surface. Thus it is very important to know the catalyst
surface area in order to get a measure of the reaction time. One expresses this in terms of volume specific surface area, i.e., surface
area/volume although in industry it is quite common to express it as surface area per unit mass of catalyst, e.g., m2/g.

Overview of NMR
Nuclear magnetic resonance (NMR) is the study of the nuclei of the response of an atom to an external magnetic field. Many nuclei
have a net magnetic moment with I ≠ 0, along with an angular momentum in one direction where I is the spin quantum number of
the nucleus. In the presence of an external magnetic field, a nucleus would precess around the field. With all the nuclei precessing
around the external magnetic field, a measurable signal is produced. NMR can be used on any nuclei with an odd number of
protons or neutrons or both, like the nuclei of hydrogen (1H), carbon (13C), phosphorous (31P), etc. Hydrogen has a relatively large
magnetic moment (μ = 14.1 x 10-27 J/T) and hence it is used in NMR logging and NMR rock studies. The hydrogen nucleus
composes of a single positively charged proton that can be seen as a loop of current generating a magnetic field. It is may be
considered as a tiny bar magnet with the magnetic axis along the spin axis itself as shown in Figure 8.6.1. In the
absence of any external forces, a sample with hydrogen alone will have the individual magnetic moments
randomly aligned as shown in Figure 8.6.2.
Nuclear magnetic resonance (NMR) is the study of the nuclei of the response of an atom to an external magnetic field. Many nuclei
have a net magnetic moment with I≠0, along with an angular momentum in one direction where I is the spin quantum number of
the nucleus. In the presence of an external magnetic field, a nucleus would precess around the field. With all the nuclei precessing
around the external magnetic field, a measurable signal is produced.
NMR can be used on any nuclei with an odd number of protons or neutrons or both, like the nuclei of hydrogen (1H), carbon
(13C), phosphorous (31P), etc. Hydrogen has a relatively large magnetic moment (μ = 14.1 x 10-27 J/T) and hence it is used in
NMR logging and NMR rock studies. The hydrogen nucleus composes of a single positively charged proton that can be seen as a
loop of current generating a magnetic field. It is may be considered as a tiny bar magnet with the magnetic axis along the spin axis
itself as shown in Figure. In the absence of any external forces, a sample with hydrogen alone will have the individual magnetic
moments randomly aligned as shown in Figure 8.6.2.

Figure 8.6.1 A simplistic representation of a spinning nucleus as bar magnet. Copyright: Halliburton Energy Services, Duncan, OK
(1999).

8.6.1 https://chem.libretexts.org/@go/page/55922
Figure 8.6.2 Representation of randomly aligned hydrogen nuclei. Copyright: Halliburton Energy Services, Duncan, OK (1999).

Advantages of NMR over BET Technique


BET measurements follow the BET (Brunner-Emmet-Teller) adsorption isotherm of a gas on a solid surface. Adsorption
experiments of a gas of known composition can help determine the specific surface area of the solid particle. This technique has
been the main source of surface area analysis used industrially for a long time. However BET techniques take a lot of time for the
gas-adsorption step to be complete while one shall see in the course of this module that NMR can give you results in times
averaging around 30 minutes depending on the sample. BET also requires careful sample preparation with the sample being in dry
powder form, whereas NMR can accept samples in the liquid state as well.

NMR Relaxation Mechanism in Solid Suspensions


Calculations
From an atomic stand point, T1 relaxation occurs when a precessing proton transfers energy with its surroundings as the proton
relaxes back from higher energy state to its lower energy state. With T2 relaxation, apart from this energy transfer there is also
dephasing and hence T2 is less than T1 in general. For solid suspensions, there are three independent relaxation mechanisms
involved:-
1. Bulk fluid relaxation which affects both T1 and T2 relaxation.
2. Surface relaxation, which affects both T1 and T2 relaxation.
3. Diffusion in the presence of the magnetic field gradients, which affects only T2 relaxation
These mechanisms act in parallel so that the net effects are given by 8.6.1 and 8.6.2.
1 1 1 1
=  +  + (8.6.1)
T2 T2,bulk T2,surf ace T2,dif f usion

1 1 1
=  +  (8.6.2)
T1 T1,bulk T1,surf ace

The relative importance of each of these terms depend on the specific scenario. For the case of most solid suspensions in liquid, the
diffusion term can be ignored by having a relatively uniform external magnetic field that eliminates magnetic gradients. Theoretical
analysis has shown that the surface relaxation terms can be written as
8.6.3 and 8.6.4.
1 S
= ρ1 ( )particle (8.6.3)
T1,surf ace V

1 S
= ρ2 ( )particle (8.6.4)
T2,surf ace V

Thus one can use T1 or T2 relaxation experiment to determine the specific surface area. We shall explain the case of the T2
technique further as 8.6.5.
1 1 S
= + ρ2 ( )particle (8.6.5)
T2 T2,bulk V

8.6.2 https://chem.libretexts.org/@go/page/55922
One can determine T2 by spin-echo measurements for a series of samples of known S/V values and prepare a calibration chart as
shown in Figure 8.6.3, with the intercept as 1/T2,bulk and the slope as ρ2, one can thus find the specific surface area of an unknown
sample of the same material.

Figure 8.6.3 Example of a calibration plot of 1/T2 versus specific surface area (S/V) of a sample.

Sample Preparation and Experimental Setup


The sample must be soluble in the solvent. For proton NMR, about 0.25-1.00 mg/mL are needed depending on the sensitivity of the
instrument.
The solvent properties will have an impact of some or all of the spectrum. Solvent viscosity affects obtainable resolution, while
other solvents like water or ethanol have exchangeable protons that will prevent the observation of such exchangeable protons
present in the solute itself. Solvents must be chosen such that the temperature dependence of solute solubility is low in the
operation temperature range. Solvents containing aromatic groups like benzene can cause shifts in the observed spectrum compared
to non-aromatic solvents.
NMR tubes are available in a wide range of specifications depending on specific scenarios. The tube specifications need to be
extremely narrow while operating with high strength magnetic fields. The tube needs to be kept extremely clean and free from dust
and scratches to obtain good results, irrespective of the quality of the tube. Tubes can cleaned without scratching by rinsing out the
contents and soaking them in a degreasing solution, and by avoiding regular glassware cleaning brushes. After soaking for a while,
rinse with distilled water and acetone and dry the tube by blowing filterened nitrogen gas through a pipette or by using a swob of
cotton wool.
Filter the sample solution by using a Pasteur pipette stuffed with a piece of cotton wool at the neck. Any suspended material like
dust can cause changes in the spectrum. When working with dilute aqueous solutions, sweat itself can have a major effect and so
gloves are recommended at all times.
Sweat contains mainly water, minerals (sodium 0.9 g/L, potassium 0.2 g/L, calcium 0.015 g/L, magnesium 0.0013 g/L and other
trace elements like iron, nickel, zinc, copper, lead and chromium), as well as lactate and urea. In presence of a dilute solution of the
sample, the proton-containing substances in sweat (e.g., lactate and urea) can result in a large signal that can mask the signal of the
sample.
The NMR probe is the most critical piece of equipment as it contains the apparatus that must detect the small NMR signals from
the sample without adding a lot of noise. The size of the probe is given by the diameter of the NMR tube it can accommodate with
common sizes 5, 10 and 15 mm. A larger size probe can be used in the case of less sensitive samples in order to get as much solute
into the active zone as possible. When the sample is available in less quantity, use a smaller size tube to get an intrinsically higher
sensitivity.

NMR Analysis
A result sheet of T2 relaxation has the plot of magnetization versus time, which will be linear in a semi-log plot as shown in Figure
8.6.4. Fitting it to the equation, we can find T2 and thus one can prepare a calibration plot of 1/T2 versus S/V of known samples.

8.6.3 https://chem.libretexts.org/@go/page/55922
Figure 8.6.4 Example of T2 relaxation with magnetization versus time on a semi-log plot.

Limitations of the T2 Technique


The following are a few of the limitations of the T2 technique:
One can’t always guarantee no magnetic field gradients, in which case the T1 relaxation technique is to be used. However this
takes much longer to perform than the T2 relaxation.
There is the requirement of the odd number of nucleons in the sample or solvent.
The solid suspension should not have any para- or ferromagnetic substance (for instance, organics like hexane tend to have
dissolved O2 which is paramagnetic).
The need to prepare a calibration chart of the material with known specific surface area.

Example of Usage
A study of colloidal silica dispersed in water provides a useful example. Figure 8.6.5 shows a representation of an individual silica
particle.

Figure 8.6.5 A representation of the silica particle with a thin water film surrounding it.
A series of dispersion in DI water at different concentrations was made and surface area calculated. The T2 relaxation technique
was performed on all of them with a typical T2 plot shown in Figure 8.6.6 and T2 was recorded at 2117 milliseconds for this
sample.

Figure 8.6.6 T2 measurement for 2.3 wt% silica in DI water.


A calibration plot was prepared with 1/T2 – 1/T2,bulk as ordinate (the y-axis coordinate) and S/V as abscissa (the x-axis
coordinate). This is called the surface relaxivity plot and is illustrated in Figure 8.6.7.

8.6.4 https://chem.libretexts.org/@go/page/55922
Figure 8.6.7 Calibration plot of (1/T2 – 1/T2,Bulk) versus specific surface area for silica in DI water.
Accordingly for the colloidal dispersion of silica in DI water, the best fit resulted in 8.6.6, from which one can see that the value of
surface relaxivity, 2.3 x 10-8, is in close accordance with values reported in literature.
1 1 −8
S
 −    =  2.3 × 10 ( )  −  0.0051 (8.6.6)
T2 T2,bulk V

The T2 technique has been used to find the pore-size distribution of water-wet rocks. Information of the pore size distribution helps
petroleum engineers model the permeability of rocks from the same area and hence determine the extractable content of fluid
within the rocks.
Usage of NMR for surface area determination has begun to take shape with a company, Xigo nanotools, having developed an
instrument called the Acorn AreaTM to get surface area of a suspension of aluminum oxide. The results obtained from the
instrument match closely with results reported by other techniques in literature. Thus the T2 NMR technique has been presented as
a strong case to obtain specific surface areas of nanoparticle suspensions.

8.6: Characterization of Graphene by Raman Spectroscopy is shared under a CC BY 4.0 license and was authored, remixed, and/or curated by
Pavan M. V. Raja & Andrew R. Barron via source content that was edited to conform to the style and standards of the LibreTexts platform; a
detailed edit history is available upon request.

8.6.5 https://chem.libretexts.org/@go/page/55922
8.7: Characterization of Graphene by Raman Spectroscopy
Graphene is a quasi-two-dimensional material, which comprises layers of carbon atoms arranged in six-member rings (Figure
8.7.1). Since being discovered by Andre Geim and co-wokers at the University of Manchester, graphene has become one of the

most exciting topics of research because of its distinctive band structure and physical properties, such as the observation of a
quantum hall effect at room temperature, a tunable band gap, and a high carrier mobility.

Figure 8.7.1 Idealized structure of a single graphene sheet. Copyright: Chris Ewels (www.www.ewels.info).
Graphene can be characterized by many techniques including atomic force microscopy (AFM), transmission electron microscopy
(TEM) and Raman spectroscopy. AFM can be used to determine the number of the layers of the graphene, and TEM images can
show the structure and morphology of the graphene sheets. In many ways, however, Raman spectroscopy is a much more important
tool for the characterization of graphene. First of all, Raman spectroscopy is a simple tool and requires little sample preparation.
What’s more, Raman spectroscopy can not only be used to determine the number of layers, but also can identify if the structure of
graphene is perfect, and if nitrogen, hydrogen or other fuctionalization is successful.

Raman Spectrum of Graphene


While Raman spectroscopy is a useful technique for characterizing sp2 and sp3 hybridized carbon atoms, including those in
graphite, fullerenes, carbon nanotubes, and graphene. Single, double, and multi-layer graphenes have also been differentiated by
their Raman fingerprints.
Figure 8.7.2 shows a typical Raman spectrum of N-doped single-layer graphene. The D-mode, appears at approximately 1350 cm-
1, and the G-mode appears at approximately 1583 cm-1. The other Raman modes are at 1620 cm-1 (D’- mode), 2680 cm-1 (2D-
mode), and 2947 cm-1 (D+G-mode).

Figure 8.7.2 Raman spectrum with a 514.5 nm excitation laser wavelength of N-doped single-layer graphene.

The G-band
The G-mode is at about 1583 cm-1, and is due to E2g mode at the Γ-point. G-band arises from the stretching of the C-C bond in
graphitic materials, and is common to all sp2 carbon systems. The G-band is highly sensitive to strain effects in sp2 system, and
thus can be used to probe modification on the flat surface of graphene.

Disorder-induced D-band and D'-band


The D-mode is caused by disordered structure of graphene. The presence of disorder in sp2-hybridized carbon systems results in
resonance Raman spectra, and thus makes Raman spectroscopy one of the most sensitive techniques to characterize disorder in sp2

8.7.1 https://chem.libretexts.org/@go/page/55923
carbon materials. As is shown by a comparison of Figure 8.7.2 and Figure 8.7.3 there is no D peak in the Raman spectra of
graphene with a perfect structure.

Figure 8.7.3 Raman spectrum with a 514.5 nm excitation laser wavelengthof pristine single-layer graphene.
If there are some randomly distributed impurities or surface charges in the graphene, the G-peak can split into two peaks, G-peak
(1583 cm-1) and D’-peak (1620 cm-1). The main reason is that the localized vibrational modes of the impurities can interact with
the extended phonon modes of graphene resulting in the observed splitting.

The 2D-band
All kinds of sp2 carbon materials exhibit a strong peak in the range 2500 - 2800 cm-1 in the Raman spectra. Combined with the G-
band, this spectrum is a Raman signature of graphitic sp2 materials and is called 2D-band. 2D-band is a second-order two-phonon
process and exhibits a strong frequency dependence on the excitation laser energy.
What’s more, the 2D band can be used to determine the number of layer of graphene. This is mainly because in the multi-layer
graphene, the shape of 2D band is pretty much different from that in the single-layer graphene. As shown in Figure 8.7.4, the 2D
band in the single-layer graphene is much more intense and sharper as compared to the 2D band in multi-layer graphene.

Figure 8.7.4 Raman spectrum with a 514.5 nm excitation laser wavelength of pristine single-layer and multi-layer graphene.

8.7: Characterization of Graphene by Raman Spectroscopy is shared under a CC BY 4.0 license and was authored, remixed, and/or curated by
Pavan M. V. Raja & Andrew R. Barron via source content that was edited to conform to the style and standards of the LibreTexts platform; a
detailed edit history is available upon request.

8.7.2 https://chem.libretexts.org/@go/page/55923
8.8: Characterization of Bionanoparticles by Electrospray-Differential Mobility
Analysis
Characterization of nanoparticles in general, and carbon nanotubes in particular, remains a technical challenge even though the
chemistry of covalent functionalization has been studied for more than a decade. It has been noted by several researchers that the
characterization of products represents a constant problem in nanotube chemistry. A systematic tool or suites of tools are needed for
adequate characterization of chemically functionalized single-walled carbon nanotubes (SWNTs), and is necessary for declaration
of success or failure in functionalization trials.
So far, a wide range of techniques have been applied to characterize functionalized SWNTs: infra red (IR), Raman, and UV/visible
spectroscopies, thermogravimetric analysis (TGA), atomic force microscopy (AFM), transmission electron microscopy (TEM), X-
ray photoelectron spectroscopy (XPS), etc. A summary of the attribute of each of the characterization method is given in Table
8.8.1.

Table 8.8.1 Common characterization methodology for functionalized SWNTs.


Method Sample Information Limitations

no evidence for covalent


TGA Solid Functionalization ratio
functionalization, not specific
no evidence of covalent
XPS solid elements, functionalization ratio functionalization, not specific
quantification complicated
not specific, quantification not
Raman solid sp3 indicated by D mode
reliable
no direct evidence for covalent
Infrared (IR) solid for ATR-IR or solution substituent groups functionalization quantification
not possible
not specific or quantitative, need
UV/Visible solution sidewall functionalization
highly disperesed sample
no evidence of covalent
Solution NMR solution substituents functionalization, high solubility
of sample
high functionalization needed,
substituents sp3 molecular long time for signal acquisition,
Solid state NMR solid motions, quantification at high quantification not available for
level of functionalization samples with protons on side
chains
only a small portion of sample
characterized, no evidence of
AFM solid on substrate topography
covalent functionalization, no
chemical identity
only a small portion of sample
characterized, no evidence of
image of sample distribution
TEM solid on substrate covalent functionalization, no
dispersion
chemical identity dispersion
information complicated
no chemical identity of functional
STM solid on substrate distribution groups small portion of sample
conductive sample only

Elemental and Physical Analysis


Thermogravimetric Analysis (TGA)

8.8.1 https://chem.libretexts.org/@go/page/55924
Thermogravimetric analysis (TGA) is the mostly widely used method to determine the level of sidewall functionalization. Since
most functional groups are labile or decompose upon heating, while the SWNTs are stable up to 1200 °C under Ar atmosphere. The
weight loss at 800 °C under Ar is often used to determine functionalization ratio using this indirect method. Unfortunately,
quantification can be complicated with presence of multiple functional groups. Also, TGA does not provide direct evidence for
covalent functionalization since it cannot differentiate between covalent attachment and physical adsorption.

X-ray Photoelectron Spectroscopy (XPS)


XPS confirms the presence of different elements in functionalized SWNTs. This is useful for identification of heteroatom elements
such as F and N, and then XPS can be used for quantification with simple substituent groups and used indirectly. Deconvolution of
XPS is useful to study fine structures on SWNTs. However, the overlapping of binding energies in the spectrum complicates
quantification.

Spectroscopy
Raman Spectroscopy
Raman spectroscopy is very informative and important for characterizing functionalized SWNTs. The tangential G mode (ca. 1550
– 1600 cm-1) is characteristic of sp2 carbons on the hexagonal graphene network. The D-band, so-called disorder mode (found at
ca. 1295 cm-1) appears due to disruption of the hexagonal sp2 network of SWNTs. The D-band was largely used to characterize
functionalized SWNTs and ensure functionalization is covalent and occurred at the sidewalls. However, the observation of D band
in Raman can also be related to presence of defects such as vacancies, 5-7 pairs, or dopants. Thus, using Raman to provide
evidence of covalent functionalization needs to be done with caution. In particular, the use of Raman spectroscopy for a
determination of the degree of functionalization is not reliable.
It has been shown that quantification with Raman is complicated by the distribution of functional groups on the sidewall of
SWNTs. For example, if fluorinated-SWNTs (F-SWNTs) are functionalized with thiol or thiophene terminated moieties, TGA
shows that they have similar level of functionalization. However, their relative intensities of D:G in Raman spectrum are quite
different. The use of sulfur substituents allow for gold nanoparticles with 5 nm in diameter to be attached as a “chemical marker”
for direct imaging of the distribution of functional groups. AFM and STM suggest that the functional groups of thio-SWNTs are
group together while the thiophene groups are widely distributed on the sidewall of SWNTs. Thus the difference is not due to
significant difference in substituent concentration but on substituent distribution, while Raman shows different D:G ratio.

Infrared Spectroscopy
IR spectroscopy is useful in characterizing functional groups bound to SWNTs. A variety of organic functional groups on sidewall
of SWNTs have been identified by IR, such as COOH(R), -CH2, -CH3, -NH2, -OH, etc. However, it is difficult to get direct
functionalization information from IR spectroscopy. The C-F group has been identified by IR in F-SWNTs. However, C-C, C-N, C-
O groups associated with the side-wall functionalization have not been observed in the appropriately functionalized SWNTs.

UV/Visible Spectroscopy
UV/visible spectroscopy is maybe the most accessible technique that provides information about the electronic states of SWNTs,
and hence functionalization. The absorption spectrum shows bands at ca. 1400 nm and 1800 nm for pristine SWNTs. A complete
loss of such structure is observed after chemical alteration of SWNTs sidewalls. However, such information is not quantitative and
also does not show what type of functional moiety is on the sidewall of SWNTs.

Nuclear Magnetic Resonance


NMR can be considered as a “new” characterization technique as far as SWNTs are concerned. Solution state NMR is limited for
SWNT characterization because low solubility and slow tumbling of the SWNTs results in broad spectra. Despite this issue, there
are still solution 1H NMR reported of SWNTs functionalized by carbenes, nitrenes and azomethine ylides because of the high
solubility of derivatized SWNTs. However, proof of covalent functionalization cannot be obtained from the 1H NMR. As an
alternative, solid state 13C NMR has been employed to characterize several functionalized SWNTs and show successful
observation of sidewall organic functional groups, such as carboxylic and alkyl groups. But there has been a lack of direct evidence
of sp3 carbons on the sidewall of SWNTs that provides information of covalent functionalization.
Solid state 13C NMR has been successfully employed in the characterization of F-SWNTs through the direct observation of the
sp3C-F carbons on sidewall of SWNTs. This methodology has been transferred to more complicated systems; however, it has been

8.8.2 https://chem.libretexts.org/@go/page/55924
found that longer side chain length increases the ease to observe sp3C-X sidewall carbons.
Solid state NMR is a potentially powerful technique for characterizing functionalized SWNTs because molecular dynamic
information can also be obtained. Observation that higher side chain mobility can be achieved by using a longer side chain length
offers a method of exploring functional group conformation. In fact, there have been reports using solid state NMR to study
molecular mobility of functionalized multi-walled carbon nanotubes.

Microscopy
AFM, TEM and STM are useful imaging techniques to characterize functionalized SWNTs. As techniques, they are routinely used
to provide an “image” of an individual nanoparticle, as opposed to an average of all the particles.

Atomic Force Microscopy


AFM shows morphology on the surface of SWNTs. The height profile on AFM is often used to show presence of functional groups
on sidewall of SWNTs. Individual SWNTs can be probed by AFM and sometimes provide information of dispersion and
exfoliation of bundles. Measurement of heights along an individual SWNT can be correlated with the substituent group, i.e., the
larger an alkyl chain of a sidewall substituent the greater the height measured. AFM does not distinguish whether those functional
groups are covalently attached or physically adsorbed on the surface of SWNTs.

Transmission Electron Microscopy


TEM can be used to directly image SWNTs and at high resolution clearly shows the sidewall of individual SWNT. However, the
resolution of TEM is not sufficient to directly observe covalent attachment of chemical modification moieties, i.e., to differentiate
between sp2 and sp3 carbon atoms. TEM can be used to provide information of functionalization effect on dispersion and
exfoliation of ropes.
Samples are usually prepared from very dilute concentration of SWNTs. Sample needs to be very homogeneous to get reliable data.
As with AFM, TEM only shows a very small portion of sample, using them to characterize functionalized SWNTs and evaluate
dispersion of samples in solvents needs to be done with caution.

Scanning Tunneling Microscopy


STM offers a lot of insight on structure and surface of functionalized SWNTs. STM measures electronic structure, while sometimes
the topographical information can be indirectly inferred by STM images. STM has been used to characterize F-SWNTs gold-
marked SWNTs, and organic functionalized SWNTs. Distribution of functional groups can be inferred from STM images since the
location of a substituent alters the localized electronic structure of the tube. STM images the position/location of chemical changes
to the SWNT structure. The band-like structure of F-SWNTs was first disclosed by STM.
STM has the same problem that is inherent with AFM and TEM, that when using small sample size, the result may not be
statistically relevant. Also, chemical identity of the features on SWNTs cannot be determined by STM; rather, they have to be
identified by spectroscopic methods such as IR or NMR. A difficulty with STM imaging is that the sample has to be conductive,
thus deposition of the SWNT onto a gold (or similar) surface is necessary.

8.8: Characterization of Bionanoparticles by Electrospray-Differential Mobility Analysis is shared under a CC BY 4.0 license and was authored,
remixed, and/or curated by Pavan M. V. Raja & Andrew R. Barron via source content that was edited to conform to the style and standards of the
LibreTexts platform; a detailed edit history is available upon request.

8.8.3 https://chem.libretexts.org/@go/page/55924
8.9: Characterization of Bionanoparticles by Electrospray-Differential Mobility
Analysis
|Electrospray-differential mobility analysis (ES-DMA) is an analytical technique that uses first an electrospray to aerosolize
particles and then DMA to characterize their electrical mobility at ambient conditions. This versatil tool can be used to quantitative
characterize biomolecules and nanoparticles from 0.7 to 800 nm. In the 1980s, it was discovered that ES could be used for
producing aerosols of biomacromolecules. In the case of the DMA, its predecesor was developed by Hewitt in 1957 to analize
charging of small particles. The modified DMA, which is a type of ion mobility analyzer, was developed by Knuts}on and Whitby
(Figure 8.9.1\) in 1975 and later it was commercialized. Among the several designs, the cylindrical DMA has become the standard
design and has been used for the obtention of monodisperse aerosols, as well as for the classification of polydisperse aerosols.

Figure 8.9.1 American engineer K. T. Whitby (1925-1983).


The first integration of ES with DMA ocurred in 1996 when this technique was used to determine the size of different globular
proteins. DMA was refined over the past decade to be used in a wide range of applications for the characterization of polymers,
viruses, bacteriophages and nanoparticle-biomolecule conjugates. Although numerous publications have reported the use of ES-
DMA in medicinal and pharmaceutical applications, this present module describes the general principles of the technique and its
application in the analysis of gold nanoparticles.

How Does ES-DMA Function?


ES-DMA consits of an electrospray source (ES) that aerosolizes bionanoparticles and a class of ion mobility analyzer (DMA) that
measures their electrical mobility by balancing electrical and drag forces on the particles. DMA continously separates particles
based on their charge to size ratio. An schematic of the experimental setup for ES-DMA is shown in Figure 8.9.2 for the analysis
of gold nanoparticles.

Figure 8.9.2 Schematic of experimental setup for ES-DMA. Reprinted with permission from D. Tsai, R. A. Zangmeister, L. F.
Pease III, M. J. Tarlov, and M. R. Zachariah. Langmuir, 2008, 24, 8483. Copyright (2015) American Chemical Society.

8.9.1 https://chem.libretexts.org/@go/page/156770
The process of analyzing particles with ES-DMA involves four steps:
First, the analyte dissolved in a volatile buffer such as ammonium acetate [NH4][O2CCH3] is placed inside a pressure chamber.
Then, the solution is delivered to the nozzle through a fused silica capillary to generate multiply charged droplets. ES nebulizers
produce droplets of 100-400 nm in diameter but they are highly charged.
In the next step, the droplets are mixed with air and carbon dioxide (CO2) and are passed through the charge reducer or neutralizer
where the solvent continues to evaporate and charge distribution decreases. The charge reducer is an ionizing α radiation source
such as Po210 that ionizes the carrier gas and reduces the net charges on the particles to a Fuchs’-Boltzmann distribution. As a
result, the majority of the droplets contain single net charge particles that pass directly to the DMA. DMA separates positively or
negatively charged particles by applying a negative or positive potential. Figure 8.9.3 shows a single channel design of cylindrical
DMA that is composed of two concentric electrodes between which a voltage is applied. The inner electrode is maintained at a
controlled voltage from 1V to 10 kV, whereas the outer electrode is electrically grounded.

Figure 8.9.3 Basic principle of a general DMA. Adapted from P. Intra and N. Tippayawong. Songklanakarin J. Sci. Technol., 2008,
30, 243-256.
In the third step, the aerosol flow (Qa) enters through a slit that is adjacent to one electrode and the sheath air (air or N2) flow (Qs)
is introduced to separate the aerosol flow from the other electrode. After a voltage is applied between the inner and outer
electrodes, an electric field is formed and the charged particles with specific electrical mobility are attracted to a charged collector
rod. The positions of the charged particles along the length of the collector depend on their electrical mobility (Zp), the fluid flow
rate and the DMA geometry. In the case of particles with a high electrical mobility, they are collected in the upper part of the rod
(particles a and b, Figure 8.9.4) while particles with a low electrical mobility are collected in the lower part of the rod (particle d,
Figure 8.9.3.
(Qs   +  Qa )ln(R2 )
Zp = (8.9.1)
R1

With the value of the electrical mobility, the particle diameter (dp) can be determined by using Stokes’ law as described by 8.9.2,
where n is the number of charge units, e is the elementary unit of charge (1.61x10-19C), Cc is the Cunningham slip correction
factor and µ is the gas viscosity. Cc 8.9.3, considers the noncontinuum flow effect when dp is similar to or smaller than the mean
free path (λ) of the carrier gas.
ne Cc
dp   = (8.9.2)
3πμZ p

−1.10dp
2λ −
Cc = 1  +   [1.257  +  0.4 e 2λ ] (8.9.3)
dp

8.9.2 https://chem.libretexts.org/@go/page/156770
In the last step, the size-selected particles are detected with a condensation particle counter (CPC) or an aerosol electrometer (AE)
that determines the particle number concentration. CPC has lower detection and quantitation limits and is the most sensitive
detector available. AE is used when the particles concentrations are high or when particles are so small that cannot be detected by
CPC. Figure 8.9.4 shows the operation of the CPC in which the aerosol is mixed with butanol (C4H9OH) or water vapor (working
fluid) that condensates on the particles to produce supersaturation. Hence, large size particles (around 10 μm) are obtained, detected
optically and counted. Since each droplet is approximately of the same size, the count is not biased. The particle size distribution is
obtained by changing the applied voltage. Generally, the performance of the CPC is evaluated in terms of the minimum size that is
counted with 50% efficiency.

Figure 8.9.4 Working principle of the condensation particle counter (CPC). Reprinted from Trends in Biotechnology, 30, S. Guha,
M. Li, M. J. Tarlov, and M. R. Zachariah, Electrospray-differential mobility analysis of bionanoparticles, 291-300, Copyright
(2015), with permission from Elsevier.

What Type of Information is Obtained by ES-DMA?


ES-DMA provides information of the mobility diameter of particles and their concentration in number of particles per unit volume
of analyzed gas so that the particle size distribution is obtained as shown in Figure 8.9.10. Another form of data representation is
the differential distribution plot of ΔN/Δlogdp vs dp (Figure 8.9.11. This presentation has a logarithmic size axis that is usually
more convenient because particles are often distributed over a wide range of sizes.

Figure 8.9.5 Size distribution of human serum albumin, [p/cc]: particles per cubic centimeter. Reprinted with permission from S. T.
Kaufman, J. W. Skogen, F. D. Dorman, and F. Zarrin. Anal. Chem., 1996, 68, 1895-1904. Copyright (2015) American Chemical
Society.

How Data from ES-DMA is processed?


To obtain the actual particle size distribution (Figure), the raw data acquired with the ES-DMA is corrected for charge correction,
transfer function of the DMA and collection efficiency for CPC. Figure 8.9.6 illustrates the charge correction in which a charge
reducer or neutralizer is necessary to reduce the problem of multiple charging and simplify the size distribution. The charge
reduction depends on the particle size and multiple charging can be produced as the particle size increases. For instance, for 10 nm
particles, the percentages of single charged particles are lower than those of neutral particles. After a negative voltage is applied,
only the positive charged particles are collected. Conversely, for 100 nm particles, the percentages of single charged particles
increase and multiple charges are present. Hence, after a negative bias is applied, +1 and +2 particles are collected. The presence of
more charges in particles indicates high electrical mobility and

8.9.3 https://chem.libretexts.org/@go/page/156770
Figure 8.9.6 Data processing for the charge correction in the aerosol phase. Reprinted from Trends in Biotechnology, 30, S. Guha,
M. Li, M. J. Tarlov, and M. R. Zachariah, Electrospray-differential mobility analysis of bionanoparticles, 291-300, Copyright
(2015), with permission from Elsevier.
The transfer function for DMA modifies the input particle size distribution and affects the resolution as shown in Figure 8.9.7. This
transfer function depends on the operation conditions such as flow rates and geometry of the DMA. Furthermore, the transfer
function can be broadened by Brownian diffusion and this effect produces the actual size distribution. The theoretical resolution is
measured by the ratio of the sheath to the aerosol flow in under balance flow conditions (sheath flow equals excess flow and
aerosol flow in equals monodisperse aerosol flow out).

Figure 8.9.7 Data processing for transfer function for DMA. Reprinted from Trends in Biotechnology, 30, S. Guha, M. Li, M. J.
Tarlov, and M. R. Zachariah, Electrospray-differential mobility analysis of bionanoparticles, 291-300, Copyright (2015), with
permission from Elsevier.
The CPC has a size limit of detection of 2.5 nm because small particles are difficult to activate at the supersaturation of the working
fluid. Therefore, CPC collection efficiency is required that consists on the calibration of the CPC against an electrometer.
Applications of ES-DMADetermination of molecular weight of polymers and proteins in the range of 3.5 kDa to 2 MDa by
correlating molecular weight and mobility diameter.
Determination of absolute number concentration of nanoparticles in solution by obtaining the ES droplet size distributions and
using statistical analysis to find the original monomer concentration. Dimers or trimers can be formed in the electrospray
process due to droplet induced aggregation and are observed in the spectrum.
Kinetics of aggregation of nanoparticles in solution by analysis of multimodal mobility distributions from which distinct types
of aggregation states can be identified.
Quantification of ligand adsorption to bionanoparticles by measuring the reduction in electrical mobility of a complex particle
(particle-protein) that corresponds to an increase in mobility diameter.

Characterization of SAM-functionalized Gold Nanoparticles by ES-DMA


Citrate (Figure 8.9.8 tabilized gold nanoparticles (AuNPs)) with diameter in the range 10-60 nm and conjugated AuNPs are
analyzed by ES-DMA. This investigation shows that the formation of salt particles on the surface of AuNPs can interfere with the
mobility analysis because of the reduction in analyte signals. Since sodium citrate is a non volatile soluble salt, ES produces two
types of droplets. One droplet consists of AuNPs and salt and the other droplet contains only salt. Thus, samples must be cleaned
by centrifugation prior to determine the size of bare AuNPs. Figure 8.9.9 presents the size distribution of AuNPs of distinct
diameters and peaks corresponding to salt residues.

8.9.4 https://chem.libretexts.org/@go/page/156770
Figure 8.9.8 Structure of citrate that provides charge stabilization to AuNPs.

Figure 8.9.9 Particle size distribution of 10 nm, 30 nm and 60 nm AuNPs after centrifugation cleaning. Reprinted with permission
from D. Tsai, R. A. Zangmeister, L. F. Pease III, M. J. Tarlov and M. R. Zachariah. Langmuir, 2008, 24, 8483. Copyright (2015)
American Chemical Society.
The mobility size of bare AuNPs (dp0) can be obtained by using 8.9.4, where dp,m and ds are mobility sizes of the AuNPs
encrusted with salts and the salt NP, respectively. However, the presence of self-assembled monolayer (SAM) produces a difference
in electrical mobility between conjugated and bare AuNPs. Hence, the determination of the diameter of AuNPs (salt-free) is critical
to distinguish the increment in size after functionalization with SAM. The coating thickness of SAM that corresponds to the change
in particle size (ΔL) is calculated by using 8.9.5, where dp and dp0 are the coated and uncoated particle mobility diameters,
respectively.
−−−−−−−−
3 3 3
dp0 =  √ dp,m   −  ds (8.9.4)

ΔL  =  dp   −  dp0 (8.9.5)

In addition, the change in particle size can be determined by considering a simple rigid core-shell model that gives theoretical
values of ΔL1 higher than the experimental ones (ΔL). A modified core-shell model is proposed in which a size dependent effect
on ΔL2 is observed for a range of particle sizes. AuNPs of 10 nm and 60 nm are coated with MUA (Figure 8.9.10), a charge
alkanethiol, and the particle size distributions of bare and coated AuNPs are presented in Figure. The increment in average particle
size is 1.2 ± 0.1 nm for 10 nm AuNPs and 2.0 ± 0.3 nm for 60 nm AuNPs so that ΔL depends on particle size.

Figure 8.9.10 Structure of 11-mercaptoundecanoic acid (MUA).

Figure 8.9.11 Particle size distributions of bare versus MUA-coated AuNP for (a) 10 nm and (b) 60 nm. (c) A comparison of
predicted ΔL from experiment (diamonds) with theory (ΔL1 in dashed lines and ΔL2 in solid lines). Reprinted with permission
from D. Tsai, R. A. Zangmeister, L. F. Pease III, M. J. Tarlov, and M. R. Zachariah, Langmuir, 2008, 24, 8483. Copyright (2015)
American Chemical Society.

8.9.5 https://chem.libretexts.org/@go/page/156770
Advantages of ES-DMA
ES-DMA does not need prior information about particle type.
It characterizes broad particle size range and operates under ambient pressure conditions.
A few µL or less of sample volume is required and total time of analysis is 2-4 min.
Data interpretation and mobility spectra simple to analyze compared to ES-MS where there are several charge states.

Limitations of ES-DMA
Analysis requires the following solution conditions: concentrations of a few hundred µg/mL, low ionic strength (<100 mM) and
volatile buffers.
Uncertainty is usually ± 0.3 nm from a size range of a few nm to around 100 nm. This is not appropriate to distinguish proteins
with slight differences in molecular weight.

Related Techniques
A tandem technique is ES-DMA-APM that determines mass of ligands adsorbed to nanoparticles after size selection with DMA.
APM is an aerosol particle mass analyzer that measures mass of particles by balancing electrical and centrifugal forces. DMA-
APM has been used to analyze the density of carbon nanotubes, the porosity of nanoparticles and the mass and density differences
of metal nanoparticles that undergo oxidation.
r

8.9: Characterization of Bionanoparticles by Electrospray-Differential Mobility Analysis is shared under a not declared license and was authored,
remixed, and/or curated by LibreTexts.

8.9.6 https://chem.libretexts.org/@go/page/156770
CHAPTER OVERVIEW
Back Matter
Index

1 https://chem.libretexts.org/@go/page/243578
Index
A F N
Atomic Force Microscopy Field Effect Transistors Neutron Activation Analysis
9.2: Atomic Force Microscopy (AFM) 10.2: Measuring Key Transport Properties of FET 1.9: Neutron Activation Analysis (NAA)
Auger Electron Spectroscopy Devices neutron diffraction
1.14: Auger Electron Spectroscopy fluorescence 7.5: Neutron Diffraction
4.5: Photoluminescence, Phosphorescence, and NMR Spectroscopy
Fluorescence Spectroscopy
B 4.7: NMR Spectroscopy
Bravais lattices G
7.1: Crystal Structure O
gas chromatography Ostwald Viscometer
3.1: Principles of Gas Chromatography
C 2.6: Viscosity
graphene
Capillary Electrophoresis 8.7: Characterization of Graphene by Raman
3.6: Capillary Electrophoresis Spectroscopy
P
combustion analysis phosphorescence
1.3: Introduction to Combustion Analysis H 4.5: Photoluminescence, Phosphorescence, and
crystallography Fluorescence Spectroscopy
HPLC Photoluminescence
7.1: Crystal Structure
3.2: High Performance Liquid chromatography
cumulant expansion 4.5: Photoluminescence, Phosphorescence, and
Hyperfine Coupling Fluorescence Spectroscopy
2.4: Dynamic Light Scattering
4.8: EPR Spectroscopy
Cyclic Voltammetry
R
2.7: Electrochemistry
I Raman Spectroscopy
ICP
D 1.5: ICP-AES Analysis of Nanoparticles
4.3: Raman Spectroscopy

Desorption Mass Spectroscopy Inductively coupled plasma atomic


5.3: Temperature-Programmed Desorption Mass
S
Spectroscopy Applied in Surface Chemistry emission spectroscopy Scanning Tunneling Microscopy (STM)
diamagnetism 1.5: ICP-AES Analysis of Nanoparticles 9.3: SEM and its Applications for Polymer Science
4.1: Magnetism interferometry Semiconductors
Differential Scanning Calorimetry 9.1: Interferometry 7.2: Structures of Element and Compound
10.1: A Simple Test Apparatus to Verify the Semiconductors
2.8: Thermal Analysis
Photoresponse of Experimental Photovoltaic Materials
differential thermal analysis and Prototype Solar Cells
Spot test
1.2: Spot Tests
2.8: Thermal Analysis Ion Chromatography
dislocation 3.5: Ion Chromatography
supercritical fluid chromatography
3.4: Supercritical Fluid Chromatography
7.1: Crystal Structure IR Spectroscopy
dual polarization interferometry 4.2: IR Spectroscopy
9.1: Interferometry T
Dynamic Light Scattering L Thermogravimetric analysis
2.4: Dynamic Light Scattering 2.8: Thermal Analysis
law of constant angles
Dynamic Viscosity 7.3: X-ray Crystallography
2.6: Viscosity V
M vertical scanning interferometry
E Mössbauer spectroscopy 9.1: Interferometry
Electrical Permittivity 4.6: Mössbauer Spectroscopy
viscosity
2.9: Electrical Permittivity Characterization of 2.6: Viscosity
magnetic moments
Aqueous Solutions
4.1: Magnetism
electron spectroscopy for chemical
magnetism X
analysis 4.1: Magnetism XAFS
1.13: X-ray Photoelectron Spectroscopy Magnetization 7.6: XAFS
Electroosmotic Mobility 4.1: Magnetism XAS
3.6: Capillary Electrophoresis MEKC 1.8: A Practical Introduction to X-ray Absorption
Electrophoretic Mobility Spectroscopy
3.6: Capillary Electrophoresis 7.6: XAFS
3.6: Capillary Electrophoresis Melting Point Apparatus
Elemental analysis XPS
2.1: Melting Point Analysis 1.13: X-ray Photoelectron Spectroscopy
1: Elemental Analysis miller indicies 4.9: X-ray Photoelectron Spectroscopy
EPR 7.1: Crystal Structure
4.8: EPR Spectroscopy
Z
ESCA
Zeta Potential
1.13: X-ray Photoelectron Spectroscopy
3.6: Capillary Electrophoresis

1 https://chem.libretexts.org/@go/page/243579
CHAPTER OVERVIEW
9: Surface Morphology and Structure
9.1: Interferometry
9.2: Atomic Force Microscopy (AFM)
9.3: SEM and its Applications for Polymer Science
9.4: Catalyst Characterization Using Thermal Conductivity Detector
9.5: Nanoparticle Deposition Studies Using a Quartz Crystal Microbalance

9: Surface Morphology and Structure is shared under a CC BY 4.0 license and was authored, remixed, and/or curated by Pavan M. V. Raja &
Andrew R. Barron via source content that was edited to conform to the style and standards of the LibreTexts platform; a detailed edit history is
available upon request.

1
9.1: Interferometry
The Application of Vertical Scanning Interferometry to the Study of Crystal Surface Processes
The processes which occur at the surfaces of crystals depend on many external and internal factors such as crystal structure and
composition, conditions of a medium where the crystal surface exists and others. The appearance of a crystal surface is the result of
complexity of interactions between the crystal surface and the environment. The mechanisms of surface processes such as
dissolution or growth are studied by the physical chemistry of surfaces. There are a lot of computational techniques which allows
us to predict the changing of surface morphology of different minerals which are influenced by different conditions such as
temperature, pressure, pH and chemical composition of solution reacting with the surface. For example, Monte Carlo method is
widely used to simulate the dissolution or growth of crystals. However, the theoretical models of surface processes need to be
verified by natural observations. We can extract a lot of useful information about the surface processes through studying the
changing of crystal surface structure under influence of environmental conditions. The changes in surface structure can be studied
through the observation of crystal surface topography. The topography can be directly observed macroscopically or by using
microscopic techniques. Microscopic observation allows us to study even very small changes and estimate the rate of processes by
observing changing the crystal surface topography in time.
Much laboratory worked under the reconstruction of surface changes and interpretation of dissolution and precipitation kinetics of
crystals. Invention of AFM made possible to monitor changes of surface structure during dissolution or growth. However, to detect
and quantify the results of dissolution processes or growth it is necessary to determine surface area changes over a significantly
larger field of view than AFM can provide. More recently, vertical scanning interferometry (VSI) has been developed as new tool
to distinguish and trace the reactive parts of crystal surfaces. VSI and AFM are complementary techniques and practically well
suited to detect surface changes.
VSI technique provides a method for quantification of surface topography at the angstrom to nanometer level. Time-dependent VSI
measurements can be used to study the surface-normal retreat across crystal and other solid surfaces during dissolution process.
Therefore, VSI can be used to directly and nondirectly measure mineral dissolution rates with high precision. Analogically, VSI
can be used to study kinetics of crystal growth.
Physical Principles of Optical Interferometry
Optical interferometry allows us to make extremely accurate measurements and has been used as a laboratory technique for almost
a hundred years. Thomas Young observed interference of light and measured the wavelength of light in an experiment, performed
around 1801. This experiment gave an evidence of Young's arguments for the wave model for light. The discovery of interference
gave a basis to development of interferomertry techniques widely successfully used as in microscopic investigations, as in
astronomic investigations.
The physical principles of optical interferometry exploit the wave properties of light. Light can be thought as electromagnetic wave
propagating through space. If we assume that we are dealing with a linearly polarized wave propagating in a vacuum in z direction,
electric field E can be represented by a sinusoidal function of distance and time.
E(x, y, z, t)  =  a cos[2π(vt  −  z/λ)] (9.1.1)

Where a is the amplitude of the light wave, v is the frequency, and λ is its wavelength. The term within the square brackets is called
the phase of the wave. Let’s rewrite this equation in more compact form,
E(x, y, z, t)  =  a cos(ωt  −  kz) (9.1.2)

where ω=2πv is the circular frequency, and k=2π/λ is the propagation constant. Let’s also transform this second equation into a
complex exponential form,
i(ψ+ωt) iωt
E(x, y, z, t)  =  Re(a e )  =  Re(a e ) (9.1.3)

where ϕ=2πz/λ and A=e−iϕ is known as the complex amplitude. If n is a refractive index of a medium where the light propagates,
the light wave traverses a distance d in such a medium. The equivalent optical path in this case is
p  =  n  ⋅  d (9.1.4)

When two light waves are superposed, the result intensity at any point depends on whether reinforce or cancel each other (Figure
9.1.1). This is well known phenomenon of interference. We will assume that two waves are propagating in the same direction and

9.1.1 https://chem.libretexts.org/@go/page/55935
are polarized with their field vectors in the same plane. We will also assume that they have the same frequency. The complex
amplitude at any point in the interference pattern is then the sum of the complex amplitudes of the two waves, so that we can write,
A  =  A1   +  A2 (9.1.5)

where A1=a1exp(−iϕ 1) and A2=a2exp(−iϕ 2) are the complex amplitudes of two waves. The resultant intensity is, therefore,
2 1/2
I   =  |A|   =  I1   +  I2   +  2(I1 I2 ) cos(Δψ) (9.1.6)

where I1 and I2 are the intensities of two waves acting separately, and Δϕ=ϕ 1−ϕ 2 is the phase difference between them. If the two
waves are derived from a common source, the phase difference corresponds to an optical path difference,
Δp  =  (λ/2π)Δψ (9.1.7)

Figure 9.1.1 The scheme of interferometric wave interaction when two waves interact with each other, the amplitude of resulting
wave will increase or decrease. The value of this amplitude depends on phase difference between two original waves.
If Δϕ, the phase difference between the beams, varies linearly across the field of view, the intensity varies cosinusoidally, giving
rise to alternating light and dark bands or fringes (Figure 9.1.1). The intensity of an interference pattern has its maximum value:
1/2
Imax   =  I1   +  I2   +  2(I1 I2 ) (9.1.8)

when Δϕ=2mπ, where m is an integer and its minimum value i determined by:
1/2
Imin   =  I1   +  I2   −  2(I1 I2 ) (9.1.9)

when Δϕ=(2m+1)π The principle of interferometry is widely used to develop many types of interferometric set ups. One of the
earliest set ups is Michelson interferometry. The idea of this interferometry is quite simple: interference fringes are produced by
splitting a beam of monochromatic light so that one beam strikes a fixed mirror and the other a movable mirror. An interference
pattern results when the reflected beams are brought back together. The Michelson interferometric scheme is shown in Figure
9.1.2.

9.1.2 https://chem.libretexts.org/@go/page/55935
Figure 9.1.2 Schematic representation of a Michelson interferometry set-up.
The difference of path lengths between two beams is 2x because beams traverse the designated distances twice. The interference
occurs when the path difference is equal to integer numbers of wavelengths,
Δp  =  2x mλ(m = 0, ±1, ±2. . . ) (9.1.10)

Modern interferometric systems are more complicated. Using special phase-measurement techniques they capable to perform much
more accurate height measurements than can be obtained just by directly looking at the interference fringes and measuring how
they depart from being straight and equally spaced. Typically interferometric system consist of lights source, beamsplitter,
objective system, system of registration of signals and transformation into digital format and computer which process data. Vertical
scanning interferometry is contains all these parts. Figure 9.1.3 shows a configuration of a VSI interferometric system.

Figure 9.1.3 Schematic representation of the Vertical scanning interferometry (VSI) system.
Many of modern interferometric systems use Mirau objective in their constructions. Mireau objective is based on a Michelson
interferometer. This objective consists of a lens, a reference mirror and a beamsplitter. The idea of getting interfering beams is
simple: two beams (red lines) travel along the optical axis. Then they are reflected from the reference surface and the sample
surface respectively (blue lines). After this these beams are recombined to interfere with each other. An illumination or light source
system is used to direct light onto a sample surface through a cube beam splitter and the Mireau objective. The sample surface
within the field of view of the objective is uniformly illuminated by those beams with different incidence angles. Any point on the
sample surface can reflect those incident beams in the form of divergent cone. Similarly, the point on the reference symmetrical
with that on the sample surface also reflects those illuminated beams in the same form.

9.1.3 https://chem.libretexts.org/@go/page/55935
The Mireau objective directs the beams reflected of the reference and the sample surface onto a CCD (charge-coupled device)
sensor through a tube lens. The CCD sensor is an analog shift register that enables the transportation of analog signals (electric
charges) through successive stages (capacitors), controlled by a clock signal. The resulting interference fringe pattern is detected by
CCD sensor and the corresponding signal is digitized by a frame grabber for further processing with a computer.
The distance between a minimum and a maximum of the interferogram produced by two beams reflected from the reference and
sample surface is known. That is, exactly half the wavelength of the light source. Therefore, with a simple interferogram the
vertical resolution of the technique would be also limited to λ/2. If we will use a laser light as a light source with a wavelength of
300 nm the resolution would be only 150 nm. This resolution is not good enough for a detailed near-atomic scale investigation of
crystal surfaces. Fortunately, the vertical resolution of the technique can be improved significantly by moving either the reference
or the sample by a fraction of the wavelength of the light. In this way, several interferograms are produced. Then they are all
overlayed, and their phase shifts compared by the computer software Figure. This method is widely known as phase shift
interferometry (PSI).

Figure 9.1.4 : Sketch illustrating phase-shift technology. The sample is continuously moved along the vertical axes in order to scan
surface topography. All interferograms are automatically overlayed using computer software.
Most optical testing interferometers now use phase-shifting techniques not only because of high resolution but also because phase-
shifting is a high accuracy rapid way of getting the interferogram information into the computer. Also usage of this technique
makes the inherent noise in the data taking process very low. As the result in a good environment angstrom or sub-angstrom surface
height measurements can be performed. As it was said above, in phase-shifting interferometry the phase difference between the
interfering beams is changed at a constant rate as the detector is read out. Once the phase is determined across the interference
field, the corresponding height distribution on the sample surface can be determined. The phase distribution φ(x, y) is recorded by
using the CCD camera.
Let’s assign A(x, y), B(x, y), C (x, y) and D(x, y) to the resulting interference light intensities which are corresponded to phase-
shifting steps of 0, π/2, π and 3π/2. These intensities can be obtained by moving the reference mirror through displacements of λ/8,
λ/4 and 3λ/8, respectively. The equations for the resulting intensities would be:
A(x, y)  =  I1 (x, y)  +  I2 (x, y) cos(α(x, y)) (9.1.11)

B(x, y)  =  I1 (x, y)  −  I2 (x, y) sin(α(x, y)) (9.1.12)

C (x, y)  =  I1 (x, y)  −  I2 (x, y) cos(α(x, y)) (9.1.13)

D(x, y)  =  I1 (x, y)  +  I2 (x, y) sin(α(x, y)) (9.1.14)

where I (x, y) and I (x, y) are two overlapping beams from two symmetric points on the test surface and the reference
1 2

respectively. Solving Equations 9.1.11 - 9.1.14, the phase map φ(x, y) of a sample surface will be given by the relation:
B(x, y)  −  D(x, y)
ψ(x, y)  =   (9.1.15)
A(x, y)  −  C (x, y)

9.1.4 https://chem.libretexts.org/@go/page/55935
Once the phase is determined across the interference field pixel by pixel on a two-dimensional CCD array, the local height
distribution/contour, h(x, y), on the test surface is given by
λ
h(x, y)  =   ψ(x, y) (9.1.16)

Normally the resulted fringe can be in the form of a linear fringe pattern by adjusting the relative position between the reference
mirror and sample surfaces. Hence any distorted interference fringe would indicate a local profile/contour of the test surface.
It is important to note that the Mireau objective is mounted on a capacitive closed-loop controlled PZT (piezoelectric actuator) as to
enable phase shifting to be accurately implemented. The PZT is based on piezoelectric effect referred to the electric potential
generated by applying pressure to piezoelectric material. This type of materials is used to convert electrical energy to mechanical
energy and vice-versa. The precise motion that results when an electric potential is applied to a piezoelectric material has an
importance for nanopositioning. Actuators using the piezo effect have been commercially available for 35 years and in that time
have transformed the world of precision positioning and motion control.
Vertical scanning interferometer also has another name; white-light interferometry (WLI) because of using the white light as a
source of light. With this type of source a separate fringe system is produced for each wavelength, and the resultant intensity at any
point of examined surface is obtained by summing these individual patterns. Due to the broad bandwidth of the source the coherent
length L of the source is short:
2
λ
L  =   (9.1.17)
nΔλ

where λ is the center wavelength, n is the refractive index of the medium, ∆λ is the spectral width of the source. In this way good
contrast fringes can be obtained only when the lengths of interfering beams pathways are closed to each other. If we will vary the
length of a pathway of a beam reflected from sample, the height of a sample can be determined by looking at the position for which
a fringe contrast is a maximum. In this case interference pattern exist only over a very shallow depth of the surface. When we vary
a pathway of sample-reflected beam we also move the sample in a vertical direction in order to get the phase at which maximum
intensity of fringes will be achieved. This phase will be converted in height of a point at the sample surface.
The combination of phase shift technology with white-light source provides a very powerful tool to measure the topography of
quite rough surfaces with the amplitude in heights about and the precision up to 1-2 nm. Through a developed software package for
quantitatively evaluating the resulting interferogram, the proposed system can retrieve the surface profile and topography of the
sample objects Figure 9.1.5.

Figure 9.1.5 : Example of muscovite surface topography, obtained by using VSI- 50x objective.
A Comparison of Common Methods to Determine Surface Topography: SEM, AFM and VSI
Except the interferometric methods described above, there are a several other microscopic techniques for studying crystal surface
topography. The most common are scanning electron microscopy (SEM) and atomic force microscopy (AFM). All these techniques

9.1.5 https://chem.libretexts.org/@go/page/55935
are used to obtain information about the surface structure. However they differ from each other by the physical principles on which
they based.

Scanning Electron Microscopy


SEM allows us to obtain images of surface topography with the resolution much higher than the conventional light microscopes do.
Also it is able to provide information about other surface characteristics such as chemical composition, electrical conductivity etc,
see Figure 9.1.6. All types of data are generated by the reflecting of accelerated electron beams from the sample surface. When
electrons strike the sample surface, they lose their energy by repeated random scattering and adsorption within an outer layer into
the depth varying from 100 nm to 5 microns.

Figure 9.1.6 Scheme of electron beam-sample interaction at SEM analysis


The thickness of this outer layer also knows as interactive layer depends on energy of electrons in the beam, composition and
density of a sample. Result of the interaction between electron beam and the surface provides several types of signals. The main
type is secondary or inelastic scattered electrons. They are produced as a result of interaction between the beam of electrons and
weakly bound electrons in the conduction band of the sample. Secondary electrons are ejected from the k-orbitals of atoms within
the surface layer of thickness about a few nanometers. This is because secondary electrons are low energy electrons (<50 eV), so
only those formed within the first few nanometers of the sample surface have enough energy to escape and be detected. Secondary
backscattered electrons provide the most common signal to investigate surface topography with lateral resolution up to 0.4 - 0.7
nm.
High energy beam electrons are elastic scattered back from the surface. This type of signal gives information about chemical
composition of the surface because the energy of backscattered electrons depends on the weight of atoms within the interaction
layer. Also this type of electrons can form secondary electrons and escape from the surface or travel father into the sample than the
secondary. The SEM image formed is the result of the intensity of the secondary electron emission from the sample at each x,y data
point during the scanning of the surface.

Atomic Force Microscopy


AFM is a very popular tool to study surface dissolution. AFM set up consists of scanning a sharp tip on the end of a flexible
cantilever which moves across a sample surface. The tips typically have an end radius of 2 to 20 nm, depending on tip type. When
the tip touch the surface the forces of these interactions leads to deflection of a cantilever. The interaction between tip and sample
surface involve mechanical contact forces, van der Waals forces, capillary forces, chemical bonding, electrostatic forces, magnetic
forces etc. The deflection of a cantilever is usually measured by reflecting a laser beam off the back of the cantilever into a split
photodiode detector. A schematic drawing of AFM can be seen in Figure 9.1.7. The two most commonly used modes of operation
are contact mode AFM and tapping mode AFM, which are conducted in air or liquid environments.

9.1.6 https://chem.libretexts.org/@go/page/55935
Figure 9.1.7 Schematic drawing of an AFM apparatus.
Working under the contact mode AFM scans the sample while monitoring the change in cantilever deflection with the split
photodiode detector. Loop maintains a constant cantilever reflection by vertically moving the scanner to get a constant signal. The
distance which the scanner goes by moving vertically at each x,y data point is stored by the computer to form the topographic
image of the sample surface. Working under the tapping mode AFM oscillates the cantilever at its resonance frequency
(typically~300 kHz) and lightly “taps” the tip on the surface during scanning. The electrostatic forces increase when tip gets close
to the sample surface, therefore the amplitude of the oscillation decreases. The laser deflection method is used to detect the
amplitude of cantilever oscillation. Similar to the contact mode, feedback loop maintains a constant oscillation amplitude by
moving the scanner vertically at every x,y data point. Recording this movement forms the topographical image. The advantage of
tapping mode over contact mode is that it eliminates the lateral, shear forces present in contact mode. This enables tapping mode to
image soft, fragile, and adhesive surfaces without damaging them while work under contact mode allows the damage to occur.

Comparison of Techniques
All techniques described above are widely used in studying of surface nano- and micromorphology. However, each method has its
own limitations and the proper choice of analytical technique depends on features of analyzed surface and primary goals of
research.
All these techniques are capable to obtain an image of a sample surface with quite good resolution. The lateral resolution of VSI is
much less, then for other techniques: 150 nm for VSI and 0.5 nm for AFM and SEM. Vertical resolution of AFM (0.5 Ǻ) is better
then for VSI (1 - 2 nm), however VSI is capable to measure a high vertical range of heights (1 mm) which makes possible to study
even very rough surfaces. On the contrary, AFM allows us to measure only quite smooth surfaces because of its relatively small
vertical scan range (7 µm). SEM has less resolution, than AFM because it requires coating of a conductive material with the
thickness within several nm.
The significant advantage of VSI is that it can provide a large field of view (845 × 630 µm for 10x objective) of tested surfaces.
Recent studies of surface roughness characteristics showed that the surface roughness parameters increase with the increasing field
of view until a critical size of 250,000 µm is reached. This value is larger then the maximum field of view produced by AFM (100
× 100 µm) but can be easily obtained by VSI. SEM is also capable to produce images with large field of view. However, SEM is
able to provide only 2D images from one scan while AFM and VSI let us to obtain 3D images. It makes quantitative analysis of
surface topography more complicated, for example, topography of membranes is studied by cross section and top view images.
Table 9.1.1 A comparison of VSI sample and resolution with AFM and SEM.
VSI AFM SEM

Lateral resolution 0.5 - 1.2µm 0.5 nm 0.5 - 1 nm

Vertical Resolution 2 nm 0.5 Å Only 2D images

Field of View 845 x 630 µm (10x objective) 100 x 100 µm 1 - 2 mm

Vertical Range of Scan 1 mm 10 µm -

Required coating of a conducted


Preparation of Sample - -
material

Required environment Air Air, liquid Vacuum

9.1.7 https://chem.libretexts.org/@go/page/55935
The Experimental Studying of Surface Processes Using Microscopic Techniques
The limitations of each technique described above are critically important to choose appropriate technique for studying surface
processes. Let’s explore application of these techniques to study dissolution of crystals.
When crystalline matter dissolves the changes of the crystal surface topography can be observed by using microscopic techniques.
If we will apply an unreactive mask (silicon for example) on crystal surface and place a crystalline sample into the experiment
reactor then we get two types of surfaces: dissolving and remaining the same or unreacted. After some period of time the crystal
surface starts to dissolve and change its z-level. In order to study these changes ex situ we can pull out a sample from the reaction
cell then remove a mask and measure the average height difference Δh bar between the unreacted and dissolved areas. The average
heights of dissolved and unreacted areas are obtained through digital processing of data obtained by microscopes. The velocity of
normal surface retreat vSNR during the time interval ∆t is defined by 9.1.18
Δℏ
νSN R   =   (9.1.18)
Δt

Dividing this velocity by the molar volume (cm3/mol) gives a global dissolution rate in the familiar units of moles per unit area
per unit time:
νSN R
R  =   (9.1.19)
¯
V

This method allows us to obtain experimental values of dissolution rates just by precise measuring of average surface heights.
Moreover, using this method we can measure local dissolution rates at etch pits by monitoring changes in the volume and density
of etch pits across the surface over time. VSI technique is capable to perform these measurements because of large vertical range of
scanning. In order to get precise values of rates which are not depend on observing place of crystal surface we need to measure
enough large areas. VSI technique provides data from areas which are large enough to study surfaces with heterogeneous
dissolution dynamics and obtain average dissolution rates. Therefore, VSI makes possible to measure rates of normal surface
retreat during the dissolution and observe formation, growth and distribution of etch pits on the surface.
However, if the mechanism of dissolution is controlled by dynamics of atomic steps and kink sites within a smooth atomic surface
area, the observation of the dissolution process need to use a more precise technique. AFM is capable to provide information about
changes in step morphology in situ when the dissolution occurs. For example, immediate response of the dissolved surface to the
changing of environmental conditions (concentrations of ions in the solution, pH etc.) can be studied by using AFM.
SEM is also used to examine micro and nanotexture of solid surfaces and study dissolution processes. This method allows us to
observe large areas of crystal surface with high resolution which makes possible to measure a high variety of surfaces. The
significant disadvantage of this method is the requirement to cover examine sample by conductive substance which limits the
resolution of SEM. The other disadvantage of SEM is that the analysis is conducted in vacuum. Recent technique, environmental
SEM or ESEM overcomes these requirements and makes possible even examine liquids and biological materials. The third
disadvantage of this technique is that it produces only 2D images. This creates some difficulties to measure Δhbar within the
dissolving area. One of advantages of this technique is that it is able to measure not only surface topography but also chemical
composition and other surface characteristics of the surface. This fact is used to monitor changing in chemical composition during
the dissolution.

Dual Polarization Interferometry for Thin Film Characterization


Overview

As research interests begin to focus on progressively smaller dimensions, the need for nanoscale characterization techniques has
seen a steep rise in demand. In addition, the wide scope of nanotechnology across all fields of science has perpetuated the
application of characterization techniques to a multitude of disciplines. Dual polarization interferometry (DPI) is an example of a
technique developed to solve a specific problem, but was expanded and utilized to characterize fields ranging surface science,
protein studies, and crystallography. With a simple optical instrument, DPI can perform label-free sensing of refractive index and
layer thickness in real time, which provides vital information about a system on the nanoscale, including the elucidation of
structure-function relationships.
History
DPI was conceived in 1996 by Dr. Neville Freeman and Dr. Graham Cross (Figure 9.1.8) when they recognized a need to measure
refractive index and adlayer thickness simultaneously in protein membranes to gain a true understanding of the dynamics of the

9.1.8 https://chem.libretexts.org/@go/page/55935
system. They patented the technique in 1998, and the instrument was commercialized by Farfield Group Ltd. in 2000.

Figure 9.1.8 English physicist Graham Cross. Copyright: Durham University.


Freeman and Cross unveiled the first full publication describing the technique in 2003, where they chose to measure well-known
protein systems and compare their data to X-ray crystallography and neutron reflection data. The first system they studied was
sulpho-NHS-LC-biotin coated with streptavidin and a biotinylated peptide capture antibody, and the second system was BS3 coated
with anti-HSA. Molecular structures are shown in Figure 9.1.9. Their results showed good agreement with known layer
thicknesses, and the method showed clear advantages over neutron reflection and surface plasmon resonance. A schematic and
picture of the instrument used by Freeman and Cross in this publication is shown in Figure 9.1.10 and Figure 9.1.11, respectively.

Figure 9.1.9 Molecular structures of (a) sulpho-NHS-LC-biotin and (b) bis-(sulphosuccinimydyl) suberate (BS3). Reprinted with
permission from G. H. Cross, A. A. Reeves, S. Brand, J. F. Popplewell, L. L. Peel, M. J. Swann, and N. J. Freeman, Biosens.
Bioelectron., 2003, 19, 383. Copyright: Biosensors & Bioelectronics (2003).

Figure 9.1.10 The first DPI schematic and instrument. Reprinted with permission from G. H. Cross, A. A. Reeves, S. Brand, J. F.
Popplewell, L. L. Peel, M. J. Swann, and N. J. Freeman, Biosens. Bioelectron., 2003, 19, 383. Copyright: Biosensors &
Bioelectronics (2003).

9.1.9 https://chem.libretexts.org/@go/page/55935
Figure 9.1.11 Picture of the DPI instrument used by Freeman and Cross.
Instrumentation

Theory
The optical power of DPS comes from the ability to measure two different interference fringe patterns simultaneously in real time.
Phase changes in these fringe patterns result from changes in refractive index and layer thickness that can be detected by the
waveguide interferometer, and resolving these interference patterns provides refractive index and layer thickness values.

Optics
A representation of the interferometer is shown in Figure 9.1.12. The interferometer is composed of a simplified slab waveguide,
which guides a wave of light in one transverse direction without scattering. A broad laser light is shone on the side facet of stacked
waveguides separated with a cladding layer, where the waveguides act as a sensing layer and a reference layer that produce an
interference pattern in a decaying (evanescent) electric field.

Figure 9.1.12 Basic representation of a slab waveguide interferometer. Reprinted with permission from M. Wang, S. Uusitalo, C.
Liedert, J. Hiltunen, L. Hakalahti, and R. Myllyla, Appl. Optics, 2012, 12, 1886. Copyright: Applied Optics (2012).
A full representation of DPI theory and instrumentation is shown in Figure 9.1.13 and Figure 9.1.14, respectively. The layer
thickness and refractive index measurements are determined by measuring two phase changes in the system simultaneously
because both transverse-electric and transverse-magnetic polarizations are allowed through the waveguides. Phase changes in each
polarization of the light wave are lateral shifts of the wave peak from a given reference peak. The phase shifts are caused by
changes in refractive index and layer thickness that result from molecular fluctuations in the sample. Switching between transverse-
electric and transverse-magnetic polarizations happens very rapidly at 2 ms, where the switching mechanism is performed by a
liquid crystal wave plate. This enables real-time measurements of the parameters to be obtained simultaneously.

9.1.10 https://chem.libretexts.org/@go/page/55935
Figure 9.1.13 DPI sensing apparatus and fringe pattern collection from transverse-magnetic and transverse-electric polarizations of
light. Adapted from J. Escorihuela, M.A. Gonzalez-Martinez, J.L. Lopez-Paz, R. Puchades, A. Maquieira, and D. Gimenez-
Romero, Chem. Rev., 2015, 115, 265. Copyright: Chemical Reviews, (2015).

Figure 9.1.14 Fringe pattern detection of the waveguides and phase change determination between the sensing and reference
interference patterns. Adapted from J. Escorihuela, M.A. Gonzalez-Martinez, J.L. Lopez-Paz, R. Puchades, A. Maquieira, and D.
Gimenez-Romero, Chem. Rev., 2015, 115, 265. Copyright: Chemical Reviews, (2015).
Comparison of DPI with Other Techniques

Initial DPI Evaluations


The first techniques rigorously compared to DPI were neutron reflection (NR) and X-ray diffraction. These studies demonstrated
that DPI had a very high precision of 40 pm, which is comparable to NR and superior to X-ray diffraction. Additionally, DPI can
provide real time information and conditions similar to an in-vivo environment, and the instrumental requirements are far simpler
than those for NR. However, NR and X-ray diffraction are able to provide structural information that DPI cannot determine.

DPI Comparison with orthogonal Analytical Techniques

9.1.11 https://chem.libretexts.org/@go/page/55935
Comparisons between DPI and alternative techniques have been performed since the initial evaluations, with techniques including
surface plasmon resonance (SPR), atomic force microscopy (AFM), and quartz crystal microbalance with dissipation monitoring
(QCM-D).
SPR is well-established for characterizing protein adsorption and has been used before DPI was developed. These techniques are
very similar in that they both use an optical element based on an evanescent field, but they differ greatly in the method of
calculating the mass of adsorbed protein. Rigorous testing showed that both tests give very accurate results, but their strengths
differ. Because SPR uses spot-testing with an area of 0.26 mm2 while DPI uses the average measurements over the length of the
entire 15 mm chip, SPR is recommended for use in kinetic studies where diffusion in involved. However, DPI shows much greater
accuracy than SPR when measuring refractive index and layer thickness.
Atomic Force Microscopy is a very different analytical technique than DPI because it is a type of microscopy used for high-
resolution surface characterization. Hence, AFM and DPI are well-suited to be used in conjunction because AFM can provide
accurate molecular structures and surface mapping while DPI provides layer thickness that AFM cannot determine.
QCM-D is a technique that can be used with DPI to provide complementary data. QCM-D differs from DPI by calculating both
mass of the solvent and the mass of the adsorbed protein layer. These techniques can be used together to determine the amount of
hydration in the adsorbed layer. QCM-D can also quantify the supramolecular conformation of the adlayer using energy dissipation
calculations, while DPI can detect these conformational changes using birefringence, thus making these techniques orthogonal. One
way that DPI is superior to QCM-D is that the latter techniques loses accuracy as the film becomes very thin, while DPI retains
accuracy throughout the angstrom scale.
A tabulated representation of these techniques and their ability to measure structural detail, in-vivoconditions, and real time data is
shown in Table 9.1.2.
Table 9.1.2 : Comparison of DPI with other analytical techniques. Data reproduced from J. Escorihuela, M. A. Gonzalez-Martinez, J. L.
Lopez-Paz, R. Puchades, A. Maquieira, and D. Gimenez-Romero, Chem. Rev., 2015, 115, 265.a Close to in-vivo means that the sensor can
provide information that is similar to those experiences under in-vivo conditions. Copyright: Chemical Reviews, (2015).
Technique Real Time Close to In-vivo Structural Details

QCM-D Yes Yes Medium

SPR Yes Yes Low

X-ray No No Very high

AFM No No High

NR No Yes High

DPI Yes Yes Medium

Applications of DPI

Protein Studies
DPI has been most heavily applied to protein studies. It has been used to elucidate membrane crystallization, protein orientation in
a membrane, and conformational changes. It has also been used to study protein-protein interactions, protein-antibody interactions,
and the stoichiometry of binding events.

Thin Film Studies


Since its establishment using protein interaction studies, DPI has seen its applications expanded to include thin film studies. DPI
was compared to ellipsometry and QCM-D studies to indicate that it can be applied to heterogeneous thin films by applying revised
analytical formulas to estimate the thickness, refractive index, and extinction coefficient of heterogeneous films that absorb light. A
non-uniform density distribution model was developed and tested on polyethylenimine deposited onto silica and compared to
QCD-M measurements. Additionally, this revised model was able to calculate the mass of multiple species of molecules in
composite films, even if the molecules absorbed different amounts of light. This information is valuable for providing surface
composition. The structure of polyethylenimine used to form an adsorbing film is shown in Figure 9.1.15.

9.1.12 https://chem.libretexts.org/@go/page/55935
Figure 9.1.15 Structure of polyethylenimine used to form a thin film for DPI measurements.
A challenge of measuring layer thickness in thin films such as polyethylenimine is that DPI’s evanescent field will create
inaccurate measurements in inhomogeneous films as the film thickness increases. An error of approximately 5% was seen when
layer thickness was increased to 90 nm. Data from this study determining the densities throughout the polyethylenimine film are
shown in Figure 9.1.16.

Figure 9.1.16 Density distribution of a polyethylenimine film using heterogeneous layer equations for DPI and QCM-D.
Reproduced from P. D. Coffey, M. J.Swann, T. A. Waigh, Q. Mua, and J. R. Lu, RSC Adv., 2013, 3, 3316.
Thin Layer Adsorption Studies
Similar to thin film characterization studies, thin layers of adsorbed polymers have also been elucidated using DPI. It has been
demonstrated that two different adsorption conformations of polyacrylamide form on resin, which provides useful information
about adsorption behaviors of the polymer. This information is industrially important because polyacrylamide is widely used
through the oil industry, and the adsorption of polyacrylamide onto resin is known to affect the oil/water interfacial stability.
Initial adsorption kinetics and conformations were also illuminated using DPI on bottlebrush polyelectrolytes. Bottlebrush
polyelectrolytes are show in Figure 9.1.17. It was shown that polyelectrolytes with high charge density initially adsorbed in layers
that were parallel to the surface, but as polyelectrolytes were replaced with low charge density species, alignment changed to prefer
perpendicular arrangement to the surface.

Figure 9.1.17 A representation of bottlebrush polyelectrolytes and how they adsorb to a layer differently over time as determined
by DPI. Reproduced from G. Bijelic, A. Shovsky, I. Varga, R. Makuska, and P. M. Claesson, J. Colloid Interf. Sci., 2010, 348, 189.
Copyright: Journal of Colloid and Interface Science, (2010).

9.1.13 https://chem.libretexts.org/@go/page/55935
Hg2+ Biosensing Studies
In 2009, it was shown by Wang et al. that DPI could be used for small molecule sensing. In their first study describing this use of
DPI, they used single stranded DNA that was rich in thymine to complex Hg2+ ions. When DNA complexed with Hg2+, the DNA
transformed from a random coil structure to a hairpin structure. This change in structure could be detected by DPI at Hg2+
concentrations smaller than the threshold concentration allowed in drinking water, indicating the sensitivity of this label-free
method for Hg2+ detection. High selectivity was indicated when the authors did not observe similar structural changes for Mg2+,
Ca2+, Mn2+, Fe3+, Cd2+, Co2+, Ni2+, Zn2+ or Pb2+ ions. A graphical description of this experiment is shown in Figure. Wang et
al. later demonstrated that biosensing of small molecules and other metal cations can be achieved using other forms of
functionalized DNA that specifically bind the desired analytes. Examples of molecules detected in this manner are shown in Figure
9.1.18.

Figure 9.1.18 Selective Hg2+ detection using single strand DNA to complex the cation and measure the conformational changes in
the DNA with DPI. Reproduced from J. Escorihuela, M. A. Gonzalez-Martinez, J. L. Lopez-Paz, R. Puchades, A. Maquieira, and
D. Gimenez-Romero, Chem. Rev., 2015, 115, 265. Copyright: Chemical Reviews, (2015).

Figure 9.1.19 Small molecules detected using DPI measurements of functionalized DNA biosensors. Reproduced from J.
Escorihuela, M. A. Gonzalez-Martinez, J. L. Lopez-Paz, R. Puchades, A. Maquieira, and D. Gimenez-Romero, Chem. Rev., 2015,
115, 265. Copyright: Chemical Reviews, (2015).

9.1: Interferometry is shared under a CC BY 4.0 license and was authored, remixed, and/or curated by Pavan M. V. Raja & Andrew R. Barron via
source content that was edited to conform to the style and standards of the LibreTexts platform; a detailed edit history is available upon request.

9.1.14 https://chem.libretexts.org/@go/page/55935
9.2: Atomic Force Microscopy (AFM)
Atomic force microscopy (AFM) is a high-resolution form of scanning probe microscopy, also known as scanning force
microscopy (SFM). The instrument uses a cantilever with a sharp tip at the end to scan over the sample surface (Figure 9.2.1). As
the probe scans over the sample surface, attractive or repulsive forces between the tip and sample, usuually in the form of Van Der
Waal forces but also can be a number of others such as electrostatic and hydrophobic/hydrophilic, cause a deflection of the
cantilever. The deflection is measured by a laser (Figure 9.2.2) which is reflected off the cantilever into photodiodes. As one of the
photodiodes collects more light, it creates an output signal that is processed and provides information about the vertical bending of
the cantilever. This data is then sent to a scanner that controls the height of the probe as it moves across the surface. The variance in
height applied by the scanner can then be used to produce a three-dimensional topographical representation of the sample.

Figure 9.2.1 Simple schematic of atomic force microscope (AFM) apparatus. Adapted from H. G. Hansma, Department of Physics,
University of California, Santa Barbara. (Public Domain; Nobelium via Wikipedia)

Modes of Operation
Contact Mode
The contact mode method utilizes a constant force for tip-sample interactions by maintaining a constant tip deflection (Figure
9.2.2.The tip communicates the nature of the interactions that the probe is having at the surface via feedback loops and the scanner

moves the entire probe in order to maintain the original deflection of the cantilever. The constant force is calculated and maintained
by using Hooke's Law, 9.2.2. This equation relates the force (F), spring constant (k), and cantilever deflection (x). Force constants
typically range from 0.01 to 1.0 N/m. Contact mode usually has the fastest scanning times but can deform the sample surface. It is
also only the only mode that can attain "atomic resolution."
F   =   − kx (9.2.1)

Figure 9.2.2 Schematic diagram of probe and surface interaction in contact mode.

Tapping Mode
In the tapping mode the cantilever is externally oscillated at its fundamental resonance frequency (Figure 9.2.3). A piezoelectric on
top of the cantilever is used to adjust the amplitude of oscillation as the probe scans across the surface. The deviations in the
oscillation frequency or amplitude due to interactions between the probe and surface are measured, and provide information about
the surface or types of material present in the sample. This method is gentler than contact AFM since the tip is not dragged across
the surface, but it does require longer scanning times. It also tends to provide higher lateral resolution than contact AFM.

Figure 9.2.3 Diagram of probe and surface interaction in tapping mode.

9.2.1 https://chem.libretexts.org/@go/page/55928
Noncontact Mode
For noncontact mode the cantilever is oscillated just above its resonance frequency and this frequency is decreased as the tip
approaches the surface and experiences the forces associated with the material (Figure 9.2.4). The average tip-to-sample distance is
measured as the oscillation frequency or amplitude is kept constant, which then can be used to image the surface. This method
exerts very little force on the sample, which extends the lifetime of the tip. However, it usually does not provide very good
resolution unless placed under a strong vacuum.

Figure 9.2.4 Diagram of probe and surface interaction in noncontact mode.

Experimental Limitations
A common problem seen in AFM images is the presence of artifacts which are distortions of the actual topography, usually either
due to issues with the probe, scanner, or image processing. The AFM scans slowly which makes it more susceptible to external
temperature fluctuations leading to thermal drift. This leads to artifacts and inaccurate distances between topographical features.
It is also important to consider that the tip is not perfectly sharp and therefore may not provide the best aspect ratio, which leads to
a convolution of the true topography. This leads to features appearing too large or too small since the width of the probe cannot
precisely move around the particles and holes on the surface. It is for this reason that tips with smaller radii of curvature provide
better resolution in imaging. The tip can also produce false images and poorly contrasted images if it is blunt or broken.
The movement of particles on the surface due to the movement of the cantilever can cause noise, which forms streaks or bands in
the image. Artifacts can also be made by the tip being of inadequate proportions compared to the surface being scanned. It is for
this reason that it is important to use the ideal probe for the particular application.
Sample Size and Preparation
The sample size varies with the instrument but a typical size is 8 mm by 8 mm with a typical height of 1 mm. Solid samples present
a problem for AFM since the tip can shift the material as it scans the surface. Solutions or dispersions are best for applying as
uniform of a layer of material as possible in order to get the most accurate value of particles’ heights. This is usually done by spin-
coating the solution onto freshly cleaved mica which allows the particles to stick to the surface once it has dried.
Applications of AFM

AFM is particularly versatile in its applications since it can be used in ambient temperatures and many different environments. It
can be used in many different areas to analyze different kinds of samples such as semiconductors, polymers, nanoparticles,
biotechnology, and cells amongst others. The most common application of AFM is for morphological studies in order to attain an
understanding of the topography of the sample. Since it is common for the material to be in solution, AFM can also give the user an
idea of the ability of the material to be dispersed as well as the homogeneity of the particles within that dispersion. It also can
provide a lot of information about the particles being studied such as particle size, surface area, electrical properties, and chemical
composition. Certain tips are capable of determining the principal mechanical, magnetic, and electrical properties of the material.
For example, in magnetic force microscopy (MFM) the probe has a magnetic coating that senses magnetic, electrostatic, and
atomic interactions with the surface. This type of scanning can be performed in static or dynamic mode and depicts the magnetic
structure of the surface.

AFM of Carbon Nanotubes


Atomic force microscopy is usually used to study the topographical morphology of these materials. By measuring the thickness of
the material it is possible to determine if bundling occurred and to what degree. Other dimensions of the sample can also be
measured such as the length and width of the tubes or bundles. It is also possible to detect impurities, functional groups (Figure
9.2.5), or remaining catalyst by studying the images. Various methods of producing nanotubes have been found and each

demonstrates a slightly different profile of homogeneity and purity. These impurities can be carbon coated metal, amorphous
carbon, or other allotropes of carbon such as fullerenes and graphite. These facts can be utilized to compare the purity and
homogeneity of the samples made from different processes, as well as monitor these characteristics as different steps or reactions
are performed on the material. The distance between the tip and the surface has proven itself to be an important parameter in

9.2.2 https://chem.libretexts.org/@go/page/55928
noncontact mode AFM and has shown that if the tip is moved past the threshold distance, approximately 30 μm, it can move or
damage the nanotubes. If this occurs, a useful characterization cannot be performed due to these distortions of the image.

Figure 9.2.5 AFM image of a polyethyleneimine-functionalized single walled carbon nanotube (PEI-SWNT) showing the presence
of PEI “globules” on the SWNT. Adapted from E. P. Dillon, C. A. Crouse, and A. R. Barron, ACS Nano, 2008, 2, 156.

AFM of Fullerenes
Atomic force microscopy is best applied to aggregates of fullerenes rather than individual ones. While the AFM can accurately
perform height analysis of individual fullerene molecules, it has poor lateral resolution and it is difficult to accurately depict the
width of an individual molecule. Another common issue that arises with contact AFM and fullerene deposited films is that the tip
shifts clusters of fullerenes which can lead to discontinuities in sample images.

A Practical Guide to Using the Nanoscope Atomic Force Microscopy


The following is intended as a guide for use of the Nanoscope AFM system within the Shared Equipment Authority at Rice
University (http://sea.rice.edu/). However, it can be adapted for similar AFM instruments.
Please familiarize yourself with the Figures. All relevant parts of the AFM setup are shown.
Initial Setup
Sign in.
Turn on each component shown in Figure 9.2.6.
1. The controller that powers the scope (the switch is at the back of the box)
2. The camera monitor
3. The white light source
Select imaging mode using the mode selector switch is located on the left hand side of the atomic force microscope (AFM)
base (Figure 9.2.7, there are three modes:
1. Scanning tunneling microscopy (STM)
2. Atomic force microscopy/lateral force microscopy (AFM/LFM)
3. Tapping mode atomic force microscopy (TM-AFM)

9.2.3 https://chem.libretexts.org/@go/page/55928
Figure 9.2.6 Schematic representation of the AFM computer, light source, camera set-up, and sample puck.

Figure 9.2.7 Schematic representation of the AFM.


Sample Preparation
Most particulate samples are imaged by immobilizing them onto mica sheet, which is fixed to a metal puck (Figure 9.2.6).
Samples that are in a solvent are easily deposited. To make a sample holder a sheet of Mica is punched out and stuck to double-
sided carbon tape on a metal puck. In order to ensure a pristine surface, the mica sheet is cleaved by removing the top sheet
with Scotch™ tape to reveal a pristine layer underneath. The sample can be spin coated onto the mica or air dried.
The spin coat method;
Use double-sided carbon sticky tape to secure the puck on the spin coater.
Load the sample by drop casting the sample solution onto the mica surface.
The sample must be dry to ensure that the tip remains clean.
Puck Mounting
1. Place the sample puck in the magnetic sample holder, and center the sample.
2. Verify that the AFM head is sufficiently raised to clear the sample with the probe. The sample plane is lower than the plane
defined by the three balls. The sample should sit below the nubs. Use the lever on the right side of the J-scanner to adjust
the height. (N.B. the labels up and down refer to the tip. “Tip up” moves the sample holder down to safety, and tip down
moves the sample up. Use caution when moving the sample up.)
3. Select the appropriate cantilever for the desired imaging mode. The tips are fragile and expensive (ca.$20 per tip) so handle
with care.
Contact AFM use a silicon nitride tip (NP).
Tapping AFM use a silicon tip (TESP).

9.2.4 https://chem.libretexts.org/@go/page/55928
Tip Mounting and Alignment
1. Mount a tip using the appropriate fine tweezers. Use the tweezers carefully to avoid possible misalignment. Work on a
white surface (a piece of paper or a paper towel) so that the cantilever can be easily seen. The delicate part of the tip the
cantilever is located at the beveled end and should not be handled at that end (shown in Figure 9.2.8). The tips are stored on
a tacky gel tape. Use care, as dropping the tip will break the cantilever. Think carefully about how you approach the tip
with the tweezers. Generally gripping it from the side is the best option. Once the tip is being held by the tweezers it needs
to be placed in the tip holder clamp. With one hand holding the tweezers, use the other hand to open the clip by pressing
down on the whole holder while it is lying on a flat hard surface. Once the clip is raised by downward pressure insert the tip
(Figure 9.2.9a). Make sure the tip is seated firmly and that the back end is in contact with the end of the probe groove, there
is a circular hole in the clamp. When the clamp holds the tip the hole should look like a half moon, with half filled with the
back straight end of the tip. The groove is larger than the tip, so try to put the tip in the same place each time you replace it
to improve reproducibility.
2. Carefully place the tip holder onto the three nubs to gently hold it in place. Bring the holder in at angle to avoid scraping it
against the sample (Figure 9.2.9 b).
3. Tighten the clamping screw located on the back of the AFM head to secure the cantilever holder and to guarantee electrical
contact. The screw is on the back of the laser head, at the center.
4. Find the cantilever on the video display. Move the translational stage to find it.
5. Adjust the focusing knob of the optical microscope (located above AFM head) to focus on the cantilever tip. Tightening the
focus knob moves the camera up. Focus on the dark blob on the right hand side of the screen as that is the cantilever.
6. Focus on the top mica surface, being careful not to focus on the bottom surface between the top of the double-sided carbon
tape and the mica surface. Generally you will see a bubble trapped between the carbon tape and the mica surface. If you are
focused on the top surface you can frequently see the reflection of the tip on the mica surface. The real focus is half way
between the two cantilever focus points.
7. Slowly lower the tip down to the surface, if the camera is focused properly onto the surface the cantilever tip will gradually
come into view. Keep lowering until the two tips images converge into one. Please note that you can crash the tip into the
surface if you go past this point. This is damaging to the tip and may not be possible to obtain an image if it happens, and
the tip may have to be replaced. You will know if this happens when looking at the cantilever tip if it goes from black to
bright white. At this point the tip is in contact with the surface and turns white as it is not reflecting light back into the
photo-diode , but instead into the camera.
8. Find the laser spot, it the spot is not visible on the camera screen look at the cantilever holder and see if it was visible. It
helps to lower the brightness of the white light, use the translational stage again to search for it.
9. Once the laser spot has been located use the X and Y laser adjustment knobs to align the laser spot roughly onto the tip of
the cantilever.
10. Maximize the sum signal using the photo-detector mirror lever located on the back of the head and the laser X and Y
translation. As long as the sum signal value is above 3.6 V, the instrument will work, but keep adjusting the X and Y
directions of the laser until the sum signal is as high as possible.
11. To ensure that the laser is centered on the photodiode, zero the detector signals using the mirror adjustment knobs located
on the top and back of the head. The knob on the top of the head adjusts TMAFM mode, and the knob at the rear of the
head adjusts AFM/LFM mode. The range is -9.9 V to 9.9 V in both modes. The number will change slowly at the extremes
of the range and quickly around 0 V. Ideally, the zeroed signal should be between ±0.1 V. Do this first in TMAFM mode,
then switch to AFM/LFM mode and try to zero the detector. Flip back and forth between the two modes a few times
(adjusting each time) until the value in both modes is as close to 0 V as possible. It will fluctuate during the experiment. If
there is steady drift, you can adjust it during the experiment. If the number won’t settle down, the laser could be at a bad
position on the cantilever. Move the laser spot and repeat (Figure 9.2.10). Always end this step in TMAFM mode.
12. Focus again on the sample surface.
13. The sample surface can still be moved with respect to the camera via the sample stage. In choosing a place to image
nanoparticles, avoid anything that you can see on the sample surface. The scale on the screen is 18 µm per cm.

Figure 9.2.8 Schematic views of the AFM tip.

9.2.5 https://chem.libretexts.org/@go/page/55928
Figure 9.2.9 Schematic view of (a) the tip holder and (b) the tip holder location in the AFM.

Figure 9.2.10 Schematic of the laser set-up.


Tip Tuning
1. Log onto computer.
2. The software is called Nanoscope. Close the version dialog box. Typically the screen on the left will allow adjustment of
software parameters, and the screen on the right will show the data.
3. On the adjustment screen, the two icons are to adjust the microscope (a picture of a microscope) and to perform data
analysis (a picture of a rainbow). Click the microscope icon.
4. Under the microscope pull down menu, choose profile and select tapping AFM. Don’t use another users profile. Use the
“tapping” AFM.
5. Before beginning tapping mode, the cantilever must be tuned to ensure correct operation. Each tip has its own resonance
frequency. The cantilever can be blindly auto-tuned or manually tuned. However the auto-tuning scheme can drive the
amplitude so high as to damage the tip.
Auto Tuning
1. Click on the cantilever tune icon.
2. Click the auto-tune button. The computer will enter the tuning procedure, automatically entering such parameters as set
point and drive amplitude. If tuned correctly, the drive frequency will be approximately 300 Hz.
Manually Tuning
1. Click on the cantilever tune icon.
2. Select manual tuning under the sweep controls menu.
3. The plot is of amplitude (white) and phase (yellow) versus frequency. The sweep width is the X-range. The central
frequency is the driving frequency which should be between 270-310 Hz. Typically the initial plot will not show any peaks,
and the X and Y settings will need to be adjusted in order to see the resonance plot.
4. Widen the spectral window to about 100 Hz. The 270 – 310 Hz window where the driving frequency will be set needs to be
visible.
5. To zoom in use the green line (this software is not click and drag!):
1. Left click separation
2. Left click position
3. Right click to do something
4. Right click to clear lines
6. If a peak is clipped, change the drive amplitude. Ideally this will be between 15 and 20 mV, and should be below 500 mV. If
a white line is not visible (there should be a white line along the bottom of the graph), the drive amplitude must be
increased.

9.2.6 https://chem.libretexts.org/@go/page/55928
7. Ideally the peak will have a regular shape and only small shoulders. If there is a lot of noise, re-install the tip and things
could improve. (Be careful as the auto-tuning scheme can drive the amplitude so high as to damage the tip.)
8. At this point, auto-tuning is okay. We can see that the parameters are reasonable. To continue the manual process, continue
following these steps.
9. Adjust the drive amplitude so that the peak is at 2.0 V.
10. Amplitude set point while tuning corresponds to the vertical off set. If it is set to 0, the green line is 0.
11. Position the drive frequency not at the center of the peak, but instead at 5% toward the low energy (left) of the peak value.
This offset is about 4/10th of a division. Right click three times to execute this change. This accounts for the damping that
occurs when the tip approaches the sample surface.
12. Left monitor - channel 2 dialogue box - click zero phase.
Image Acquisition
1. Click the eyeball icon for image mode.
2. Parameter adjustments.
1. Other controls.
2. Microscope mode: tapping.
3. Z-limit max height: 5.064 µm. This can be reduced if limited in Z-resolution.
4. Color table: 2.
5. Engage set point: 1.00.
6. Serial number of this scanner (double check since this has the factory parameter and is different from the other AFM).
7. Parameter update retract; disabled.
3. Scan Controls
1. Scan size: 2 µm. Be careful when changing this value – it will automatically go between µm and nm
2. (reasonable values are from 200 nm to 100 µm).
3. Aspect ratio: 1 to 1.
4. X and Y offset: 0.
5. Scan angle (like scan rotation): raster on the diagonal.
6. Scan rate: 1.97 Hz is fast, and 100 Hz is slow.
4. Feedback Control:
1. SPM: amplitude.
2. Integral gain: 0.5 (this parameter and the next parameter may be changed to improve image).
3. Proportional gain: 0.7.
4. Amplitude set point: 1 V.
5. Drive frequency: from tuning.
6. Drive amplitude: from tuning.
Once all parameters are set, click engage (icon with green arrow down) to start engaging cantilever to sample surface and to
begin image acquisition. The bottom of the screen should be “tip secured”. When the tip reaches the surface it automatically
begins imaging.
If the amplitude set point is high, the cantilever moves far away from the surface, since the oscillation is damped as it
approaches. While in free oscillation (set amplitude set point to 3), adjust drive amplitude so that the output voltage (seen on
the scope) is 2 V. Big changes in this value while an experiment is running indicate that something is on the tip. Once the
output voltage is at 2 V, bring the amplitude set point back down to a value that puts the z outer position line white and in the
center of the bar on the software (1 V is very close).
Select channel 1 data type – height. Select channel 2 data type - amplitude. Amplitude looks like a 3D image and is an
excellent visualization tool or for a presentation. However the real data is the height data.
Bring the tip down (begin with amplitude set point to 2). The goal is to tap hard enough to get a good image, but not so hard as
to damage the surface of the tip. Set to 3 clicks bellow jus touching by further lowering amplitude set point with 3 left arrow
clicks on the keyboard. The tip Z-center position scale on the right hand screen shows the extension on the piezo scanner.
When the tip is properly adjusted, expect this value to be near the center.

9.2.7 https://chem.libretexts.org/@go/page/55928
Select view/scope mode (the scope icon). Check to see if trace and retrace are tracking each other. If so, the lines should look
the same, but they probably will not overlap each other vertically or horizontally. If they are tracking well, then your tip is
scanning the sample surface and you may return to view/image mode (the image icon). If they are not tracking well, adjust the
scan rate, gains, and/or set point to improve the tracking. If tracing and retrace look completely different, you may need to
decrease the set point to improve the tracking. If trace and retrace look completely different, you may need to decrease the set
point one or two clicks with the left arrow key until they start having common features in both directions. Then reduce the scan
rate: a reasonable value for scan sizes of 1-3 µm would be 2 Hz. Next try increasing the integral gain. As you increase the
integral gain, the tracking should improve, although you will reach a value beyond which the noise will increase as the
feedback loop starts to oscillate. If this happens, reduce gains, if trace and retrace still do not track satisfactorily, reduce the set
point again. Once the tip is tracking the surface, choose view/image mode.
Integral gain controls the amount of integrated error signal used in the feedback calculation. The higher this parameter is set,
the better the tip will track the same topography. However, if it is set too high, noise due to feedback oscillation will be
introduced into the scan.
Proportional gain controls the amount of proportional arrow signal used in the feedback calculation.
Once amplitude set point is adjusted with the phase data, change channel 2 to amplitude. The data scale can be changed (it is
the same as for display as it does not affect the data). In the amplitude image, lowering the voltage increases the contrast.
Move small amounts on the image surface with X and Y offset to avoid large, uninteresting objects. For example, setting the Y
offset to -2 will remove features at the bottom of the image, thus shifting the image up. Changing it to -3 will then move the
image one more unit up. Make sure you are using µm and not nm if you expect to see a real change.
To move further, disengage the tip (click the red up arrow icon so that the tip moves up 25 µm and secures). Move upper
translational stage to keep the tip in view in the light camera. Re-engage the tip.
If the shadow in the image is drawn out, the amplitude set point should be lowered even further. The area on the image that is
being drawn is controlled by the frame pull-down menu (and the up and down arrows). Lower the set point and redraw the
same neighborhood to see if there is improvement. The proportional and integral gain can also be adjusted.
The frame window allows you to restart from the top, bottom, or a particular line.
Another way to adjust the amplitude set point value is to click on signal scope to ensure trace and retrace overlap. To stop Y
rastering, slow scan axis.
To take a better image, increase the number of lines (512 is max), decrease the speed (1 Hz), and lower the amplitude set point.
The resolution is about 10 nm in the X and Y directions due to the size of the tip. The resolution in the Z direction is less than 1
nm.
Changing the scan size allows us to zoom in on features. You can zoom in on a center point by using zoom in box (left clicking
to toggle between box position and size), or you can manually enter a scan size on the left hand screen.
Click on capture (the camera icon) to grab images. To speed things up, restart the scan at an edge to grab a new image after
making any changes in the scan and feedback parameters. When parameters are changed, the capture option will toggle to “
next”. There is a forced capture option, which allows you to collect an image even if parameters have been switched during the
capture. It is not completely reliable.
To change the file name, select capture filename under the capture menu. The file will be saved in the!directory which is
d:\capture. To save the picture, under the utility pull-down menu select TIFF export. The zip drive is G:.
Image Acquisition
Analysis involves flattening the image and measuring various particle dimensions, click the spectrum button.
Select the height data (image pull-down menu, select left or right image). The new icons in the “analysis” menu are:
Thumbnails
Top view
Side view
Section analysis
Roughness
Rolling pin (flattening)

9.2.8 https://chem.libretexts.org/@go/page/55928
Plane auto-fit
To remove the bands (striping) in the image, select the rolling pin. The order of flattening is the order of the baseline
correction. A raw offset is 0 order, a straight sloping line is order 1. Typically a second order correction is chosen to remove
“scanner bow” which are the dark troughs on the image plane.
To remove more shadows, draw exclusion boxes over large objects and then re-flatten. Be sure to save the file under a new
name. The default is t overwrite it.
In section analysis, use the multiple cursor option to measure a particle in all dimensions. Select fixed cursor. You can save
pictures of this information, but things must be written down! There is also a particle analysis menu.
Disengage the cantilever and make sure that the cantilever is in secure mode before you move the cantilever to the other spots
or change to another sample.
Loosen the clamp to remove the tip and holder.
Remove the tip and replace it onto the gel sticky tape using the fine tweezers.
Recover the sample with tweezers.
Close the program.
Log out of the instrument.
After the experiment, turn off the monitor and the power of the light source. Leave the controller on.
Sign out in the log book.

AFM - Scanning Probe Microscopy


Atomic force microscopy (AFM) has become a powerful tool to investigate 2D materials and the related 2D materials (e.g.,
graphene) for both the nano-scale imaging as well as the measurement and analysis of the frictional properties.
The basic structure and function of the typical Nanoscope AFM system is discussed in the section on the practical guide.
For the contact mode of AFM, a schematic is shown in Figure 9.2.11 The tip scans at the surface of the sample, the cantilever will
have a shift of Δz, which is a function of the position of the tip. If we know the mechanical constant of the tip C, the interaction
force, or the normal load of between the tip and sample can be calculated by 9.2.2, where C is determined by the material and
intrinsic properties of the tip and cantilever. As shown in Figure 9.2.11 a, we usually treat the back side of the cantilever as a
mirror to reflect the laser, so the change of the position will change the path length of the laser, and then detected by the quadrant
detector.
F   =  C ⋅ Δz (9.2.2)

Figure 9.2.11 (a) Schematic of the contact mode AFM. Adapted from https://commons.wikimedia.org/wiki/F...matic_(EN).svg. (b)
Agilent 5500 Atomic Force Microscope.
We can get the topography, height profile, phase and lateral force channel while measuring through the contact mode AFM.
Comparing the tapping mode, the lateral force, also known as the friction, appears very crucial. The direct signal acquired is the
current change caused due to the lateral force on the sample interacting with the tip, so the unit is usually nA. To calculate the real
friction force in Newton (N) or nano-Newton (nN), you need to let this current signal time a friction coefficient, which is also
determined by the intrinsic properties of the materials that makes the tip.

9.2.9 https://chem.libretexts.org/@go/page/55928
A typical AFM is shown in Figure 9.2.11 b. The sample stage is at the inside of the bottom chamber. You can blow the gas into the
chamber or pump the vacuum in need for the testing under different ambient. That is especially important in testing the frictional
properties of materials.
For the sample preparation part, the sample fixed on the mica mentioned earlier in the guide is for the synthesized chemical
powders. For graphene, it can be simply placed on any flat substrate, such as mica, SiC, sapphire, silica, etc. Just placing the solid
state sample on substrate onto the sample stage and the further work can be conducted.
Data Collection

For data collection, the topography and height profile are acquired using the same method in the tapping mode. However, there are
two additional pieces of information that are necessary in order to determine the frictional properties of the material. First, the
normal load. The normal load is described in 9.2.2; however, what we directly get here proportional to the normal load is the
setpoint we give it for the tip to the sample. It is a current. So we need a vertical force coefficient (CVF) to get what the normal
load we apply to the material, as illustrated in 9.2.3
F   =  Isetpoint ⋅ CV F (9.2.3)

For data collection, the topography and height profile are acquired using the same method in the tapping mode. However, there are
two additional pieces of information that are necessary in order to determine the frictional properties of the material. First, the
normal load. The normal load is described in 9.2.4, where K is the stiffness of the tip, it can be get through the vibrational model of
the cantilever, and usually we can get it if we buy the commercial AFM tip. L is the optical coefficient of the cantilever, it can be
acquired by calibrate the force-displacement curve of the tip, as shown in Figure 9.2.12. Then L can be acquired by getting the
slope of process 1 or 6 in Figure 9.2.13.
K
CV P   =   (9.2.4)
L

Figure 9.2.12 Force-displacement curve calibration of the tip.


Figure 9.2.13 is a typical friction image, it is composed of n*n lines by scanning. Each point is the friction force value
corresponding to that point. All we need to do is to get the average friction for the area we are interested in. Then use this current
signal multiplied by the lateral force coefficient then we can obtain the actual friction force.

Figure 9.2.13 The friction image by AFM of the CVD grown monolayer graphene. Adapted from P. Egberts, G. H. Han, X. Z. Liu,
A. T. C. Johnson, and R. W. Carpick, ACS Nano, 2014, 8, 5012. Copyright: American Chemical Society (2014).
During the process of collecting the original data of the lateral force (friction), for every line in the image, the friction information
is actually composed of two data line: trace and retrace (see Figure 9.2.13). The average of results for trace (Figure 9.2.13, black
line) and retrace (Figure 9.2.13, red line) as the friction signal of the certain point on the line. That is to say, the actual friction is

9.2.10 https://chem.libretexts.org/@go/page/55928
determined from 9.2.5, where the Iforward and Ibackward are data points we can derive from the trace and retrace from the friction
image, and CLF is the lateral force coefficient.
If orward   −  Ibackward
Ff   =   ⋅ CLF (9.2.5)
2

Data Analysis
There are several ways to compare the details of the frictional properties at the nanoscale. Figure 9.2.14 is an example comparing
the friction on the sample (in this case, few-layer graphene) and the friction on the substrate (SiO2). As illustrated in 9.2.5,
qualitatively we can easily see the friction on the graphene is way smaller than it on the SiO2 substrate. As graphene is a great
lubricant and have low friction, the original data just enable us to confirm that.

Figure 9.2.14 AFM image of few-layer graphene (a) and the friction profile (b) along the selected (yellow) line in (a).
Figure 9.2.15 shows multi-layers of graphene on a mica. By selecting a certain cross section line and comparing both height profile
and friction profile, it will provide us some information of the friction related to the structure behind this section. The friction-
distance curve is a typical important path for the data analysis.

Figure 9.2.15 The topography of graphene on mica (a) and the corresponding height and friction profile (b) of the selected section
defined by the red line in (a). Adapted from H. Lee, J. H. Ko, J. S. Choi, J. H. Hwang, Y. H. Kim, M. Salmeron and J. Y. Park, J.
Phys. Chem. Lett., 2017, 8, 3483. Copyright: American Chemical Society (2017).
We can also take the average of friction signal for an area and compare that from the region to the region. Figure 9.2.16 shows a
region of the graphene with the layer numbers from 1-4. Figure 9.2.16 a and b are also the topography and the friction image
respectively. By compare the average friction from the area to the area, we can obviously see the friction on graphene decreases as
the number of layers increases. Though Figure 9.2.16 c and d we can obviously see this average friction change on the surface
from 1 to 4 layers of graphene. But for a more general statistical way, getting the normalized signal of the average friction and
comparing them can be more straightforward.

9.2.11 https://chem.libretexts.org/@go/page/55928
Figure 9.2.16 (a) The topography image of graphene from 1 to 4 layers on SiOx. (b) The corresponding friction image of (a). (c)
and (d) are the corresponding Friction-Normal Load curves of the area. Adapted from P. Gong, Z. Ye, L.Yuan, and P. Egberts,
Carbon, 2018, 132, 749. Copyright: Elsevier (2018).
Another way to compare the frictional properties is that, to apply different normal load and see how the friction change, then get
the information on friction-normal load curve. This is important because we know too much normal load for the materials can
easily break or wear the materials. Examples and details will be discussed below.
The effect of H2O: a cautionary tale
During the process of using tip approach to graphene and applying the normal load (increasing normal load, loading process) and
withdrawing the tip gradually (decreasing normal load, unloading process), the friction on graphene exhibits hysteresis, which
means a large increment of the friction while we drag off the tip. This process can be analyzed from friction-normal load curve, as
shown in Figure 9.2.17. It was thought that this effect may be due to the detail of interacting behavior of the contact area between
the tip and graphene. However, if you test this in different ambient conditions, for example if nitrogen was blown into the chamber
while testing occured, this hysteresis disappears.

Figure 9.2.17 Friction hysteresis on the surface of graphene/Cu. Adapted from P. Egberts, G. H. Han, X. Z. Liu, A. T. C.
Johnson, and R. W. Carpick, ACS Nano, 2014, 8, 5012. Copyright: American Chemical Society (2014).
In order to explore the mechanism of such a phenomenon, a series of friction test under different conditions. A key factor here is
the humidity in the testing environment. Figure 9.2.18 is a typical friction measurement on monolayer and 3-layer graphene on
SiOx. We can see the friction hysteresis is very different under dry nitrogen gas (0.1% humidity) and the ambient (24% humidity)
from Figure 9.2.19.

9.2.12 https://chem.libretexts.org/@go/page/55928
Figure 9.2.18 Friction behavior of monolayer and 3-layer graphene under different humidity conditions. Adapted from P. Gong, Z.
Ye, L.Yuan, and P. Egberts, Carbon, 2018, 132, 749. Copyright: Elsevier (2018).
Simulation on this system suggests this friction hysteresis on the surface of graphene is due to the water interacting with the surface
of graphene. The contact angle between the tip/water molecule-graphene interfaces is the key component. The further study
suggests once you put the graphene samples in air and expose them for a long period of times (several days), the chemical bonding
at the surface can change due to the water molecule in the air so that the friction properties at nanoscale can be very different.
The bond between the material under investigation and the substrate can be very vital for the friction behavior at the nanoscale. The
studies during the years suggest that the friction of the graphene will decrease as the number of layers increase. This is adaptable
for suspended graphene (with nothing to support it), and graphene on most of substrates (such as SiOx, Cu foil and so on).
However, if the graphene is supported by fresh cleaved mica surface, there’s no difference for the frictional properties of different-
layer graphene, this is due to the large surface dissipation energy, so the graphene is very firmly fixed to the mica.
However, on the other hand, the surface of mica is also hydrophilic, this is causal to the water distribution on the surface of mica,
and the water intercalation between the graphene and mica bonding. Through the friction measurement of the graphene on mica,
we can analyze this system quantitatively, as shown in Figure 9.2.18.
Summary
This case study just gives an example that, contact-mode Atomic Force Microscopy, or Frictional Force Microscopy is a powerful
tool to investigate the frictional properties of materials, for the use both in scientific research as well as chemical industry.
The most important lesson for researchers is that in analyzing any literature data it is important to know what the relative humidity
conditions are for the particular experiment, such that various experiments may be compared directly.

9.2: Atomic Force Microscopy (AFM) is shared under a CC BY 4.0 license and was authored, remixed, and/or curated by Pavan M. V. Raja &
Andrew R. Barron via source content that was edited to conform to the style and standards of the LibreTexts platform; a detailed edit history is
available upon request.

9.2.13 https://chem.libretexts.org/@go/page/55928
9.3: SEM and its Applications for Polymer Science
SEM and its Applications for Polymer Science
Introduction
The scanning electron microscope (SEM) is a very useful imaging technique that utilized a beam of electrons to acquire high
magnification images of specimens. Very similar to the transmission electron microscope (TEM), the SEM maps the reflected
electrons and allows imaging of thick (~mm) samples, whereas the TEM requires extremely thin specimens for imaging; however,
the SEM has lower magnifications. Although both SEM and TEM use an electron beam, the image is formed very differently and
users should be aware of when each microscope is advantageous.
Microscopy Physics

Image Formation
All microscopes serve to enlarge the size of an object and allow people to view smaller regions within the sample. Microscopes
form optical images and although instruments like the SEM have extremely high magnifications, the physics of the image
formation are very basic. The simplest magnification lens can be seen Figure 9.3.1. The formula for magnification is shown in
9.3.1, where M is magnification, f is focal length, u is the distance between object and lens, and v is distance from lens to the

image.
f v−f
M  =     = (9.3.1)
u −f f

Figure 9.3.1 Basic microscope diagram illustrating inverted image and distances u, f, and v.
Multistage microscopes can amplify the magnification of the original object even more as shown in Figure. Where magnification is
now calculated from 9.3.2, where f1, f2 are focal distances with respect to the first and second lens and v1, v2are the distances
from the lens to the magnified image of first and second lens, respectively.
(v1   −  f1 )(v2   −  f2 )
M  =   (9.3.2)
f1 f2

Figure 9.3.2 A schematic diagram of the optics used in a multistage microscope.


In reality, the objects we wish to magnify need to be illuminated. Whether or not the sample is thin enough to transmit light divides
the microscope into two arenas. SEM is used for samples that do not transmit light, whereas the TEM (transmission electron
microscope) requires transparent samples. Due to the many frequencies of light from the introduced source, a condenser system is
added to control the brightness and narrow the range of viewing to reduce aberrations, which distort the magnified image.

9.3.1 https://chem.libretexts.org/@go/page/55929
Electron Microscopes
Microscope images can be formed instantaneous (as in the optical microscope or TEM) or by rastering (scanning) a beam across
the sample and forming the image point-by-point. The latter is how SEM images are formed. It is important to understand the basic
principles behind SEM that define properties and limitations of the image.

Resolution
The resolution of a microscope is defined as the smallest distance between two features that can be uniquely identified (also called
resolving power). There are many limits to the maximum resolution of the SEM and other microscopes, such as imperfect lenses
and diffraction effects. Each single beam of light, once passed through a lens, forms a series of cones called an airy ring (see Figure
9.3.3). For a given wavelength of light, the central spot size is inversely proportional to the aperture size (i.e., large aperture yields

small spot size) and high resolution demands a small spot size.

Figure 9.3.3 Airy ring illustrating center intensity (left) and intensity as a function of distance (right).
Aberrations distort the image and we try to minimize the effect as much as possible. Chromatic aberrations are caused by the
multiple wavelengths present in white light. Spherical aberrations are formed by focusing inside and outside the ideal focal length
and caused by the imperfections within the objective lenses. Astigmatism is because of further distortions in the lens. All
aberrations decrease the overall resolution of the microscope.

Electrons
Electrons are charged particles and can interact with air molecules therefore the SEM and TEM instruments require extremely high
vacuum to obtain images (10-7 atm). High vacuum ensures that very few air molecules are in the electron beam column. If the
electron beam interacts with an air molecule, the air will become ionized and damage the beam filament, which is very costly to
repair. The charge of the electron allows scanning and also inherently has a very small deflection angle off the source of the beam.
The electrons are generated with a thermionic filament. A tungsten (W) or LaB6 filament is chosen based on the needs of the user.
LaB6 is much more expensive and tungsten filaments meet the needs of the average user. The microscope can be operated as field
emission (tungsten filament).

Electron Scattering
To accurately interpret electron microscopy images, the user must be familiar with how high energy electrons can interact with the
sample and how these interactions affect the image. The probability that a particular electron will be scattered in a certain way is
either described by the cross section, σ, or mean free path, λ, which is the average distance which an electron travels before being
scattered.

Elastic Scatter
Elastic scatter, or Rutherford scattering, is defined as a process which deflects an electron but does not decrease its energy. The
wavelength of the scattered electron can be detected and is proportional to the atomic number. Elastically scattered electrons have
significantly more energy that other types and provide mass contrast imaging. The mean free path, λ, is larger for smaller atoms
meaning that the electron travels farther.

9.3.2 https://chem.libretexts.org/@go/page/55929
Inelastic Scatter
Any process that causes the incoming electron to lose a detectable amount of energy is considered inelastic scattering. The two
most common types of inelastic scatter are phonon scattering and plasmon scattering. Phonon scattering occurs when a primary
electron looses energy by exciting a phonon, atomic vibrations in a solid, and heats the sample a small amount. A Plasmon is an
oscillation within the bulk electrons in the conduction band for metals. Plasmon scattering occurs when an electron interacts with
the sample and produces plasmons, which typically have 5 - 30 eV energy loss and small λ.

Secondary Effects
A secondary effect is a term describing any event which may be detected outside the specimen and is essentially how images are
formed. To form an image, the electron must interact with the sample in one of the aforementioned ways and escape from the
sample and be detected. Secondary electrons (SE) are the most common electrons used for imaging due to high abundance and are
defined, rather arbitrarily, as electrons with less than 50 eV energy after exiting the sample. Backscattered electrons (BSE) leave
the sample quickly and retain a high amount of energy; however there is a much lower yield of BSE. Backscattered electrons are
used in many different imaging modes. Refer to Figure 9.3.4 for a diagram of interaction depths corresponding to various electron
interactions.

Figure 9.3.4 Diagram illustrating the depths at which various sample interactions occur.
SEM Construction
The SEM is made of several main components: electron gun, condenser lens, scan coils, detectors, specimen, and lenses (see Figure
9.3.5). Today, portable SEMs are available but the typical size is about 6 feet tall and contains the microscope column and the

control console.

Figure 9.3.5 Schematic drawing of the SEM illustrating placement of electron generation, collimation process, sample interaction
and electron detection.
A special feature of the SEM and TEM is known as depth of focus, dv/du the range of positions (depths) at which the image can be
viewed with good focus, see 9.3.3. This allows the user to see more than a singular plane of a specified height in focus and

9.3.3 https://chem.libretexts.org/@go/page/55929
essentially allows a range of three dimensional imaging.
2
dv −v 2
  =    =  −M (9.3.3)
2
du u

Electron Detectors (image formation)


The secondary electron detector (SED) is the main source of SEM images since a large majority of the electrons emitted from the
sample are less than 50 eV. These electrons form textural images but cannot determine composition. The SEM may also be
equipped with a backscatter electron detector (BSED) which collects the higher energy BSE’s. Backscattered electrons are very
sensitive to atomic number and can determine qualitative information about nuclei present (i.e., how much Fe is in the sample).
Topographic images are taken by tilting the specimen 20 - 40° toward the detector. With the sample tilted, electrons are more likely
to scatter off the top of the sample rather than interact within it, thus yielding information about the surface.
Sample Preparation
The most effective SEM sample will be at least as thick as the interaction volume; depending on the image technique you are using
(typically at least 2 µm). For the best contrast, the sample must be conductive or the sample can be sputter-coated with a metal
(such as Au, Pt, W, and Ti). Metals and other materials that are naturally conductive do not need to be coated and need very little
sample preparation.
SEM of Polymers
As previously discussed, to view features that are smaller than the wavelength of light, an electron microscope must be used. The
electron beam requires extremely high vacuum to protect the filament and electrons must be able to adequately interact with the
sample. Polymers are typically long chains of repeating units composed primarily of “lighter” (low atomic number) elements such
as carbon, hydrogen, nitrogen, and oxygen. These lighter elements have fewer interactions with the electron beam which yields
poor contrast, so often times a stain or coating is required to view polymer samples. SEM imaging requires a conductive surface, so
a large majority of polymer samples are sputter coated with metals, such as gold.
The decision to view a polymer sample with an SEM (versus a TEM for example) should be evaluated based on the feature size
you expect the sample to have. Generally, if you expect the polymer sample to have features, or even individual molecules, over
100 nm in size you can safely choose SEM to view your sample. For much smaller features, the TEM may yield better results, but
requires much different sample preparation than will be described here.
Polymer Sample Preparation Techniques

Sputter Coating
A sputter coater may be purchased that deposits single layers of gold, gold-palladium, tungsten, chromium, platinum, titanium, or
other metals in a very controlled thickness pattern. It is possible, and desirable, to coat only a few nm’s of metal onto the sample
surface.

Spin Coating
Many polymer films are depositing via a spin coater which spins a substrate (often ITO glass) and drops of polymer liquid are
dispersed an even thickness on top of the substrate.

Staining
Another option for polymer sample preparation is staining the sample. Stains act in different ways, but typical stains for polymers
are osmium tetroxide (OsO4), ruthenium tetroxide (RuO4) phosphotungstic acid (H3PW12O40), hydrazine (N2H4), and silver
sulfide (Ag2S).

9.3.4 https://chem.libretexts.org/@go/page/55929
Examples

Comp-block Copolymer (Microstructure of Cast Film)


Cast polymer film (see Figure 9.3.6).
To view interior structure, the film was cut with a microtome or razor blade after the film was frozen in liquid N2 and fractured.
Stained with RuO4 vapor (after cutting).
Structure measurements were averaged over a minimum of 25 measurements.

Figure 9.3.6 SEM micrograph of comb block copolymer showing spherical morphology and long range order. Adapted from M. B.
Runge and N. B. Bowden, J. Am. Chem. Soc., 2007, 129, 10551. Copyright: American Chemical Society (2007).

Polystyrene-polylactide Bottlebrush Copolymers (Lamellar Spacing)


Pressed polymer samples into disks and annealed for 16 h at 170 °C.
To determine ordered morphologies, the disk was fractured (see Figure 9.3.7).
Used SEM to verify lamellar spacing from USAXS.

Figure 9.3.7 SEM image of a fractured piece of polymer SL-1. Adapted from J. Rzayev, Macromolecules, 2009, 42, 2135.
Copyright: American Chemical Society (2009).

SWNTs in Ultrahigh Molecular Weight Polyethylene


Dispersed SWNTs in interactive polymer.
Samples were sputter-coated in gold to enhance contrast.
The films were solution-crystallized and the cross-section was imaged.
Environmental SEM (ESEM) was used to show morphologies of composite materials.
WD = 7 mm.
Study was conducted to image sample before and after drawing of film.
Images confirmed the uniform distribution of SWNT in PE (Figure 9.3.8).
MW = 10,000 Dalton.
Study performed to compare transparency before and after UV irradiation.

9.3.5 https://chem.libretexts.org/@go/page/55929
Figure 9.3.8 SEM images of crystallized SWNT-UHMWPE films before (left) and after (right) drawing at 120 °C. Adapted from
Q. Zhang, D. R. Lippits, and S. Rastogi, Macromolecules, 2006, 39, 658. Copyright: American Chemical Society (2006).

Nanostructures in Conjugated Polymers (Nanoporous Films)


Polymer and NP were processed into thin films and heated to crosslink.
SEM was used to characterize morphology and crystalline structure (Figure 9.3.9).
SEM was used to determine porosity and pore size.
Magnified orders of 200 nm - 1 μm.
WD = 8 mm.
MW = 23,000 Daltons
Sample prep: spin coating a solution of poly-(thiophene ester) with copper NPs suspended on to ITO coated glass slides. Ziess,
Supra 35

Figure 9.3.9 SEM images of thermocleaved film loaded with nanoparticles with scale bar 1 μm. Adapted from J. W. Andreasen, M.
Jorgensen, and F. C. Krebs, Macromolecules, 2007, 40, 7758. Copyright: American Chemical Society (2007).

Cryo-SEM Colloid Polystyrene Latex Particles (Fracture Patterns)


Used cryogenic SEM (cryo-SEM) to visualize the microstructure of particles (Figure 9.3.10)
Particles were immobilized by fast-freezing in liquid N2 at –196 °C.
Sample is fractured (-196 °C) to expose cross section.
3 nm sputter coated with platinum.
Shapes of the nanoparticles after fracture were evaluated as a function of crosslink density.

9.3.6 https://chem.libretexts.org/@go/page/55929
Figure 9.3.10 Cryo-SEM images of plastically drawn polystyrene and latex particles. Adapted from H. Ge, C. L. Zhao, S. Porzio,
L. Zhuo, H. T. Davis, and L. E. Scriven, Macromolecules, 2006, 39, 5531. Copyright: American Chemical Society (2006).

9.3: SEM and its Applications for Polymer Science is shared under a CC BY 4.0 license and was authored, remixed, and/or curated by Pavan M.
V. Raja & Andrew R. Barron via source content that was edited to conform to the style and standards of the LibreTexts platform; a detailed edit
history is available upon request.

9.3.7 https://chem.libretexts.org/@go/page/55929
9.4: Catalyst Characterization Using Thermal Conductivity Detector
Introduction
A catalyst is a "substance that accelerates the rate of chemical reactions without being consumed". Some reactions, such as the
hydrodechlorination of TCE, 9.4.1, don't occur spontaneously, but can occur in the presence of a catalyst.
C2 C l3 H   +  4 H2 → C2 H6   +  3H C l (9.4.1)
PD

Metal dispersion is a commong term within the catalyst industry. The term refers to the amount of metal that is active for a specific
reaction. Let’s assume a catalyst material has a composition of 1 wt% palladium and 99% alumina (Al2O3) (Figure 9.4.1) Even though
the catalyst material has 1 wt% of palladium, not all the palladium is active. The material might be oxidized due to air exposure or some
of the material is not exposed to the surface (Figure 9.4.2), hence it can’t participate in the reaction. For this reason it is important to
characterize the material.

Figure 9.4.1 A photograph of a sample of commercially available 1 wt% Pd/Al2O3.

Figure 9.4.2 Representation of Pd nanoparticles on Al2O3. Some palladium atoms are exposed to the surface, while some other lay below
the surface atoms and are not accessible for reaction.
In order for Pd to react according to 9.4.1, it needs to be in the metallic form. Any oxidized palladium will be inactive. Thus, it is
important to determine the oxidation state of the Pd atoms on the surface of the material. This can be accomplished using an experiment
called temperature programmed reduction (TPR). Subsequently, the percentage of active palladium can be determined by hydrogen
chemisorption. The percentage of active metal is an important parameter when comparing the performance of multiple catalyst. Usually
the rate of reaction is normalized by the amount of active catalyst.

Principle of Thermal Conductivity


Thermal conductivity is the ability of a chemical specie to conduct heat. Each gas has a different thermal conductivity. The units of
thermal conductivity in the international system of units are W/m·K. Table 9.4.1 shows the thermal conductivity of some common gasses.

Figure 9.4.3 A simplified circuit diagram of a thermal conductivity detector.


This detector is part of a typical commercial instrument such as a Micromeritics AutoChem 2920 (Figure 9.4.4). This instrument is an
automated analyzer with the ability to perform chemical adsorption and temperature-programmed reactions on a catalyst, catalyst support,

9.4.1 https://chem.libretexts.org/@go/page/55930
or other materials.

Figure 9.4.4 A photograph of a Micromeritics AutoChem 2920.

Temperature Programmed Reduction (TPR)


TPR will determine the number of reducible species on a catalyst and will tell at what temperature each of these species was reduced. For
example palladium is ordinarily found as Pd(0) or Pd(II), i.e., oxidation states 0 and +2. Pd(II) can be reduced at very low temperatures (5
- 10 °C) to Pd(0) following 9.4.2.

P dO  +  H2 → P d(0)  +  H2 O (9.4.2)

A 128.9 mg 1wt% Pd/Al2O3 samples is used for the experiment, Figure 9.4.5. Since we want to study the oxidation state of the
commercial catalyst, no pre-treatment needs to be executed to the sample. A 10% hydrogen-argon mixture is used as analysis and
reference gas. Argon has a low thermal conductivity and hydrogen has a much higher thermal conductivity. All gases will flow at 50
cm3/min. The TPR experiment will start at an initial temperature of 200 K, temperature ramp 10 K/min, and final temperature of 400 K.
The H2/Ar mixture is flowed through the sample, and past the detector in the analysis port. While in the reference port the mixture doesn’t
become in contact with the sample. When the analysis gas starts flowing over the sample, a baseline reading is established by the detector.
The baseline is established at the initial temperature to ensure there is no reduction. While this gas is flowing, the temperature of the
sample is increased linearly with time and the consumption of hydrogen is recorded. Hydrogen atoms react with oxygen atoms to form
H2O.

Figure 9.4.5 A sample of Pd/Al2O3 in a typical sample holder.


Water molecules are removed from the gas stream using a cold trap. As a result, the amount of hydrogen in the argon/hydrogen gas
mixture decreases and the thermal conductivity of the mixture also decrease. The change is compared to the reference gas and yields to a
hydrogen uptake volume. Figure 9.4.6 is a typical TPR profile for PdO.

Figure 9.4.6 A typical TPR profile of PdO. Adapted from R. Zhang, J. A. Schwarz, A. Datye, and J. P. Baltrus, J. Catal., 1992, 138, 55.

Pulse Chemisorption
Once the catalyst (1 wt% Pd/Al2O3) has been completely reduced, the user will be able to determine how much palladium is active. A
pulse chemisorption experiment will determine active surface area, percentage of metal dispersion and particle size. Pulses of hydrogen

9.4.2 https://chem.libretexts.org/@go/page/55930
will be introduced to the sample tube in order to interact with the sample. In each pulse hydrogen will undergo a dissociative adsorption
on to palladium active sites until all palladium atoms have reacted. After all active sites have reacted, the hydrogen pulses emerge
unchanged from the sample tube. The amount of hydrogen chemisorbed is calculated as the total amount of hydrogen injected minus the
total amount eluted from the system.

Data Collection for Hydrogen Pulse Chemisorption


The sample from previous experiment (TPR) will be used for this experiment. Ultra high-purity argon will be used to purge the sample at
a flow rate of 40 cm3/min. The sample will be heated to 200 °C in order to remove all chemisorbed hydrogen atoms from the Pd(0)
surface. The sample is cooled down to 40 °C. Argon will be used as carrier gas at a flow of 40 cm3/min. Filaments temperature will be
175 °C and the detector temperature will be 110 °C. The injection loop has a volume of 0.03610 cm3 @ STP. As shown in Figure 9.4.6,
hydrogen pulses will be injected in to the flow stream, carried by argon to become in contact and react with the sample. It should be noted
that the first pulse of hydrogen was almost completely adsorbed by the sample. The second and third pulses show how the samples is been
saturated. The positive value of the TCD detector is consistent with our assumptions. Since hydrogen has a higher thermal conductivity
than argon, as it flows through the detector it will tend to cool down the filaments, the detector will then apply a positive voltage to the
filaments in order to maintain a constant temperature.

Figure 9.4.7 A typical hydrogen pulse chemisorption profile of 1 wt% Pd/Al2O3.

Pulse Chemisorption Data Analysis


Table 9.4.1 shows me the integration of the peaks from Figure 9.4.7. This integration is performed by an automated software provided
with the instrument. It should be noted that the first pulse was completely consumed by the sample, the pulse was injected between time 0
and 5 minutes. From Figure 9.4.7 we observe that during the first four pulses, hydrogen is consumed by the sample. After the fourth
pulse, it appears the sample is not consuming hydrogen. The experiment continues for a total of seven pulses, at this point the software
determines that no consumption is occurring and stops the experiment. Pulse eight is denominated the "saturation peak", meaning the
pulse at which no hydrogen was consumed.
Table 9.4.1 Hydrogen pulse chemisorption data.
Pulse n Area

1 0

2 0.000471772

3 0.00247767

4 0.009846683

5 0.010348201

6 0.10030243

7 0.009967717

8 0.010580979

Using 9.4.3 the change in area (Δarean) is calculated for each peak pulse area (arean)and compared to that of the saturation pulse area
(areasaturation = 0.010580979). Each of these changes in area is proportional to an amount of hydrogen consumed by the sample in each
pulse. Table 9.4.2 shows the calculated change in area.

ΔArean   =  Areasaturation   −  Arean (9.4.3)

Table 9.4.2 Hydrogen pulse chemisorption data with ΔArea.


Pulse n Arean ΔArean

1 0 0.010580979

9.4.3 https://chem.libretexts.org/@go/page/55930
2 0.000471772 0.0105338018

3 0.00247767 0.008103309

4 0.009846683 0.000734296

5 0.010348201 0.000232778

6 0.010030243 0.000550736

7 0.009967717 0.000613262

8 0.010580979 0

The Δarean values are then converted into hydrogen gas consumption using 9.4.4, where Fc is the area-to-volume conversion factor for
hydrogen and SW is the weight of the sample. Fc is equal to 2.6465 cm3/peak area. Table 9.4.3 shows the results of the volume adsorbed
and the cumulative volume adsorbed. Using the data on Table 9.4.3, a series of calculations can now be performed in order to have a
better understanding of our catalyst properties.
ΔArean × Fc
Vadsorbed   =   (9.4.4)
SW

Table 9.4.3 Includes the volume adsorbed per pulse and the cumulative volume adsorbed
Cumulative quantity
Pulse n arean Δarean Vadsorbed (cm3/g STP)
(cm3/g STP)

1 0 0.0105809790 0.2800256 0.2800256

2 0.000471772 0.0105338018 0.2787771 0.558027

3 0.00247767 0.0081033090 0.2144541 0.7732567

4 0.009846683 0.0007342960 0.0194331 0.7926899

5 0.010348201 0.0002327780 0.0061605 0.7988504

6 0.010030243 0.0005507360 0.0145752 0.8134256

7 0.009967717 0.000613262 0.0162300 0.8296556

8 0.010580979 0 0.0000000 0.8296556

Gram Molecular Weight

Gram molecular weight is the weighted average of the number of moles of each active metal in the catalyst. Since this is a monometallic
catalyst, the gram molecular weight is equal to the molecular weight of palladium (106.42 [g/mol]). The GMCCalc is calculated using
9.4.5, where F is the fraction of sample weight for metal N and WatomicN is the gram molecular weight of metal N (g/g-mole). 9.4.6 shows

the calculation for this experiment.


1
GM WC alc   =   (9.4.5)
F1 F2 FN
( )  +  ( )  +  . . .   +  ( )
Wa tomic  1 Wa tomic  2 Wa tomic  N

g
106.42
1 Watomic P D g−mole g
GM WC alc   =     =    =    =  106.42 (9.4.6)
F1
( ) F1 1 g − mole
Wa tomic  Pd

Metal Dispersion
The metal dispersion is calculated using 9.4.7, where PD is the percent metal dispersion, Vs is the volume adsorbed (cm3 at STP), SFCalc
is the calculated stoichiometry factor (equal to 2 for a palladium-hydrogen system), SW is the sample weight and GMWCalc is the
calculated gram molecular weight of the sample [g/g-mole]. Therefore, in 9.4.8 we obtain a metal dispersion of 6.03%.
Vs × S FC alc
P D  =  100  ×  ( ) × GM WC alc (9.4.7)
SW × 22414

3
0.8296556[c m ]  ×  2 g
P D  =  100  ×  ( )  ×  106.42[ ]  =  6.03% (9.4.8)
3
cm g − mol
0.1289[g]  × 22414[ ]
mol

9.4.4 https://chem.libretexts.org/@go/page/55930
Metallic Surface Area per Gram of Metal
The metallic surface area per gram of metal is calculated using 9.4.9, where SAMetallic is the metallic surface area (m2/g of metal),
SWMetal is the active metal weight, SFCalc is the calculated stoichiometric factor and SAPd is the cross sectional area of one palladium
atom (nm2). Thus, in 9.4.10 we obtain a metallic surface area of 2420.99 m2/g-metal.
VS
23
S AMetallic   =  ( )  ×  (S FC alc )  ×  (6.022  ×  10 )  ×  S AP d (9.4.9)
S WMetal   ×  22414

3 2
0.8296556 [c m ] atoms nm
23
S AMetallic   =  ( )  ×  (2)  ×  (6.022  ×  10  [ ])  ×  0.07 [ ]  (9.4.10)
3
cm mol atom
0.001289 [ gmetal ]  ×  22414 [ ]
mol

2
m
=  2420.99 [ ]
g − metal

Active Particle Size


The active particle size is estimated using 9.4.11, where DCalc is palladium metal density (g/cm3), SWMetal is the active metal weight,
GMWCalc is the calculated gram molecular weight (g/g-mole), and SAPd is the cross sectional area of one palladium atom (nm2). As seen
in 9.4.12 we obtain an optical particle size of 2.88 nm.
6
AP S  =   (9.4.11)
WS 2
DC alc   ×  ( )  ×  (6.022  ×  10 3)  ×  S AMetallic
GMWC a lc

600
AP S  =     =  2.88 nm (9.4.12)
gPd 0.001289 [g] atoms m
2
−20 23
(1.202  ×  10 [ 3
])  ×  ( g
)  ×  (6.022  ×  10  [ ])  ×  (2420.99 [ ])
nm Pd mol g−P d
106.42 [ ]
mol

In a commercial instrument, a summary report will be provided which summarizes the properties of our catalytic material. All the
equations used during this example were extracted from the AutoChem 2920-User's Manual.
Table 9.4.4 Summary report provided by Micromeritics AuthoChem 2920.
Properties Value

Palladium atomic weight 106.4 g/mol

Atomic cross sectional area 0.0787 nm2

Metal Density 12.02 g/cm3

Palladium loading 1 wt %

Metal dispersion 6.03 %

Metallic surface area 2420.99 m2/g-metal

Active particle diameter (hemisphere) 2.88 nm

9.4: Catalyst Characterization Using Thermal Conductivity Detector is shared under a CC BY 4.0 license and was authored, remixed, and/or curated by
Pavan M. V. Raja & Andrew R. Barron via source content that was edited to conform to the style and standards of the LibreTexts platform; a detailed edit
history is available upon request.

9.4.5 https://chem.libretexts.org/@go/page/55930
9.5: Nanoparticle Deposition Studies Using a Quartz Crystal Microbalance
Overview
The working principle of a quartz crystal microbalance with dissipation (QCM-D) module is the utilization of the resonance
properties of some piezoelectric of materials. A piezoelectric material is a material that exhibits an electrical field when a
mechanical strain is applied. This phenomenon is also observed in the contrary where an applied electrical field produce a
mechanical strain in the material. The material used is α-SiO2 that produces a very stable and constant frequency. The direction and
magnitude of the mechanical strain is directly dependent of the direction of the applied electrical field and the inherent physical
properties of the crystal.
A special crystal cut is used, called AT-cut, which is obtain as wafers of the crystal of about 0.1 to 0.3 mm in width and 1 cm in
diameter. The AT-cut is obtained when the wafer is cut at 35.25° of the main crystallographic axis of SiO2. This special cut allows
only one vibration mode, the shear mode, to be accessed and thus exploited for analytical purposes. When a electrical field is
applied to the crystal wafer via metal electrodes, that are vapor-deposited in the surface, a mechanical shear is produced and
maintained as long as the electrical field is applied. Since this electric field can be controlled by opening and closing an electrical
circuit, a resonance within the crystal is formed (Figure 9.5.1).

Figure 9.5.1 Schematic representation of the piezoelectric material: (a) a baseline is obtained by running the sensor without any
flow or sample; (b) sample is starting to flow into the sensor; (c) sample deposited in the sensor change the frequency.
Since the frequency of the resonance is dependent of the characteristics of the crystal, an increase of mass, for example when the
sample is loaded into the sensor would change the frequency change. This relation 9.5.1 was obtained by Sauerbrey in 1959, where
Δm (ng.cm-2) is the areal mass, C (17.7 ngcm-2Hz-1) is the vibrational constant (shear, effective area, etc.), n in Hz is the resonant
overtone, and Δf is the change in frequency. The dependence of the change in the frequency can be related directly to the change in
mass deposited in the sensor only when three conditions are met and assumed:
The mass deposited is small compared to the mass of the sensor
It is rigid enough so that it vibrates with the sensor and does not suffer deformation
The mass is evenly distributed among the surface of the sensor
1
Δm  =   − C Δf (9.5.1)
n

An important incorporation in recent equipment is the use of the dissipation factor. The inclusion of the dissipation faster takes into
account the weakening of the frequency as it travels along the newly deposited mass. In a rigid layer the frequency is usually
constant and travels through the newly formed mass without interruption, thus, the dissipation is not important. On the other hand,
when the deposited material has a soft consistency the dissipation of the frequency is increased. This effect can be monitored and
related directly to the nature of the mass deposited.
The applications of the QCM-D ranges from the deposition of nanoparticles into a surface, from the interaction of proteins within
certain substrates. It can also monitors the bacterial amount of products when feed with different molecules, as the flexibility of the
sensors into what can be deposited in them include nanoparticle, special functionalization or even cell and bacterias!

Experimental Planning
In order to use QCM-D for studing the interaction of nanoparticles with a specific surface several steps must be followed. For
demonstration purposes the following procedure will describe the use of a Q-Sense E4 with autosampler from Biolin Scientific. A
summary is shown below as a quick guide to follow, but further details will be explained:
Surface election and cleaning according with the manufacturer recommendations
Sample preparation including having the correct dilutions and enough samplke for the running experiment
Equipment cleaning and set up of the correct aparameters for the experiment
Data acquisition

9.5.1 https://chem.libretexts.org/@go/page/55931
Data interpretation
Surface Election
The decision of what surface of the the sensor to use is the most important decision to make fore each study. Biolin has a large
library of available coatings ranging from different compositions of pure elements and oxides (Figure 9.5.2) to specific binding
proteins. It is important to take into account the different chemistries of the sensors and the results we are looking for. For example
studying a protein with high sulfur content on a gold sensor can lead to a false deposition results, as gold and sulfur have a high
affinity to form bonds. For the purpose of this example, a gold coated sensor will be used in the remainder of the discussion.

Figure 9.5.2 From left to right, silica (SiO2), gold (Au), and iron oxide (Fe2O3) coated sensors. Each one is 1 cm in diameter.
Sensor Cleaning
Since QCM-D relies on the amount of mass that is deposited into the surface of the sensor, a thorough cleaning is needed to ensure
there is no contaminants on the surface that can lead to errors in the measurement. The procedure the manufacturer established to
clean a gold sensor is as follows:
1. Put the sensor in the UV/ozone chamber for 10 minutes
2. Prepare 10 mL of a 5:1:1 solution of hydrogen peroxide:ammonia:water
3. Submerge in this solution at 75 C for 5 minutes
4. Rinse with copious amount of milliQ water
5. Dry with inert gas
6. Put the sensor in the UV/ozone chamber for 10 minutes as shown in Figure 9.5.3.

Figure 9.5.3 Gold sensors in loader of the UV/ozone chamber in the final step of the cleaning process.
Once the sensors are clean, extreme caution should be taken to avoid contamination of the surface. The sensors can be loaded in the
flow chamber of the equipment making sure that the T-mark of the sensor matches the T mark of the chamber in order to make sure
the electrodes are in constant contact. The correct position is shown in Figure 9.5.4.

9.5.2 https://chem.libretexts.org/@go/page/55931
Figure 9.5.4 Correct position of the sensor in the chamber.
Sample Preparation
As the top range of mass that can be detected is merely micrograms, solutions must be prepared accordingly. For a typical run, a
buffer solution is needed in which the deposition will be studied as well as, the sample itself and a solution of 2% of sodium
dodecylsulfate [CH3(CH2)10CH2OSO3Na, SDS]. For this example we will be using nanoparticles of magnetic iron oxide (nMag)
coated with PAMS, and as a buffer 8% NaCl in DI water.
For the nanoparticles sample it is necessary to make sure the final concentration of the nanoparticles will not exceed 1 mM.
For the buffer solution, it is enough to dissolve 8 g of NaCl in DI water.
For the SDS solution, 2 g of SDS should be dissolved very slowly in approximate 200 mL of DI water, then 100 mL aliquots of
DI water is added until the volume is 1 L. This is in order to avoid the formation of bubbles and foam in the solution.
Instrument Preparation
Due to the sensitivity of the equipment, it is important to rinse and clean the tubing before loading any sample or performing any
experiments. To rinse the tubing and the chambers, use a solution of 2% of SDS. For this purpose, a cycle in the autosampler
equipment is program with the steps shown in Table 9.5.1.
Table 9.5.1 Summary of cleaning processes.
Step Duration (min) Speed (μL/min) Volume (mL)

DI water (2:2) 10 100 1

SDS (1:1) 20 300 6

DI water (1:2) 10 100 1

Once the equipment is cleaned, it is ready to perform an experiment, a second program in the autosampler is loaded with the
parameters shown in Table 9.5.2.
Table 9.5.2 Experimental set-up
Step Duration (min) Speed (μL/min) Volume (mL)

Buffer (1:3) 7 100 0.7

Nanoparticles 30 100 3.0

The purpose of flowing the buffer in the beginning is to provide a background signal to take into account when running the
samples. Usually a small quantity of the sample is loaded into the sensor at a very slow flow rate in order to let the deposition take
place.
Data Acquisition
Example data obtained with the above parameters is shown in Figure 9.5.5. The blue squares depict the change in the frequency.
As the experiment continues, the frequency decreases as more mass is deposited. On the other hand, shown as the red squares, the
dissipation increases, describing the increase of both the height and certain loss of the rigidity in the layer from the top of the
sensor. To illustrate the different steps of the experiment, each section has been color coded. The blue part of the data obtained
corresponds to the flow of the buffer, while the yellow part corresponds to the deposition equilibrium of the nanoparticles onto the

9.5.3 https://chem.libretexts.org/@go/page/55931
gold surface. After certain length of time equilibrium is reached and there is no further change. Once equilibrium indicates no
change for about five minutes, it is safe to say the deposition will not change.

Figure 9.5.5 Data of deposition of nMag in a gold surface.


Instrument Clean-up
As a measure preventive care for the equipment, the same cleaning procedure should be followed as what was done before loading
the sample. Use of a 2% solution of SDS helps to ensure the equipment remains as clean as possible.

Data Modeling
Once the data has been obtained, QTools (software that is available in the software suit of the equipment) can be used to convert
the change in the frequency to areal mass, via the Sauerbrey equation, 9.5.1. The correspondent graph of areal mass is shown in
9.5.1. From this graph we can observe how the mass is increasing as the nMag is deposited in the surface of the sensor. The blue

section again illustrates the part of the experiment where only buffer was been flown to the chamber. The yellow part illustrates the
deposition, while the green part shows no change in the mass after a period of time, which indicates the deposition is finished. The
conversion from areal mass to mass is a simple process, as gold sensors come with a definite area of 1 cm2, but a more accurate
measure should be taken when using functionalized sensors.

Figure 9.5.6 Areal mass of deposition of nMag into gold surface.


It is important to take into account the limitations of the Saubery equation, because the equation accounts for a uniform layer on top
of the surface of the sensor. Deviations due to clusters of material deposited in one place or the formation of partial multilayers in
the sensor cannot be calculated through this model. Further characterization of the surface should be done to have a more accurate
model of the phenomena.

9.5: Nanoparticle Deposition Studies Using a Quartz Crystal Microbalance is shared under a CC BY 4.0 license and was authored, remixed,
and/or curated by Pavan M. V. Raja & Andrew R. Barron via source content that was edited to conform to the style and standards of the
LibreTexts platform; a detailed edit history is available upon request.

9.5.4 https://chem.libretexts.org/@go/page/55931
CHAPTER OVERVIEW
10: Device Performance
10.1: A Simple Test Apparatus to Verify the Photoresponse of Experimental Photovoltaic Materials and Prototype Solar Cells
10.2: Measuring Key Transport Properties of FET Devices

10: Device Performance is shared under a CC BY 4.0 license and was authored, remixed, and/or curated by Pavan M. V. Raja & Andrew R.
Barron via source content that was edited to conform to the style and standards of the LibreTexts platform; a detailed edit history is available upon
request.

1
10.1: A Simple Test Apparatus to Verify the Photoresponse of Experimental
Photovoltaic Materials and Prototype Solar Cells
Introduction
One of the problems associated with testing a new unproven photovoltaic material or cell design is that significant processing
required in order to create a fully functioning solar cell. If it is desired to screen a wide range of materials or synthetic conditions it
can be time consuming (and costly of research funds) to prepare fully functioning devices. In addition, the success of each
individual cell may be more dependent on fabrication steps not associated with the variations under study. For example, lithography
and metallization could cause more variability than the parameters of the materials synthesis. Thus, the result could be to give no
useful information as to the viability of each material under study, or even worse a false indication of research direction.
So-called quick and dirty qualitative measurements can be employed to assess not only the relative photoresponse of new absorber
layer materials, but also the relative power output of photovoltaic devices. The measurement procedure can provide a simple,
inexpensive and rapid evaluation of cell materials and structures that can help guide the development of new materials for solar cell
applications.

Equipment Needs
Everything needed for the measurements can be purchased at a local electronics store and a hardware or big box store. Needed
items are:
Two handheld digital voltmeter with at least ±0.01 mV sensitivity (0.001 mV is better, of course).
A simple breadboard and associated wiring kit.
A selection of standard size and wattage resistors (1/8 - 1 Watt, 1 - 1000 ohms).
A selection of wire wound potentiometers (0 - 10 ohms; 0 - 100 ohms; 0 - 1000 ohms) if I-V tracing is desired.
A light source. This can be anything from a simple flood light to an old slide projector.
A small fan or other cooling device for “steady state” (i.e., for measurements that last more than a few seconds such as tracing
an I-V curve).
9 volt battery and holder or simple ac/dc low voltage power supply.

Measurement of the Photo-response of an Experimental Solar Cell


A qualitative measurement of a solar cell’s current-voltage (I-V) characteristics can be obtained using the simple circuit diagram
illustrated in Figure 10.1.1. Figure 10.1.2 shows an I-V test setup using a household flood lamp for the light source. A small fan
sits to the right just out of the picture.

Figure 10.1.1 Simple circuit diagram for I-V measurement of a prototype solar cell.

10.1.1 https://chem.libretexts.org/@go/page/55927
Figure 10.1.2 Simple test apparatus for qualitative measurement of the current-voltage output from an experimental thin film solar
cell.
Driving the potentiometer to its maximum value will place the cell close to open circuit operation, depending on the potentiometer
range, so that the open circuit voltage can be simply extrapolated from the I versus V curve. If desired, the circuit can simply be
opened to make the actual measurement once the rest of the data have been recorded. Data in this case were simply recorded by
hand and later entered into a spreadsheet so an I-V plot could be generated. A sample plot is shown in Figure 10.1.3. Keep in mind
that cell efficiency cannot be determined with this technique unless the light source has been calibrated and color corrected to
match terrestrial sunlight. The fact that the experimental device actually generated net power was the result sought. The shape of
the curve and the very low voltage are the result of very large resistive losses in the device along with a very “leaky” junction.

Figure 10.1.1 and 10.1.2


One improvement that can be made to the above system is to replace the floodlight with a simple slide projector. The floodlight
will typically have a spectrum very heavily weighted in the red and infrared and will be deficient in the shorter wavelengths.
Though still not a perfect match to the solar spectrum, the slide projector does at least have more output at the shorter wavelengths;
at the same time it will have less IR output compared to the floodlight and the combination should give a somewhat more
representative response. A typical set up is shown in Figure 10.1.4.

10.1.2 https://chem.libretexts.org/@go/page/55927
Figure 10.1.4 Test setup using a slide projector.
The mirror in Figure 10.1.5 serves two purposes. First, it turns the beam so the test object can be laid flat a measurement bed and
second it serves to collimate and concentrate the beam by focusing it on a smaller area, giving a better approximation of terrestrial
solar intensity over a range of intensities such as AM2 (air mass 2) through AM0 (Figure 10.1.5). An estimate of the intensity can
be made using a calibrated silicon solar cell of the sort that can be purchased online from any of several scientific hobby shops such
as Edmunds Scientific. While still far from enabling a quantitative measurement of device output, the technique will at least
provide indications within a ballpark range of actual cell efficiency.

Figure 10.1.5 Solar irradiance spectrum at AM 0 (yellow) and AM2 (red). Adapted from M. Pagliaro, G. Palmisano, and R.
Ciriminna, Flexible Solar Cells, John Wiley, New York (2008).
Figure 10.1.6 shows a measurement made with the test device placed at a distance from the mirror for which the intensity was
previously determined to be equivalent to AM1 solar intensity, or 1000 watts per square meter. Since the beam passes through the
projector lens and reflects from the second surface of the slightly concave mirror, there is essentially no UV light left in the beam
that could be harmful to the naked eye. Still, if this technique is used, it is recommended that observations be made through a piece
of ordinary glass such as eyeglasses or even a small glass shield inserted for that purpose. The blue area in the figure represents the
largest rectangle that can be drawn under the curve and gives the maximum output power of the cell, which is simply the product of
the current and voltage at maximum power.
Figure 10.1.6 is a plot of current density, obtained by dividing the current from the device by its area. It is common to normalize
the output is this manner.
If the power density of the incident light (P0) is known in W/cm2, the device efficiency can be obtained by dividing the maximum
power (as determined from Im and Vm) by the incident power density times the area of the cell (Acell), 10.1.1.

η  =  Im Vm / P0 Acell (10.1.1)

10.1.3 https://chem.libretexts.org/@go/page/55927
Figure 10.1.6 The picture shows the relative brightness of the light beam at an approximate intensity of 1000 W/m2. A small
concave mirror serves to both turn the beam and to concentrate it a small amount to reach that level.

Measurement of the Photoconductivity of Experimental Photovoltaic Materials


In many cases it is beneficial to determine the photoconductivity of a new material prior to cell fabrication. This allows for the
rapid screening of materials or synthesis variable of a single material even before issues of cell design and construction are
considered.
Figure 10.1.7 shows the circuit diagram of a simple photoconductivity test made with a slightly different set up compared to that
shown above. In this case a voltage is placed across the sample after it has been connected to a resistor placed in series with the
sample. A simple 9 V battery secured with a battery holder or a small ac to dc power converter can be used to supply the voltage.
The sample and resistor sit inside a small box with an open top.

Figure 10.1.7 Circuit diagram for simple photoconductance test.


The voltage across (in this case) the 10 ohm resister was measured with a shutter held over the sample (a simple piece of cardboard
sitting on the top of the box) and with the shutter removed. The difference in voltage is a direct indication of the change in the
photoconductance of the sample and again is a very quick and simple test to see if the material being developed does indeed have a
photoresponse of some sort without having to make a full device structure. Adjusting the position of the light source so that the
incident light power density at the sample surface is 200 or 500 or 1000 W/m2 enables an approximate numerical estimate of the
photocurrent that was generated and again can help guide the development of new materials for solar cell applications. The results
from such a measurement are shown in Figure 10.1.8 for a sample of carbon nanotubes (CNT) coated with CdSe by liquid phase
deposition (LPD).

10.1.4 https://chem.libretexts.org/@go/page/55927
Figure 10.1.8 Photoresponse of a carbon nanotube (CNT) carpet coated with CdSe by liquid phase deposited.

10.1: A Simple Test Apparatus to Verify the Photoresponse of Experimental Photovoltaic Materials and Prototype Solar Cells is shared under a
CC BY 4.0 license and was authored, remixed, and/or curated by Pavan M. V. Raja & Andrew R. Barron via source content that was edited to
conform to the style and standards of the LibreTexts platform; a detailed edit history is available upon request.

10.1.5 https://chem.libretexts.org/@go/page/55927
10.2: Measuring Key Transport Properties of FET Devices
Field Effect Transistors
Arguably the most important invention of modern times, the transistor was invented in 1947 at Bell Labs by John Bardeen, William
Shockley, and Walter Brattain. The result of efforts to replace inefficient and bulky vacuum tubes in current regulation and
switching functions. Further advances in transistor technology led to the field effect transistors (FETs), the bedrock of modern
electronics. FETs operate by utilizing an electric field to control the flow of charge carriers along a channel, analogous to a water
valve to control the flow of water in your kitchen sink. The FET consists of 3 terminals, a source (S), drain (D), and gate (G). The
region between the source and drain is called the channel. The conduction in the channel depends on the availability of charge
carriers controlled by the gate voltage. Figure depicts a typical schematic and Figure 10.2.1 the associated cross-section of a FET
with the source, draing and gate terminals labeled. FETs come in a variety of flavors depending on their channel doping (leading to
enhancement and depletion modes) and gate types, as seen in Figure 10.2.2. The two FET types are junction field effect transistors
(JFETs) and metal oxide semiconductor field effect transistors (MOSFETs).

Figure 10.2.1 The n-channel enhancement mode MOSFET symbol.

Figure 10.2.2 A typical cross-section of a n-channel enhancement mode MOSFET.

Figure 10.2.3 Field effect transistor family tree. Adapted from P. Horowitz and W. Hill, in Art of Electronics, Cambridge
University Press, New York, 2nd Edn., 1994.
JFET Fundamentals
Junction field effect transistors (JFETs) as their name implies utilize a PN-junction to control the flow of charge carriers. The PN-
junction is formed when opposing doping schemes are broght together on both sides of the channel. The doping schemes can be
made to be either n-type (electrons) or p-type (holes) by doping with boron/gallium or phosphorus/arsenic respectively. The n-
channel JFETs consists of pnp junctions where the source and drain are n-doped and the gate is p-doped. Figure 10.2.4 shows the
cross section of a n-channel JFET in the “ON” state obtained by applying a positive drain-source voltage in the absence of a gate-
source voltage. Alternatively the p-channel JFET consists of npn junctions where the source and drain are p-doped and the gate is
n-doped. For p-channel a negative drain-source voltage is applied in the absence of a gate voltage to turn “ON” the npn device, as
seen in Figure 10.2.5. Since JFETs are “ON” when no gate-source voltage is applied they are called depletion mode devices.
Meaning that a depletion region is required to turn “OFF” the device. This is where the PN-junction comes into play. The PN-
junction works by enabling a depletion region to form where electrons and holes combine leaving behind positive and negative ions
which inhibit further charge transfer as well as depleting the availability of charge carriers at the interface. This depletion region is
pushed further into the channel by applying a gate-source voltage. If the voltage is sufficient the depletion region on either side of
the channel will “pinch off” the flow through the channel and the device will be “OFF”. This voltage is called the pinch off voltage,

10.2.1 https://chem.libretexts.org/@go/page/55937
VP. The n-channel VP is obtained by increasing the gate-source voltage in the negative direction, while the p-channel VP is
obtained by increasing the gate-source voltage in the positive direction.

Figure 10.2.4 Cross-section of a n-channel JFET in the "ON" state.

Figure 10.2.5 Cross-section of a p-channel JFET in the "ON" state.


MOSFET Fundamentals
The metal oxide semiconductor field effect transistor (MOSFET) utilizes an oxide layer (typically SiO2) to isolate the gate from
the source and drain. The thin layer of oxide prevents flow of current to the gate, but enables an electric field to be applied to the
channel which regulates the flow of charge carriers through the channel.MOSFETs unlike JFETs can operate in depletion or
enhancement mode characterized by their ON or OFF state at zero gate-source voltage, VGS.
For depletion mode MOSFETs the device is “ON” when the VGS is zero as a result of the devices structure and doping scheme.
The n-channel depletion mode MOSFET consists of heavily n-doped source and drain terminals on top of a p-doped substrate.
Underneath an insulating oxide layer there is a thin layer of n-type silicon which allows charge carriers to flow in the absence of a
gate voltage. When a negative voltage is applied to the gate a depletion region forms inside the channel, as seen in Figure. If the
gate voltage is sufficient the depletion region pinches off the flow of electrons.

Figure 10.2.6 Cross-section of a n-channel depletion mode MOSFET when a negative gate voltage is applied with the resultant
depletion layer.
For enhancement mode MOSFETs the ON state is attained by applying a gate voltage in the direction of the drain voltage; a
positive voltage for n-channel enhancement MOSFETs, and a negative voltage for p-channel enhancement MOSFETs. The term
“enhancement” is derived from the increase in conductivity seen by applying a gate voltage. This increase in conductivity is
enabled by an inversion layer induced by the applied electric field at the gate as shown in Figure 10.2.7 for n-channel enhancement
mode MOSFETs and Figure 10.2.8 for p-channel enhancement mode MOSFETs respectively.

10.2.2 https://chem.libretexts.org/@go/page/55937
Figure 10.2.7 A depiction of the induced inversion layer with n-type charge carriers in a n-channel enhancement mode MOSFET.

Figure 10.2.8 A depiction of the induced inversion layer with p-type charge carriers in a p-channel enhancement mode MOSFET.
The thickness of this inversion layer is controlled by the magnitude of the gate voltage. The minimum voltage required to form the
inversion layer is called the gate-to-source threshold voltage, VT. In the case of n-channel enhancement mode MOSFETs, the “ON”
state is reached when VGS > VT and a positive drain-source voltage, VDS, is applied. If the VGS is too low, then increasing the VDS
further results only in increasing the depletion region around the drain. The p-channel enhancement mode MOSFETs operate
similarly except that the voltages are reversed. Specifically, the “ON” state occurs when VGS < VT and a negative drain-source
voltage is applied.

Measurement of key FET Parameters


In both an academic and industrial setting characterization of FETs is beneficial for determining device performance. Identifying
the quality and type of FET can easily be addressed by measuring the transport characteristics under different experimental
conditions utilizing a semiconductor characterization system (SCS). By analyzing the V-I characteristics through what are called
voltage sweeps, the following key device parameters can be determined:
Pinch off Voltage Vp
The voltage needed to turn “OFF” a JFET. When designing circuits it is essential that the pinch-off voltage be determined to avoid
current leakage which can dramatically reduce performance.
Threshold Voltage VT
The voltage needed to turn “ON” a MOSFET. This is a critical parameter in effective circuit design.
Channel Resistance RDS
The resistance between the drain and source in the channel. This influences the amount of current being transferred between the
two terminals.
Power Dissipation PD
The power dissipation determines the amount of heat generated by the transistor. This becomes a real problem since the transport
properties deteriorate as the channel is heated.
Effective Charge Carrier Mobility µn
The charge carrier mobility determines how quickly the charge carrier can move through the channel. In most cases higher mobility
leads to better device performance. The mobility can also be used to gauge the impurity, defect, temperature, and charge carrier
concentrations.

10.2.3 https://chem.libretexts.org/@go/page/55937
Transconductance gain gm (transfer admittance)
The gm is a measure of gain or amplification of a current for a given change in gate voltage. This is critical for amplification type
electronics.
Equipment Needs
PC with Keithley Interactive Test Environment (KITE) software.
Semiconductor characterization system (Keithley 4200-SCS or equivalent).
Probe station.
Probe tips.
Protective gloves.

Measurement (V-I) Characteristics


The Semiconductor Characterization System is an automated system that provides both (V-I) and (V-C) characterization of
semiconductor devices and test structures. The advanced digital sweep parameter analyzer provides sub-micron characterization
with accuracy and speed. This system utilizes the Keithley Interactive Test Environment (KITE) software designed specifically for
semiconductor characterization.

Procedure
1. Connect the probe tips to the probe station. Then attach the banana plugs from the probe station to the BNC connector, making
sure not to connect to ground.
2. Select the appropriate connections for your test from Table 10.2.1
3. Place your transistor sample on the probe station, but don’t let the probe tips touch the sample to prevent possible electric
shock(during power up, the SMU may momentarily output high voltage).
4. Turn on power located on the lower right of the front panel. The power up sequence may take up to 2 minutes.
5. Start KITE software. Figure 10.2.9 shows the interface window.
6. Select the appropriate setup from the Project Tree drop down (top left).
7. Match the Definition tab terminal connections to the physical connections of probe tips. If connection is not yet matched you
can assign/reassign the terminal connections by using the arrow key next to the instrument selection box that displays a list of
possible connections. Select the connection in the instrument selection box that matches the physical connection of the device
terminal.
8. Set the Force Measure settings for each terminal. Fill in the necessary function parameters such as start, stop, step size, range,
and compliance. For typical voltage sweeps you’ll want to force the voltage between the drain and source while measuring the
current at the drain. Make sure to conduct several voltage sweeps at various forced gate voltages to aid in the analysis.
9. Check the current box/voltage box if you desire the current/voltage to be recorded in the Sheet tab Data worksheet and be
available for plotting in the Graph tab.
10. Now make contact to your sample with the probe tips
11. Run the measurement setup by clicking the green Run arrow on the tool bar located above the Definition tab. Make sure the
measuring indicator light at bottom right hand corner of the front panel is lit.
12. Save data by clicking on the Sheet tab then selecting the Save As tab. Select the file format and location.
Table 10.2.1 Connection selection.
Connection Description

SMU1 Medium power with low noise preamplifier

SMU2 Medium power source without preamplifier

SMU3 High Power

GNRD For large currents

10.2.4 https://chem.libretexts.org/@go/page/55937
Figure 10.2.9 Keithley Interactive Test Environment (KITE) interface window.

Measurement Analysis
Typical V-I Characteristics of JFETs
Voltage sweeps are a great way to learn about the device. Figure 10.2.10 shows a typical plot of drain-source voltage sweeps at
various gate-source voltages while measuring the drain current, ID for a n-channel JFET. The V-I characteristics have four distinct
regions. Analysis of these regions can provides critical information about the device characteristics such as the pinch off voltage,
VP, transcunductance gain, gm, drain-source channel resistance, RDS, and power dissipation, PD.

Figure adapted from Electronic Tutorials (www.electronic-tutorials.ws).

Ohmic Region (Linear Region)


This region is bounded by VDS < VP. Here the JFET begins to flow a drain current with a linear response to the voltage, behaving
like a variable resistor. In this region the drain-source channel resistance, RDS is modeled by 10.2.1, where ΔVDS is the change in
drain-source voltage, ΔID is the change in drain current, and gm is the transcunductance gain. Solving for gm results in 10.2.2.
ΔVDS 1
RDS   =     =  (10.2.1)
ΔID gm

ΔID 1
gm   =     =  (10.2.2)
ΔVDS RDS

Saturation Region

10.2.5 https://chem.libretexts.org/@go/page/55937
This is the region where the JFET is completely “ON”. The maximum amount of current is flowing for the given gate-source
voltage. In this region the drain current can be modeled by the 10.2.3, where ID is the drain current, IDSS is the maximum current,
VGS is the gate-source voltage, and VP is the pinch off voltage. Solving for the pinch off voltage results in 10.2.4.
VGS
ID   =  IDSS (1  −   ) (10.2.3)
VP

VGS
VP   =  1  −   −
− −
− (10.2.4)
ID

ID SS

Breakdown Region
This region is characterized by the sudden increase in current. The drain-source voltage supplied exceeds the resistive limit of the
semiconducting channel, resulting in the transistor to break down and flow an uncontrolled current.

Pinch-off Region (Cutoff Region)


In this region the gate-source voltage is sufficient to restrict the flow through the channel, in effect cutting off the drain current. The
power dissipation, PD, can be solved utilizing Ohms law (I = V/R) for any region using 10.2.5.
2 2
PD   =  ID   ×  VDC   =  (ID )   ×  RDS   =  (VDS ) / RDS (10.2.5)

The p-channel JFET V-I characteristics behave similarly except that the voltages are reversed. Specifically, the pinch off point is
reached when the gate-source voltage is increased in a positive direction, and the saturation region is met when the drain-source
voltage is increased in the negative direction.
Typical V-I Characteristics of MOSFETs
Figure 10.2.11 shows a typical plot of drain-source voltage sweeps at various gate-source voltages while measuring the drain
current, ID for an ideal n-channel enhancement MOSFET. Like JFETs, the V-I characteristics of MOSFETS have distinct regions
that provide valuable information about device transport properties.

Figure adapted from Electronic Tutorials (www.electronic-tutorials.ws).

Ohmic Region (Linear Region)


The n-channel enhanced MOSFET behaves linearly, acting like a variable resistor, when the gate-source voltage is greater than the
threshold voltage and the drain-source voltage is greater than the gate-source voltage. In this region the drain current can be
modeled by 10.2.6, where ID is the drain current, VGS is the gate-source voltage, VT is the threshold voltage, VDS is the drain-
source voltage, and k is the geometric factor described by 10.2.7, where µn is the charge-carrier effective mobility, COX is the gate
oxide capacitance, W is the channel width, and L is the channel length.
2
ID   =  2k(VGS − VT )VDS   −  [(VDS ) /2] (10.2.6)

10.2.6 https://chem.libretexts.org/@go/page/55937
W
k  =  μn COX (10.2.7)
L

Saturation Region
In this region the MOSFET is considered fully “ON”. The drain current for the saturation region is modeled by 10.2.8 . The drain
current is mainly influenced by the gate-source voltage, while the drain-source voltage has no effect.
2
ID   =  k(VGS   −  VT ) (10.2.8)

Solving for the threshold voltage VT results in 10.2.9.


−−

ID
VT   =  VGS   −  √ (10.2.9)
k

Pinch-off Region (Cutoff Region)


When the gate-source voltage, VGS, is below the threshold voltage VT the charge carriers in the channel are not available “cutting
off” the charge flow. Power dissipation for MOSFETs can also be solved using equation 6 in any region as in the JFET case.
FET V-I Summary
The typical I-V characteristics for the whole family of FETs seen in Figure 10.2.11 are plotted in Figure 10.2.12.

Figure 10.2.12 Plot of V-I characteristics for the various FET types. Adapted from P. Horowitz and W. Hill, in Art of Electronics,
Cambridge University Press, New York, 2nd Edn., 1994.
From Figure 10.2.12 we can see how the doping schemes that lead to enhancement and depletion are displaced along the VGS
axis. In addition, from the plot the ON or OFF state can be determined for a given gate-source voltage, where (+) is positive, (0) is
zero, and (-) is negative, as seen in Table 10.2.1.
Table 10.2.1 : The ON/OFF state for the various FETs at a given gate-source voltages where (-) is a negative voltage and (+) is a positive
voltage.
FET Type VGS = (-) VGS = 0 VGS = (+)

n-channel JFET OFF ON ON

p-channel JFET ON ON OFF

n-channel depletion MOSFET OFF ON ON

p-channel depletion MOSFET ON ON OFF

n-channel enhancement MOSFET OFF OFF ON

p-channel enhancement MOSFET ON ON OFF

10.2: Measuring Key Transport Properties of FET Devices is shared under a CC BY 4.0 license and was authored, remixed, and/or curated by
Pavan M. V. Raja & Andrew R. Barron via source content that was edited to conform to the style and standards of the LibreTexts platform; a

10.2.7 https://chem.libretexts.org/@go/page/55937
detailed edit history is available upon request.

10.2.8 https://chem.libretexts.org/@go/page/55937
Index
A F N
Atomic Force Microscopy Field Effect Transistors Neutron Activation Analysis
9.2: Atomic Force Microscopy (AFM) 10.2: Measuring Key Transport Properties of FET 1.9: Neutron Activation Analysis (NAA)
Auger Electron Spectroscopy Devices neutron diffraction
1.14: Auger Electron Spectroscopy fluorescence 7.5: Neutron Diffraction
4.5: Photoluminescence, Phosphorescence, and NMR Spectroscopy
Fluorescence Spectroscopy
B 4.7: NMR Spectroscopy
Bravais lattices G
7.1: Crystal Structure O
gas chromatography Ostwald Viscometer
3.1: Principles of Gas Chromatography
C 2.6: Viscosity
graphene
Capillary Electrophoresis 8.7: Characterization of Graphene by Raman
3.6: Capillary Electrophoresis Spectroscopy
P
combustion analysis phosphorescence
1.3: Introduction to Combustion Analysis H 4.5: Photoluminescence, Phosphorescence, and
crystallography Fluorescence Spectroscopy
HPLC Photoluminescence
7.1: Crystal Structure
3.2: High Performance Liquid chromatography
cumulant expansion 4.5: Photoluminescence, Phosphorescence, and
Hyperfine Coupling Fluorescence Spectroscopy
2.4: Dynamic Light Scattering
4.8: EPR Spectroscopy
Cyclic Voltammetry
R
2.7: Electrochemistry
I Raman Spectroscopy
ICP
D 1.5: ICP-AES Analysis of Nanoparticles
4.3: Raman Spectroscopy

Desorption Mass Spectroscopy Inductively coupled plasma atomic


5.3: Temperature-Programmed Desorption Mass
S
Spectroscopy Applied in Surface Chemistry emission spectroscopy Scanning Tunneling Microscopy (STM)
diamagnetism 1.5: ICP-AES Analysis of Nanoparticles 9.3: SEM and its Applications for Polymer Science
4.1: Magnetism interferometry Semiconductors
Differential Scanning Calorimetry 9.1: Interferometry 7.2: Structures of Element and Compound
10.1: A Simple Test Apparatus to Verify the Semiconductors
2.8: Thermal Analysis
Photoresponse of Experimental Photovoltaic Materials
differential thermal analysis and Prototype Solar Cells
Spot test
1.2: Spot Tests
2.8: Thermal Analysis Ion Chromatography
dislocation 3.5: Ion Chromatography
supercritical fluid chromatography
3.4: Supercritical Fluid Chromatography
7.1: Crystal Structure IR Spectroscopy
dual polarization interferometry 4.2: IR Spectroscopy
9.1: Interferometry T
Dynamic Light Scattering L Thermogravimetric analysis
2.4: Dynamic Light Scattering 2.8: Thermal Analysis
law of constant angles
Dynamic Viscosity 7.3: X-ray Crystallography
2.6: Viscosity V
M vertical scanning interferometry
E Mössbauer spectroscopy 9.1: Interferometry
Electrical Permittivity 4.6: Mössbauer Spectroscopy
viscosity
2.9: Electrical Permittivity Characterization of 2.6: Viscosity
magnetic moments
Aqueous Solutions
4.1: Magnetism
electron spectroscopy for chemical
magnetism X
analysis 4.1: Magnetism XAFS
1.13: X-ray Photoelectron Spectroscopy Magnetization 7.6: XAFS
Electroosmotic Mobility 4.1: Magnetism XAS
3.6: Capillary Electrophoresis MEKC 1.8: A Practical Introduction to X-ray Absorption
Electrophoretic Mobility Spectroscopy
3.6: Capillary Electrophoresis 7.6: XAFS
3.6: Capillary Electrophoresis Melting Point Apparatus
Elemental analysis XPS
2.1: Melting Point Analysis 1.13: X-ray Photoelectron Spectroscopy
1: Elemental Analysis miller indicies 4.9: X-ray Photoelectron Spectroscopy
EPR 7.1: Crystal Structure
4.8: EPR Spectroscopy
Z
ESCA
Zeta Potential
1.13: X-ray Photoelectron Spectroscopy
3.6: Capillary Electrophoresis

1 https://chem.libretexts.org/@go/page/212260
Index
A F N
Atomic Force Microscopy Field Effect Transistors Neutron Activation Analysis
9.2: Atomic Force Microscopy (AFM) 10.2: Measuring Key Transport Properties of FET 1.9: Neutron Activation Analysis (NAA)
Auger Electron Spectroscopy Devices neutron diffraction
1.14: Auger Electron Spectroscopy fluorescence 7.5: Neutron Diffraction
4.5: Photoluminescence, Phosphorescence, and NMR Spectroscopy
Fluorescence Spectroscopy
B 4.7: NMR Spectroscopy
Bravais lattices G
7.1: Crystal Structure O
gas chromatography Ostwald Viscometer
3.1: Principles of Gas Chromatography
C 2.6: Viscosity
graphene
Capillary Electrophoresis 8.7: Characterization of Graphene by Raman
3.6: Capillary Electrophoresis Spectroscopy
P
combustion analysis phosphorescence
1.3: Introduction to Combustion Analysis H 4.5: Photoluminescence, Phosphorescence, and
crystallography Fluorescence Spectroscopy
HPLC Photoluminescence
7.1: Crystal Structure
3.2: High Performance Liquid chromatography
cumulant expansion 4.5: Photoluminescence, Phosphorescence, and
Hyperfine Coupling Fluorescence Spectroscopy
2.4: Dynamic Light Scattering
4.8: EPR Spectroscopy
Cyclic Voltammetry
R
2.7: Electrochemistry
I Raman Spectroscopy
ICP
D 1.5: ICP-AES Analysis of Nanoparticles
4.3: Raman Spectroscopy

Desorption Mass Spectroscopy Inductively coupled plasma atomic


5.3: Temperature-Programmed Desorption Mass
S
Spectroscopy Applied in Surface Chemistry emission spectroscopy Scanning Tunneling Microscopy (STM)
diamagnetism 1.5: ICP-AES Analysis of Nanoparticles 9.3: SEM and its Applications for Polymer Science
4.1: Magnetism interferometry Semiconductors
Differential Scanning Calorimetry 9.1: Interferometry 7.2: Structures of Element and Compound
10.1: A Simple Test Apparatus to Verify the Semiconductors
2.8: Thermal Analysis
Photoresponse of Experimental Photovoltaic Materials
differential thermal analysis and Prototype Solar Cells
Spot test
1.2: Spot Tests
2.8: Thermal Analysis Ion Chromatography
dislocation 3.5: Ion Chromatography
supercritical fluid chromatography
3.4: Supercritical Fluid Chromatography
7.1: Crystal Structure IR Spectroscopy
dual polarization interferometry 4.2: IR Spectroscopy
9.1: Interferometry T
Dynamic Light Scattering L Thermogravimetric analysis
2.4: Dynamic Light Scattering 2.8: Thermal Analysis
law of constant angles
Dynamic Viscosity 7.3: X-ray Crystallography
2.6: Viscosity V
M vertical scanning interferometry
E Mössbauer spectroscopy 9.1: Interferometry
Electrical Permittivity 4.6: Mössbauer Spectroscopy
viscosity
2.9: Electrical Permittivity Characterization of 2.6: Viscosity
magnetic moments
Aqueous Solutions
4.1: Magnetism
electron spectroscopy for chemical
magnetism X
analysis 4.1: Magnetism XAFS
1.13: X-ray Photoelectron Spectroscopy Magnetization 7.6: XAFS
Electroosmotic Mobility 4.1: Magnetism XAS
3.6: Capillary Electrophoresis MEKC 1.8: A Practical Introduction to X-ray Absorption
Electrophoretic Mobility Spectroscopy
3.6: Capillary Electrophoresis 7.6: XAFS
3.6: Capillary Electrophoresis Melting Point Apparatus
Elemental analysis XPS
2.1: Melting Point Analysis 1.13: X-ray Photoelectron Spectroscopy
1: Elemental Analysis miller indicies 4.9: X-ray Photoelectron Spectroscopy
EPR 7.1: Crystal Structure
4.8: EPR Spectroscopy
Z
ESCA
Zeta Potential
1.13: X-ray Photoelectron Spectroscopy
3.6: Capillary Electrophoresis

1 https://chem.libretexts.org/@go/page/350618
Glossary
Sample Word 1 | Sample Definition 1

1 https://chem.libretexts.org/@go/page/279537

You might also like