Bioinstrumentation
Bioinstrumentation
This book describes measurement methods in medicine and biology. While many books on
medical instrumentation cover only hospital instrumentation, this book also encompasses
measurements in the growing fields of molecular biology and biotechnology.
As a first course in bioinstrumentation for undergraduate biomedical engineers, this
book describes measurement methods used in the growing fields of cell engineering, tissue
engineering, and biomaterials. It will also serve as a text for medical and biological personnel
who wish to learn measurement techniques.
Chapter 1 covers the basics of instrumentation systems, errors, and the statistics
required to determine how many subjects are required in a research study. Because many
biomedical engineers are not specialists in electronics, chapter 2 provides the necessary
background in circuits, amplifiers, filters, converters, and signal processing.
Chapter 3 describes clinical measurements of molecules, such as oxygen and glucose,
to biotechnology measurements such as DNA sequencing. For the fields of biomaterials and
tissue engineering, chapter 4 covers measurements on polymers, using surface analysis,
protein adsorption and molecular size.
Measurements on blood components are the commonest measurements on cells, as
described in chapter 5. Cells are counted and identified using changes in impedance and light
scattering. Chapter 6 covers cellular measurements in biomaterials and tissue engineering,
such as cellular orientation, rolling velocity, pore size, deformation, shear stress, adhesion,
migration, uptake, protein secretion, proliferation, differentiation, signaling and regulation.
Chapter 7 describes measurements on the nervous system—action potentials, EEG,
ERG, EOG, EMG and audiometry—and brain imaging using X ray, CT, MRI, nuclear
imaging, SPECT, PET, and biomagnetism. Chapter 8 covers heart and circulation, with
measurements of cardiac biopotentials, pressures, output, sounds, viability—as well as blood
flow and pressure in the periphery.
Measurements of pulmonary volume, flow, diffusion, and airway resistance are
described in chapter 9, which also includes kidney clearance, bone mineral, and skin water
loss. Chapter 10 covers measurements of body temperature, fat, and movement.
Each chapter has references for further study as well as homework problems to test
comprehension. A web site contains complete laboratory instructions for 12 laboratories,
examination questions, and quiz questions.
Suggested prerequisite courses are calculus, biology, chemistry, and physics. I would
welcome suggestions for improvement of subsequent editions.
John G. Webster
Webster@engr.wisc.edu
January 2002
1
Measurement Systems
Kevin Hugo
1-1
1-2 Chapter 1 Measurement Systems
The patient in the above story has noninsulin-dependent diabetes mellitus (NIDDM), a
disease in which afflicted persons have high levels of blood glucose. Diabetes mellitus is
a very common chronic disease with many complications, and is one of the most expen-
sive diseases to treat in the United States. However, with early detection and diligent
treatment, the risk of complications can be reduced (Diabetes, 1993).
In diagnosing the patient’s disease and managing the future treatment, it is nec-
essary to take measurements using biomedical instruments, which is the subject of this
book. One way to classify the various types of bioinstrumentation is to consider the level
of biological organization involved in taking the measurement. For example, it may in-
volve molecules, cells, tissues, organ systems, or the entire body. The chapters of this
book are presented in the order of these biological levels.
This chapter discusses how and why measurements are taken. For example,
every patient has vital signs taken not only because they may be useful in analyzing the
problem that the patient is concerned with, but also because they indicate that person’s
general health. We must also understand that the results of measurements will vary due to
errors in taking the measurement as well as individual differences.
An automated thermometer measured Ms. Jones’ temperature in just several
seconds. This is accomplished by using a temperature-sensing element as well as elec-
tronics to derive the temperature based on a mathematical algorithm. This type of elec-
tronic signal processing is explained in Chapter 2.
Measurements of concentrations of glucose and other molecules are performed
by spectrophotometry, as described in Chapter 3. Ms. Jones may have coronary artery
disease and need an implanted graft vessel. Chapter 4 describes the electron microscopes
used to test proposed biomaterials and the adhesion of protein molecules to them.
Methods used to analyze the blood sample that was drawn are described in
Chapter 5. Diabetes causes a number of complications. In order to study the urinary casts
(cylindrical protein molds) passed in the urine, Chapter 6 describes the optical micro-
scopes used. Ms. Jones reported a loss of feeling in her legs. Chapter 7 describes the
nervous system and the test used to diagnose nerve disease.
Chapter 8 describes the electrocardiogram, which is used to diagnose Ms. Jones’
cardiac problems, as well as methods to obtain blood pressure measurements. Ms. Jones’
kidney function is diagnosed by measuring creatinine clearance, as described in Chapter
9. NIDDM patients are frequently overweight. Chapter 10 describes measurements on the
body, including measurements of body fat.
lowing paragraphs in this section. The Whitaker Foundation describes various fields of
biomedical engineering at http://www.whitaker.org/glance/acareer.html.
Bioinstrumentation
Biomaterials
Biomechanics
Biosignals
Biosystems
Biotransport
Cellular engineering
Clinical engineering
Tissue engineering
Rehabilitation engineering
Table 1.2 shows that instrumentation can be organized by the biological discipline that
utilizes it. Each of these disciplines can be divided into several subspecialties. For exam-
ple, medicine can be subdivided into anatomy, anesthesiology, biomolecular chemistry,
biostatistics and medical informatics, family medicine, medical microbiology and immu-
nology, medical physics, neurology, neurophysiology, obstetrics-gynecology, oncology,
ophthalmology, pathology and laboratory science, pediatrics, physiology, preventive
medicine, psychiatry, radiology, rehabilitation medicine, and surgery.
Table 1.2 Biomedical engineers work in a variety of disciplines. One example of instrumentation is
listed for each discipline.
Biomedical Engineers may become a part of the workplace in a variety of different areas,
as shown in Table 1.3. Industry has a demand for biomedical engineers for development
of new technologies and also for testing the theories that others develop. They could also
be technology representatives for existing products. Positions in the government often
deal with testing of proposed products for safety and reliability, to safeguard the public
from harm. In a clinical setting, the biomedical engineer may be the only individual with
working knowledge of some of the instruments, and is needed to allow for their efficient
use. Research institutions are also the place of developing new ideas, and often involve
the dissemination of techniques and practices to others to further the general understand-
ing.
Industry
Government
Clinical institutions
Academic research
Figure 1.1 shows that researchers often need to use instrumentation to obtain data as part
of the scientific method. For example, we might hypothesize that exercise reduces high
blood pressure yet experimentation and analysis are needed to support or refute the hy-
pothesis. Experiments are normally performed multiple times. Then the results can be
analyzed statistically to determine the probability (hopefully less than 0.05) that the re-
sults might have been produced by chance. Results are reported in scientific journals with
enough detail so that others can replicate the experiment to confirm them.
Figure 1.1 In the scientific method, a hypothesis is tested by experiment to determine its validity.
Figure 1.2 shows that physicians often need instrumentation to obtain data as part of the
scientific method. For example, a physician obtaining the history of a patient with a com-
plaint of poor vision would list diabetes as one possibility on a differential diagnosis.
1.2 The Need for Bioinstrumentation 1-7
Figure 1.2 The physician obtains the history, examines the patient, performs tests to determine the
diagnosis and prescribes treatment.
Figure 1.3 shows that the measurand is measured by a sensor, which converts the variable
to an electrical signal, which can undergo signal processing. Sometimes the measurement
system provides feedback through an effector to the subject. As an example, light flashes
stimulate the eye when measuring visual evoked responses from electrodes placed over
the visual cortex of the brain.
Outputs
Figure 1.3 A typical measurement system uses sensors to measure the variable, has signal process-
ing and display, and may provide feedback.
Figure 1.4(a) shows that a patient reading an instrument usually lacks sufficient
knowledge to achieve the correct diagnosis. Figure 1.4(b) shows that by adding the clini-
cian to form an effective feedback system, the correct diagnosis and treatment result.
In certain circumstances, proper training of the patient and a well-designed in-
strument can lead to self-monitoring and self-control, one of the goals of bioinstrumenta-
tion. This is shown in Figure 1.5. An example of such a situation is the day-to-day moni-
toring of glucose by people suffering from diabetes. Such an individual will contact a
clinician if there is an alert from the monitoring instrument.
1-8 Chapter 1 Measurement Systems
Patient Instrument
(a) (b)
Figure 1.4 (a) Without the clinician, the patient may be operating in an ineffective closed loop
system. (b) The clinician provides knowledge to provide an effective closed loop system.
Abnormal
readings
Clinician Patient Instrument
Figure 1.5 In some situations, a patient may monitor vital signs and notify a clinician if abnormali-
ties occur.
1.3 Instrumentation
1.3.1 Measurands
The measurand is the measured quantity. It may be voltage as when measuring the elec-
trocardiogram, pressure as when measuring the blood pressure, force as when measuring
weight, or other quantities. Table 1.4 lists some common biomedical measurands.
Measurands can be obtained by invasive or noninvasive methods. For example,
if we want to accurately measure the blood pressure in the aorta in vivo (i.e. in a living
organism), we must place a pressure sensor on the tip of a catheter (a long narrow tube),
cut into a peripheral vessel, and advance the catheter to the aorta. An alternative invasive
method is to place the pressure sensor at the distal end of the catheter outside the body.
1.3 Instrumentation 1-9
1.3.2 Sensors
A variety of sensors have been adapted from industry for use in measuring biomedical
variables such as pressure, flow, and temperature. Table 1.5 shows the specifications for
a typical blood pressure sensor. Note that several specifications are designed to prevent
patient electrocution.
Specification Value
Pressure range –30 to +300 mmHg
Overpressure without damage –400 to +4000 mmHg
Maximum unbalance ±75 mmHg
Linearity and hysteresis ±2% of reading or ±1 mmHg
Risk current at 120 V 10 µA
Defibrillator withstand 360 J into 50 Ω
Table 1.5 Sensor specifications for blood pressure sensors are determined by a committee com-
posed of individuals from academia, industry, hospitals, and government.
Linearity is the degree to which the sensor stays within its theoretical output for
some input. Section 1.3.3 discusses linearity in more detail. Hysteresis is a phenomenon
in which increasing the pressure the sensor gives one output but decreasing the pressure
1-10 Chapter 1 Measurement Systems
to the same previous value gives a different output, as shown in Figure 1.6. Risk current
is the amount of current flowing through the sensor at 120 V. Defibrillator withstand is
the amount of energy (in joules) the sensor can withstand if a defibrillator is used while
the sensor is taking measurements.
Sensor
signal
Measurand
Figure 1.6 A hysteresis loop. The output curve obtained when increasing the measurand is different
from the output obtained when decreasing the measurand.
Optimal sensors are designed specifically for the desired measurand, immune to
interference from other variables, and have high gain as shown in Figure 1.7(b). A sensor
with low gain, as shown in Figure 1.7(a), would not change its signal as much for a small
change in the measurand as one with a high gain.
Most sensors are analog and provide a continuous range of amplitude values for
output, as shown in Figure 1.8(a). Other sensors yield the digital output shown in Figure
1.8(b). Digital output has poorer resolution, but does not require conversion before being
input to digital computers and is more immune to interference. Digital and analog signals
will be examined with more detail in Chapter 2.
Sensor Sensor
signal signal
Measurand Measurand
(a) (b)
Figure 1.7 (a) A low-sensitivity sensor has low gain. (b) A high sensitivity sensor has high gain.
Amplitude Amplitude
Time Time
(a) (b)
Figure 1.8 (a) Analog signals can have any amplitude value. (b) Digital signals have a limited
number of amplitude values.
1.3 Instrumentation 1-11
Specific characteristics
Bioinstrumentation should be designed with a specific signal in mind. Table 1.6 shows a
few specifications for an electrocardiograph (a device that records the electrical activity
generated by the heart.) The values of the specifications, which have been agreed upon by
a committee, are drawn from research, hospitals, industry, and government.
Specification Value
Input signal dynamic range ±5 mV
Dc offset voltage ±300 mV
Slew rate 320 mV/s
Frequency response 0.05 to 150 Hz
Input impedance at 10 Hz 2.5 MΩ
Dc lead current 0.1 µA
Return time after lead switch 1s
Overload voltage without damage 5000 V
Risk current at 120 V 10 µA
Table 1.6 Specification values for an electrocardiograph are agreed upon by a committee.
The electrocardiogram (ECG), the recording of the electrical activity of the heart,
can have a voltage value of anywhere between 0.5 mV and 5 mV (Table 1.4). The elec-
trocardiograph should be able to handle this range, therefore its dynamic input signal
range is designed to handle signals from –5 mV to 5 mV. If the measurand exceeds the
input range of the instrument, the output will saturate and cause an error. Figure 1.9
shows an input signal that exceeds the ±5 mV input range (a) and the resulting amplified
signal (b).
Direct current (dc) offset voltage is the amount a signal may be moved from its
baseline and still be amplified properly by the system. Figure 1.10 shows an input signal
without (a) and with (b) offset.
Slew rate is the maximum rate at which the system can observe a changing volt-
age per unit time without distortion. In the specifications above, if the signal changes
more than 320 mV/s then the resulting output signal will be distorted.
The ECG signal is composed of components that vary in frequency from 0.05
Hz to 150 Hz (Table 1.4) The electrocardiograph is designed to accommodate this as is
shown by its frequency response range. The frequency response of a device is the range
of frequencies of a measurand that it can handle. Frequency response is usually plotted as
gain versus frequency. Figure 1.11 shows the plotted frequency response of this electro-
cardiograph. Frequency response is discussed in greater detail in Section 2.2.6.
Input impedance is the amount of resistance the signal sees coming into the elec-
trocardiograph. It is desirable to have very high input impedances, in this case 2.5 MΩ,
1-12 Chapter 1 Measurement Systems
so that the majority of the signal will enter the system to be processed and not be lost in
the skin impedance.
Amplitude
5 mV
Dynamic
range
Time
–5 mV
(a)
Amplitude
1V
Time
–1 V (b)
Figure 1.9 (a) An input signal that exceeds the dynamic range. (b) The resulting amplified signal is
saturated at ±1 V.
Amplitude
Time
(a)
Amplitude
Dc offset
Time
(b)
Figure 1.10 (a) An input signal without dc offset. (b) An input signal with dc offset.
Dc lead current is the amount of current that flows through the leads to the pa-
tient being observed by the electrocardiograph. It is important that current flowing to the
patient not exceed very miniscule amounts (i.e. on the order of microamperes). This is
because larger currents can polarize electrodes, resulting in a large offset voltage at the
amplifier input.
1.3 Instrumentation 1-13
Amplitude
1.0
0.1
0.05 Hz 150 Hz
Frequency
General Characteristics
Another characteristic that all systems have is whether or not they are linear. Linearity is
highly desirable for simplifying signal processing. For example, if you measure some
input (measurand) you get some output. Now if you double that input you would expect
to double the output. This is not the case for nonlinear systems. Doubling the input in a
nonlinear system might yield triple the output. Figure 1.12 shows the input/output rela-
tion of a linear system (a) and of a nonlinear system (b).
1-14 Chapter 1 Measurement Systems
Output Output
Input Input
(a) (b)
Figure 1.12 (a) A linear system fits the equation y = mx + b. Note that all variables are italic. (b) A
nonlinear system does not fit a straight line.
Finally, Figure 1.13 shows that all bioinstrumentation observes the measurand
either continuously or periodically. Continuous signal measurement is most desirable
because no data are lost. However, computer-based systems require periodic measure-
ments since by their nature, computers can only accept discrete numbers at discrete inter-
vals of time.
Amplitude Amplitude
Time Time
(a) (b)
Figure 1.13 (a) Continuous signals have values at every instant of time. (b) Discrete-time signals
are sampled periodically and do not provide values between these sampling times.
1.3.4 Outputs
Bioinstrumentation can direct its output three ways: display(s), storage, and communica-
tion. A display is simply a monitor or some other device attached to the instrumentation
that gives the user a sense of what the input is doing. The term display is a bit misleading
since output does not necessarily have to be visible (e.g. a stethoscope uses sound for its
output.) Storage is increasingly common as computers are integrated into bioinstrumenta-
tion. Storage usually takes the form of saving data onto a hard drive or some other type of
permanent record. Communication involves transmitting the output to some other loca-
tion for analysis. Modern imaging systems make extensive use of communication via
computer networks and wireless transmission.
Certain groups of measurements are often ordered together because they are very com-
monly used or because they are related. This may occur even if the measurements are
1.3 Instrumentation 1-15
based on different principles or are taken with different sensors. Table 1.7 is an example
of one of these groups of measurements, which are called panels or series.
When we measure a variable, for example the voltage of the electrocardiogram, we seek
to determine the true value, as shown in Figure 1.14(a). This true value may be corrupted
by a variety of errors. For example movement of electrodes on the skin may cause an
undesired added voltage called an artifact. Electric and magnetic fields from the power
lines may couple into the wires and cause an undesired added voltage called interference,
as shown in Figure 1.14(b). Thermal voltages in the amplifier semiconductor junctions
may cause an undesired added random voltage called noise. Temperature changes in the
amplifier electronic components may cause undesired slow changes in voltage called drift.
We must evaluate each of these error sources to determine their size and what we can do
to minimize them. Usually we try to minimize the largest error, since that is the weakest
link in the measurement.
1-16 Chapter 1 Measurement Systems
(a) (b)
Figure 1.14 (a) Signals without noise are uncorrupted. (b) Interference superimposed on signals
causes error. Frequency filters that can be used to reduce noise and interference are discussed in
Chapter 2.
Figure 1.15 (a) Original waveform. (b) An interfering input may shift the baseline. (c) A modifying
input may change the gain.
Figure 1.15(b) shows that an interfering input may be added to the original
waveform and shift it away from the baseline. Figure 1.15(c) shows that a modifying
input may multiply the original waveform to change its gain.
Resolution is the smallest incremental quantity that can be reliably measured. For exam-
ple, a voltmeter with a larger number of digits has a higher resolution than one with fewer
digits. A speedometer with more tick marks between each numbered speed has a higher
resolution than one with no tick marks. However, high resolution does not imply high
accuracy.
Figure 1.16 shows that precision is the quality of obtaining the same output from
repeated measurements from the same input under the same conditions. High resolution
implies high precision.
Repeatability is the quality of obtaining the same output from repeated meas-
urements from the same input over a period of time.
Accuracy is the difference between the true value and the measured value di-
vided by the true value. Figure 1.17(a) shows low accuracy and high precision. Figure
1.17(b) shows high accuracy and high precision.
Obtaining the highest possible precision, repeatability, and accuracy is a major
goal in bioinstrumentation design.
1.4 Errors in Measurements 1-17
(a) (b)
Figure 1.16 Data points with (a) low precision and (b) high precision.
(a) (b)
Figure 1.17 Data points with (a) low accuracy and (b) high accuracy.
1.4.3 Calibration
1-18 Chapter 1 Measurement Systems
Output Output
Input Input
(a) (b)
Figure 1.18 (a) The one-point calibration may miss nonlinearity. (b) The two-point calibration may
also miss nonlinearity.
1.5 Statistics
1.5.1 Distributions
Mean
∑ xi
x=
n
(1.1)
where i = 1, 2,…n.
1.5 Statistics 1-19
Standard Deviation
A measure of the spread of data about the mean is the estimated sample standard devia-
tion
∑ ( xi − x )
2
s=
n −1
(1.2)
s
x=
n −1 (1.3)
Example 1.1 If Person 1 weighs 68 kg, Person 2 weighs 90 kg and Person 3 weighs 95
kg, find the mean, the standard deviation, and the standard deviation of the mean.
Here n = 3. The mean of the data is
∑ xi 68 + 90 + 95
x= = = 84.3 kg
n 3
s 14.4
x= = = 10.2 kg
n −1 3 −1
Gaussian Distribution
The spread (distribution) of data may be rectangular, skewed, Gaussian, or other. The
Gaussian distribution is given by
e −( X −µ ) (2σ )
2 2
f (X ) =
2π σ (1.4)
1-20 Chapter 1 Measurement Systems
where µ is the true mean and σ is the true standard deviation of a very large number of
measurements. Figure 1.19 shows the familiar “bell curve” shape of the Gaussian distri-
bution.
δ
α
–σ µ +σ x
Mean
Figure 1.19 For the normal distribution, 68% of the data lie within ±1 standard deviation. By
measuring samples and averaging, we obtain the estimated mean x s , which has a smaller standard
deviation sx. α is the tail probability that xs does not differ from µ by more than δ.
Poisson Probability
The Poisson probability density function is another type of distribution. It can describe,
among other things, the probability of radioactive decay events, cells flowing through a
counter, or the incidence of light photons. The probability that a particular number of
events K will occur in a measurement (or during a time) having an average number of
events m is
–m
p ( K ; m) = e mK
K!
(1.5)
The standard deviation of the Poisson distribution is m . Figure 1.20 shows a typical
Poisson distribution.
1.5 Statistics 1-21
p 0.2
0.1
0.0
0 1 2 K 3 4 5
Example 1.2 Assume that the average number of cells passing through a counter in one
sampling period is 10. Calculate the probability that 15 cells will be counted in a sam-
pling period.
Since 10 is the average number counted cells, m = 10. We want to find the prob-
ability of 15 cells being counted so K = 15.
–m
p ( K ; m) = e mK = e −10 1015 = 0.035
K! 15!
Suppose we want to know the average weight of adults. It is not feasible to weigh every
single adult and then take the average of all the weights. All adults are called the popula-
tion. Instead, we decide to take a small fraction of the adults, say 1 out of every 1000, and
average these weights. This small fraction is called our sample population. Now we have
an average weight for our sample population. We want to know if our sample population
average weight is a good estimation of the population average weight. There are two
ways for us to find out.
If we represent our sample population average weight as a single number, xs ,
then we have made a point estimation of the population average weight. Let’s call µ the
population average weight and the difference between µ and x s is δ. The standard devia-
tion of the population is σ. The tail probability that x s will not differ from µ by more
than δ is α. We can control the probability of how much x s differs from µ by adjusting
the sample size, N, using the formula
2
σ
N = Z α2
δ (1.6)
where Zα is a multiple of the standard deviation. This number is obtained from a statisti-
cal table (Distribution Tables, 2002).
1-22 Chapter 1 Measurement Systems
The confidence is how sure one wants to be that the sample size, N, calculated will actu-
ally meet the specifications for δ
1− percent confidence
α=
2
(1.7)
If we know the standard deviation, σ, of the population average weight, we can use it.
Often, the standard deviation of a population can be found in the literature or from some
other source. If it cannot, then the sample standard deviation, s, must be used instead.
Also, the Student t distribution, tα, must be used instead of Zα.
2
s
N = tα2
δ (1.8)
Example 1.3 Suppose we know that the standard deviation of the adult average weight is
3 kg. Using Eq. (1.6), we will set δ to 2, which means our estimate does not deviate from
the population average weight by more than ±2 kg. Lastly, we want a 95% confidence
that the sample size N we calculate will meet our specifications for δ.
2
3
N = (1.96) 2 = 8.6 ≈ 9
2
which indicates that we would need to take the average of nine weights to get an estima-
tion of the population average weight within ±2 kg with a confidence of 95%. We could
then say that the average weight of the population was some value, w.
Another way to represent estimations is with a confidence interval. When deal-
ing with confidence intervals, δ is twice as large because it represents the difference be-
tween the upper limit and lower limit instead of between µ and x s . The sample size
needed for an interval estimation is twice that needed for a point estimation and can be
found with the formula
2
σ
N = 2Z α2
δ (1.9)
Using the conditions stated above, we could solve for N and find it equal to approxi-
mately 18. We would need 18 samples to be 95% confident that our weight represented
the average weight of the population. We could then report our average weight, w, as
1.5 Statistics 1-23
In the event that we do not know the standard deviation of the population, we can use the
sample standard deviation, s, and the Student t distribution tα.
In hypothesis testing, there are two hypotheses. H0, the null hypothesis, is a hypothesis
that assumes that the variable in the experiment will have no effect on the result and Ha is
the alternative hypothesis that states that the variable will affect the results. For any popu-
lation, one of the two hypotheses must be true. The goal of hypothesis testing is to find
out which hypothesis is true by sampling the population.
In reality, H0 is either true or false and we draw a conclusion from our tests of
either true or false. This leads to four possibilities, as shown in Table 1.8.
Usually, the probability of making a Type I error is designated as α and the
probability of making a Type II error is designated as β. Common values for α, the sig-
nificance level, are 0.01, 0.05, and 0.1. The statistical power of a test is the probability
that H0 will be rejected when it is false. Statistical power is given by 1 – β. Common
power levels are above 0.8, indicating that when H0 is false it will be correctly rejected
more than 80% of the time.
Many scientific studies involving a hypothesis test will report the p value of the
test. A very small p value means that the null hypothesis is unlikely to be true. When p is
below an arbitrary cut-off value, e.g. 0.05, the result is called statistically significant. A
study that reports p = 0.05 means that since the result obtained would happen only once
in 20 times if H0 were true, then H0 should be rejected.
Table 1.9 Equivalent table of Table 1.8 for results relating to a condition or disease.
Test results about a condition or disease use the terminology in Table 1.9. In this
case H0 states that an individual does not have a condition or a disease, whereas Ha states
that an individual does have a condition or disease.
1-24 Chapter 1 Measurement Systems
The terms in Table 1.9 are useful for defining measures that describe the propor-
tion of, for example, a disease in a population and the success of a test in identifying it.
Incidence is the number of cases of a disease during a stated period, such as x
cases per 1000 per year.
Prevalence is the number of cases of a disease at a given time such as y cases per
1000. It is all diseased persons divided by all persons.
TP + FN
Prevalence =
TN + TP + FN + FP
(1.10)
Sensitivity is the probability of a positive test result when the disease is present.
Among all diseased persons, it is the percent who test positive.
TP
Sensitivity = 100%
TP + FN
(1.11)
TN
Specificity = 100%
TN + FP
(1.12)
Considering only those who test positive, positive predictive value (PPV) is the
ratio of patients who have the disease to all who test positive.
TP
PPV = 100%
TP + FP
(1.13)
Considering only those who test negative, negative predictive value (NPV) is the
ratio of nondiseased patients to all who test negative.
TN
NPV = 100%
TN + FN
(1.14)
Tests to diagnose disease are not perfect. Figure 1.21 shows that positive test
results may occur for only a portion of the normal population and only a portion of the
diseased population. We set the threshold for the test to maximize true results and mini-
mize false results.
Normal Threshold
population
True
negative
False positive,
p=α
Diseased
population
True
positive
False negative,
p=β
Figure 1.21 The test result threshold is set to minimize false positives and false negatives.
The sample size needed to test a hypothesis depends on the level of significance
and power desired. We can increase the threshold to make it very difficult to reject H0 in
order to reduce Type I errors (false positives). But then we risk not rejecting H0 when it
is false and making a Type II error (false negative). If we decrease the threshold, the con-
verse is also true.
For a test comparing the means m1, m2 of two Gaussian populations with known
and equal standard deviations, σ, the size of each sample needed to test H0: |m1 – m2| < δ,
Ha: |m1 – m2| > δ, is (Selwyn, 1996, Appendix B)
2
(
N = 2 Zα + Z β )2 σδ
(1.15)
where Zα and Zβ are the values for the standard Gaussian distribution corresponding to,
respectively, the tail probabilities of size α and β. For example, if α = 0.05, β = 0.1, and
1 – β = 0.9, Zα = 1.96 and Zβ = 1.65 and N = 26.06(σ/δ)2. If the result of the test suggests
1-26 Chapter 1 Measurement Systems
that we accept H0, then there is a 5% probability that H0 is false. If the result of the test
suggests that we reject H0, then there is a 10% probability that H0 is true.
For a test about the mean m of a normal population, with H0: |m| < δ and Ha: |m|
> δ, the sample size needed is
2
(
N = Zα + Z β )2 σδ
(1.16)
where Zα and Zβ have the same meaning as in Eq. (1.15). Kanji (1993) provides a fine
summary of statistical tests.
A final word of caution about N in the above equations. All the sample sizes
calculated are estimates based on several assumptions, some of which may be difficult to
prove. N is not an exact number. Hence, it may be convenient to modify some of the rele-
vant factors involved in the equations in order to seek their effect on N. In any case, a
high value for N indicates the need for pilot studies aimed to reduce the experimental
variability.
1.7 References
Biomedical Engineering: IEEE EMBS [Online] www.eng.unsw.edu.au/embs/index.html.
Dawson-Saunders, B. and Trapp, R. G. 1990. Basic and Clinical Biostatistics. Norwalk,
CT: Appleton & Lange.
1.7 References 1-27
Doebelin, E. O. 1990. Measurement Systems: Application and Design. 4th ed. New York:
McGraw-Hill.
Kanji, G. K. 1993. 100 Statistical Tests. Newbury Park, CA: SAGE Publications.
Nathan, D. M. 1993. Long-term complications of diabetes mellitus. N. Engl. J. Med., 328
(23): 1676–85.
Distribution Tables 2002. [Online] www.statsoft.com/textbook/stathome.html
The Diabetes Control and Complications Trial Research Group. 1993. The effect of in-
tensive treatment of diabetes on the development and progression of long-term com-
plications in insulin-dependent diabetes mellitus. N. Engl. J. Med., 329 (14): 977–86.
Selwyn, M. R. 1996. Principles of Experimental Design for the Life Sciences. Boca Raton:
CRC Press.
Webster, J. G. (ed.) 1998. Medical Instrumentation: Application and Design. 3rd ed. New
York: John Wiley & Sons.
The Whitaker Foundation [Online] www.whitaker.org
1.8 Problems
1.1 Compare the scientific method with that of making a diagnosis. List three things
these processes have in common and three differences.
1.2 Give an example of a problem requiring different areas of biomedical engineering
to be integrated in order to solve it.
1.3 Describe why a differential diagnosis is important.
1.4 Apply the steps of diagnosing diseases to diabetes.
1.5 For a system with a continuous, constant drift, explain what type of calibration
would work best.
1.6 Your samples from a population are 1, 1, 3, 5, 5. Estimate the mean x and stan-
dard deviation s.
1.7 Your samples from a population are 1, 1, 3, 5, 5. Estimate the standard deviation of
the mean sx.
1.8 For the Poisson probability density distribution, calculate and plot p(K;m) for K = 0,
1, 2, 3, 4, and m = 2.
1.9 An X-ray image has 1000 × 1000 picture elements (pixels). Calculate the number
of X-ray photons required so that the signal-to-noise ratio (SNR) = (average num-
ber of photons)/(standard deviation of the number of photons) = 8.
1.10 The results of test X are based on a threshold; positive results are greater than the
threshold and negative results are less than the threshold. If the threshold is in-
creased, explain how the sensitivity and specificity of test X change.
1.11 Find the prevalence, sensitivity, specificity, positive predictive value, and negative
predictive value for a test where TP = 80, FN = 20, TN = 30, FP =10.
1.12 A test had 90% sensitivity and 80% specificity. If there are 5 people actually af-
fected by the condition and 60 not affected, then find the number of true positives,
true negatives, false positives, and false negatives.
1-28 Chapter 1 Measurement Systems
1.13 A test has 83.3% sensitivity and 91.7% specificity. If the tests report that 15 people
are positive and 57 people are negative, calculate the number of TP, TN, FP, and
FN.
1.14 Express positive predictive value and negative predictive value in terms of the
prevalence, sensitivity, and specificity.
1.15 For the blood pressure of adults, the mean µ = 120 mmHg and the standard devia-
tion σ = 5 mmHg. Calculate the number N of adults we must test to obtain a prob-
ability p = 0.95 that our sample population average blood pressure x s does not dif-
fer from µ by more than δ = 3 mmHg.
1.16 Find a magazine or journal article about a new technology in biomedical engineer-
ing. Provide the complete reference. Write a one-paragraph summary of what it
may be used for and how it works.
1.17 Find web information about a new technology in biomedical engineering. Provide
the complete reference. Write a one-paragraph summary of what it may be used for
and how it works.
2
Basic Concepts of Electronics
Hong Cao
Medical instruments are widely used in clinical diagnosis, monitoring, therapy, and
medical research. They provide a quick and precise means by which physicians can
augment their five senses in diagnosing disease. These instruments contain electric
components such as sensors, circuits, and integrated circuit (IC) chips. Modern
electronics technology, which includes transistors, ICs, and computers, has
revolutionized the design of medical instruments.
Biomedical engineers should have a fundamental understanding of their
operations and a basic knowledge of their component electric and electronic s ystems.
Using this knowledge provides a better understand the principles of various
measurements or even develop new measurements and instruments.
Electrical engineering is too large a topic to cover completely in one chapter.
Thus, this chapter presents some very basic concepts in several fields of electrical
engineering. It discusses analog components such as resistors, capacitors, and inductors.
It then goes on to basic circuit analysis, amplifiers, and filters. From this it moves to the
digital domain, which includes converters, sampling theory, and digital signal processing.
It then discusses the basic principles of microcomputers, programming languages,
algorithms, database systems, display components, and recorders.
2.1.1 Current
An atom contains a nucleus surrounded by electrons. Most of the electrons are tightly
bound to the atom, while some electrons in the outer orbits are loosely bound and can
move from one atom to another. In conductors, there are many free electrons. This
process of electron transfer occurs in random directions in materials.
1
2 Chapter No. Title
I
+
− − +
Electron
− +
(a) (b)
Figure 2.1 Electric current within a conductor. (a) Random movement of electrons generates no
current. (b) A net flow of electrons generated by an external force.
I av = ∆Q ∆t (2.1)
The rate at which current flows varies with time as does charge. We therefore can define
the instantaneous current, I, as the differential limit of Eq. (2.1).
I = dQ dt (2.2)
The unit of current is the ampere (A), which represents a net flow of one coulomb (1 C),
of charge or 6.242 × 1018 electrons across the plane per second (s).
The electron has a negative charge. When a negative charge moves in one
direction, it yields the same result as a positive charge moving in the opposite direction.
Conventionally, we define the direction of positive charge movement to be the direction
of the electric current. Figure 2.1(b) shows that the direction of current is opposite to that
of the flow of electrons.
Section No. Title 3
WAB
∆V = VB − VA = (2.3)
q0
where VB and VA are the potential at points B and A, respectively. The unit of potential is
the volt (V), where 1 J/C = 1 V. If we choose the potential at infinity to be zero, the
absolute potential of a point in an electric field can be defined as the total work per unit
charge that has been done to move the charge from infinity to the point.
It is important that potential difference not be confused with difference in
potential energy. The potential difference is proportional to the change in potential
energy, where the two are related by ∆U = q0∆V.
From Eq. (2.3), we can determine the work needed to move a charge from A to
B if we know the potential difference between the two points. We are more interested in
potential difference than the absolute potential for a single point. Notice that the potential
is a property of the electric field, whether there is a charge in the field or not.
Voltage can be generated in a circuit by a voltage source or by a current source
and resistor in parallel. Even though they exist, current sources are much less common
than voltage sources in real circuits.
Example 2.1 How much work is needed to move an electron (a charge of 1.6×10–19 C)
from 0 V to 4 V?
Rearranging Eq. (2.3)
When free electrons move in a conductor, they tend to bump into atoms. We call this
property resistance and use a resistor as the electric component implementation. The unit
of resistance is the Ohm (Ω), in honor of Georg Simon Ohm, who discovered what is
now known as Ohm’s law. Ohm’s law states that for many materials, when a potential
difference is maintained across the conductor, the ratio of the current density J (current
4 Chapter No. Title
per unit area) to the electric field that is producing the current, E, is a constant, σ, that is
independent of the electric field. This constant is called the conductivity of the conductor.
J = σE (2.4)
It is useful to consider Ohm’s law in the case of a straight wire with cross-sectional area a
and length l, as in Figure 2.2. V = Vb – Va is the potential difference maintained across the
wire, which creates an electric field and current in the wire. The potential difference can
be related to the electric field by the relationship
V = El (2.5)
Vb Va
E
Figure 2.2 A model of a straight wire of length l and cross-sectional area a. A potential difference
of Vb – Va is maintained across the conductor, setting up an electric field E. This electric field
produces a current that is proportional to the potential difference.
Since the magnitude of the current density is related as J = σE, and current
density J is defined as I/a, the potential difference can be rewritten as
V = (l σ )J = (l σa )I (2.6)
The quantity (l/σa) is defined as the resistance, R, of the conductor. The relationship can
also be defined as a ratio of the potential difference to the current in the conductor
R = (l σ a ) = V I (2.7)
l
R=ρ (2.8)
a
where ρ, the reciprocal of conductivity σ, is the resistivity (Ω⋅m), l is the length of the
conductor in meters (m), and a is the cross sectional area of the conductor (m2).
Ohm’s law is written as
V = IR (2.9)
where V is the voltage across the resistor and I is the current through the resistor.
Section No. Title 5
Example 2.2 What is the resistance of a 12.0 cm long piece of copper wire, if the wire
has a resistivity of 1.7×10–8 Ω⋅m and a radius of 0.321 mm?
The cross-sectional area of the wire is
(
a = πr 2 = π 0.321 × 10 − 3 m )2 = 3.24 ×10−7 m2
The resistance can be found using Eq. (2.8)
l 0.12 m
R=ρ = 1.7 × 10 − 8 Ω m −7 m 2
= 6.30 × 10 − 3 Ω
a 3 . 24 × 10
Most electric circuits make use of resistors to control the currents within the
circuit. Resistors often have their resistance value in ohms color coded. This code serves
as a quick guide to constructing electric circuits with the proper resistance. The first two
colors give the first two digits in the resistance value. The third value, called the
multiplier, represents the power of ten that is multiplied by the first two digits. The last
color is the tolerance value of the resistor, which is usually 5%, 10% or 20% (Figure 2.3).
In equation form, the value of a resistor’s resistance can be calculated using the following
AB × 10 C ± D% (2.10)
where A is the first color representing the tens digit, B is the second color representing
the ones digit, C is the third color or multiplier and D is the fourth color or tolerance.
Table 2.1 gives the color code for each of these four categories, and their corresponding
digit, multiplier, or tolerance value.
6 Chapter No. Title
Figure 2.3 The colored bands that are found on a resistor can be used to determine its resistance.
The first and second bands of the resistor give the first two digits of the resistance, and the third
band is the multiplier, which represents the power of ten of the resistance value. The final band
indicates what tolerance value (in %) the resistor possesses. The resistance value written in
equation form is AB×10C ± D%.
Table 2.1 The color code for resistors. Each color can indicate a first or second digit, a multiplier,
or, in a few cases, a tolerance value.
Example 2.3 What is the resistance value of a resistor where the first band is orange,
second band is blue, third band is yellow, and fourth band is gold?
Using Table 2.1, the first band, orange, represents the first digit, 3. The second
band gives the second digit, 6. The third band, is yellow, 4. The fourth band, or tolerance,
is gold, 5%. Therefore, the resistance value is AB×10C ± D% or 36×104 = 360 kΩ, with a
tolerance value of 5% or 18 kΩ.
P = VI (2.11)
The unit of power is the watt (W). V is the voltage between the two terminals of the
resistor and I is the current through the resistor.
Using Eq. (2.7) we can rewrite Eq. (2.11) as
Section No. Title 7
V2
P= = I 2R (2.12)
R
Although power is most often used to describe the rate at which work can be done by
mechanical objects, in electronics it represents the amount of energy dissipated by a
component. The power lost as heat in a conductor, for example, is called joule heat and is
referred to as a loss of I2R.
2.4 Assume the chest is a cylinder 10 cm in diameter and 40 cm long with a resistivity of
0.8 Ω⋅m. For a voltage of 2 kV during defibrillation, calculate the current and power
dissipated by the chest.
Calculate the current and power by using Eqs. (2.8), (2.9) and (2.11).
l 0.4
R=ρ = 0.8 = 40.8 Ω
a π
× (0.1) 2
4
V 2000
I= = = 49 A
R 40.8
P = VI = 2000 × 49 = 98 kW
An electric system generally contains many components such as voltage and current
sources, resistors, capacitors, inductors, transistors, and others. Some IC chips, such as
the Intel Pentium, contain more than one million components.
Electric systems composed of components are called networks, or circuits. We
can analyze the performance of simple or complex circuits using one theorem and two
laws of linear circuit analysis.
Superposition theorem: The current in an element of a linear network is the sum
of the currents produced by each energy source.
In other words, if there is more than one source in the network, we can perform
circuit analysis with one source at a time and then sum the results. The superposition
theorem is useful because it simplifies the analysis of complex circuits.
One concept we need to introduce is the polarity of a voltage drop.
Conventionally, if the direction of current flow in a circuit is known or assumed, then the
direction of current flow across a circuit element is from + to –. In other words, by
current flowing through an element, a voltage drop is created (as shown by Ohm’s law)
by that element with the polarity of + to – in the direction of current flow (Figure 2.4(a))
8 Chapter No. Title
R1 R2
+ 10 Ω − + 20 Ω −
I R V1 V2
+ V −
I VS = 30 V
(a) (b)
Figure 2.4 (a) The voltage drop created by an element has the polarity of + to – in the direction of
current flow. (b) Kirchhoff’s voltage law.
∑V = 0 (2.13)
or the sum of voltage drops around a loop is zero. In other words, starting at any point on
the closed loop, we can use the voltage law to form an equation by adding each voltage
drop across every element (resistors and sources). We can assume the direction of current,
but if our calculations give a negative result then the actual current flows in the opposite
direction. This law follows from conservation of energy. A charge that moves in a closed
loop must gain as much energy as it loses if a potential can be found at each point in the
circuit.
Figure 2.4(b) shows a circuit to which we can apply Kirchhoff’s voltage law. If
we sum the voltages counterclockwise around the loop we get
−VS + V1 + V2 = 0
VS = V1 + V2
Notice that the voltage source, VS, has a polarity opposite that of the other voltage drops.
A resistor that is traversed in the direction of the current gives a drop in potential
across the resistor equal to –IR, while a resistor that is traversed in the direction opposite
of a current gives an increase in potential across the resistor equal to +IR. This is directly
related to Eq. (2.9), where the direction of current flow across a resistor affects the
potential change across that resistor. Also a voltage source that is traversed from the +
to – terminal gives a drop in potential of –V, while a voltage source that is traversed from
the – to + terminal gives a increase in potential of +V.
In circuit analysis, a node is a junction of two or more branches. Kirchhoff’s
current law (KCL) states that at any node
Section No. Title 9
∑I = 0 (2.14)
or the sum of currents entering or leaving any node is zero, this follows the law of
conservation of charge. In other words, the currents entering a node must equal the
currents leaving the node (Figure 2.5)
I1 3A 6A
I2
I3 I=?
(a) (b)
Figure 2.5 (a) Kirchhoff’s current law states that the sum of the currents entering a node is 0. (b)
Two currents entering and one “negative entering”, or leaving.
When we use Kirchhoff’s current law, we can arbitrarily label the current
entering the node + and leaving the node –. The sum of the currents entering the node in
Figure 2.5(b) is
+3 + 6 − I = 0
I =9A
Kirchhoff’s voltage law and current law are basic laws for circuit analysis.
There are two types of circuit analysis, each based on each of Kirchhoff’s laws.
We assume unknown currents in the loops and set up the equations using
Kirchhoff’s voltage law and then solve these equations simultaneously.
We can use Figure 2.4(b) to illustrate loop analysis. We assume a current flows
through the circuit in the direction already indicated in the figure. If we assumed the
current in the other direction, our result would just be negative. Recalling Ohm’s law, we
have for the sum of the voltages through the clockwise loop
VS = R1I + R2 I
30 = 10 I + 20 I
I = 1A
Now that we know the current through the loop, we can find the voltage drop through
either resistor using Ohm’s law. In this example, the voltage drop for R1 is 10 V.
10 Chapter No. Title
4W
a b e
-
10 V
+ +
2W 14 V
I1 6W -
d c f
I3 I2
Example 2.5 Find the currents I1, I2 and I3 in the circuit shown in Figure 2.6.
Applying Kirchhoff’s current law to the current junction at node c, we have
I1 + I 2 = I3 (1)
All together there are three loops, abcda, befcb, and aefda (the outer loop) to which
Kirchhoff’s voltage law can be applied. Applying these Kirchhoff’s voltage law to the
two inside loops in the clockwise direction, we have
Expressions (1), (2) and (3) represent three equations with three unknowns, and we can
solve this system of three equations, first by substituting (1) into (2), giving
10 = 8I1 + 2 I 2 (4)
22 = 11I1
I1 = 2 A
Section No. Title 11
Substituting this value into (5), we find I2 = –3 A., and substituting both these values in
(1), we find I3 = –1 A.
+ 20 W - + 20 W -
A C B
V
+ +
+
30 V 20 W 20 V
- -
-
30 − V 20 − V V − 0 50 − 3V
+ − = =0
20 20 20 20
V = 16.7 V
When solving for voltages or currents in a circuit using the loop method or the
nodal method, it is important to note that the number of independent equations needed is
equal to the number of unknowns in the circuit problem. With an equal number of
equations and unknowns, large numbers of unknowns can be solved for using a matrix
and linear algebra.
2.1.5 Attenuators
When we amplify the 1 mV electrocardiogram signal from the heart, the signal may be
undesirably reduced (attenuated) by the input resistance of the amplifier. For example,
Figure 2.8 shows that the resistance of the skin, Rs, may be 100 kΩ and the input
resistance, Ri (input resistance of the oscilloscope used as an amplifier), is 1 MΩ.
12 Chapter No. Title
Rs
1 mV
100 kΩ Vo
Electrocardiogram 1 MΩ Ri
Figure 2.8 The 1 mV signal from the electrocardiogram is attenuated by the resistive divider
formed by the 100 kΩ skin resistance and the 1 MΩ input resistance of the oscilloscope.
V 1 mV
i= = = 0.91 nA
R 100 kΩ + 1 MΩ
Thus the input voltage (ECG) has been undesirably attenuated by an amount that we can
calculate from the voltage divider equation
Vo Ri
= (2.15)
Vi Rs + Ri
Vo Ri 1 MΩ
= = = 0.91
Vi Rs + Ri 100 kΩ + 1 MΩ
Figure 2.8 is a typical example of voltage divider. With two resistors in series,
the output voltage is part of the input voltage. With adjusting the values of Ri and Rs, one
can flexibly obtain any percentage of the input voltage.
A potentiometer is a three-terminal resistor with an adjustable sliding contact
that functions as an adjustable voltage divider or attenuator. Figure 2.9 shows that if the
slider is at the top, vo = vi. If the slider is at the bottom, vo = 0. If the slider is in the
middle, vo = 0.5vo. Potentiometers are usually circular with a rotatable shaft that can be
turned by hand or a screwdriver. The potentiometer is useful to provide a variable gain
for an amplifier or for the volume control on a radio. Alternatively, a potentiometer can
be used as a two-terminal variable resistor by using the variable resistance between the
slider and only one end.
Section No. Title 13
vi
Slider
vo
Figure 2.9 A potentiometer is a three-terminal resistor with an adjustable sliding contact shown by
the arrow. The input signal vi is attenuated by the potentiometer to yield an adjustable smaller
voltage vo.
Galvanometer
The galvanometer is a main component that is used in creating ammeters (devices that
measure current in a circuit) and voltmeters (devices that measure potential difference in
a circuit), as the galvanometer can be used in conjunction with other circuit elements to
measure current or potential differences of an electronic signal.
A common type of galvanometer consists of a coil of wire mounted in such a
way that it is free to deflect or rotate in the presence of a magnetic field provided by a
permanent magnet. The main principal behind the galvanometer makes use of a torque
that acts on the loop of current when a magnetic field is applied. The torque is
proportional to the amount of current that passes through the galvanometer, such that the
larger the current, the greater amount of deflection or rotation of the coiled wire.
Typically, an off-the-shelf galvanometer is not directly used as an ammeter,
since it has a large resistance (about 60 Ω) that would considerably reduce the amount of
current in the circuit in which the galvanometer is placed. Also, the fact that the
galvanometer gives a full-scale deflection for low currents (1 mA or less) makes it
unusable for currents greater in magnitude. If a galvanometer is placed in parallel with a
shunt resistor, Rp, with a relatively small resistance value compared with that of the
galvanometer, the device can be used effectively as an ammeter. Most of the current that
is measured will pass through this resistor (Figure 2.10(a)).
If an external resistor, Rs, is placed in series with the galvanometer, such that its
resistance value is relatively larger than that of the galvanometer, the device can be used
as a voltmeter. This ensures that the potential drop across the galvanometer doesn’t alter
the voltage in the circuit to which it is connected (Figure 2.10(b)).
Older ammeters and voltmeters that used galvanometers with moving coils have
been largely replaced by digital multimeters with no moving parts.
14 Chapter No. Title
Galvanometer
Galvanometer
Rp Rs
(a) (b)
Figure 2.10 (a) When a shunt resistor, Rp, is placed in parallel with a galvanometer, the device can
be used as an ammeter. (b) When a resistor, Rs, is connected in series with the galvanometer, it can
be used as a voltmeter.
Wheatstone bridge
I1 I2
+
R1 R1
−
a b
R3 Rx
Figure 2.11 A circuit diagram for a Wheatstone bridge. The circuit is often used to measure an
unknown resistance Rx, when the three other resistances are known. When the bridge is balanced,
no current passes from node a to node b.
I1R1 = I 2 R2 (1)
I1R3 = I 2 Rx (2)
Dividing (2) by (1), we can eliminate the current, and solve for Rx
R2 R3
Rx = (2.16)
R1
From Eq. (2.16), we can calculate the value of the unknown resistor.
The fact that a Wheatstone bridge is valuable for measuring small changes of a
resistance makes it also suitable to measure the resistance change in a device called a
strain gage. This device, commonly used in biomedical instruments to measure
experimental stresses, often consists of a thin wire matrix attached to a flexible plastic
backing and glued to the stretched metal. Stresses are measured by detecting changes in
resistance of the strain gage as it bends. The resistance measurement is made with the
strain gage as one or more elements in the Wheatstone bridge. Strain gages diffused into
16 Chapter No. Title
the diaphragm of an silicon integrated circuit blood pressure sensor are formed into a
Wheatstone bridge.
2.1.7 Capacitors
From Ohm’s law we know the voltage across a resistor is related to the current through
the resistor by the resistance value. Now we investigate two elements with different
relationships between voltage and current.
Conventionally, we use capital letters to denote variables that do not change
with time and small letters to denote variables that change with time.
A capacitor is a two terminal element in which the current, i, is proportional to
the change of voltage with respect to time, dv/dt, across the element (Figure 2.12), or
dv
i=C (2.19)
dt
where C is capacitance and is measured in farads (F). One farad is quite large so in
practice we often see µF (10–6 F), nF (10–9 F), and pF (10–12 F). Common capacitor
values are 10 pF to 1 µF. Capacitors are commonly used in a variety of electric circuits
and are a main component of electronic filtering systems (introduced in 2.3) in
biomedical instruments.
i
i
C +
1 v C
dv/dt −
(a) (b)
Figure 2.12 (a) Capacitor current changes as the derivative of the voltage. (b) Symbol of the
capacitor.
1 t
v(t ) = v(t0 ) + ∫ idt (2.20)
Ct
0
Section No. Title 17
where t0 and t are the beginning and ending times over which we observe the current, v(t0)
is the initial voltage across the capacitor. We usually choose a t0 of zero.
A capacitor usually consists of two conductors that are separated by an insulator.
The capacitance of a device depends on its geometry and the dielectric material. A
parallel-plate capacitor consists of two parallel plates of equal area A, separated by a
distance d (Figure 2.13). One plate has a charge +Q and the other –Q, where the charge
per unit area on either plate is σ = Q/A. If the charged plates are very close together,
compared to their length and width, we can assume that the electric field is uniform
between the plates and zero everywhere else.
+Q
−Q
d Area = A
Figure 2.13 Diagram of a parallel plate capacitor. The component consists of two parallel plates of
area A separated by a distance d. When charged, the plates carry equal charges of opposite sign.
We can then see that the electric field between the plates is
E = σ εo = Q εo A (2.21)
where ε0 is the permittivity of free space (8.85 pF/m)). Since the potential difference
between the plates is equal to Ed, we find that the capacitance is
C = ε0 A d (2.22)
−1
1 1 CC
Ce = + = 1 2 2 capacitors
C1 + C2
C1 C 2 (2.23)
1 1 1 1
= + + + mm multiple capacitors
Ce C1 C2 C3
For the parallel capacitances in Figure 2.14(b), the equivalent capacitance, Ce, can be
calculated by
Ce = C1 + C2 2 capacitors
(2.24)
Ce = C1 + C2 + C3 + mm multiple capacitors
C1
C1 C2
C2
(a) (b)
Figure 2.14 (a) A series combination of two capacitors. (b) A parallel combination of two
capacitors
2.1.8 Inductors
di
v=L (2.25)
dt
where L is the inductance. Its unit is the henry (H). Like the farad, this unit is large so we
use more convenient units such as mH (10–3 H) and µH (10–6 H). This is usually taken to
be the defining equation for the inductance of any inductor, regardless of its shape, size
or material characteristics. Just as resistance is a measure of opposition to a current,
inductance is a measure of the opposition to any change in current. The inductance of an
element often depends on its geometry.
Section No. Title 19
V
i
L +
1 v L
di/dt −
(a) (b)
Figure 2.15 (a) Inductor voltage changes as the derivative of the current. (b) Symbol of the
inductor.
? 0e? t i L=2H
1 t
i (t ) = i (t0 ) + ∫ vdt
Lt
0
1 t 1 − 10 −5t
i (t ) = 0 + ∫ − 10e−5t V dt = e
2H0 2 −5
i (t ) = e −5t A
20 Chapter No. Title
For multiple inductors in series, the equivalent inductance Le, can be calculated
by
Le = L1 + L2 + L3 + mm (2.27)
1 1 1 1
= + + + ll (2.28)
Le L1 L2 L3
Consider the simple resistor and capacitor (RC) circuit shown in Figure 2.17(a).
Suppose that at time t = 0 the voltage across the capacitor is vC(0). For vC(t)
when t ≥ 0, we have
VC(t)
1 VC(0)
+
R C
− −
iR iC
τ t
(a) (b)
Figure 2.17 (a) Simple RC circuit with v(0) on capacitor at time t = 0. (b) Normalized voltage
across the capacitor for t ≥ 0 (normalized means the largest value is 1).
vC dv
vR = vC and iR + iC = 0 ⇒ +C C =0 (2.29)
R dt
dvC 1
+ v =0 (2.30)
dt RC C
This equation contains the variable vC and its first derivative dvC dt . This is a
first-order differential equation. Many complex systems can be reduced to multiple first-
order systems, which are more easily to be solved and the high order systems.
Section No. Title 21
t
−
vC (t ) = vC (0)e RC (2.31)
t
vC (t ) v (0) −
iC (t ) = −iR (t ) = =− C e RC (2.32)
R R
From these results we can see that voltage and current in the circuit decrease as
time increases and approach zero as time goes to infinity. The voltage and current in this
circuit can be described by equations of the form
t
−
f (t ) = Ke τ (2.33)
where K and τ are constants. The constant τ is called the time constant of the first-order
system. In this case, τ = RC. When time t reaches τ, the amplitude of vC (t = τ) is vC(0)E–
1. When t = 3τ, the value of vC(t) is down to only 5% of the value of vC(0). We can
assume the value of vC(t) is practically zero after five time constants.
In the circuit shown in Figure 2.17(a), there was no independent source (or
input), only an initial state of the circuit. For this reason, we call the voltage and current
we obtained the natural response.
Example 2.7 Consider the circuit in Figure 2.17(a). Suppose R = 1 kΩ, C = 40 µF and
the initial voltage across the capacitor at t = 0 is 10 V. How long will it take for the
capacitor to discharge to half of its original voltage?
This simple RC circuit can be solved with a first-order differential equation.
vC(0) = 10 V, so half of that voltage would give us vC(t) = 5 V. Using Eq (2.31) and
algebra we can solve for t.
t
−
vC (t ) = vC (0)e RC
v (t ) t
ln C = −
vC (0) RC
vC (t )
t = − RC ln = −(1 kΩ )(40 µF)ln (0.5) = 27.7 ms
vC (0)
Now let us consider the case of nonzero input using the circuit shown in Figure
2.18(a). From t = –∞ to just before t = 0 (t = 0–), the switch has been open. Therefore, no
22 Chapter No. Title
current has been flowing and no charge is in the capacitor. At t = 0, we close the switch
which causes current to flow and charge to accumulate in the capacitor.
+ vR −
I
+ R +
V vC
C
− −
VC(t)
V 1
VC(t)
τ t
(a) (b)
Figure 2.18 (a) Series RC circuit with voltage step input at time 0. (b) Normalized voltage across
the capacitor.
dvC vC
V =C + for t ≥ 0
dt R
or (2.34)
dvC 1 V
+ v = for t ≥ 0
dt RC C RC
t
−
vC = V − Ve RC (2.35)
t
−
vR = Ve RC
t (2.36)
V −
i= e RC
R
The voltage vC(t) is shown in Figure 2.18(b). The voltage across the resistor
decreases exponentially and the voltage across the capacitor increases exponentially.
From the definition of a capacitor, we know the voltage across it cannot change abruptly
or the current will become infinite. When the switch is closed, the voltage on the
capacitor is still zero and the voltage drop across the resistor is V, the source voltage. As
time increases, the capacitor becomes charged and as time approaches infinity, all the
source voltage is across the capacitor while the voltage drop across the resistor is zero.
Example 2.8 Consider the circuit in Figure 2.18(a). Suppose that V = 12 V, R = 1 kΩ and
C = 10 µF. At time t = 0 the switch closes. Sketch the voltage across the capacitor versus
time. How long will it take for the capacitor to become 90% of the voltage source?
Using Eq (2.35):
t
−
vC = V − Ve RC
t
−
vC = 12 − 12e 0.01 V
90% of the voltage source is 10.8 V. Substituting this value in for vC in the
above equation and solving for t yields a time of 23 ms with a time constant of 10 ms.
12
10.8
vC (V)
0.023 0.05
Time (s)
We can also use a resistor and an inductor to make a first-order circuit where τ =
L/R.
2.1.10 Frequency
Sinusoidal waves are widely used in electrical circuits. Figure 2.20 shows two sinusoidal
waveform, which can be represented as
2π
A sin( 2πft + θ ) = A sin(ωt + θ ) = A sin( t +θ) (2.37)
T
sine
A
Amplitude
0 t
T
−Α cosine
Note the sine function lags 90º in phase behind the cosine function.
Example 2.9 Figures 2.21 and 2.22 show a comparison of a sinusoidal waveform with
different frequency and different phase angle.
In Figure 2.21 the solid line shows the function y = sin(ωt), the dashed line
shows the function with the frequency doubled, y´ = sin(2ωt). By doubling the frequency
the period is reduced by half by T = 1/f.
Section No. Title 25
0.5
v (V)
0
–0.5
–1
0 1 2 3 4 5 6
t (s)
In Figure 2.22 the solid line again shows the function y = sin(ωt), the dashed
line shows the function with a phase shift of 180° (π in radians), y´ = sin(ωt − π). The
minus sign in front of the phase angle shifts this waveform to the right and can be said
that it is leading the original waveform. Also with a phase shift of 180°, the two
waveforms are said to be completely out of phase.
0.5
v (V)
–0.5
–1
0 1 2 3 4 5 6
t (s)
26 Chapter No. Title
Figure 2.22 Sinusoidal waveforms with 0º phase angle (solid) and 180º phase angle (dashed).
This representation comes from Euler's identity and has the same amplitude, frequency,
and phase angle as the noncomplex form. Let us denote the complex number
A = Ae jφ (2.39)
by
A = A∠φ (2.40)
which is known as a phasor, with bold type meaning it is a vector. Equation (2.40) can be
read as, "The vector A has a magnitude of A and an angle of φ."
Let us consider a resistor. If the current and voltage of the resistor are all
complex sinusoids, say
i = Ie j (ω t +θ ) (2.41)
v = Ve j (ω t +φ )
v = Ri
Ve j (ω t +θ ) = RIe j (ω t +φ ) (2.42)
Ve jθ = RIe jφ or V∠θ = RI∠φ
Eq. (2.42) shows that θ = φ, which means the voltage and the current are in
phase across the resistor (Figure 2.23(a)). We can re-write this as the phasor
V = RI (2.43)
di
v=L (2.44)
dt
Therefore
V = jωLI (2.46)
and θ = φ + 90°, which shows that current lags the voltage by 90° (Figure 2.23(b)).
Using the methods presented in Eq. (2.44) through (2.46) yields
1
V= I (2.47)
jωC
and θ = φ – 90°, which shows that current leads the voltage by 90° (Figure 2.23(c)).
t t t
iL
(a) (b) (c)
Figure 2.23 Plots of the current and voltage as a function of time. (a) With a resistor both the
current and the voltage vary as sin(ωt), the current is in phase with the voltage, meaning that when
the current is at a maximum, the voltage is also. (b) For an inductor, the current lags behind the
voltage 90°. (c) For a capacitor, the current leads the voltage by 90°.
From the equations above, we can see that there is a general equation in phasor
notation for the resistor, capacitor, and inductor. That equation is
V = ZI (2.48)
1
ZR = R ZC = Z L = jωL (2.49)
jωC
With impedance form of Eq. (2.49) and the method introduced in section 2.1.4,
the relationship of voltage and current is expressed in algebraic form instead of
differential form. Phasor equations simplify the circuit analysis and can be used by
computers to solve the problem.
Note that the impedance of capacitors and inductors changes with frequency.
Therefore, the performance of the circuit will also change with frequency. This change in
circuit performance is called the frequency response of the circuit.
28 Chapter No. Title
Example 2.10 The current through a 50 mH inductor is 100∠0° mA. If ω = 1000 rad/s,
what is the voltage across the inductor?
Using the phasor equation in Eq. (2.46), we can solve for V.
V = jωLI
V = j1000(0.050 )(100∠0° mA ) = 5∠90° V
Now that the impedances for the three passive circuit elements are defined (Eq. 2.49), it
is time to study two types of configurations for them in a circuit. Figure 2.24(a) shows
two impedances in series. These two impedances, and in general infinitely many of them,
can be combined into one equivalent impedance. For the series impedances in Figure
2.24(a), the equivalent impedance, Ze, can be calculated by
Z e = Z1 + Z 2 (2.50a)
Z e = Z1 + Z 2 + Z 3 + (2.50b)
For the parallel impedances in Figure 2.24(b), the equivalent impedance, Ze, can be
calculated by
1 ZZ
Ze = = 1 2 (2.51a)
1 1 Z1 + Z2
+
Z1 Z 2
1 1 1 1
= + + + (2.51b)
Z e Z1 Z 2 Z3
A shorthand notation for indicating parallel circuit elements as in Eq. (2.51) is to use a
double vertical bar between the parallel elements, such as Ze = Z1||Z2.
Using these definitions, it is possible to reduce large complicated circuits down
to (ideally) one or two impedances and a voltage source. Figure 2.24(c) shows the
equivalent impedance of the combined series or combined parallel circuit.
Section No. Title 29
Z1
Z1 Z2
Z2 Ze
Figure 2.24 (a) Series circuit. (b) Parallel circuit. (c) Single impedance equivalent.
Because safety is paramount in considering biomedical devices in the real world setting,
great emphasis is placed on currents and voltages that exist in an apparatus and what
dangers these may pose to a patient or operator who might use the device.
Electric shock may result in fatal burns, or can cause muscles or even the heart
to malfunction. Often the degree of damage depends on the magnitude of current that is
applied, how long the current acts, and through which point on the body the current
passes. Typically, skin currents of 5 mA or less may cause a light sensation of shock, but
usually do no damage. Currents larger than 10 mA tend to cause muscles to contract,
making the person unable to let go of a live wire. If currents of 100 mA pass through the
body, even for a few seconds, respiratory muscles can become paralyzed and breathing
stops. The heart can go into ventricular fibrillation, which is fatal (macroshock). If an
electrode or catheter is within the heart, current must not exceed 10 µA because the
current density is high (microshock). Contact with live wires or voltages above 24 V is
not recommended.
Often in biomedical engineering, currents and voltage outputs must be amplified,
attenuated, or filtered safely in order to achieve the best possible electronic signal with
minimal harm to the patient or operator. In the following sections, we will explore ways
in which this can be done.
2.2 Amplifiers
Most bioelectric signals have a very small magnitude (on the order of millivolts or
microvolts) and therefore require amplification so that users can process them. This
section will emphasize the operational amplifier (op amp) for use in amplifier design.
30 Chapter No. Title
RS
v1 vo V+
v1 −
A vo
+
Rd A(v2 − v1) v2 +
−
V−
v2
(a) (b)
Figure 2.25 (a) Equivalent circuit for op amp. (b) Symbol of op amp. Many times V+ and V– are
omitted in the op amp symbol, but it is understood that they are present.
When designing a circuit, we assume that we are using an ideal op amp. Then we
investigate whether or not the nonideal characteristics affect the performance. If they do,
we revise the initial design to accommodate the nonideal characteristics.
For an ideal op amp, we make the following assumptions
1. A = ∞ (gain, or amplification of the input, is infinity)
2. vo = 0 when v1 = v2 (no offset voltage)
3. Rd = ∞ (differential input impedance is infinity)
4. Rs = 0 (output impedance is zero)
5. Bandwidth = ∞. (Bandwidth will be discussed in greater detail in sections
2.27 and 2.3.)
Two basic rules should be stated before we discuss designing circuits with op amps.
Section No. Title 31
2.2.3 Inverter
Figure 2.26 shows an inverting amplifier, or inverter. It is the most widely used circuit in
instrumentation. Note that Rf is connected from the output to the negative input terminal
and thus provides negative feedback. The negative feedback increases the bandwidth and
lowers the output impedance of the op amp circuit.
i Rf
i
vi −
Ri A vo
+
In Figure 2.26, the positive input terminal of the op amp is connected to ground
and is always at 0 V. By rule 1, the negative input of the op amp is also at 0 V and
remains at 0 V no matter what happens to the other components of the circuit. This
clamping of the negative input to 0 V is called a virtual ground.
By rule 2, we know no current flows into the op amp. Therefore, the current
flows through Ri and then through Rf to the output terminal, then through the output
circuit of the op amp to ground.
Using KCL at the “–” terminal
vi − 0 0 − vo Rf v R
iin = iout ⇒ = ⇒ vo = − vi ⇒ o = − f (2.52)
Ri Rf Ri vi Ri
32 Chapter No. Title
The output vo has the opposite sign of the input vi and is amplified by an amount Rf Ri .
When using op amps in circuits it is important to consider the output impedance
of any source that they are connected to, as this value will affect the total input resistance
and the gain of the op amp.
Rf
Rs vi
−
Ri A vo
vs= 1 V +
Figure 2.27 Inverter circuit attached to a generator that contains an internal resistance.
vo Rf
=− = −3.48
vs Ri + Rs
Figure 2.28(a) shows the circuit for a noninverting amplifier. From rule 1, we know the
voltage of the negative input always equals the voltage of the positive input terminal,
which in this case is vi. From Figure 2.28(a) we have
vi − 0 vo − vi
=
Ri Rf
(2.53)
vo Rf + Ri R + Ri R
= or vo = f vi = vi 1 + f
vi Ri Ri Ri
Section No. Title 33
i i Rf
Ri − −
A vo vo
vi + vi +
(a) (b)
Figure 2.28 (a) A noninverting amplifier has a gain of (Rf + Ri)/Ri. (b) A follower, or buffer, with
unity gain.
Notice the output vo has the same sign as the input voltage vi. The amplifier gain
is determined by the ratio (Rf + Ri)/Ri.
If we choose Ri = ∞ (open circuit), the circuit can be simplified as Figure
2.28(b). The gain, (Rf + Ri)/Ri, would be unity. Therefore vo = vi, or the output voltage
follows the input voltage. At first glance, this circuit does nothing, since the output equals
the input. However, this circuit is very useful as a buffer, to prevent a high source
resistance from being loaded by a low-resistance load.
Loading can be expressed by the voltage divider equation (Eq. 2.15). In Figure
2.8, if Ri is a significant fraction of Rs, we would get an attenuated signal for Vo. We say
that the source resistance, Rs, is being loaded by Ri. This can be seen mathematically in
Eq. 2.15as Rs increases to infinity, Vo/Vi approaches zero. In general, we like to have
Ri >> Rs, which is what the buffer circuit does. It provides an extremely high input
impedance and an extremely low output impedance.
Example 2.12 Suppose the signal generator in Figure 2.27 were connected in a
noniverting configuration (Figure 2.29), will the generator’s output impedance affect the
gain?
No, since no current will flow into the positive input terminal, vi = 1 V and the
gain would remain (Rf + Ri)/Ri. If Rf = 3 kΩ and Ri = 1 kΩ, the gain would be 4,
independent of the value of Rs.
34 Chapter No. Title
Ri Rf
−
Rs A vo
+
vi
vs= 1 V
Figure 2.29 The gain of the noninverting amplifier is not affected by the addition of the impedance
Rs due to the generator.
Figure 2.30 is a simple one-op-amp differential amplifier. Current flows from v2 through
R1 and R2 to the ground and no current flows into the op amp. Thus, R1 and R2 act as a
voltage divider.
R1 R2
v1
v3 −
R1 A vo
v3
v2 +
R2
Figure 2.30 A differential amplifier uses two active inputs and a common connection.
The voltage that appears at the positive input terminal determines the voltage at
the negative terminal. The top part of the amplifier works like an inverter.
The voltage at the positive input terminal is
R2
v3 = v (2.54)
R1 + R2 2
Section No. Title 35
v1 − v3 v3 − vo
i= = (2.55)
R1 R2
R2
vo = (v − v ) (2.56)
R1 2 1
This is the equation for a differential amplifier. If the two inputs are hooked
together and driven by a common source, then the common-mode voltage is equal to Vc
(v1 = v2). The differential amplifier circuit gives an output of 0 and the differential
amplifier common-mode gain (Gc) is 0. When v1 ≠ v2, the differential voltage (v2 – v1)
produces a differential gain (Gd), which equals R2/R1.
An important application of differential amplifier is to remove interference from
the input signal. Take ECG for example, the wires from patient to ECG machine are
about 1 ~ 2 m in length and are exposed to different interference such as power line
interference. As there wires have the almost same amount of noise since they are exposed
to the same common noise sources, this common noise is rejected by the differential
amplifier according to Eq. (2.56). The output of the amplifier solely depends on the input
from the patient.
In some instances, the signal on both input lines (such ECG input terminals) is
corrupted by a surrounding interference (such as power line interference) and is not
acceptable for us to process. From Eq. (2.56) we know that only different signals input
into the differential amplifier will be amplified. Therefore, the common interference on
both inputs will not be amplified.
No differential amplifier perfectly rejects the common-mode voltage. To
quantify the performance of a differential amplifier, we use the term common-mode
rejection ratio (CMRR), which is defined as
Gd
CMRR = (2.57)
Gc
- R1 R2
Rs
_ +
vd/2 -
+ A vo
_ - R1 +
vd/2 Rs
_ + +
vc R2
+
Figure 2.31 Differential amplifier attached to a differential voltage that contains varying
impedances. Adding buffers ensure that fluctuations in Rs do not affect the gain.
R4 = 1 kΩ R5 =1.5 kΩ
4V
v3 −
R1 = 1 kΩ A
v3 vo
+
8V
R3 = 1 kΩ
R2 = 500 Ω
Example 2.13 For the differential amplifier in Figure 2.32, solve for vo.
Using KCL, we can solve for vo. First, the voltage at the positive terminal of the
amplifier is part of a voltage divider.
R3 1 kΩ
v3 = 8V = 8V =6V
(R1 || R2 ) + R3 (1 kΩ || 500 Ω ) + 1 kΩ
Section No. Title 37
Then because there is no flow of current into the op amp, using Kirchhoff’s
current law, we can solve for vo.
4 V − v3 v3 − vo
i= =
R4 R5
4 V − 6 V 6 V − vo
=
1 kΩ 1.5 kΩ
vo = 9 V
2.2.6 Comparators
A comparator is a circuit that compares the input voltage with some reference voltage.
The voltage output flips from one saturation limit to the other as the negative input of the
op amp passes through 0 V. If the input is greater than the reference, the output is the
negative saturation voltage. Otherwise, the output is the positive saturation voltage. Note
that the saturation voltage is always smaller than the voltage of the power supply for the
op amp.
The simplest comparator is the op amp itself. If a reference voltage is connected
to the positive input and vi is connected to the negative input, the circuit is complete.
However the circuit in Figure 2.33 is more widely used.
vo
R1 Saturated
voltage −vref = vi
vi
−
vref A vo vi
R1 +
(a) (b)
Figure 2.33 (a) A comparator. (b) The input–output characteristic of the comparator in (a).
a T wave. The QRS complex normally has a much larger amplitude than the other two
waves. A threshold is set, which is larger than the P and T waves but smaller than the R
waves. Every time a QRS complex passes through the comparator, the comparator gives
out a pulse, which corresponds to the heart beat. These pulses are sent to a counter and
the heart rate can be calculated.
R
Threshold
Q S
Figure 2.34 Heart beat detector uses a comparator to determine when the R wave exceeds a
threshold.
An ideal op amp has unlimited bandwidth; its gain is same for low frequency and high
frequency. However, an op amp is made of several stages of circuitry to obtain very high
gain. Each of the stages has some stray capacitance or junction capacitance associated
with it. This capacitance reduces the high frequency gain of the op amp. Figure 2.35 is
the typical frequency response of an open loop op amp.
It would appear that an open-loop op amp has very poor frequency response,
since its gain is reduced when the frequency exceeds 10 Hz. Because of this, we never
use an open loop op amp to build an amplifier. Instead, we introduce large negative
feedback to control the gain of the amplifier circuit. For example, if we have an op amp
with a circuit gain of 10 (Figure 2.35), the frequency response is flat up to 100 kHz and is
reduced above that frequency.
Section No. Title 39
Circuit bandwidth
Ideal
gain
Gain
Typical open
loop gain
10 Circuit gain of 10
1
100 102 104 106
Frequency (Hz)
Figure 2.35 The typical op amp open loop gain is much larger, but less constant, than the circuit
gain. However the in circuit bandwidth is larger than the open loop bandwidth.
2.3 Filters
The changes made to a signal depend on the frequencies contained in the signal and the
design of the filter. Generally a filter is inserted between an input source and an output.
Suppose the input, vi, is a signal with a frequency of ω
40 Chapter No. Title
vi = Vm sin(ωt + φ )
(2.58)
vo = Vm T (ω ) sin(ωt + φ + θ (ω ))
where Vm is the amplitude of the input signal, |T( ω)| is the amplitude of the transfer
function, or filter, and θ(ω) is the angle of T(ω).
Thus the filter multiplies the amplitude of the input sine wave by |T( ω)| and adds
to the phase of the sine wave θ(ω). Since |T( ω)| and θ( ω) are functions of frequency, they
are named the amplitude response and phase response, respectively, of the filter.
For example, the ideal low-pass filter has a magnitude response of
1 if f < f c
T( f ) = (2.59)
0 if f ≥ f c
where f = ω/2π.
The low-pass filter passes all frequency components lower than fc and rejects all
frequency components higher than fc. Realistically, building a filter with such a sharp
transition is not possible since the transition from passband (T(f) = 1), to stopband (T(f) =
0) takes some finite number of frequencies.
Figure 2.36 shows the magnitude characteristics of four widely used filters: low-
pass, high-pass, bandpass and bandstop filter. All these filters have a passband, where the
filter passes the frequency components, and a stopband, where the filter rejects or
attenuates the frequency components.
Section No. Title 41
T(f) T(f)
PB PB
1.0 1.0
0.1 0.1
0.01 0.01
1 fc 10 100 f 1 10 fc 100 f
(a) (b)
T(f) T(f)
PB PB PB
1.0 1.0
0.1 0.1
0.01 0.01
1 f1 10 f2 100 f 1 f1 10 f2 100 f
(c) (d)
Figure 2.36 (a) Low-pass filter. (b) High-pass filter. (c) Bandpass filter. (d) Bandstop filter. (PB
denotes passband)
Circuits can be used to approximate ideal filters. The closer the actual circuit
characteristics approach the ideal response (i.e. the steeper the transition band), the better
the approximation. However, the complexity of the circuit increases as the ideal response
is more closely approximated.
Low-pass filters pass low-frequency signals and attenuate high-frequency signals. The
corner frequency, ωc, distinguishes the passband from the stopband.
The first-order RC and RL networks are approximations of a low-pass filter.
Figure 2.37 shows the RC and RL filter.
42 Chapter No. Title
R L
+
+
+
+
vi C vo vi R vo
− − − −
(a) (b)
For the RC filter in Figure 2.37(a), the frequency response is calculated using
the voltage divider principle for complex impedance
1/ jωC 1 1
T (ω ) = = = (2.60)
R + 1 / jωC 1 + jωRC 1 + j ω ωc
where
1
ωc = (2.61)
RC
R 1
T (ω ) = = (2.62)
R + jωL 1 + j ω ωc
where
R
ωc = (2.63)
L
The magnitude and phase responses of these filters are similar to those in Figure
2.36(a). At ω = ωc (f = fC), the magnitude is 1 2 times the magnitude at ω = 0. We
designate ωc the corner frequency. This frequency is also known as the half power
frequency because at ω = ωc the power delivered to the load is half of what it is at ω = 0.
A typical use of a low-pass filter in bioinstrumentation is to remove the radio
frequency (RF) interference from an ECG signal. A combination of unshielded ECG
electrode leads and long electrode-to-amplifier connecting wires causes RF signals from
other medical equipments to couple to the amplifier. The RF signal is rectified by the
transistor junctions and causes interference below 150 Hz. However, frequency
components of the ECG also lie below 150 Hz. Therefore, the ECG is always sent
Section No. Title 43
High-pass filters also be can realized with first-order RC or RL circuits (Figure 2.38).
R
C
+
+
+
vi L vo
vi R vo
− −
− −
(a) (b)
1
T (ω ) = (2.64)
1 − j ωc ω
where
1
ωc = for the RC circuit
RC (2.65)
R
ωc = for the RL circuit
L
The frequency response of the first order high-pass filter is similar to that shown
in Figure 2.36(b). The corner frequency of the high pass filter is ωc.
The ECG signal always has some dc offset and low-frequency artifacts due to
the electrodes. A high-pass filter with a corner frequency of 0.05 Hz is used to filter the
signal. A resistor of 3.2 MΩ and a capacitor of 1 µF can be used to make a high-pass
filter that passes frequencies only higher than 0.05 Hz.
More complex circuits, such as second, third, or higher order circuits, can be used to
make the changes in magnitude from passband to stopband more abrupt.
44 Chapter No. Title
Second order circuits, such as RLC circuits, can be used to realize bandpass and
bandstop filters. Or a combination of a low-pass circuit and a high-pass circuit can be
used together to achieve the same filtering effect.
For example, a device requires a bandpass filter with corner frequencies ω1 and
ω2 ( ω1 < ω2). In this case, a low-pass filter with a corner frequency of ω2 and a high-pass
filter with corner frequency of ω1 are suitable. These two filters can be cascaded so the
signal will first go through the low-pass filter and then through the high-pass filter. The
high-pass filter rejects the frequency components smaller than ω1 while the low-pass
filter rejects frequency components greater than ω2. As a result, only the frequency
components between ω1 and ω2 will pass through the cascaded filter. This combination
of a high-pass and low-pass filter is used as an approximation of a bandpass filter (Figure
2.39).
Figure 2.39 A low-pass filter and a high-pass filter are cascaded to make a bandpass filter.
Digital circuits play a very important role in today’s electronic systems. They
are employed in almost every aspect of electronics, including communications, control,
instrumentation, and computing. This widespread usage is due to the development of
inexpensive integrated circuit (IC) packages. The circuit complexity of a digital IC ranges
from a small number of transistors (~100 to 10,000) to a complete computer chip (42
million transistors for a Pentium 4 chip).
The most commonly used numbering system is the decimal system, which was devised
by Hindu mathematicians in India around 400 A.D. Although other numbering systems
have been developed, the decimal system is the most natural one and is widely accepted.
A decimal number is represented by a string of digits. Depending on its position
in the string, each digit has an associated value of an integer raised to the power of 10.
For example, the decimal number of 215, can be computed as
Conventionally, we only write the digits and infer the corresponding powers of
10 from their positions. In general, a decimal number can be written as a string of digits
as
An An −1 An − 2 ... A1 A0 (2.66)
or in position form
( An An −1 An − 2 ... A1 A0 )r (2.68)
(314)5 = 3 × 52 + 1 × 51 + 4 × 50 = 75 + 5 + 4 = (84)10
The binary number system is a base-2 system that uses the digits 0 and 1. Two
numbers are chosen because they can represent the two states of the digital system (e.g.
voltage high and low, or switch closed and open). A binary number can be converted to a
decimal number using Eq. (2.67). For example
(100101) 2 = 1 × 25 + 0 × 24 + 0 × 23 + 1 × 22 + 0 × 21 + 1 × 20
= 32 + 4 + 1 = (37)10
Binary digital impulses, alone, appear as long strings of ones and zeros, and
have no apparent meaning to a human observer. However, if a digital-to-analog converter
(DAC) is used to decode the binary digital signals, meaningful output appears. Examples
of this output might be a computerized voice, a picture, or a voltage to control
mechanical motion.
The digital representation of a continuous analog signal is discrete both in time and
amplitude. Often an analog signal must be sampled discretely with an analog-to-digital
converter (ADC). Digital information is stored in a memory device for real time or future
processing. A digital signal needs to be converted to an analog signal with a digital-to-
analog converter (DAC), since many devices need an analog signal for controlling,
monitoring or feedback. For example, modems use ADCs and DACs in converting digital
information in computers to analog signals for phone lines and back again to
communicate with Internet.
The DAC changes digital input to analog output. Figure 2.40 shows the static
behavior of a 3-bit DAC. All the combinations of the digital input word are on the
horizontal axis, while the corresponding analog outputs are on the vertical axis. The
maximum analog output is 7/8 of the reference voltage Vref. For each digital input, there
is a unique analog output. The difference between the consecutive analog outputs is
Section No. Title 47
1 1
Vref = 3 Vref
8 2
We call this value the resolution of the DAC. In general, for an n-bit DAC or ADC, the
resolution is
1
resolution = Vref (2.69)
2n
7/8 Vref
6/8 Vref
5/8 Vref
Analog output
4/8 Vref
3/8 Vref
2/8 Vref
1/8 Vref
0V
000 001 010 011 100 101 110 111
Figure 2.40 The ideal static behavior of a 3-bit DAC. For each digital input, there is a unique
analog output.
Equation (2.69) shows that if an application calls for high resolution, the DAC
needs many bits. DACs are available in 4, 8, 10, 12, 16, 20 or more bits. The more bits a
DAC has, the more complex and expensive it is.
A variety of circuit configurations are suitable for a DAC. We present the
simplest one: a voltage scaling DAC. Figure 2.41 is a simple 3-bit voltage scaling DAC.
48 Chapter No. Title
The Vref in Figure 2.41 is the reference voltage of the DAC. The analog output is derived
from this reference. The accuracy and stability of the reference should be carefully
considered in the circuit design.
The reference voltage is divided by the series resistor network to obtain the
desired discrete voltages on each node. The digital inputs are decoded by the decoder and
the corresponding switch is closed. This voltage is then sent out as the analog output.
Voltage scaling is suited to IC technology because this technology is optimized
for making many copies of the same structure. The value of R may have an error up to
50%, but as long as the difference between two resistors is still below 1%, the discrete
output retains the accuracy of 1%.
The output of the DAC is discrete and not continuous. Interpolation should be
used to make the output voltage smoother. Interpolation could be implemented by an
electronic circuit with a voltage holder, linear interpolation, or low-pass filter.
Section No. Title 49
b2 b1 b0
3-to 8 decoder
Vref
R 7 6 5 4 3 2 1 0
R
Vo
R
characteristic of a 3-bit ADC. Figure 2.42 shows that the resolution of an ADC depends
on the number of bits of the ADC. We can determine resolution using the same Eq. (2.69)
as the DAC.
111
110
101
Digital output
100
011
010
001
000
0/8 1/8 2/8 3/8 4/8 5/8 6/8 7/8 Vref
Clock
Digital
control
logic
Vin +
_
Digital
output
DAC converter
Figure 2.44 shows the possible conversion paths for a 3-bit converter. The total
number of clock cycles required to convert an n-bit word is n.
Vout
111
111
110
110
101
101
100
1/2 Vref 100 011
011
010
010
001
001
000
0 1 2 3 4
Conversion cycle
Figure 2.44 The possible conversion paths of a 3-bit successive approximation ADC.
Analog signals are continuous in time and amplitude. Digital signals are discrete in time
and amplitude. They are represented by a sequence of numbers, x, in which the nth
number in the sequence is denoted x[n], where n is an integer. Such sequences are usually
obtained by the periodic sampling of an analog signal xa(t). In the sampled sequence, the
value of the nth sample in the sequence is equal to the value of xa(t) (rounded to the
precision of the ADC) at time nT, or
where T is the sampling period and its reciprocal is the sampling frequency. As an
example, Figure 2.45(a) shows a segment of a sine wave and Figure 2.45(b) is the
sampled sequence of the sine wave.
Section No. Title 53
Amplitude
0
−1 Time
0 1 2 3 4 5 6
(a)
1
Amplitude
Sample
−1
0 5 10 15 20 25 30 numbers
(b)
Figure 2.45 (a) Continuous signal. (b) Sampled sequence of the signal in (a) with a sampling
period of 0.2 s.
To convert a continuous time signal into a discrete time signal and process it with a
microprocessor, we are assuming that we can represent a continuous time signal by its
instantaneous amplitude values taken at periodic points in time. We are also assuming
that we will be able to reconstruct the original signal perfectly with just these sampled
points.
The sampling theorem, developed by Shannon, guarantees that the original
signal can be reconstructed from its samples without any loss of information. It states that,
for a continuous band-limited signal that contains no frequency components higher than
fc, the original signal can be completely recovered without distortion if it is sampled at a
frequency of at least 2fc. A sampling frequency fs, which is twice the highest frequency in
a signal, is called the Nyquist frequency.
Suppose the original signal can be represented by function Xc(f) in the frequency
domain and its highest frequency component is fc. When it is sampled periodically at
frequency fs, then the sampled signal can be represented as (Oppenheim and Schafer,
1989)
+∞
X s ( f ) = fs ∑ X c ( f − kfs ) (2.71)
k = −∞
54 Chapter No. Title
where Xs(f) is the frequency representation of the sampled sequence. Figure 2.45 is a
sketch of the sampling process. We know the sampled signal consists of periodically
repeating copies of Xc(f). These copies are shifted by integer multiples of the sampling
frequency and then superimposed to produce the periodic samples. Figure 2.46 shows
that if we use a low-pass filter in which fn is the corner frequency, we can reconstruct the
original by passing the sampled signal through the low-pass filter (fn is between fc and
fs – fc).
In Figure 2.46, if the frequency fs – fc is less than fc, the copies of the original
signal overlap for some of the frequency components. This overlap will corrupt the
original signal and we will not be able to reconstruct the original signal from the sampled
signal. To guarantee that we can reconstruct the original signal, we must meet the
condition of the sampling theorem
Here we present an example of a moving window digital filter for an ECG signal.
In some ECG signals, there is the higher frequency electromyogram (EMG)
which interferes with the ECG signal. We can use a low-pass digital filter to attenuate the
EMG using the following equation
1
y[n] = ( x[n] + 2 x[n − 1] + x[n − 2]) (2.73)
4
We can also change the coefficients of the digital filter to get different results or
to use it for other applications. This is more flexible than the analog filter, which would
require a change in components or even construction of a new circuit.
Section No. Title 55
Magnitude
Original
fC Frequency
(a)
Magnitude
Sampling
function
fS 2fS Frequency
(b)
Magnitude
Sampled
signal
Low-pass
filter
fn Frequency
(d)
Magnitude
Reconstructed
signal
fC Frequency
(e)
Figure 2.46 (a) Spectrum of original signal. (b) Spectrum of sampling function. (c) Spectrum of
sampled signal. (d) Low-pass filter for reconstruction. (e) Reconstructed signal, which is the same
as the original signal.
2.6 Microcomputers
The computer is one of the most important inventions of the twentieth century. In the last
twenty years, computer technology, especially microcomputer technology, has had a
significant impact on bioinstrumentation.
Computers are divided, by their size and use, into several categories:
supercomputers and mainframe computers, workstations, microcomputers (or personal
56 Chapter No. Title
A microcomputer is composed of its CPU (central processing unit), memory, I/O devices,
and buses, as shown in Figure 2.47.
Figure 2.47 General block diagram of a microcomputer system (arrows represent the data flow
direction).
The CPU plays an important role in a computer. It fetches, decodes and executes
instructions and controls the performance of other modules. The most commonly used
CPU in microcomputers is the Intel Pentium 4, which contains 42 million transistors on
one chip.
Read-only memory (ROM) can only be read by the CPU and random-access
memory (RAM) can be read and written to by the CPU. The core programs, such as start-
up instructions and interface instructions for the CPU, are generally stored in ROM.
ROM retains these programs even if there is no power. Because it holds basic programs,
a modern microcomputer contains tens of kilobytes of ROM.
The system software and applications are loaded to RAM for processing. Data
and instructions are also stored in RAM. With advances in VLSI technology, the capacity
of RAM becomes greater while the price keeps decreasing. In 2002, most computers have
more than 256 megabytes (MB) of RAM, with some having several gigabyte (GB)
memory .
The bus is a communication path used to transfer information (such as
instructions, data, addresses or control signals) between the modules that make up the
system. There may be one or more functionally separate buses linking the various
microcomputer modules.
The clock is an oscillation signal generated to synchronize the performance of
the modules in the system (which is not shown in Figure 2.47). The faster the clock speed,
the faster the computer system. The first IBM PC had an Intel 8088 with a clock speed of
4.8 MHz. In 2002, the Intel Pentium 4 has a clock speed of 2.2 GHz. Intel or other
companies will manufacture much faster chips in the future.
I/O (input/output) devices are used by the computer to input and output
information. These devices can be divided into several categories.
The most common input devices for a microcomputer are the keyboard and
mouse. A user gives the computer instructions or data through these devices. The
computer can display the result on the output devices (monitors or printers) for the user.
Section No. Title 57
Mass storage devices include floppy disk drives, hard disk drives, CD-ROM
drives and tape drives. They are used to permanently store software and data on mass
storage media.
A computer can be connected to other computers via a network. Network
devices include a modem or network card, and a network. When connected to the
network, computers can share resources (such as devices, software and databases), which
improves the performance of connected computers.
In bioinstrumentation, ADC and DAC devices are widely used to collect the
data for processing and to send out signals to control other equipment.
Problem
Application software
Editor Complier Linker
Operating System
CPU
Figure 2.48 Three levels of software separate the hardware of microcomputer from the real
problem.
Two aspects should be considered for developing a biomedical system: (1) the
choice of operating system to support the task, and (2) the choice of the programming
language to implement the application. There are many combinations of operating
systems and programming languages that are able to address the problem, but some
choices are better than others for certain applications.
The operating system (OS) is a set of procedures that handle the resources of the
computer system. The resources include the CPU, file system, display, printer and so on.
The OS manages the file system control, CPU control, memory control and system
security. It provides a virtual machine so that the user just deals with the operating system
instead of the computer hardware. The operating system also provides some frequently-
used functions so that users can work more efficiently.
An example of the type of task an OS handles is file system management. In a
computer, there are several storage media such as a floppy disk, hard disk, or tape. When
saving files, we have to consider the following: (1) the mechanisms of these devices are
different, so we should use different procedures to write our files to each one, (2) for a
device, we have to decide the format of the file and location of the file, (3) we have to
determine a method for file indexing so that we can obtain access to the file. The OS
provides an interface between the user and media. A user just asks the operating system
to save the file. The OS takes care of the details automatically. Without the OS, computer
users have to communicate with the computer hardware directly by writing some code,
which would take a lot of effort.
Below is a brief discussion of three common operating systems: Unix,
DOS/Windows, and Mac.
Unix is widely used on mainframe computers and workstations, although some
versions of Unix are available on microcomputers. It is very flexible, provides the
capability to maximally manipulate a computer system, and gives excellent control of
input/output and other facilities. It also provides for multiple simultaneous users to do
multiple simultaneous tasks (i.e. it is a multiuser, multitasking operating system).
Because of this great flexibility, Unix requires considerable expertise to use all of its
Section No. Title 59
capabilities. Different hardware companies (such as HP, IBM and others) provide their
own versions of Unix for their products. However, all versions of Unix are very similar.
The Macintosh is a computer hardware/software system by Apple. It is designed
for ease of use and for graphics-oriented applications. Macintosh is equipped with a very
easy to use graphical user interface (GUI) by sacrificing a great deal of direct user control
of the hardware system. The concept was to produce a personal computer that would be
optimal for running applications instead of writing new ones. Without training, an
individual can sit down and quickly learn to use the Macintosh because of its intuitive OS.
MS-DOS was designed by Microsoft for the IBM PC and compatibles. It
became the most popular operating system on personal computers because of the wide
use of the IBM PC. In the 90s, MS-DOS evolved to Windows system (3.1, NT, 2000,
XP), which has graphical user interface. In DOS/Windows, a user can still control the
hardware of the system, but it is not as flexible as the Unix system.
A typical need for bioinstrumentation is the ability for real-time processing. In
other words, we need the results when we collect the data. Sometimes this is critical,
especially in the intensive care unit of a hospital. In most cases, biomedical applications
require a single-user system that gives maximal performance from the computing power,
and the DOS/Windows system satisfies this requirement. The Unix system is not
desirable because of its multi-user/multitask property. The Macintosh system is not
desirable, as the user cannot control the hardware structure.
The OS takes care of talking to the hardware but the user still has to give the computer
instructions on how to implement the solution (i.e. program the computer).
The instructions for the computer are in binary format and are machine
dependent, which means that binary codes can be used on only one kind of computer. For
example, the binary code for PC machines that use DOS cannot be used on Macintosh
machines.
Since instructions for the CPU are binary, the final codes for the computer
should be in binary format. When the computer was first invented, engineers had to
remember instructions in binary code and write their own programs in binary format.
After 40 years of development, there are now many programming languages, some of
which are similar to our native languages. The codes developed in the programming
language are called source codes. A complier is used to translate the source codes to
binary codes that computers can understand. One advantage of programming languages is
ease of use. Another advantage is transportability, which means user can use the source
codes on different machines with different operating system provided that the language
compiler is available in the different operating systems.
There are numerous programming languages, but all languages can be classified
in one of two categories.
One category is low-level languages, which includes assembly language. Every
instruction in assembly language corresponds to a binary instruction code for the CPU.
Therefore, it is known as a low-level language. Assembly language can achieve the
greatest run-time performance because it provides the direct manipulation of the
60 Chapter No. Title
processor. However it is also the most difficult language to write since we have to think
about the problem in the way the processor does. Another drawback is that the program is
machine dependent. Once we want to use another system, we have to write the code
again from scratch.
The other category of programming languages is high-level languages. The
instructions are similar to the way we think of the problem and programs are relatively
easy to write. High-level languages make developing programs much easier. The tradeoff
is that the programs usually are not as efficient as those in assembly language, especially
for real-time processing.
The most widely used high-level languages are FORTRAN, Pascal and C/C++.
FORTRAN was developed for scientific calculation (FORmula TRANslation). Pascal
was designed for teaching students structural programming techniques.
The C language is the most popular language and is used to develop Unix
systems. It has the characteristics of a high-level language, but also contains functions
and interfaces for low-level programming. C++ is a modern version of the C language
and allows object-oriented structure. C/C++ is the current language of choice for real-
time programming, with some parts of the code possibly written in assembly language for
increased efficiency.
There are two other special languages worth mentioning. LabVIEW is a visual
computing language available for many operating systems. It is optimized for laboratory
applications. Programming is accomplished by interconnecting functional blocks that
represent processes such as data acquisition and analysis or virtual instruments. Thus,
unlike traditional programming by typing command statements, LabVIEW is a purely
graphical, block diagram language. In most cases, LabVIEW is used to acquire and
process data in the laboratory and in some industrial applications.
MATLAB is general purpose software for matrix manipulation. It is used for
matrix calculation and also has many packages containing various functions such as
signal processing, image processing, data analysis and so on. MATLAB is very useful for
digital signal processing, which in turn is very important for biomedical instrumentation.
2.7.3 Algorithms
An algorithm is a list of steps that solve a given problem. Such a description should be
precise enough to be used in a completely automatic manner (i.e. by a computer). It is the
programmer’s responsibility to ensure the correctness of the algorithm.
There are several aspects of the correctness of an algorithm. For every satisfying
input, the results of the algorithm should be correct. Also for every satisfying input, the
algorithm should execute to the end and not abort in the middle. If these conditions are
not met, the system will produce incorrect results, especially for real-time processing.
The algorithm should also terminate in a finite number of steps. An algorithm that needs
an infinite number of steps is useless.
Another consideration for an algorithm is its efficiency. Execution of an
algorithm on a computer requires a certain amount of computer resources as well as a
certain amount of computing time. Before implementing an algorithm on a computer, we
have to investigate whether it is feasible to use the algorithm to solve the problem. A lot
Section No. Title 61
of known algorithms grow too computationally expensive as the size of the input
increases. For many real-time processing applications in bioinstrumentation, execution
time and memory requirement are important considerations in system design.
Cathode
Deflection
control
Screen
Figure 2.49 Sketch for cathode ray tube (CRT). There are two pairs of electrodes to control the
deflection of the electron, but only one pair is shown.
Section No. Title 63
2.10 References
Alexandridis, N. A. 1984. Microprocessor System Design Concepts. Rockville, MD:
Computer Science Press.
Banachowski, L., Kreczmar, A. and Rytter, W. 1991. Analysis of Algorithms and Data
Structure. Reading, MA: Addison-Wesley.
Bobrow, L. S. 1987. Elementary Linear Circuit Analysis. New York: Holt, Rinehart and
Winston.
Date, C. J. 1977. An Introduction to Database Systems. Reading MA: Addison-Wesley.
Hansen, P. B. 1973. Operating System Principles. Englewood Cliffs, NJ: Prentice-Hall.
Hilburn, J. L. and Julich, P. M. 1979. Microcomputers/Microprocessors: Hardware,
Software and Applications. Englewood Cliffs, NJ: Prentice-Hall.
Lurch, E. N. 1963. Electric Circuits. New York: John Wiley & Sons.
Oppenheim, A. V. and Schafer, R. W. 1989. Discrete-time Signal Processing. Englewood
Cliffs, NJ: Prentice-Hall.
Tompkins, W. J. (ed.). 1993. Biomedical Digital Signal Processing. Englewood Cliffs,
NJ: Prentice-Hall.
64 Chapter No. Title
Webster, J. G. (ed.) 1998. Medical Instrumentation: Application and Design. 3rd ed. New
York: John Wiley & Sons.
2.11 Problems
2.1 One ampere flows through an electrosurgical dispersive electrode that has a
resistance of 50 Ω. Calculate the power dissipated. Discuss the possibility of burns
under the electrode.
2.2 Calculate the equivalent resistance for a 400 Ω and 1 kΩ resistor in series.
2.3 Calculate the equivalent resistance for a 400 Ω and 1 kΩ resistor in parallel.
2.4 Design a resistive attenuator to convert a 20 V voltage from a signal generator to a
10 mV voltage for an amplifier.
2.5 An operational amplifier has limited current output of 20 mA. Calculate the time to
charge a 1 µF capacitor from 0 to 10 V.
2.6 From a 2 kV source in series with a 20 kΩ resistor, calculate the time required to
charge a 100 µF defibrillator capacitor to 1.9 kV.
2.7 Calculate the time response of a 10 kΩ, 30 mH high-pass and low-pass RL circuits
(similar to the two cases of RC circuit, but τ = L/R) to a 1 V step input.
2.8 Design an inverting amplifier with a gain of –10. If a voltage of 2sin(2ωt) is
applied, sketch the waveform of the output and label axes.
2.9 Design a noninverting amplifier with a gain of +10. If a voltage of 2sin(2ωt) is
applied, sketch the waveform of the output and label axes.
2.10 In Figure 2.30, R1 = 1 kΩ, the two R2 are 10 kΩ and 9 kΩ respectively (the one
connected to ground is 9 kΩ). Calculate the CMRR.
2.11 Design a comparator to detect an amplified electrocardiogram whose Q wave is 0
V and R wave is 4 V.
2.12 A low-pass filter with R = 10 kΩ and C = 300 pF is placed before an ECG
amplifier. Calculate by what factor it reduces 1 MHz interference from an
electrosurgical unit.
2.13 Calculate the number of bits to obtain a resolution of 1/1000 Vref for an ADC.
2.14 Explain the advantages and disadvantages of digital signals versus analog signals.
2.15 Define Nyquist frequency. For a signal that contains frequency components up to 1
kHz, calculate the minimum sampling frequency.
2.16 Several students are attempting to digitally process a sinusoidal signal that repeats
every 250 µs. The students randomly select a DSP that takes 360,000 samples per
minute. Will they be successful in processing the signal correctly? Why or why not?
2.17 Calculate the minimum bit size and sampling speed for a DSP to handle an
incoming signal with a range from 0 to 5 V, frequency components between 25 and
250 Hz and sensitive to ±1 mV.
2.18 Explain the function of the following devices in a microcomputer: CPU, RAM,
ROM, I/O port, data bus.
Section No. Title 65
2.19 Give two examples for each category of software: application software package,
high-level programming language, low-level programming language. Also state on
which platform each operates.
2.20 Describe how a CRT functions.
2.21 Go to www.bmenet.org/BMEnet/. Click on jobs. Find an industrial (not university)
job and describe in a few sentences.
2.22 A radio-frequency (RF) catheter is introduced into the heart to inject power at 500
kHz to ablate (destroy) tissue that causes tachycardia. This causes interference in
the electrocardiogram (ECG). Design a filter that reduces this interference to
0.0005 of its former level.
3
Analysis of Molecules in Clinical Medicine
Mat Klein
3.1 Spectrophotometry
Photometry is based on measurements of electromagnetic energy emitted, absorbed, or
transmitted under controlled conditions. Spectrophotometry and flame photometry are
two different types of photometry commonly used to determine the concentration of light
absorbing molecules in a solution. Spectrophotometry is based on the use of light inten-
sity measurements at a particular wavelength or a discrete portion of wavelengths to de-
termine the concentration of the desired molecules in solution by the amount of radiant
energy absorbed by the desired molecules.
3.1.1 Components
The purpose of the light source in a spectrophotometer is to provide incident light of suf-
ficient intensity to the sample. The wavelength and intensity of incident light desired de-
termines what light source to use. The most common source for substances that absorb
3-1
3-2 Chapter 3 Analysis of Molecules in Clinical Medicine
radiation in the visible, near infrared, and near ultraviolet regions, is a glass enclosed
tungsten filament. These distinctions between the spectrum are made since silica, used to
make cuvettes, transmits light effectively at wavelengths greater than 220 nm. To in-
crease the lifetime of the tungsten filament, there is usually low-pressure iodine or bro-
mine vapor in the bulb. The tungsten light bulb does not supply enough radiant energy for
measurements below 320 nm. High-pressure hydrogen or deuterium discharge lamps are
sufficient for measurements in the near ultraviolet region. At wavelengths below this, the
emission is no longer continuous. Two advantages deuterium lamps have is that they pro-
duce about three times the light intensity of hydrogen lamps and have a longer lifetime.
Xenon arc or high-pressure mercury vapor lamps provide high levels of continuous ultra-
violet illumination. These lamps may require thermal insulation with or without auxiliary
cooling to protect surrounding components since they become extremely hot during op-
eration.
Exit slit
Prism Violet
Cuvette
Figure 3.1 Block diagram of a single beam spectrophotometer. The prism serves as the dispersing
device while the monochromator refers to the dispersing device (prism), entrance slit, and exit slit.
The exit slit is moveable in the vertical direction so that those portions of the power spectrum pro-
duced by the power source (light source) that are to be used can be selected.
Monochromator
with a polished aluminum or aluminum–copper alloy. When radiation strikes the grating,
light rays bend around the sharp edges of the closely spaced parallel lines. The amount of
bending is dependent upon the wavelength of the light. A tiny spectra is produced for
each line of the grating. Wave fronts are formed as the light waves move past the corners.
Wave fronts that are in phase when they cross reinforce each other, while wave fronts
that are out of phase cancel out, leaving a complete spectrum from which to choose a
narrow band of wavelengths or particular wavelength.
Cuvette
There are two requirements for a detector of radiant energy in a spectrophotometer. The
photosensitive detector must have a linear response and be sensitive enough in the part of
the spectrum it is being used in. The most common devices used to detect the amount of
radiant energy leaving the cuvette include barrier layer cells, photodiode arrays, and pho-
tomultiplier tubes. These devices convert electromagnetic energy to electric energy,
which can then be measured.
A readout device displays the electric energy from the detector onto some type
of scale such as absorbance or transmittance. The two types of readout devices are the
direct reading system and the null point system. In a direct readout meter, there is a linear
relationship between milliamperes and %T (percent transmittance) and a log relationship
between millivolts and absorbance. This characteristic makes direct readout systems fast
and simple. In a null point system, the meter is calibrated (zeroed) by a potentiometer. In
a null point system, the absorbance, transmittance, or any other arbitrary scale is fitted to
the potentiometer scale.
3.1.2 Theory
than 10–2 M (molar)] is directly related to its absorbance. An amalgamation of these two
discoveries is known as Beer’s Law, as stated by Bouguer: “Equal thickness of an ab-
sorbing material will absorb a constant fraction of the energy incident upon it.” (Wheeler,
1998) This relationship is
where
Transmittance is the ratio of light intensity leaving the cuvette I to light intensity entering
the cuvette I0
I
T= (3.2)
I0
I 100
A = log 0 = log = 2 − log(%T ) (3.3)
I
%T
As the concentration of the substance in solution increases (or decreases), the transmit-
tance varies logarithmically and inversely. The fraction of incident light absorbed by a
substance in solution at a particular wavelength is a constant characteristic of the sub-
stance called absorptivity, a. Let the length of the light path (usually the outer diameter of
the cuvette) be a constant, L, and the concentration of the substance in solution be c. The
relation of these parameters to the total absorbance is expressed as:
A = aLc (3.4)
where a is in liters per gram times centimeters, L is in cm, and c is in grams per liter. Ab-
sorptivity is often replaced by molar absorptivity, ε, expressed in liters per mole times
centimeters. Epsilon is a constant corresponding to a 1 molar solution of the absorbing
substance, with a light path of L = 1 cm and a given wavelength. Hence, absorbance can
be expressed as
A = εLc (3.5)
3.1 Spectrophotometry 3-5
cu Au
= (3.6)
cs As
where Au is the absorbance of the unknown concentration and cu is the unknown concen-
tration.
3.1.3 Calibration
Table 3.1 lists common molecules measured in the clinical laboratory and their normal
and toxic levels
Table 3.1 Normal and toxic levels of various molecules in the body given in both US and European
(SI) units (Pagana and Pagana, 1995). NA denotes “not available”.
3-6 Chapter 3 Analysis of Molecules in Clinical Medicine
3.1.5 Microdialysis
Microdialysis is a technique for sampling the chemistry of the individual tissues and or-
gans of the body, and is applicable to both animal and human studies (de la Peña, 2000).
The basic principle is to mimic the function of a capillary blood vessel by perfusing a thin
dialysis tube implanted into the tissue with a physiological liquid. The perfusate is ana-
lysed chemically and reflects the composition of the extracellular fluid with time due to
the diffusion of substances back and forth over the membrane. Microdialysis is thus a
technique whereby substances may be both recovered from and supplied to a tissue. The
most important features of microdialysis are as follows: it samples the extracellular fluid,
which is the origin of all blood chemistry; it samples continuously for hours or days
without withdrawing blood; and it purifies the sample and simplifies chemical analysis by
excluding large molecules from the perfusate. However, the latter feature renders the
technique unsuitable for sampling large molecules such as proteins. The technique has
been extensively used in the neurosciences to monitor neurotransmitter release, and is
now finding application in monitoring of the chemistry of peripheral tissues in both ani-
mal and human studies.
[HbO 2 ]
HbO 2SAT= (3.7)
[RHb] + [HbO 2 ]
3.3 Bilirubin
Red blood cells are replaced approximately every 100 days, which means that every day
one percent of the body’s red blood cells, which are produced in bone marrow, are re-
placed. Bilirubin is waste resulting from the removal of old red blood cells.
Hemoglobin consists of four subunits. Each subunit has one chain of protein,
also referred to as globin, and one molecule of heme. Heme is made up of a single iron
atom and porphyrin, a ring shaped molecule to which is attached the iron atom. When a
red blood cell is destroyed, the body recycles the iron. The ring shaped molecule is toxic
and consequently is broken down into bilirubin. Unconjugated bilirubin is produced in
the spleen when the porphyrin is broken down. The unconjugated bilirubin enters the
blood stream and travels to the liver where it is converted to conjugated bilirubin and
subsequently excreted.
Unconjugated bilirubin is produced when red blood cells are destroyed. Abnor-
mally high levels of conjugated bilirubin in the bloodstream result from liver disease and
can turn the whites of a person’s eyes yellow as well as their skin. This condition is
known as jaundice. Neonatal jaundice, a common problem that occurs after birth, is a
result of the mother’s antibodies attacking the baby’s red blood cells. In general, blood
samples can be taken and used to measure bilirubin concentration when diagnosing liver
and/or biliary disease.
The most common way to quantitatively measure bilirubin is based on the diazo
reaction and spectrophotometry. This technique is used to measure total and conjugated
bilirubin in blood serum or plasma. A detailed discussion of this method, also known as
the Jendrassik and Grof Technique, can be found in (Burtis and Ashwood, 1994).
3.4 Lactate
Lactate, also known as blood lactate in medicine, is the anionic form of lactic acid pre-
sent in the blood. Lactic acid is a metabolic intermediate involved in many biochemical
processes including glycolysis and gluconeogenesis (the formation of new glucose from
3-8 Chapter 3 Analysis of Molecules in Clinical Medicine
lactate oxidase
L-lactate + O2 → pyruvate + H2O2 (3.9)
LDH
Pyruvate + NADH + H+ → L-lactate + NAD+ (3.10)
3.5 Creatinine
Phosphocreatine is found in muscle cells and can be used to convert adenosine diphos-
phate (ADP) back to adenosine triphosphate (ATP), thereby replenishing the muscle
cell’s energy. During this conversion, creatine is produced. Creatine is usually converted
back to phosphocreatine and the cycle starts over. However, when creatine needs to be
excreted, it is dehydrated and converted to creatinine. Creatinine has the property of not
being reabsorbed when going through the kidneys during urine production. Therefore, it
is an ideal candidate for being used to measure the condition of the kidneys.
Creatinine can be measured in the blood (serum) and in urine. Elevated levels of
creatinine in the blood result from muscle damage or as a result of strenuous physical
activity. In a person with kidney disease, creatinine builds up in the blood since it is being
produced faster than it is being eliminated. Therefore, measurement of creatinine in blood
(serum) is a rough estimate of the health of the kidneys. Measurement of creatinine from
the urine is a much better measure of kidney function (see Chapter 9) and is the most
prevalent clinical test for approximating glomerular filtration rate (the rate of filtration by
the kidneys).
The most common method used to measure creatinine is based on the Jaffe reac-
tion
alkaline
Creatinine + Picric Acid → Picrate–creatinine Complex (3.11)
3.5 Creatinine 3-9
In an alkaline (basic) medium, creatinine and picric acid form a red–orange compound
whose structure has been postulated but not confirmed. The hypothesized picrate–
creatinine complex forms from a creatinine to picric acid ratio of 1:1. The complex can
be measured spectrophotometrically at wavelengths between 505 to 520 nm. The concen-
tration of the hydroxyl ion (alkaline medium) affects the rate of the complex formation.
Most methods use a 0.5 M concentration of sodium hydroxide and picric acid in excess of
stochiometric amounts so that picric acid is not the limiting reagent. One of the main
problems with the Jaffe reaction is that it is nonspecific when used to measure creatinine
in plasma. Molecules that interfere with the specificity of the Jaffe reaction include glu-
cose, protein, ascorbic acid, acetone, and pyruvate. As a result, several modifications
exist which increase the specificity of the reaction (Burtis and Ashwood, 1994). However,
many interference problems are still unresolved.
3.6 Urea
Urea, NH2–CO–NH2, a nitrogen containing molecule, is a metabolic product of breaking
down proteins. Figure 3.2 shows how urea is formed in the liver.
Liver NH3
Protein Amino acids Ammonia Urea
Figure 3.2 Origin of urea in the body (Burtis and Ashwood, 1994). The liver produces urea by first
breaking proteins down into their building blocks, amino acids, by a process that breaks the peptide
bonds between the amino acids (proteolysis). The amino group (NH2) of amino acids is removed
and ultimately used to form ammonia (NH3) and urea.
Over 90% of urea is excreted through the kidneys to the urine. The body produces urea to
rid itself of excess nitrogen.
Ammonia is produced by the liver as a waste product of gluconeogenesis and
the liver converts it into urea. Urea is then transported in the blood to the kidneys as
blood urea nitrogen (BUN). Although the urea nitrogen measurement is often referred to
as BUN, it is never measured from whole blood. Urea nitrogen is most often measured
from blood serum (watery fluid separated from coagulated blood) and sometimes plasma.
An above normal amount of urea in the blood is an indicator of decreased kidney func-
tion, and therefore possibly kidney disease.
The two primary methods of measuring urea nitrogen are spectrophotometric.
The first method measures urea indirectly by quantifying the concentration of the ammo-
nium ion spectrophotometrically. Urea is hydrolyzed by water in the presence of urease
(an enzyme) and the resulting ammonium ion is quantified
urease
→ 2 NH 4 + + CO3 2 − (3.12)
NH 2 − CO − NH 2 + H 2 O → 2NH 2 + CO 2
3-10 Chapter 3 Analysis of Molecules in Clinical Medicine
Two ways of quantifying the ammonium ion concentration are Berthoelot’s reaction and
the enzymatic assay with glutamate dehydrogenase. An electrochemical approach exists
but is too slow to be used repeatedly (Burtis and Ashwood, 1994).
The second method measures urea directly. It is based on the Fearon reaction,
and can measure urea in urine and plasma. It has been used in several automated systems,
such as the Dupont Dimension 380. The reaction is
H + , Heat
Urea + Diacetyl → Diazine + H 2 O (3.13)
3.7 Glucose
Glucose is the main source of energy for all organisms. Diabetes mellitus is a group of
metabolic disorders of carbohydrate metabolism in which glucose is underutilized, pro-
ducing hyperglycemia (high blood sugar levels). The two types of diabetes mellitus are
insulin dependent diabetes mellitus (type I) and non-insulin dependent diabetes mellitus
(type II). When a person has high blood glucose, glucose shows up in the urine. Normally,
sugar is not present in the urine.
Insulin is one of the hormones that controls whether or not glucose is going to
be taken out of storage and put into the blood, or vice versa. In type I diabetes, the body
does not produce insulin because there is some destruction of the pancreatic islets that
produce it. Type II patients produce insulin, but cells do not recognize it (i.e. they have
defective insulin receptors). Type I is a more severe form of diabetes mellitus, is gener-
ally seen in children, and has to be treated with insulin. Type II is seen in much older
people, and for many, careful control of diet and exercise will be enough for treatment. In
more severe cases, insulin is taken.
Self-monitoring of blood glucose is required for diabetic patients, especially
insulin dependent, in order to maintain normal blood glucose levels (glycemia). When an
abnormal amount of glucose is present in the blood, the individual needs to correct the
abnormality to avoid short and long term health complications. By regulating blood glu-
cose levels, patients are mimicking the body by providing themselves with the correct
amount of insulin. Insulin dependent patients need to measure their blood glucose levels
about three times a day.
One reason the level of blood glucose needs to be tightly regulated is that glu-
cose is the only source of energy neurons can consume (they do not have the enzymes to
consume anything else). Low blood glucose levels result in hypoglycemia. When this
occurs, an individual’s neurons have no source of energy and if low enough, the person
will go into a coma and die. High blood glucose levels result in hyperglycemia. When
blood glucose becomes too high, the glucose molecules will denature (alter the shape of)
proteins, such as collagen and hemoglobin, throughout the body. Collagen attaches to the
3.7 Glucose 3-11
lining of blood vessels (the basement membrane). When glucose denatures collagen,
blood vessels are destroyed. This leads to decreased blood flow in the arms and legs
(lower perfusion). When glucose denatures proteins associated with neurons, nerve dam-
age occurs and results in a person’s inability to feel. Diabetes mellitus is the leading
cause of amputation because decreased blood flow causes tissue to die and damaged
nerves hinder the sensation thus making widespread damage much more likely.
The most common techniques for measuring blood glucose are enzymatic. The
glucose oxidase method is a very popular manual procedure used for self-monitoring.
The Hexokinase method is widely used in laboratories since the procedures for it are car-
ried out by automated equipment.
The glucose oxidase method is used in a large number of commercially available strip
tests. These simple strip tests allow easy and quick blood glucose measurements. A strip
test product, One Touch II (Lifescan, Milpitas, CA), depends on the glucose oxidase–
peroxidase chromogenic reaction. After a drop of blood is combined with reagents on the
test strip, the reaction shown in Eq. (3.15) occurs.
glucoseoxidase
Glucose + 2H 2 O + O 2 → GluconicAcid + 2H 2 O 2 (3.15)
Addition of the enzyme peroxidase and o-dianiside, a chromogenic oxygen results in the
formation of a colored compound that can be evaluated visually.
peroxidase
o - dianisine + H 2 O 2
→oxidized o - dianisine + H 2 O (3.16)
Much of the automated equipment is based on the hexokinase method. The general reac-
tions are
hexokinase
Glucose + ATP → glucose–6–phosphate + ADP (3.17)
G − 6 − PD
Glucose–6–phosphate + NAD+ → 6–phospogluconate + NADH + H+ (3.18)
Front Back
LIFESCAN
109 mg/dL
Sure Step
Apply a drop The dot turns blue
of blood to when enough
the strip blood is applied
Figure 3.3 LifeScan, Inc., a system by Johnson and Johnson for self-monitoring glucose levels.
4Ag + 4Cl −
→ 4AgCl + 4e − (3.19)
O 2 + 2H 2 O + 4e −
→ 2H 2 O 2 + 4e −
→ 4OH − (3.20)
4OH − + 4KCl
→ 4KOH + 4Cl −
The electrode at which the reduction–oxidation (redox) reaction involving the molecule
of interest occurs is often referred to as the working electrode. According to the Nernst
3.8 Amperometric Biosensors for Oxygen and Glucose 3-13
equation (Eq. 3.21), a cell reaction will spontaneously proceed to the right if the potential
of the cell is greater than zero (E > 0) because the concentration of O2 is greater than H20.
A cell reaction will spontaneously proceed to the left if the potential for the cell is less
than zero (E < 0). When the potential for the cell is zero (E = 0), the redox reaction is at
equilibrium.
Platinum cathode
Glass rod
0.7 V
Sample outlet
e– – +
Dissolved
O2
e– Cl–
Sample inlet e–
Ag/AgCl anode
O2 permeable membrane
Phosphate buffer Ammeter
Figure 3.4 In the PO2 electrode, O2 dissolved in the blood diffuses through a permeable membrane.
Current is proportional to PO2 (Hicks et al., 1987).
0.059 CH O
E = E0 + log 2
(3.21)
n CO
2
where n is the valence of the electrode (cathode) material. E0 is the standard half-cell
potential. It is the half-cell potential when CO = CH 0. A dc voltage between 600 and 800
2 2
mV allows the electron transfer reaction between O2 and H2O to occur. The entire reac-
tion at the working electrode consists of several steps, each of which has its own inherent
rate. These steps include the transport of O2 from the solution to the working electrode,
the transfer of electrons from O2 to H2O, interactions between O2 and H2O (the oxidized
and reduced forms of the species of interest), and the transport of H2O back to the solu-
tion. The amount of current produced is controlled by the slowest of these steps. Because
O2 is consumed at the cathode, there is a concentration gradient from the dissolved O2 to
the cathode. It takes about 60 s to establish this concentration gradient for a stable meas-
urement. The cathode may be vibrated to ensure that fresh O2 is available at the cathode.
The current produced by this reaction is directly proportional to the rate of elec-
trons transferred from O2 to H2O. The rate of electrons transferred from O2 to H2O is
directly proportional to the concentration of O2.
3.8.2 Glucose
The first enzyme electrode to measure glucose was developed by Clark and Lyons in
1962 (Figure 3.5). These biosensors still serve as the classical example for measuring
3-14 Chapter 3 Analysis of Molecules in Clinical Medicine
glucose because their operation explains the basic concepts for measuring glucose with
an amperometric enzymatic electrode (Hall, 1991).
O2 Glucose oxidase O2
Pt cathode
0.7 V
Ag/AgCl –+
anode
Ammeter
Figure 3.5 In the glucose enzyme electrode, when glucose is present, it combines with O2, thus
decreasing the O2 that reaches the cathode.
It is often referred to as a first generation biosensor due to its structure and the
level at which its components are integrated (Taylor and Schultz, 1996). This glucose
sensor uses the PO2 electrode and the enzyme glucose oxidase immobilized on a mem-
brane. The reaction for the basic glucose sensor is
glucoseoxidase
Glucose + O 2 + 2H 2 O → Gluconicacid + 2H 2 O 2 (3.22)
Glucose and oxygen react in the presence of the enzyme glucose oxidase. The amount of
O2 consumed, indicated by a decrease in PO2, in the reaction is proportional to the con-
centration of glucose in the solution. (The concentration of glucose is indicated by the
consumption of O2 by the enzyme glucose oxidase.) The disadvantage of the single oxy-
gen electrode glucose sensor is that it is sensitive to variations of PO2 in the blood.
A dual oxygen electrode glucose sensor (Peura, 1998) eliminates the problem of
variations in PO2 that the single oxygen electrode is sensitive to. As the name implies,
the dual oxygen electrode uses two oxygen electrodes to compensate for the PO2
sensitivity. An enzyme within a gel covers one electrode, thus depleting oxygen when
glucose is present. The other electrode, not covered by the enzyme, senses only oxygen.
The amount of glucose is determined as a function of the difference between the
electrode responses. The disadvantage of enzymatic glucose sensors is that they can only
3.8 Amperometric Biosensors for Oxygen and Glucose 3-15
responses. The disadvantage of enzymatic glucose sensors is that they can only be used in
vivo for a few months and the enzyme is unstable.
A more recent approach to amperometric enzyme glucose electrodes involves
the use of the electron transfer mediator dimethyl ferrocene. Electron transfer mediators,
such as dimethyl ferrocene, are low molecular weight species, which shuttle electrons
between the working electrode and the oxidation–reduction center of the enzyme (Taylor
and Schultz, 1996). There are several advantages in using mediators. First, mediators are
much more readily reduced than the substrate co-factor (in this case O2). Second, they
allow glucose measurements that are independent of the variations in PO2 of the sample
(Burtis and Ashwood, 1994). In the series of reactions shown in Figure 3.6, a reduced
form of the mediator dimethyl ferrocene reduces the oxidized form of glucose oxidase.
Measurement using this reaction can be made using whole blood and the current pro-
duced is directly proportional to the glucose concentration. One of the disadvantages of
this technique is that dimethyl ferrocene (as well as all mediators) is not very soluble and
ends up adsorbing to the electrode surface.
Electrode surface
Dimethyl ferrocene Glucose oxidase
Glucose
oxidized form reduced form
e–
Figure 3.6 The sequence of reactions involved in the mediated reaction of a glucose sensor (Taylor
and Schultz, 1996). Dimethyl ferrocene expedites the transfer of electrons from the electrode sur-
face to the redox center of the enzyme glucose oxidase. The use of the mediator dimethyl ferrocene
reduces the sensor’s dependence on oxygen tension.
Ion selective electrodes are used to measure the activity of a specific ion in a cell or solu-
tion. An ion selective electrode contains a membrane that is permeable to a specific
cation or anion. In the presence of the specific ion, a membrane potential results which
indicates the amount of ion activity. pH is the negative logarithm to the base 10 of the
hydrogen ion concentration [H+] in mol/L.
3-16 Chapter 3 Analysis of Molecules in Clinical Medicine
pH = –log10[H+] (3.23)
The measurement of pH is obtained by using an ion selective electrode (ISE) with a glass
membrane. To prevent pH measurement errors that occur due to other ions, a highly se-
lective H+ glass membrane is desired. One such glass composite consists of silicon diox-
ide, lithium oxide, and calcium oxide in the ratio 68:25:7 (Burtis and Ashwood, 1994).
The glass membrane separates the test solution on side 2 from the reference so-
lution, usually hydrochloric acid of known pH, on side 1 of Eq. (3.24). When the pH of
the solutions on side 1 and 2 differ, a potential across the membrane develops, V = V1 –
V2. The potentials of solution 1 (hydrochloric acid) can be recorded with a reference elec-
trode such as a silver/silver chloride electrode. The potential of solution 2 can be re-
corded with a calomel electrode. Ideally, the pH electrode obeys the Nernst equation
RT
E = K + (2.303 ) log(a i) (3.25)
ZF
where E is the potential, R the universal gas constant, F is the Faraday constant (= 96487
C/mol, C = coulomb), T the absolute temperature, Z the ionic charge, ai the activity of the
ion, and K is a constant containing contributions from various sources. The activity of an
ion, ai, equals the activity coefficient λi times its concentration c.
ai = λi c (3.26)
Factors such as other ions in the solution and the strength of H+ in the solution
influence λi. λi decreases as c increases. λi for plasma is about 0.75 for Na+, 0.74 for K+,
and 0.31 for Ca2+. In solutions where the primary ion (H+) is of very low concentration
(µM), the membrane potential is not directly proportional to the logarithm of the activity
of the diffusible ion H+ in solution 1 or 2. The relationship is better expressed by the Ni-
kolski–Eisenman equation.
Or the potential can also be calculated from
[H + ]i
E = −0.0615 log10 (3.27)
[ H + ]o
where [H+]i is the concentration of H+ inside the cell and [H+]o is the concentration of H+
ions outside the cell.
IQ Scientific Instruments (2001) offers pH measurement without glass. A
stainless steel probe has a silicon chip sensor and can be stored dry. The ISFET consists
of a silicon semiconductor substrate with two electrical contacts (source and drain) a
small distance apart. Deposited on the substrate between the source and drain is a silicon
electrical insulator. Hydrogen ions at or near the surface of the insulator cause a variable
voltage potential between the insulator and the underlying semiconductor material be-
3.9 Ion-Selective Electrodes for PH and CO2 3-17
tween the source and drain. The variable voltage potential is proportional to the relative
concentration of hydrogen ions in the sample solution. The pH can then be derived from
this voltage to a very high level of accuracy.
The measurement pH is the measure of the concentration of H+ protons. The pH
level in the body is more strictly regulated than glucose. The normal range of pH is 7.36
to 7.44. A pH of less than 7.36 is called acidosis. If the pH is less than 7.0, a person will
go into a coma and die. A pH greater than 7.44 is called alkalosis, and a pH greater than
7.8 causes the body’s muscles to seize up, a condition known as tetany. The explanations
for these phenomena lie in acid–base chemistry. Molecules in the body, particularly pro-
teins, have sites where protons can associate or dissociate, and when the body’s pH is
altered, the charge of that particular portion of the molecule is essentially changed. The
electrostatic charges on the parts of the molecule may attract or repel in different ways
than they should. As a result, proteins do not fold properly and DNA does not form as it
should, among other things.
The metabolism of substances in the body tends to produce protons. With the
exception of vomiting, protons leave the body via the kidney. The kidney can excrete
acids directly or it can produce bicarbonate, which combines with a proton to form car-
bonic acid. Carbonic acid spontaneously breaks down into water and carbon dioxide.
Another way the body eliminates acid is by exhaling carbon dioxide through the lungs.
An individual can exhale too much or too little carbon dioxide, thus changing the acidity
of the blood. Or, if the kidney is prevented from eliminating acid, acid buildup can occur,
causing a condition known as acidosis. Both the lungs and the kidneys regulate the
body’s pH level. Having two ways to eliminate acid allows one method to compensate if
the other becomes impaired. If the pH level is abnormal, there are three things that can be
measured: carbon dioxide concentration, pH, and bicarbonate concentration. By measur-
ing these three concentrations, we can determine if the kidneys or lungs are functioning
abnormally. If the problem is abnormal CO2 concentration, then we know the problem is
with the lungs. If the bicarbonate concentration is abnormal, then a condition called
metabolic acidosis exists and the cause might be either the kidney or something within
the metabolic pathway. Further studies are required for abnormal bicarbonate concentra-
tion such as sodium concentration or chloride concentration measurements of the blood.
In the CO2 electrode, blood comes in contact with a plastic membrane. Holes in
the membrane permit gas to diffuse through, but block everything else. The CO2 diffuses
to a chamber containing a pH electrode, forms carbonic acid, and increases the acidity.
The reaction for this process is
technique for quantifying these elements, because compared to calcium and magnesium,
they have low excitation potentials. Figure 3.7 shows a block diagram of a flame pho-
tometer. The purpose of the flame is to provide energy to the elements such that their
valence electrons will be excited from ground states to higher excited states. Electrons in
the excited states are unstable and emit energy as photons of a particular wavelength as
they return to their ground states. The wavelength(s) of the light emitted by an element is
characteristic of that element. When sodium, potassium, and lithium are present as
cations in an aqueous solution and exposed to a flame, sodium produces yellow, potas-
sium violet, and lithium red colors. The intensity of the light given off is directly propor-
tional to the number of photons being emitted, which in turn is directly proportional to
the number of atoms, or concentration of cations, in the solution.
Collimating lens
Detector Readout
Fuel inlet Fuel inlet
Capillary
Oxygen inlet
Aqueous sample
Figure 3.7 The flame photometer aspirates a sample containing metal ions and heats it to incandes-
cence. Detector output is proportional to concentration.
In flame photometry, the monochromator and detector are similar to those found
in the spectrophotometer, although the detector is measuring emission as opposed to ab-
sorption. The distinguishing features of the system are the flame and the atomizer. The
atomizer draws the sample solution containing the cations through an aspirator and into
the flame. This is achieved by passing a gas over the upper outlet of a capillary tube at a
high velocity, while the lower end of the capillary tube is inserted into the sample. The
fuel used to generate the flame can be either propane, natural gas, or acetylene mixed
with compressed air or oxygen. Lithium can be used in calibration because it does not
naturally occur in the body (provided the patient is not taking lithium medication). A
known concentration of lithium is added to the sample. Then a ratiometric technique can
be used to compare the ratio of sodium and potassium to lithium.
3.10 Flame Photometry 3-19
3.10.1 Lithium
Lithium is not produced in the body and does not naturally appear in the body. The pur-
pose of lithium measurement is to measure lithium carbonate, which is used in the treat-
ment of the psychiatric disorder manic depression (also known as bipolar disorder).
Sodium and potassium are used by the body to maintain concentration gradients in nerve
and muscle cells, which allows them to conduct action potentials. To keep these cells
working properly, the amount of sodium and potassium in the body must be regulated. It
is the responsibility of individual cells to do the regulation. Sodium and potassium enter
the body via eating. For people who do not eat properly, it is the kidneys’ job to regulate
and compensate the levels of sodium and potassium after eating. For example, if an indi-
vidual eats a lot of salt, the excess salt in their blood will be passed in the urine. Potas-
sium chloride can be injected into the blood stream to raise the level of potassium in the
body, although it is important to be extremely careful with the amount that is injected.
Too much potassium can kill a person by stopping their heart as a result of decreasing the
concentration gradient, and hence the ability to generate an action potential in the heart
muscle cells.
Figure 3.8 shows that the specimen containing known or unknown molecules enters the
mass spectrometer through the sample inlet, where it is vaporized under a high vacuum.
Several types of sample inlet systems exist that accommodate certain aspects of the sam-
ple such as its temperature, vapor pressure, and its volatility. An important inlet system is
3-20 Chapter 3 Analysis of Molecules in Clinical Medicine
Figure 3.9 shows the ion source where the sample gets ionized. Many different methods
to ionize the sample exist, one of which is the heated filament. A filament, (rhenium or
tungsten) when heated to greater than 2000 °C by electric current, emits electrons under
the influence of an electric field. Typical current flow is about 500 µA. The electrons are
focused in a magnetic field, which causes them to spiral toward the positively charged
target. Electron energy is typically 70 to 80 eV. When a sample is bombarded by the
high-energy electrons, the ionization process is initiated.
Vacuum ADC
Interface system
Figure 3.8 The mass spectrometer separates molecules in a high vacuum after they have passed
through a porous plug.
e–
Repeller
To mass
analyzer
Electron trap
(positively S
charged target)
Pole of magnet
Figure 3.9 An electron impact ion source bombards the molecules and carrier gas from a gas
chromatography column with electrons. The electron collisions with ion ABC (Eq. (3.28)) produce
both positive and negative ions. The positive ions are directed by focusing elements to the mass
analyzer.
For a time span of 0.1 to 10 ns after the collision of the electron with the ion,
several of these molecular ions disintegrate, forming radicals (•) and positively charged
fragment ions.
ABC•+ → AB+ + C +
A + + BC +
AB•+ + C (Loss of neutral) (3.29)
+ •
AC + B (Rearrangement)
etc.
Not all of the positively charged molecular ions undergo fragmentation. These
unfragmented positively charged molecular ions are drawn out of the ion source by an
electric field while the vacuum system pumps the negative and neutral fragments away.
An electric field then appropriately focuses and accelerates them into the mass analyzer.
The intensity of the electron beam and size of the sample determine the number of frag-
ments and molecular ions, whereas the chemical stability of the molecule determines the
relative amounts of molecular ions and fragment ions. Figure 3.10 is an example of a
mass spectrum for a sample and its resulting parts. A mass spectrum is a plot that repre-
sents the relative ion abundance at each value of m/z (mass to charge ratio).
3-22 Chapter 3 Analysis of Molecules in Clinical Medicine
AB+•
Sample abundance
C+
ABC+•
A+
AC+
m/z
Figure 3.10 The mass spectrum displays the relative abundance of charged molecular ions and
fragments.
The function of the mass analyzer is to separate the stream of ions produced from the ion
source by their mass to charge ratio (m/z). This is usually done with a magnetic field.
Figure 3.11 shows the main components of a quadrupole mass analyzer.
Detector
–
RF voltage
+ supply
Ion
source –
DC voltage
+ supply
Figure 3.11 A quadrupole mass analyzer accelerates ions from the ion source. The dc and RF volt-
ages that are applied to the four ceramic rods stabilize the paths of ions with selected m/z values.
Ions that do not stabilize collide with the rods and do not reach the detector.
3.11.4 Detector
The three most commonly used detection schemes are the electron multiplier, photo-
graphic emulsions, and Faraday cups (Haven et al., 1995). Figure 3.12 shows the main
components of an electron multiplier. Positively charged ions collide violently into the
3.11 Mass Spectrometry 3-23
cathode, thus releasing secondary electrons. These secondary electrons are attracted to
the first dynode, which is 100 V more positive than the cathode. A larger number of terti-
ary electrons are released and attracted to the second dynode, etc. to build up a current of
about 1 µA, which is measured at the anode.
Dynodes
Cathode
Anode
Horn
Collector
Ion+
Dynodes
Quadrupole
mass analyzer
rods
X-ray shield
Figure 3.12 Block diagram of an electron multiplier. An electron multiplier is sensitive enough to
detect the presence of a single ion. A 10 to 12-dynode electron multiplier produces a signal ampli-
fication of 105 to 106. (Haven et al., 1995).
and noting the power of the infrared light transmitted to the sample, and the power of the
infrared light transmitted through the sample. Using infrared light of a particular wave-
length characteristic for measuring CO2, the concentration of CO2 in a mixture of gases
can be determined through Beer’s Law
where Pt is the power per unit area of transmitted infrared light received by the detector,
P0 is the power per unit area of infrared light entering the sample, a is the absorption co-
efficient, L is the path length of light through the gas, and c is the concentration of gas.
Figure 3.13 shows the basic components of an infrared spectroscopy system (Primiano,
1998). The infrared source is modulated by rotating vanes similar to a windmill. The de-
tector compares transmission through the sample cell with transmission through a refer-
ence cell.
Figure 3.13 An infrared transmission spectroscope measures the absorption of infrared light by a
sample drawn through the sample cell by a vacuum pump.
Needle valve
Electrodes
Reflecting surface
Ionization
chamber
Optical filter
To vacuum pump
Figure 3.14 Emission spectroscopy measures ultraviolet light emitted by ionized nitrogen
(Primiano, 1998).
3.14.1 Fluorometry
Fluorescence is the emission of energy in the form of light as a result of the electrons of a
molecule returning to their ground state. When certain molecules absorb energy (in the
form of electromagnetic radiation) their electrons are raised to higher energy levels. As
the electrons of the molecule return to their ground states, they fluoresce. A fluorometer
is an instrument that measures the intensity of light produced when the electrons of a
molecule return to their ground state. Thus, fluorometry is defined as measuring the rela-
tionship between the concentration of a substance and the intensity of the fluorescence
produced by that compound when it is excited by radiation. Fluorometry is used to meas-
ure therapeutic drugs such as phenobarbital (treatment of epilepsy), enzymes such as pro-
teases (used to break down proteins), hormones such as cortisol, and analytes such as
catecholamines (e.g. epinepherine and norepinepherine) and bilirubin. Figure 3.15 shows
the components of a fluorometer, which include an excitation source, a primary (excita-
tion) filter, a secondary (emission) filter, a cuvette sample, a detector, and a readout de-
vice.
Although the components of a fluorometer are basically the same as those of a
spectrophotometer, fluorometry is up to four orders of magnitude more sensitive than
spectrophotometry (Wheeler, 1998). In spectrophotometry, the concentration of the sub-
3-26 Chapter 3 Analysis of Molecules in Clinical Medicine
stance is determined by the difference of the absorbance between a solution that has zero
absorbance (the solution used to calibrate the spectrophotometer) and the absorbance of
the unknown solution. When the absorbance of the unknown solution is relatively small,
small errors in the measurement of the zero absorbance solution can cause relatively large
errors in the final determination. Fluorometry avoids this problem by measuring fluores-
cence directly from the sample without reference to another measurement. Though a dis-
advantage of fluorometry is that measurements are also susceptible to inner-filter effects.
These effects include excessive absorption of the excitation radiation (prefilter effect) and
self-absorption of atomic resonance fluorescence (postfilter effect).
Cuvette
Primary filter
I0
If
Secondary filter
Detector
Readout
Figure 3.15 Block diagram of a fluorometer. The primary filter passes only wavelengths that excite
the fluorescent molecule. The secondary filter blocks all scattered excitation wavelengths and
passes only the scattered fluorescent wavelengths. The secondary filter and detector are at a right
angle to the primary beam in order to avoid direct transmission of the light source through the sam-
ple to the detector.
Theory of fluorometry
The theory behind fluorometry can be derived from the Beer–Lambert law
I = I010–aLc (3.31)
where
I0 – I = I0(1 – 10–aLc)
We can expand through a Taylor’s series, rearrange, and convert the logarithm base to
yield
IF = φI0 (2.23aLc)
Thus the fluorescence intensity is directly proportional to the concentration of the fluo-
rescent molecules in the sample and the excitation intensity.
Fluorometry is limited by several factors, including concentration of the sample.
For more details, see (Burtis and Ashwood, 1994).
Spectrofluorometers replace the fixed secondary filter with an emission mono-
chromator and sweep it at 5 to 50 nm increments to yield a spectrum of emission magni-
tude versus wavelength. Some spectrofluorometers also sweep an excitation monochro-
mator instead of having a fixed primary filter. In time-resolved fluorometers, long-lived
fluorescent molecules are excited by pulsed light and their decay is monitored over 0.6 to
100 µs. Laser-induced fluorometers measure emission from individual fluorescent or
fluorescent-labeled molecules, cells, or particles passing through a flow cell cuvette.
3.14.2 Chromatography
Gas chromatography (GC) is a type of chromatography that uses an inert gas as the mo-
bile phase. GC is used to separate compounds that are volatile at the temperature in the
column of the chromatograph. These mixtures of volatile organic compounds, often small
organic molecules such as therapeutic drugs, are measured from a sample of urine or
sometimes blood. Figure 3.16 shows a gas chromatograph consisting of several compo-
nents: a carrier gas, a column, a column oven, an injector, a detector, a recorder, and a
flowmeter.
Injector
Flow control
Carrier gas
Figure 3.16 In a gas chromatography system, the sample is injected into a carrier gas and flows
through a long column to the detector.
There are several advantages of GC: it is fast, extremely sensitive, and requires a very
small sample.
Carrier gas
The inert carrier gas, such as argon, helium, or nitrogen, is the mobile phase in the system
because it moves the components in the evaporated solute through the column. The car-
rier gas has very little effect on the rate of retention of components in the solute.
3.14 Drugs by Fluorometry and Chromatography 3-29
Column
Detector
The purpose of the detector is to quantify the components in a sample. It does this by
providing an electric signal proportional to the amount of the component coming off the
column. Some of the commonly used detectors in GC include the flame ionization detec-
tor, the thermal conductivity detector, the electron capture detector, the nitrogen phospho-
rus detector, and detectors based on mass spectroscopy. The combination of mass spec-
trometry and GC is ubiquitous in clinical chemistry laboratories in hospitals and is a very
powerful tool.
Resolution
Figure 3.17 shows that the detector yields a chromatogram. For a flow rate F, the peak of
a solute, A, appears after retention time tr. The volume that has passed before the peak is
the peak volume and is given by Vr(A) = trF. Band spreading will broaden the peak to a
Gaussian shape. An efficient column minimizes band spreading. The degree of efficiency
is the number of theoretical plates, N = [Vr(A)/s(A)]2, where s(A) is the standard devia-
tion of the peak volume. The width of a Gaussian peak at 1/2 its height, w1/2, is 2.354
standard deviations, from which N = 5.54[Vr(A)/w1/2(A)]2.
3-30 Chapter 3 Analysis of Molecules in Clinical Medicine
w 1/2 1/2
height
INJ
tr w
Figure 3.17 In chromatography, the peak appears after retention time tr.
Vr (B) − Vr (A)
Rs = (3.33)
[ w(A) + w(B)]/2
where w(A) and w(B) are the peak widths measured at the corresponding base. Inade-
quate separations occur for Rs < 0.8 and baseline separation occurs for Rs > 1.25 (Bowers
et al., 1994).
Example 3.1 For a gas chromatograph with a flow rate of 1.5 L/min, solute A has a
retention time of 1 min and solute B has a retention time of 2.5 min. If the peak width for
solute A is 0.5 L and the peak width for solute B is 1.5 L, is the resolution sufficient?
Since the flow rate of the solutes and peak widths are known, we only need to
find the peak volume before we can use Eq. (3.33) to determine if the resolution is suffi-
cient.
w(A) = 0.5 L
w(B) = 1.5 L
Coupling GC to MS
Liquid chromatography (LC) is a type of chromatography that uses a liquid of low vis-
cosity as the mobile phase. The stationary phase through which this liquid flows is an
immiscible liquid coated on either a porous support or absorbent support. One commonly
used type of liquid chromatography is high-performance liquid chromatography (HPLC).
HPLC is used to separate components dissolved in solution by conducting the separation
process at a high velocity and with a pressure drop. Figure 3.18 shows a schematic of an
HPLC instrument, indicating the major components: the reservoir of mobile phase, a
high-pressure pump, the sample injection port, a separation column, and a detector.
An HPLC instrument operates by injecting a liquid sample via a syringe into the
stream of mobile phase that is being pumped into the column. A detector records the
presence of a substance in the effluent. The detector can observe changes in refractive
index, UV light absorption and fluorescence. Just as in gas chromatography, a mass spec-
trometer can be coordinated with the HPLC to help identify the components of the mix-
ture.
3-32 Chapter 3 Analysis of Molecules in Clinical Medicine
Vacuum Flow
pump splitter
Sample
Solvent 1 Solvent 2 injection port
Pressure
gage
Degasser 1 Degasser 2
Mixing
vessel Pre-column
Differential
detector
Another type of chromatography that is widely used in the separation and purification of
biological material is ion exchange chromatography. Ion exchange chromatography oper-
ates on the principle that charged molecules in a liquid phase pass through a column
where the stationary phase has a charge opposite that of the molecules in the liquid. Due
to the high affinity of charged molecules for opposite charges, this method is highly se-
lective. It requires a solution with a specific pH or ionic strength before molecules will
elute from the column. Ion exchange chromatography is used in the separation of both
proteins and amino acids.
3.15 Electrophoresis
Electrophoresis is a technique used to separate charged molecules in a liquid medium
with an electric field. It is extensively used in separation of serum proteins, separation of
proteins in urine, determination of molecular weight of proteins, DNA sequencing, ge-
netic disease diagnosis, and comparison of DNA sequences in forensic testing.
factors influence the distance and rate of migration. Electrophoresis works because when
a charged molecule is placed in an electric field, the molecule sustains a force, which is
proportional to the charge of the molecule and the strength of the electric field.
Cover
– +
Anode Power
supply
Cathode
Figure 3.19 In an electrophoresis system, charged molecules move through a support medium be-
cause of forces exerted by an electric field.
Power supply
The power supply provides a constant voltage, current, or power across the support me-
dium. Most power supplies for electrophoresis systems provide between 50 and 200 V.
The heat introduced into the electrophoretic system because of the resistance of the sup-
port medium is
Support medium
One of the most commonly used support mediums is agar. Agar is so prevalent in clinical
labs that electrophoresis is often referred to as agarose gel electrophoresis. Clinically,
electrophoresis is performed on a plastic support covered with a thin (1 mm) layer of aga-
rose gel. Agarose has the following advantages: after electrophoresis is performed and
the plate is dried, the agarose is very clear, which enables a densitometer to easily exam-
ine it; agarose absorbs very little water because it contains very few ionizable groups.
Another commonly used support medium is cellulose acetate. Cellulose acetate
membranes are produced by reacting acetic anhydride with cellulose. These membranes
consist of about 80% air in pockets in-between interconnected cellulose acetate fibers.
These air pockets fill with liquid when the membranes are placed in a buffer. Cellulose
acetate electrophoresis (CAE) has the advantage of being relatively quick (20 to 60 min).
CAE membranes can be made clear for densitometry by using a solution that dissolves
the cellulose acetate fibers.
Electrodes may be made of metal such as platinum, or of carbon. Gel trays may be made
from UV transparent acrylic to enable direct observation of the migration of different
molecules.
There are several factors that cause different migration rates among the molecules in a
mixture. They are: the net charge of the molecule, q; the size, shape and mass of the
molecule; the strength of the electric field, E; the viscosity of the medium; the tempera-
ture of the medium. A molecule in the medium experiences a force FE = qE in the direc-
tion of the electric field that is proportional to the strength of the electric field, E, and the
net charge of the molecule, q. The molecule also experiences a drag force FD = fv =
6πηrv in the opposite direction of movement which is proportional to the molecule’s re-
sistance f, velocity, v, the viscosity, η, of the buffer solution in which it is migrating, and
the ionic radius, r, of the solute molecule, Q.
FE = FD (3.35)
qE = fv
qE
vss = (3.36)
f
The mobility of a charged particle is defined as its steady state velocity divided by the
strength of the electric field.
3.15 Electrophoresis 3-35
electrophoretic mobility µ =
vss = q (3.37)
E f
One mobility unit is defined as 10–5 cm2/(V·s).
Electrophoresis plates are usually dried before analysis to prevent diffusion of the mi-
grated solutes. Most proteins, nucleic acids, and other molecules require staining in order
to be seen. There are cameras available that can photograph the electrophoresis gel, using
filters specific for different stains. A commonly used stain for visualizing DNA is
ethidium bromide, which is fluorescent under a light known as a Woods lamp. The bands
appearing in the gel, or on the plate, can be identified by inspection under the Woods
lamp, but a densitometer is required for accurate quantification.
Densitometry is commonly used in clinical laboratories. A densitometer meas-
ures the transmittance of light through a solid electrophoretic sample, similar to how a
spectrophotometer operates. An electrophoresis plate is placed over the exit slit of the
monochromator of the densitometer. The plate or sample is then slowly moved past the
exit slit while the amount of transmitted light is measured. As the sample specimen is
moved past the exit slit, the location and absorbance of each band of the electrophoretic
sample are recorded on a chart.
Electrophoresis is used to separate the proteins in blood serum. Figures 3.20 and 3.21
show that the first peak is albumin. It is produced in the liver and performs many tasks.
For example, it thickens the blood, binds steroids, and binds carbon dioxide. The other
three peaks are referred to as alpha, beta, and gamma. The proteins that make up gamma
are referred to as gamma globulins. The most prevalent gamma globulins are antibodies
(the term antibodies and immunoglobulins are synonymous since antibodies are part of
the immune system). Various diseases can be diagnosed with the information contained
in these peaks. For example, if the albumin is low, that can be a sign of liver disease. The
gamma globulin peak in Figure 3.21 is a spread of many types of antibodies as a result of
a B-lymphocyte tumor (bone tumor).
200 mm long. Amino acids in a buffer solution are pumped through the column at about
13 MPa for about 40 min. Amino acids with the least positive charge bind weakly to the
resin, move most rapidly through the column and elute first to be spectrophotometrically
detected at wavelengths less than 300 nm. Amino acids with more positive charge bind
more tightly to the resin and elute later. The area under the peak of each amino acid on
the chromatogram is proportional to the amount of that amino acid (Lehninger et al.,
1993).
Albumin
α1 α2 β γ
Migration distance
Figure 3.20 This serum protein electrophoresis demonstrates a normal pattern, with the largest
peak for albumin.
Albumin
α2 β
α1
Migration distance
Figure 3.21 This serum protein electrophoresis demonstrates a decrease in the albumin and an
increase in gamma globulins.
the cell, amino acids are synthesized into a longer polypeptide sequence (a protein)
through the translation of information encoded in messenger RNA by an RNA–protein
complex called a ribosome.
To separate and purify proteins, cells are broken open to yield a crude extract.
Differential centrifugation may yield subcellular fractions. Ion-exchange chromatography
can separate proteins by charge in the same way it separates amino acids. Size-exclusion
chromatography separates by size. Affinity chromatography separates by binding speci-
ficity to a ligand specific for the protein of interest. The purified protein is characterized
by ion-exchange chromatography to measure the amount of the protein of interest and the
contaminants.
A nucleotide consists of a nitrogenous base, a pentose sugar, and one or more
phosphate groups. The nucleic acids RNA and DNA are polymers of nucleotides. The
genetic code can be determined by sequencing the four nucleotides that form DNA: A =
Adenine, C = Cytosine, G = Guanine, and T = Thymine. A dideoxynucleoside phosphate
(ddNTP) analog specific for one of the nucleotides interrupts DNA synthesis to prema-
turely terminate the fragment at that nucleotide, for example A. Different analogs termi-
nate C, G, and T fragments. When each of these radiolabeled (with a radioactive com-
pound) four fragments is separated electrophoretically, it yields the autoradiogram (by
darkening photographic film) pattern in Figure 3.22.
A C G T
G •
T •
G •
T •
C •
A •
Figure 3.22 Nucleotide fragments ending in A, C, G, and T are injected into lanes at the top of elec-
trophoresis columns. The sequence is read from the rows of bands from the bottom up as ACTGTG.
3.17 References
Antolasic, F. 1996. What is Mass Spectrometry? [Online]
http://minyos.its.rmit.edu.au/~rcmfa/mstheory.html
Berger, S. A. 1996. Introduction to Bioengineering. New York: Oxford University Press.
Blum, L. J. 1991. Biosensor Principles and Applications. New York: Marcel Dekker.
3-38 Chapter 3 Analysis of Molecules in Clinical Medicine
3.18 Problems
3.1 Give the equation for Beer’s law, define each term, and give units.
3.2 Explain the operation of a spectrophotometer and its purpose. List the components
of a spectrophotometer.
3.3 A sample concentration of 10 mg/dL yields a spectrophotometer transmission of
35%. Assume Beer’s law holds and calculate the unknown concentration for a
transmission of 70%.
3.4 A sample of concentration 20 mg/dL has an absorbance of 0.4 in a spectropho-
tometer. The sample is then diluted and yields an absorbance of 0.25. Calculate the
new concentration.
3.5 Define oxygen saturation and state the physiological meaning of SaO2.
3.6 Search the literature for a plot of SaO2 versus PO2 and sketch it.
3.7 Describe how NADH is used to measure lactate concentration and why lactate
concentration isn’t ascertained by measuring lactate directly.
3.8 Describe why creatinine is measured and the technique used to measure it.
3.9 Describe why and how to measure urea and the body fluids it can be measured in.
3.10 Describe why and how to measure glucose from a drop of blood.
3.11 Describe how to measure glucose in automated equipment.
3.12 Describe amperometry as used in the PO2 electrode.
3.13 Describe the most common enzymatic electrode method for measuring glucose.
3.14 Calculate the pH for a hydrogen ion concentration of 10–7 mol/L.
3.15 Draw a pH electrode and explain its principle of operation. Explain why its ampli-
fier input impedance is important. Explain the relation of the CO2 electrode to the
pH electrode.
3.16 Explain the principle of operation and give an example of use for flame photometry.
3.17 Explain the principle of operation and give an example of use for mass spectrome-
try.
3.18 Explain why and how CO2 is measured by infrared transmission spectroscopy.
3.19 Explain why and how N2 is measured by emission spectroscopy.
3.20 Explain why and how fluorometry is used. Describe one of the advantages of
fluorometry.
3.21 Explain why and how chromatography is used. Explain the two principles that are
the primary factors affecting interactions in chromatography.
3.22 Explain how the glucose sensor minimzes sensitivity to PO2 variations.
3-40 Chapter 3 Analysis of Molecules in Clinical Medicine
3.23 Explain why and how electrophoresis is used. List the factors that cause differential
migration rates among the component molecules of a mixture.
3.24 Explain how protein purity is measured.
3.25 Explain how the DNA code is determined.
4
Surface Characterization in Biomaterials and Tissue
Engineering
Jorge E. Monzon
This chapter discusses measurements of molecular variables that are needed for research
and development in the rapidly growing areas of biomaterials and tissue engineering.
Over the years, a number of different types of materials have been used in medical
practice, either for treating wounds and diseases or for cosmetic purposes.
A biomaterial is a synthetic material used to replace part of a living system or to function
in intimate contact with living tissue (Park and Lakes, 1992). This definition suggests that
the development of materials for biological use can be accomplished only through an
integrated approach of several disciplines: science and engineering, especially material
sciences for detailed study and testing of the structural properties and phenomena of
materials; biology and physiology, to provide the necessary tools for experimentation
based on immunology, anatomy, cellular and molecular biology; clinical sciences, as
many medical specialties are directly related to the actual use of biomaterials (Park,
1995).
A device fabricated with a biomaterial that is placed inside the living system is called an
implant.
Materials for biological use can be classified according to their base structure as ceramics,
composites, metals, and polymers. Table 4.1 summarizes different types of biomaterials
and their most common applications (Silver, 1994; Park and Lakes, 1992). This listing is
not comprehensive, as the search for new materials and their clinical applications is a
dynamic process. Synthetic polymers constitute the vast majority of biomaterials used in
humans (Marchant and Wang, 1994).
Table 4.1 Classification of biomaterials in terms of their base structure and some of their most
common applications.
Biomaterials Applications
Figure 4.1 These titanium-alloy joint replacements are an example of the many applications for
metal biomaterials for implantations. (from
http://www.spirebiomedical.com/Biomedical/ionguard.html)
4.1.3 Polymers
Structure
CH
Figure 4.2 Polymers are made up of many monomers. This is the monomer for poly(ethylene), a
common biomaterial used for medical tubing and many other applications.
The average number of repeating units is called the degree of polymerization
(Lee et al., 1995). A larger number of these subunits (i.e. bigger molecule) leads to a
4 Surface Characterization in Biomaterials and Tissue Engineering
polymer of higher molecular weight, and thus to a more rigid material (Park and Lakes,
1992).
Synthetic polymers are obtained through two basic chemical processes: addition
polymerization and condensation polymerization. Some addition polymers are
poly(ethylene), poly(methyl methacrylate), poly(vinyl chloride), and poly(ethylene
terephtalate). Condensation polymers include poly(esters), poly(amides), and
poly(urethanes) (Marchant and Wang, 1994; Park and Lakes, 1992; Silver, 1994).
Complex chemical processes beyond the scope of this book allowed the
development of polymers of very different physical and chemical properties to find
assorted applications in health care. Table 4.2 illustrates the most common polymers and
their clinical uses.
Table 4.2 The clinical uses of some of the most common biomedical polymers relate to their
chemical structure and physical properties.
Poly(ethylene) (PE)
Low density (LDPE) Bags, tubing
High density (HDPE) Nonwoven fabric, catheter
Ultra high molecular weight
(UHMWPE) Orthopedic and facial implants
Properties
Biocompatibility
The most important requirement for an implant is its biocompatibility, that is, the
acceptance of an artificial device by the surrounding tissues and by the body as a whole
(Park, 1995). Greisler et al. (1996) suggest that biocompatibility does not mean lack of
biological response—as might be expected—but rather the presence of a desirable set of
biological responses.
Bulk and surface properties of polymer biomaterials, which are functions of
their molecular structure and organization, determine the interaction with living
organisms (Marchant and Wang, 1994).
Figure 4.3 This artificial heart valve is coated with a biocompatible material that allows the body
to accept the implant. (from
http://www.sjm.com/products/heartvalves/mechanical/mastersseries.shtm)
4.2 Molecules and Tissue Engineering
As is the case with biomaterials, molecular phenomena also play an important role in
tissue engineering research. We next present some basic aspects of this engineering field.
Tissue engineering can be defined as the application of engineering and life sciences
toward the design, growth, and maintenance of living tissues (Berthiaume et al., 1994).
This discipline uses living cells and their extracellular products to develop more
biological replacements instead of using only inert implants (Skalak et al., 1988). Tissue
engineering activities can be performed either in vivo (in the body) or in vitro (in
solution), the latter being of more interest to bioengineers.
Bioartificial tissues are constructed in vitro, are composed of biologic and
synthetic substances, and represent a valid alternative to organ transplantation in cases of
impaired natural organ function, vital organ failure, or organ donor unavailability.
Although many bioartificial tissues are still in a developmental stage, it is expected that
in the future bioartificial organs and tissues will be able to regenerate and perform
complex biochemical functions, similar to the natural organs and tissue they will replace,
a behavior that pure artificial implants cannot exhibit (Berthiaume et al., 1994).
Tissues are multiphase systems of cellular composites. We can identify three main
structural components: cells organized into functional units, the extracellular matrix, and
scaffolding architecture (Wintermantel et al., 1996).
The three-dimensional structures that characterize bioartificial tissues represent
a cell mass that is greater, by several orders of magnitude, than the two-dimensional cell
cultures traditionally developed by biologists (Berthiaume et al., 1994). Tissue
engineering also concentrates on the development of in vitro models that overcome the
absence of an adequate extracellular matrix in conventional cell cultures (Sittinger et al.,
1996).
Table 4.3 shows some current examples of applications and in vitro research in
tissue engineering (Berthiaume et al., 1994; Palsson and Hubbell, 1995; Skalak and Fox,
1988).
Table 4.3 Some examples of current applications in tissue engineering. Not all of the listed
applications are at the same developmental stage.
4.3.1 Topography
Andrade (1985c) points out that the very first characterization to be performed on any
surface is to look at it. Some properties such as color or reflectivity can be observed with
the naked eye. Microscopes are needed for observing other qualities such as roughness,
surface texture, or topography (or microtopography for a molecular scale).
Major developments in topographic studies of synthetic and biologic materials
were achieved after the invention of the electron microscope. Next we discuss the
different types of electron microscopes that are used for observation of surface
topographies.
This is the classical electron microscope. Developed several decades ago, TEM made
possible a relevant increase in magnification compared to visible light optical
microscopes. Instead of using a beam of light as the image-forming radiation, TEM relies
on an electron beam that passes through the sample in a vacuum medium to develop a
transmitted image of the specimen.
h
λ= (4.1)
mv
where h is Planck’s constant, m is the mass of the electron, and v is the velocity. Better
resolution is achieved by using lower wavelengths. It is therefore possible to increase the
resolving power of the microscope by increasing the electron velocity and thus lowering
λ. Electrons accelerated in an electric field will reach a velocity:
2eV
v= (4.2)
m
where V is the accelerating voltage, e is the charge of the electron, and m is its mass. By
substitution in Eq. (4.1), and by replacing the constants by their values (see Appendix),
we obtain
1.22
λ= nm (4.3)
V
which shows that short wavelengths can be obtained by using high voltages (on the order
of kilovolts). Equation (4.3) is not useful for calculating the resolution because lens
aberrations, refractive index of the medium, and aperture angles limit the resolution.
However, it is possible to demonstrate that the highest practical resolution of TEM is 0.4
to 1 nm (4 to 10 Å), with a practical magnification of 100,000 to 200,000.
Operation: Figure 4.4(a) shows the basic components of a TEM. Not shown is the overall
vacuum system for proper operation. The simplest electron source is a heated, pointed
tungsten wire, although higher resolution microscopes are equipped with a lanthanum
hexaboride (LaB6) cathode to which typically voltages of 40 to 120 kV are applied
(Dykstra, 1992).
10 Surface Characterization in Biomaterials and Tissue Engineering
Figure 4.4 (a) TEM microscope. The electron beam passes through the sample, generating on the
fluorescent screen a projected image of the sample, which can be recorded by photographic means.
(b) SEM microscope. Condenser lenses focus the electron beam on the specimen surface leading to
secondary electron emission that is captured by the detector and visualized on the CRT screen.
Both TEM and SEM operate in a particle free (vacuum) environment.
A set of condenser lenses is used to focus electrons onto the specimen on the
area under examination. Lenses for an electron microscope are not made of glass as for
optical microscopes, but rather they are magnetic coils (solenoids) that can bend the
electron path (Bozzola and Russell, 1991).
4.3 Surface Analysis 11
Below the sample plane, the objective lens assembly focuses and magnifies the
specimen image. By changing the aperture of the objective lens it is possible to control
the specimen contrast and to correct astigmatism. Scattered elements of the electron beam
emerging from the sample are also eliminated by the objective lens system (Dykstra,
1992).
To control the magnification of the image being projected, it is necessary to use
a projector lens, which will also focus the beam of electrons for appropriate intensity
upon the fluorescent screen. This screen can be observed through a viewing window (not
shown) or the image can be photographically reproduced by exposing a photographic
plate immediately beneath the fluorescent screen (Packer, 1967).
SEMs and TEMs have many features in common—they both use electron beams to
visualize a sample. However, they differ in the way the beam interacts with the specimen.
The principle of SEM operation makes it very useful for topographic analysis, as we will
see in the next paragraphs.
Principle: SEM is based upon the interaction of an electron beam with a specimen. The
incident beam (primary electrons) displaces orbital electrons from the sample atoms,
giving rise to secondary electron emission (Figure 4.5)
12 Surface Characterization in Biomaterials and Tissue Engineering
Incident beam
(primary electrons)
Secondary
electrons
Backscattered
electrons
Figure 4.5 Principle of SEM operation. An incident beam of primary electrons displaces orbital
electrons from the sample atoms resulting in secondary electron emission, which is detected for
image formation. Some primary electrons pass by the nucleus to become backscattered electrons.
The number of secondary electrons emitted depends on the electron density of
the specimen, which is related to its elemental composition, so suitable detection of this
secondary emission yields information on the atoms of the surface sample. Some primary
electrons may not collide with orbital electrons. Instead they pass by the nucleus to
become backscattered electrons (Dykstra, 1992).
These are not the only emissions arising at the atomic level, as we explain in
section 4.3.2 (see Figure 4.10), but these two are the ones of interest for SEM operations.
Resolution and images: Image resolution in SEM is a more complex concept than in
TEM because time is an integral part of the image formation process in scanning
microscopes (Dykstra, 1992).
Synchronization of beam and CRT plate deflectors account for simultaneous
emission of secondary electrons and their imaging on the CRT screen.
However, some backscattered electrons, as they leave the atom (Figure 4.5) and
just prior to their exit, may interact with electrons some distance from the initial point of
entry in the specimen, which will produce secondary electron emission. These extraneous
secondary electrons produced by backscattered radiation are also detected by the
collector and displayed on the CRT screen, giving a blurred or noisy signal. One way of
reducing the noise is by increasing the true signal-to-noise ratio. Some backscattered
electrons can also be collected by a special detector. By carefully adjusting the condenser
lenses it is possible to obtain high resolution with suitable depth of field for working
distances (distance from the final condenser to the specimen) of approximately 10 nm
(Bozzola and Russell, 1991).
A very attractive feature of the SEM is its ability to generate images that appear
to have three dimensions. The positioning of the primary scanning beam (incident angle)
and the location of the detector will affect the image contrast, and therefore giving a
sense of three dimensions. Secondary electrons are emitted from various parts of the
specimen, but those that face the primary beam and are in line of sight with the detector
will appear bright on the CRT screen, whereas a dark appearance corresponds to areas
not reached by the beam, and not emitting secondary electrons. Medium brightness
represents partial specimen exposure to the beam, with detection of fewer secondary
electrons. This variation in contrast contributes to the generation of an image that appears
to have depth, although that image is visualized on a two-dimensional CRT screen
(Bozzola and Russell, 1991).
Sample preparation: As for the TEM, specimen preparation for SEM viewing is also a
lengthy process that must be carefully carried out (Dykstra, 1992).
The three dimensional effect on SEM-generated images is very useful for
topographical studies. Usually low voltage SEM (~5 kV) provides similar surface
sensitivity to that obtained with XPS (see section 4.3.2). The sample can be probed to
depths of tenths of micrometers making it possible to convert a chemical structure into
topographical relief. In this way, SEM may be used to view a topography that is a
function of the structure with very high surface resolution (Goodman et al., 1988).
Samples must be stable in the high vacuum environment. Mostly dry samples
are used. Additionally, specimens should be conductive (noncharging), therefore a thin
metal coating on the sample is needed (Andrade, 1985c). The remarkable depth of focus
allows the examination of uneven surfaces with the SEM. The analysis of elemental
14 Surface Characterization in Biomaterials and Tissue Engineering
The scanning tunneling microscope, along with the scanning force microscope (SFM)
that we discuss later, belongs to the more general group of scanning probe microscopes
(SPM or SXM).
Very small sensors (probes) scan the surface of the sample in close proximity. In
this way, it is possible to obtain information on surface topography, and on some
mechanical and electronic properties at the atomic level (Marti, 1993).
The STM was initially designed to give a clearer view of the atomic nature of
the surface of solids. Properties of the surface derive from atoms not being completely
surrounded by other atoms, as they are in the interior of a solid (Dykstra, 1992).
Principle of operation: STM operation is based upon the electron tunneling effect, which
is governed by quantum mechanics. This rather complicated function of kinetic and
potential energies—which is related to the probability of electrons crossing a gap
between two conducting surfaces—can be synthesized in our discussion of the STM by
saying that if a voltage is applied between the probe tip and the sample surface, a current
will develop across the junction (tunneling current).
This is due to the fact that a cloud of electrons exists above the specimen surface
and the number of electrons decreases exponentially with the distance from the surface.
Distances of an atomic diameter can cause a relevant decrease in the tunneling current,
thus allowing precise measurements of the vertical position of atoms at the specimen
surface (Dykstra, 1992). Images are formed by monitoring voltages, tunneling currents
and/or probe positions.
The resolution of the STM is a function of the geometry of the tip, the
topography of the sample, and of their electronic structures. For large objects, in the
micrometer range, the geometry of the tip plays a leading role. However, on an atomic
scale, resolution is governed by the stability of the tunnel barrier width (i.e. the distance
between tip and sample). Lateral resolution of the STM (as the tip scans the specimen
surface) is related to the diameter of the tunneling current filament (Marti, 1993).
Calculation of STM resolution is not a straightforward process. The gap between
probe and specimen is typically 1 nm. Stable and precise STMs can measure distance
variations of 1 pm, provided the instrument is in a vibration-free environment (Dykstra,
1992).
4.3 Surface Analysis 15
(a)
(b)
Figure 4.6 (a) An STM probe tip made of tungsten magnified 4,000 times. The tip is very small,
and can be ruined on a sample, which is seen in Figure 4.6(b). (from
http://www.orc.soton.ac.uk/~wsb/stm/photos.htm)
STM instrument: Figure 4.8 shows a block diagram of the STM. Piezotranslators are used
to move the sample relative to the tip rather than vice versa. The piezoelectric effect is
the mechanical expansion and contraction of a material in response to an electric field. In
this way electric voltages are used to generate a three dimensional (3-D) movement of the
piezoscanner holding the probe. A widely used piezoscanner made of ceramic material is
the piezotube, which has inner and outer electrodes for movements along the three
Cartesian axes x, y, and z. The distance between the tip and the sample is kept at 1 nm.
The sample is connected to a tunnel voltage source. The tunneling current from the
sample to the tip is fed into a current-to-voltage converter. The output of the I/V converter
is further processed at the voltage processor block for display. This voltage is also fed
back to the system via the scanner voltage block, which provides the necessary voltages
for movements along the z axis. A scan controller generates the voltages needed for
movements along the x–y plane. Visualization and interpretation of the data provided by
16 Surface Characterization in Biomaterials and Tissue Engineering
the STM must be accomplished using image processing methods and algorithms (Marti,
1993).
A
Sliding
B surfaces
Sample
holder cage
Figure 4.7 This is a sample of a piezotube. There are different approaches, but all use the same
method of two opposing piezoelectric materials to move the sample in each axis. (from
http://www.topac.com/cryosxm.html)
I/V converter
Tip
Sample
Tunnel Voltage
voltage source processor
Piezotube
Scanner voltage
Figure 4.8 STM schematics. The tip of a probe scans the surface of the sample. Three-dimensional
movements of the sample under the tip are accomplished using a voltage-controlled piezoscanner.
The tunneling current crossing from the sample to the tip is further processed leading to a
topographical image.
4.3 Surface Analysis 17
Modes of operation: Constant current and constant height are the two modes of STM
operation. In the constant current mode, very high-gain voltage amplifiers are used, so as
to keep the tunneling current constant and to force the tip to follow the contours. By
recording the input of the scanner voltage block (in fact the z voltage amplifier) as a
function of x and y it is possible to determine the topography of the specimen surface.
Any suitable recording device can be used for such purpose: chart recorder, oscilloscope,
or computer. The constant height mode of operation is appropriate only for flat surfaces.
In this mode the tip is held at a constant height and no longer follows the sample
topography. The tunneling current is then recorded as a function of x and y (Marti, 1993).
Three disadvantages of the STM are the need for an environment free of
vibrations, very stable temperatures and the requirement for highly conductive samples to
develop tunneling currents.
The low cost and simplicity of the STM are attractive. However, careful data
acquisition and interpretation are highly recommended (Ratner, 1988).
The scanning force microscope (SFM), also known as an Atomic Force Microscope
(AFM) was developed after the STM. They share many of the same features. Other types
of force microscopes are the magnetic force microscope, the attractive mode force
microscope, the friction force microscope, and the electronic force microscope. All of
these microscopes measure forces generated between a probe and a sample surface
(Dykstra, 1992).
Theory of the SFM: As with the STM, operation of the SFM is governed by principles of
quantum mechanics. The force between a tip and the sample is used to obtain an image of
the surface topography. In the absence of other magnetic or electrostatic potential, this
small force—on the order of 100 nN—depends upon the interaction potentials between
the atoms of the tip and of the sample (Marti, 1993).
Contact SFMs require the tip to actually touch the sample. These are the most
common microscopes, although noncontact SFMs are also available (Bustamante et al.,
1996).
Instrument operation: Figure 4.9 shows the SFM. A sharp probe tip is mounted on a
flexible cantilever. As the tip moves toward the sample, two types of forces develop: at
large distances the interaction is attractive due to the van der Waals forces; at short
distances repulsive forces develop due to the exclusion principle of quantum mechanics
(Marti, 1993).
18 Surface Characterization in Biomaterials and Tissue Engineering
Laser Two-segment
photodetector
Lens
Mirror
Cantilever
Sample
Piezoscanner
Figure 4.9 Sketch of an SFM. A laser beam is focused on the cantilever, and reflected back to a
two-segment photodetector. The difference in output from each segment is proportional to the
deflection amplitude of the cantilever scanning the sample.
The attractive and repulsive forces between tip and sample lead to downward
and upward deflection of the cantilever, respectively. Although the tip radius is very
small (~10 nm), the forces acting on it can deflect the entire cantilever. As the sample
scans under the probe, the cantilever is deflected in correspondence with the surface
topography (Bustamante et al., 1996).
Several methods are used to detect cantilever deflections of less than 1 nm. The
most common method is the optical lever, shown in Figure 4.9. A laser beam is focused
on the cantilever and reflected back to a photodiode in the middle of two subdiodes (two-
segment photodetector). Any deflection of the cantilever will result in an unbalance in the
number of photons hitting each of the subdiodes, which in turn will unbalance the electric
currents in the photodiodes. This differential current signal at the output of the
photodetector is then proportional to the cantilever deflection and thus to the surface
topography. The differential current signal is amplified for further processing and also
used in a feedback loop (not shown in Figure 4.9) to control the movements of the sample
with respect to the tip according to the modes of operation discussed next. The optical
lever acts as a motion amplifier, as a 100 pm deflection of the cantilever can easily be
detected by the differential photodiode (Bustamante et al., 1996; Marti, 1993).
4.3 Surface Analysis 19
Modes of operation: Similar to the STM, piezoelectric elements are used to control the
motion of the sample or the tip.
In the constant force mode, the contact force between probe and sample is kept
constant by means of an electronic feedback loop that controls the piezoscanner. We said
previously that these forces are in the nanonewton range, but due to the action on tiny
areas of contact (about 0.1 nm2), very high pressures develop. As the tip scans the surface,
an upward deflection of the cantilever indicates a topographic obstacle and the sample
retracts, by means of the feedback loop, to keep the force constant. The movement of the
sample correlates to the surface topography. In the constant height mode the cantilever
can deflect in both directions, and the sample is kept at a constant height. In this mode,
direct recording of the cantilever deflection is performed. A combined mode, the error
signal mode, requires fast feedback response to follow the surface as in the constant force
mode, but recordings correspond to cantilever deflection and not to the sample motion.
This mode allows enhancement of sharp features (Bustamante et al., 1996).
Thermal drift is not usually a problem using optical detectors, although careful
calibration is required to focus and position the laser beam on the cantilever—an
operation that requires the help of optical microscopes. In contact SFM (repulsive force
operation), some lateral forces between the tip and the sample may develop, giving
distorted images and thus reduced resolution. These lateral tensions deform the sample.
The magnitude of the lateral forces is a function of the tip dimension and shape, so
careful design of the tip is necessary. One possible solution to the unwanted shearing
forces is offered by noncontact SFMs. Although very useful for imaging soft samples,
they exhibit some mechanical instability due to their attractive force operation
(Bustamante et al., 1996; Marti, 1993).
The AFM is widely used in surface analyses of biomaterials. It has great
advantages over the STM in that it does not need electrically conducting samples. The
AFM can work very easily in liquid (Descouts, 1995).
Researchers rely on XPS’s high surface specificity, which takes its information
from within 10 nm (100 Å) of the surface, because it allows the studying of biomaterials
interfaces with living tissue. Concentration and chemical state of every element in the
sample can be obtained using XPS (Paynter, 1988).
XPS is based on the photoelectric effect, shown in Figure 4.10 When an X-ray
photon interacts with an atomic orbital electron, if the photon carries enough energy to
overcome the binding energy of the electron, the electron is liberated from the atom as a
photoelectron. The basic equation for this photoemission process and thus for XPS is
E k = hν − E b − φ s (4.4)
(a) Photoelectron
X-ray photon
(b)
Auger electron
Figure 4.10 When an X-ray photon (a) interacts with an atomic orbital electron of the sample, a
photoelectron (b) is emitted verifying Eq. (4.4). The now unstable atom must relax to the ground
state. The relaxation process can be accomplished by either of two mechanisms: (1) an outer orbital
electron releases energy as fluorescent radiation (c) while occupying the place of the emitted
photoelectron, or (2) the excess energy is used to unbind and emit another outer orbital electron
called an Auger electron (d). These mechanisms operate for different sample depths, yielding the
Auger electron emission characteristic of the outermost surface of the sample.
4.3 Surface Analysis 21
The interaction of a monochromatic X-ray photon beam with the surface of the
sample will result in photoelectrons produced with different kinetic energies according to
their orbital position in the atom (i.e. different bonding energies). An electron analyzer,
also known as an electron spectrometer, can measure the number of these electrons.
Processing the electrons’ kinetic energies yields a spectrum of photoelectron intensity as
a function of binding energy.
Figure 4.11 shows a typical XPS spectrum. Distinct peaks correspond to specific
atomic orbitals so that the binding energy position of each peak allows elemental
identification. Tabulations of orbital binding energies are readily available and usually
preprogrammed into the data system of the instrument. All of the elements in the periodic
table exhibit binding energies in the approximate range of 0 to 1000 eV (electron volts),
therefore positive identification is possible. Further quantization can be achieved by
measuring the area under the peaks (Andrade, 1985b; Paynter, 1988). Many detected
electrons lose energy in interactions with the sample, leading to lower kinetic energies
and higher binding energies recordings (see Eq. (4.4)) adjacent to every intense peak,
giving the spectrum a stair-step shape (Andrade, 1985b).
100
Intensity (%)
Figure 4.11 A typical XPS spectrum, showing photoelectron intensity as a function of binding
energy. Each peak may correspond to a distinct element of the periodic table or to different orbital
electrons of the same element. Some peaks may also represent Auger radiation.
Figure 4.12 shows the basic schematic of an XPS instrument. The X-ray source
provides a monochromatic photon beam that hits the surface of the sample, which is
placed on a moving holder for appropriate scanning. The resulting photoelectron
22 Surface Characterization in Biomaterials and Tissue Engineering
radiation is captured by an analyzer. The excitation and emission processes take place in
an ultrahigh vacuum (UHV) chamber (not shown in Figure 4.12) for minimal sample
contamination and for easy passage of the photoelectrons to the analyzer (Paynter, 1988).
X-ray source
Retarding grid
Sample holder
Hemispherical
electrostatic
analyzer
Slit
Display Data Detector
processor
Figure 4.12 Basic schematics of an XPS instrument. An X-ray beam strikes the sample surface,
giving photoelectron radiation. These electrons enter the hemispherical analyzer where they are
spatially dispersed due to the effects of the retarding grid and of the electrostatic field of the
concentric hemispheres. Ramping voltages at the retarding grid allow kinetic energy scanning. At
the other end of the analyzer electrons are detected, counted, and a spectrum of photoelectron
intensity versus binding energy is displayed.
The analyzer consists of two concentric and electrostatically charged
hemispheres. An electrically biased grid at the entrance of the analyzer retards the kinetic
energy of the photoelectrons, which are then spatially dispersed as they initiate their
semicircular trajectory between the hemispheres. The voltage across the hemispheres is
kept constant (fixed analyzer transmission or FAT mode), therefore the energy resolution
is also constant. Ramping voltages in the pre-retardation stage allows actual kinetic
energy scanning with a constant resolution mode of typically ~0.1 to 1.0 eV.
At the other end of the analyzer an electron multiplier detector increases by
several orders of magnitude the number of electrons reaching the data processing and
display sections, which usually consist of a phosphorescent screen and a detector device
4.3 Surface Analysis 23
that counts the number of light pulses arising at the screen (Andrade, 1985b; Paynter,
1988).
The use of an X-ray monochromator in XPS eliminates undesirable radiation
coming from the X-ray source (see Figure 4.12). Although monochromated beams
minimize the radiation damage to the sample, they also decrease the intensity of the flux
to the sample (Andrade, 1985b). The detection system must compensate for this reduced
intensity in order to obtain suitable signal levels. This is done by incorporating
multichannel photoelectron detectors (Paynter, 1988).
High cost of the instrument, large area of analysis, high vacuum required and
low speed (several minutes per peak) are major disadvantages of the XPS technique
(Andrade, 1985b). Also the behavior of a biomaterial in a vacuum may differ
considerably from its behavior when in contact with living tissues (Paynter, 1988).
The lack of spatial resolution has also been an inherent weakness of XPS. A system like
ESCALAB, shown in Figure 4.13, uses the capability of parallel imaging to produce
photoelectron images with spatial resolution better than 5 mm.
In parallel imaging method, photoelectrons are simultaneously collected from
the whole field of view. A combination of lenses before and after the energy analyzer
focus the photoelectrons on a channel plate detector. The hemispherical analyzer permits
electrons with only a narrow band of energy to reach the detector. Neither the X-ray nor
the photoelectrons are scanned to produce an image. Spectroscopy and imaging use the
same analyzer to avoid and risk of misalignment. In ESCALAB, images are formed on a
channel plate detector. When using spectroscopy, an array of channel electron multipliers
is used.
The advantage of iXPS is that it produces images much more quickly than the
conventional methods (for comparable spatial resolution) because the entire field of view
is imaged simultaneously. Using an electron flood source, it is possible to get a real time
(i.e. close to TV rate) physical image of exactly the same position as the images. This
makes setting up experiments much quicker. Moreover, the introduction of a magnetic
objective lens has improved the sensitivity for a given spatial resolution enabling the new
technique of imaging of XPS to be developed.
24 Surface Characterization in Biomaterials and Tissue Engineering
Hemispherical
analyzer
Spectroscopy
detector
Lenses
Imaging
detector
Irises
Camera
X rays
Zoom
Sample microscope
AES, based on an electron beam and not an X-ray photon beam, and Auger electron
emission (refer to Figure 4.10) are complementary to XPS for surface analysis, being
especially useful for quantitative analysis of certain elements. The Auger electron can
only be characterized by its absolute kinetic energy and not by a binding energy. AES
found practical applications earlier than XPS, but due to the need of electron
bombardment to the sample (instead of photons), it is a highly destructive technique not
recommended for modern polymer surface analysis (Andrade, 1985b). AES is also
known as SAM (Scanning Auger microprobe) (Ratner, 1988).
Secondary-ion mass spectroscopy (SIMS) is another technique for studying the nature of
biomaterial surfaces. Due to its destructive nature, it has been applied to biomedical
polymers. However, it can provide useful information for the understanding of the
composition and structure of an implant material surface (Andrade, 1985c; Castner and
Ratner, 1988).
SIMS is based on the principle of mass spectrometry, a proven sensitive and
powerful analytical technique, in which sample molecules are ionized. If enough ionizing
energy is present, sample molecules are even dissociated, forming ionic fragments. The
mass-to-charge ratios of the molecular ions and fragment ions are determined and plotted
against abundance (mass-spectrum). The pattern of product ions formed is dependent on
the structure of the intact compound (Wait, 1993).
Figure 4.14 illustrates the SIMS procedure and instrument schematics. A SIMS
instrument, and in general all mass spectrometers, consists of a means of ion generation,
a mass analyzer for their separation, and an ion detector (Wait, 1993). As with XPS,
ultrahigh vacuum (UHV) chambers are used to avoid not only external contamination but
also collisions with particles that may be present in a nonvacuum environment.
26 Surface Characterization in Biomaterials and Tissue Engineering
Ion Processing
detector
Primary Secondary
ions ions
Sample
Figure 4.14 Schematic diagram of a SIMS instrument. Bombardment of primary ions on the
sample surface leads to secondary ion emission. A mass analyzer separates these ions in terms of
their mass-to-charge ratio. The ion detector converts this ionic current into an electric signal for
further processing. The display presents the SIMS spectra, consisting of the count of ions versus
their mass-to-charge ratio.
Ion generator: The generator or gun provides a beam of charged particles that once
accelerated into the surface sample, lead to the ejection of secondary ions. The size,
energy, mass, and flux of the incident beam are directly proportional to secondary ion
emission and sample damage. A good compromise between suitable secondary emission
and damage is a 2 keV Xe+ ion beam which gives a sampling depth of approximately 1
nm (10 Å). The use of such low-energy inert gas ion beams is called static SIMS and is
the most commonly used technique. It allows the analysis of the topmost surface of
insulating materials such as polymers. A high-energy and high-flux beam technique,
called dynamic SIMS, is used to profile the depth of minerals and semiconductors
(Andrade, 1985c).
The use of a charged incident beam as mentioned previously can significantly
alter the electrical characteristics of the sample surface, particularly insulators, as surface
charge buildup will affect the secondary ion detection. One way of avoiding this charge
buildup is by using a neutral atom incident beam, a technique referred to as fast atom
bombardment (FAB) SIMS (Castner and Ratner, 1988).
4.3 Surface Analysis 27
Charge neutralization: For suitable reading of secondary ion emission, the charge that
builds up on the surface of an insulating sample must be neutralized. This can be
accomplished by a variety of devices that generate electron beams for the case of
positively charged incident ion beams. Even for the case of neutral FAB SIMS, the use of
neutralizing electrons has shown to increase the quality of detected data (Castner and
Ratner, 1988)
Mass analyzer: The counts of secondary emitted ions versus their mass-to-charge ratio
(measured in Daltons, where 1 D = 1/12 the mass of C12) constitute the data obtained
through a SIMS instrument.
The well known expression for kinetic energy in terms of mass and velocity can
be applied to an ion, for which its number of charges z (in units of the charge of an
electron) and its accelerating voltage V are known, leading to Eq. (4.5)
1 2
zV = mv (4.5)
2
m 2V
= 2 (4.6)
z v
Data processing and interpretation: Once the ions are separated according to their mass-
to-charge ratio m/z, a detector is used to convert the ion current into an electric signal by
a type of electron multiplier. The electric signal is amplified, computer processed, and
visualized.
In modern instruments, computers are used to select parts of spectra for analysis
and also to control some instrument features such as beam generation and intensity,
neutralization energies, and sample scanning (Castner and Ratner, 1988; Wait, 1993).
Static SIMS is used for determining the types, structures, and quantity of the
chemical compounds on a surface. Note that relating the numerous peaks of the SIMS
spectra to the surface chemical group from which they originated is not an easy task. This
28 Surface Characterization in Biomaterials and Tissue Engineering
ISS: Another ion beam technique called Ion Scattering Spectroscopy (ISS) consists of a
low-energy inert gas beam that interacts with the sample surface atoms. Incident ions lose
energy in these elastic collisions, a loss that depends on the mass of the atom in the
surface. There is no secondary ion emission as in SIMS, and therefore primary ions are
the ones being detected in ISS (Andrade, 1985c).
Next we present some infrared spectroscopy techniques, which are usually combined to
allow surface analysis.
Infrared spectroscopy is a very versatile, fast, and inexpensive method that leads
to conclusive molecular identification of samples in any of the three phases: gas, liquid,
or solid (Grim and Fateley, 1984).
Infrared
detector
Sliding
Beamsplitter mirror
Source
Fixed mirror
Figure 4.15 Michelson interferometer. A beamsplitter transmits half of the source radiation to the
fixed mirror and the other half to the sliding mirror. A phase difference between the beams can be
induced by sliding the mirror, which causes detection of the two beams at different times. The
detector provides the interferogram, a plot of energy as a function of differences in optical paths.
Beams have been slightly shifted in the drawing to allow easy following of their path.
Fourier Transform infrared spectroscopy (FT-IR): For a polychromatic source, the
interferogram will show a center burst (maximum) only when the two path lengths are
equal. We use equal path lengths with no movement. Then if absorption from one path is
present at one wavelength, an inverted cosine wave will appear across the interferogram.
We can use a mathematical procedure: the Fourier Transform (FT) or the Fast Fourier
Transform (FFT) to measure all wavelengths simultaneously and convert from the spatial
domain of the interferogram to the absorption spectrum of spectroscopy (Grim and
Fateley, 1984).
This operator allows the display of the interferogram (and in fact of any time
signal) in the frequency domain. The interferogram represents all wavelengths in one
graph; this is an advantage over dispersive spectroscopy in which energy levels are
measured for each wavelength at a time. Calculations for the FT for displaying
interferograms in the frequency domain were a lengthy and tedious process that has since
been all but eliminated by the use of digital computers. This powerful mathematical tool
has also allowed the use of interferometry in the far infrared region, which required
almost endless calculations. It has improved the benefits of interferometry over dispersive
4.3 Surface Analysis 31
Attenuated internal total reflection (ATR): Total internal reflection is a special case of the
more general phenomenon of reflection and refraction of incident electromagnetic
radiation at an interface of two media having different indexes of refraction (Knutson and
Lyman, 1985).
Figure 4.16 shows that when an incident beam with an angle θ traveling in a
medium of refractive index ηc encounters the interface with a second medium of
different refractive index ηs, the beam will be reflected back into the first medium and
transmitted into the second medium with an angle other than the incident (refraction)
according to Snell’s Law of Refraction:
where the indexes of refraction represent relationships of the speed of light in the medium
to the speed of light in a vacuum. From Eq. (4.7) and from Figure 4.10 it becomes
evident that if the first medium is denser that the second (i.e. ηc >ηs), then φ is greater
than θ. Also, there is a value for θ that will make φ = 90° and is called the critical angle
θc. Eq. (4.7) becomes
η
θ c = sin −1 s
(4.8)
ηc
Values of θ larger than θc will cause total reflection of the incident beam (i.e. no
refractive beam develops).
32 Surface Characterization in Biomaterials and Tissue Engineering
ηc θ
φ ηs
Figure 4.16 When an incident beam traveling at an angle θ in a medium of refractive index ηc
encounters another medium of refractive index ηs, it will reflect in a direction given by θ and
refract in the direction given by φ, verifying Snell’s Law of Refraction (Eq. (4.7)).
We have highly simplified the presentation of these phenomena. A more
rigorous treatment would have required the use of tools provided by electromagnetic
wave theory.
Biomedical polymers are usually implanted within a liquid environment. The polymer-
liquid interface in terms of surface free energy and surface tension can be accurately
measured with the contact angle method.
The contact angle method requires relatively simple and inexpensive equipment, although
it is one of the most sensitive methods used to obtain information about the outermost
layer of solid surfaces (Andrade et al., 1985).
This method allows the determination of the excess free energy per unit area, the
surface tension, whose SI unit is N/m. The surface is actually a discontinuous boundary
between three different phases: liquid, gas, and solid. A surface tension develops in these
phase boundaries when the equilibrium is altered giving rise to an excess energy, which
will minimize the surface area. Adsorption (attraction of foreign materials) or
chemisorption (bonding with the adsorbent) are other means of minimizing the surface
energy (Park and Lakes, 1992).
Figure 4.17 illustrates the three surface tension components that interact to limit
the spread of a drop on top of a rigid surface. At equilibrium, Young’s equation is
γ SV − γ SL = γ LV cos θ (4.9)
where γ is the interfacial free energy for the solid-vapor, solid-liquid, and liquid-vapor
boundaries, and θ is the contact angle (Andrade et al., 1985). Rearranging Eq. (4.9) leads
to
γ SV − γ SL
cos θ = (4.10)
γ LV
γ LV
θ
γ SV γ SL
Figure 4.17 Surface tension components of a three-phase system to limit the spread of a drop on
top of a surface. γ is the interfacial free energy for each of the phases. θ is the contact angle.
34 Surface Characterization in Biomaterials and Tissue Engineering
Adsorbed proteins may exhibit a complex behavior not only related to the direct
interaction of surface and proteins, but also to the structural re-arrangements (i.e.
conformational changes, see Section below) that proteins may undergo during the
adsorption process (Boyan et al., 1996).
Proteins are polymers of biologic origin, and among organic molecules they constitute
the largest group.
Carbon, nitrogen, oxygen, and hydrogen atoms are the basic elements in a
protein molecule; in many cases, sulfur atoms are also a part of the protein molecule.
Proteins, as we mentioned, are polymeric chains in which the subunits are called amino
acids. There are 20 different amino acids. Figure 4.18 shows the structure of an amino
acid molecule, which has a central carbon atom. An amino group (—NH2), a hydrogen
atom, and an acid group (—COOH or carboxyl group) are bonded to the central carbon
atom. Also in Figure 4.18, R represents the rest of the amino acid molecule, a part that is
different for each of the 20 amino acids.
NH2 C C OH
H O
Figure 4.18 The amino acid molecule. To a central carbon atom, an amino group, a carboxyl group
and a hydrogen atom are bonded. R represents the rest of the molecule, which is different for each
of the 20 amino acids.
Different amino acids combine to form polypeptides and these in turn give rise
to the very high number of complex structures (proteins). This combination of
polypeptides will give to each protein a distinct three-dimensional shape. This spatial
relationship between the amino acid chains giving rise to the structure of a protein is
called conformation. On the surface of a protein molecule there are hydrophobic, charged
and polar domains. The particular regions of a protein to which antibodies or cells can
bind are called epitopes. The existence, arrangement, and availability of different protein
epitopes is called organization (Chinn and Slack, 2000).
For a single protein solution, the rate of adsorption to the substrate depends upon
transport of the protein to the substrate. Four primary transport mechanisms have been
identified: diffusion, thermal convection, flow convection, and combined convection-
diffusion. For constant temperature and static systems, transport is exclusively by
diffusion, and the net rate of adsorption can be described by Langmuir’s theory of gas
adsorption (Chinn and Slack, 2000).
rA = k A C b (1 − θ ) − k Dθ (4.11)
where rA is the net rate of adsorption, kA is the adsorption rate constant, Cb is the bulk
concentration of the protein in solution, θ is the fraction of surface occupied by the
adsorbed molecules and kD is the desorption rate constant. At equilibrium, the rate of
adsorption equals the rate of desorption, so the net rate rA is zero and from Eq. (4.11),
the fractional surface coverage θ is:
kA
Cb
kD
θ= (4.12)
k
1 + A Cb
kD
adsorption, electron microscopy (section 4.3.1) and microautoradiography are used. The
use of radiolabeled proteins allows direct measurement of protein adsorbed to a substrate.
Protein adsorption involves redistribution of charge, changes in the state of hydration and
rearrangements in the protein structure (Norde, 1985). The result of these processes is a
generation of heat, which can be measured in a calorimeter. Most commercial
calorimeters are of the conduction type. The heat developed in the reaction vessel is
completely transferred through a thermopile to a surrounding heat sink of large capacity.
The voltage signal generated by the thermopile is proportional to the heat flux. In a twin
calorimeter, two vessels are within a single heat sink. The reaction takes place in one
vessel and the other vessel serves as the reference. The detectors are in opposition so the
differential signal indicates the heat from the reactor vessel and rejects disturbances from
the surroundings that affect both vessels (Feng and Andrade, 1994).
Many biologically important molecules are not functionally identical to their mirror
images. They are called optically active or chiral. These asymmetric molecules exhibit
optical rotation. That is, a vertical band of polarized light will rotate to the left or right if
passed through an optically active substance. For left- and right-circularly polarized light
(LCP and RCP), they exhibit circular dichroism (CD). This is the difference in absorption
of LCP and RCP light (Hatano, 1986). Measurements can also be carried out in the
infrared (Ferraro and Basile 1985).
Most proteins fluoresce naturally due to the intrinsic fluorescence of one or more of the
aromatic amino acids comprising them: tyrosine (Tyr), phenylaline (Phe), or tryptophan
(Trp) (Hlady et al., 1985). Light from xenon, or other light source, passes through an
excitation monochromator to emit a narrow band of usually ultraviolet wavelengths. The
protein in a flow cell fluoresces at a longer wavelength. The emission passes through an
emission monochromator and is measured by a photomultiplier.
Ellipsometry
Ellipsometry operates on the principle that light changes its polarization upon reflection
from a solid–solution interface. It requires optically flat, reflective surfaces and provides
information on the thickness and optical contents of adsorbed film as a function of time.
Morrissey (1977) used ellipsometry to measure differences in adsorbed amount and
thickness of gamma globulin onto flat silica plates.
38 Surface Characterization in Biomaterials and Tissue Engineering
Autoradiography
Radiolabeling
Radiolabeling permits measurement of the affinity for a given protein onto materials
(Baquey et al., 1991). During testing, peristaltic pumps cause biological fluids to flow
through tubing segments of material samples such as PVC, silicone elastomer,
polyurethane or polyethylene. Human albumin may be labeled with 99mTc or radio-
iodinated. Detectors based on junctions of semiconductors are preferred because of their
good energy resolution. A scintillation camera (used for viewing emissions of radioactive
substances) can be used to image the amount of adsorbed protein on a given surface and
to study its distribution on this surface.
where:
G(τ) = ACF
I(t) = intensity at time t
I(t + τ) = intensity at (t + τ)
τ = delay time
< > = time average
G(τ)
0
τ
Figure 4.19 The autocorrelation function G(τ) is 1 when two signals have delay time τ = 0, then
decays to 0 for long delay time.
A PCS measurement can be thought of as a measure of the time it takes for a
particle to diffuse a distance 1/K, where K is the scattering vector
sin (θ / 2)
K = 4πn (4.14)
λ
where:
n = refractive index of the diluent (n = 1.333 for water)
θ = scattering angle
λ = wavelength (λ = 633 nm for He–Ne laser)
where:
G(τ) = ACF
Γ = decay constant of particle
Note that when G(τ) = 0.37, 1/(2Γ) = τ. We can then use Γ to calculate the
diffusion coefficient DT = Γ/K2. From the diffusion constant, we can use the Stokes–
Einstein equation to calculate the diameter of a spherical particle.
40 Surface Characterization in Biomaterials and Tissue Engineering
kT
DT = (4.16)
3πηd
where DT is the diffusion coefficient, the Boltzmann constant is k = 1.38 × 10–23 J/K, T is
the absolute temperature, η is the viscosity of the medium (η = 0.001 Pa·s for water at 20
°C), and d is the diameter of the particle.
Example 4.1 Given autocorrelation function data obtained from experiment, calculate the
particle size of polystyrene latex in the sample solution. The data were obtained at a
temperature of 20 °C with θ = 90° and λ = 632.8 nm. Table 4.4 gives the data for τ and
the corresponding intensity. The solution is considered water (n =1.33, η = 0.001 Pa·s).
τ G(τ)
0 1
0.001 0.35
0.002 0.15
0.003 0.051
0.004 0.012
The data from Table 4.4 yield an exponential plot, but it doesn’t provide a useful
mathematical model for calculation. Therefore, take a natural log of the data and then
linearize the data using mathematical software or, for example, LINEST in Microsoft
Excel to obtain a linear equation. The given data yield the linear approximation
ln (G (τ )) = −1077.2(τ ) + 0.0852
To calculate the diffusion coefficient, Γ needs to be calculated first. Since G(τ) = 0.37
when 1/(2Γ) = τ, the linear approximation can be used to find τ when G(τ) = 0.37.
ln (0.37) − 0.0852
τ= = 1.00209 * 10 −3
− 1077.2
Finally obtain the diffusion coefficient DT = Γ/K2 =1.427×10–12 and solve for the particle
diameter using the rearranged Stokes–Einstein equation, Eq. (4.16)
4.5 Molecular Size 41
Lee, H. B., Khang, G., Lee J. H. 2000. Polymeric biomaterials. In J. D. Bronzino (ed.)
The Biomedical Engineering Handbook. 2nd ed. Boca Raton, FL: CRC Press.
Marchant, R. E. and Wang, I. 1994. Physical and chemical aspects of biomaterials used in
humans. In R. S. Greco (ed.) Implantation Biology: The Host Response and
Biomedical Devices. Boca Raton, FL: CRC Press.
Marti, O. 1993. SXM: an introduction. In O. Marti and M. Amrein (eds.) STM and SFM
in Biology. San Diego, CA: Academic Press.
Morrissey, B. W. 1977. Adsorption of proteins at solid surfaces. Ann. N. Y. Acad. Sci.,
288: 50-64.
Norde, W. 1985. Probing protein adsorption via microcalorimetry. In J. D. Andrade (ed.)
Surface and Interfacial Aspects of Biomedical Polymers, Vol. 2 Protein adsorption.
New York: Plenum.
Palsson, B. O. and Hubbell, J. A. 2000. Tissue engineering. In J. D. Bronzino (ed.) The
Biomedical Engineering Handbook. 2nd ed. Boca Raton, FL: CRC Press.
Packer, L. 1967. Experiments in Cell Physiology. New York: Academic Press.
Pagana, K. D. and Pagana, T. J. 1997. Mosby’s Diagnostic and Laboratory Test
Reference. 3rd ed. St. Louis: Mosby.
Park, J. B. 2000. Biomaterials. In J. D. Bronzino (ed.) The Biomedical Engineering
Handbook. 2nd ed. Boca Raton, FL: CRC Press.
Park, J. B. and Lakes, R. S. 1992. Biomaterials: An Introduction. 2nd ed. New York:
Plenum.
Paynter, R. W. 1988. Introduction to x-ray photoelectron spectroscopy. In B. D. Ratner
(ed.) Surface Characterization of Biomaterials. Amsterdam: Elsevier Science
Publishers B.V.
Ratner, B. D. 1988. The surface characterization of biomedical materials: how finely can
we resolve surface structures? In B. D. Ratner (ed.) Surface Characterization of
Biomaterials. Amsterdam: Elsevier Science Publishers B.V.
Ratner, B. D., Hoffman, A. S., Schoen, F. L. and Lemons, J. E. (eds.) 1996. Biomaterials
Science. San Diego: Academic Press, 1996.
Sheffield, W. D. and Matlaga, B. F. 1986. Evaluation by transmission electron
microscope. In A. F. von Recum (ed.) Handbook of Biomaterial Evaluation:
Scientific, Technical and Clinical Testing of Implant Materials. New York:
Macmillan.
Silver, S. H. 1994. Biomaterials, Medical Devices and Tissue Engineering: An Integrated
Approach. London: Chapman & Hall.
Sittinger, M., Bujia, J., Rotter, N., Reitzel, D., Minuth, W. W. and Burmester, G. R. 1996.
Tissue engineering and autologous transplant formation: practical approaches with
resorbable biomaterials and new cell culture techniques. Biomaterials, 17 (3): 237–
42.
Skalak, R. and Fox, C. F. (eds.) 1988. Tissue Engineering. New York: Alan R. Liss.
Skalak, R., Fox, C. F. and Fung, B. 1988. Preface. In Skalak, R. and Fox, C. F. (eds.).
Tissue Engineering. New York: Alan R. Liss.
Spire Biomedical. Biotreatments. [Online] http://www.spirecorp.com/bio_treatments.htm
Spire Biomedical. Biotechnologies. [Online]
http://www.spirecorp.com/bio_technologies.htm
44 Surface Characterization in Biomaterials and Tissue Engineering
4.7 Problems
4.1 Search the literature to find a biomaterial listed in Table 4.1 and describe its
application.
4.2 Search the literature to find a polymer listed in Table 4.2 and describe its
application.
4.3 Describe how polymers are made more rigid.
4.4 Describe the three main structural components of cellular composites.
4.5 Search the literature to find an application listed in Table 4.3 and describe it.
4.6 State the highest practical resolution of TEM and calculate the accelerating
voltage required to achieve it. Compare with typical accelerating voltage.
4.7 Describe the principle and sample preparation for a TEM.
4.8 Describe the principle and sample preparation for a SEM.
4.9 Describe the principle and sample preparation for 2 modes of STM.
4.10 Describe the principle and sample preparation for 2 modes of SFM.
4.11 Describe the principle and sample preparation for XPS.
4.12 Describe the principle and sample preparation for AES.
4.13 Describe the principle and sample preparation for SIMS.
4.14 Describe the principle and sample preparation for ISS.
4.15 Describe the principle and sample preparation for dispersive infrared
spectroscopy.
4.16 Describe the principle and sample preparation for interferometric infrared
spectroscopy.
4.17 Describe the principle and sample preparation for FT-IR.
4.18 Describe the principle and sample preparation for ATR.
4.19 Describe the principle and sample preparation for FTIR-ATR.
4.20 Describe the principle of the contact angle method.
4.21 Describe the principle of DSC.
4.22 Describe the principle of CD.
4.23 Describe the principle of TIRF.
4.24 Describe the principle of ellipsometry.
4.25 Describe the principle of autoradiography.
4.26 Describe the principle of radiolabeling.
4.7 Problems 45
4.27 Given the data in section 4.5 and when G(τ) = 0.37,τ = 100 ns, calculate particle
diameter d.
4.28 We wish to view the topography of a biomaterial surface and the molecules that
adsorb to it in a liquid. Select an instrument for imaging. Name other
instruments you rejected and why you rejected them. Describe the principle and
operation of your selected imaging instrument.
6
Cellular Measurements in Biomaterials and Tissue
Engineering
Jeffrey S. Schowalter
This chapter discusses the different types of instrumentation used in the investigation of
biological processes on the cellular level. Major advances have been made over the past
several years allowing for the observation of cells and their structure. Many of the mys-
teries of how various inter- and intracellular processes work have been solved in recent
years due to these advances in bioinstrumentation. One area that has benefited the most is
tissue engineering and the development of biomaterials. Tissue engineering draws from a
number of different disciplines including cell biology. Researchers are using live cells,
chemicals, and other synthetic materials to develop biomaterials for implantation in the
human body (Lewis, 1995). But a key reason researchers have been able to make this
progress is due to the significant advances in the area of cellular bioinstrumentation.
A cell is the most basic structural unit that makes up plants, animals, and single cell or-
ganisms. It is the basic unit of life and all organisms are composed of one or more cells.
6-1
6-2 Chapter 6 Cellular Measurements
There are two types of cells: those with a nucleus (eukaryotic cells) and those without
(prokaryotic cells). Eukaryotic cells are composed of three main features: a nucleus, sur-
rounded by cytoplasm, which in turn is surrounded by a sack-like plasma membrane.
This membrane is made from a phospholipid bilayer and is semipermeable, which means
it allows only certain things to pass in or out of the cell. Cells are 90% fluid (cytoplasm),
which consists of amino acids, proteins, glucose, and numerous other molecules. On an
elemental level, the cell contains 59% hydrogen, 24% oxygen, 11% carbon, 4% nitrogen,
and 2% others including phosphorus and sulfur. On a molecular level, the cell contains
50% protein, 15% nucleic acid, 15% carbohydrates, 10% lipids, and 10% other. Table 6.1
shows the contents of the typical cell.
Structure Description
Cytoplasm The inside of the cell not including the organelles.
Organelles Membranous sacs within the cytoplasm.
Cytoskeleton Structural support made of microtubules, actin and intermediate fila-
ments.
Endoplasmic Reticulum Site of protein and lipid synthesis and a transport network for mole-
(ER) (two types) cules.
Rough ER Has ribosomes, and tends to be flat.
Smooth ER Does not have ribosomes and is tubular.
Golgi Apparatus Modifies molecules and packages them into small membrane bound
sacs called vesicles.
Lysosomes Main point of digestion.
Microtubules Made from tubulin and make up centrioles, cilia, cytoskeleton, etc.
Mitochondria Site of aerobic respiration and the major energy production center.
Nucleus Location of DNA; RNA transcription.
Peroxisomes Use oxygen to carry out catabolic reactions.
Ribosomes: Located on the Endoplasmic Reticulum in the cytoplasm. RNA goes
here for translation into proteins.
To provide a rough idea of the scale that cell biologists and other related re-
searchers are working on, Table 6.2 shows the typical sizes for some of the cell features.
Table 6.2 Typical sizes of cellular features
Many processes that occur within cells are still not well understood. Individual
cells are complex systems by themselves but when added to their multicellular environ-
ment, the system becomes quite intricate. The human body contains over 200 types of
cells of various shapes, sizes and functions, yet each contains the same genetic material.
The study of cells by microscope can be divided into two broad categories. The tradi-
tional method of fixed cell evaluation is through the viewing of cells that have been per-
manently affixed to a microscope slide. The other way to examine cells is while they are
living (e.g. as part of a tissue or in a controlled environment like a petri dish).
Fixed cells
In some cases, with relatively loose or transparent tissue, the object under examination
can be placed on a slide and then placed under the microscope. If the item of interest is
part of a larger object, such as an organ, or a bone, then the tissue must be sliced. If the
object is generally opaque, then the slice either must be very thin or some other method
must be used as discussed in Chapter 4. Structures inside most cells are almost colorless
and transparent to light. Therefore, dyes are used which react with some of the cell’s
structures but not others. Staining the cells may alter the structure of the cell or kill it.
Because of this, the cells are normally preserved or fixed for observation. This is nor-
mally done with some fixing agent such as formaldehyde. In fixing the cells, the general
structure of the cell and organelles may be preserved but enzymes and antigens within the
cell may be destroyed. Usually the fixing process takes at least 24 h.
6-4 Chapter 6 Cellular Measurements
Live cells
Live cells have the advantage that time is not typically a factor when viewing. The cells
are viewed in their natural environment and time studies are easily accomplished using
video or other recording devices.
One of the three general categories of live cell research is the division of cell
population studies. This is done while the cells are free to interact with one another. With
this configuration, not only the activity within the cell can be studied, but also the interac-
tion between cells and other substances or objects external to the cellular structure such
as man-made materials.
A second category of cell research involves examining the characteristics of an
individual cell. This may be the physical features of the cell, its topology, or the relative
chemical/molecular structure of the cell. Another area of study involves the response of
the cell to experimental treatments.
Single cell imaging involves one of two different techniques. The first is the 3-D
reconstruction of a scanned object. Using digital (computer) techniques, characterization
of single cells can be done by the cell researcher. The other technique involves time se-
quencing of the movement, or morphology, of a cell. This involves the use of video cam-
eras and video recording systems. For example, cell adhesion may be studied by video
and fluorescence imaging of the spreading and detachment of adherent cells subjected to
a range of fluid shear stresses in a radial flow chamber. Cell motility is assessed by time-
lapse video microscopy of individual cells crawling on sheets of polydimethylsiloxane.
6-6 Chapter 6 Cellular Measurements
1 1 1
+ = (6.1)
LO LI LF
Object
Image
F
LO
LI
LF
(a) (b)
Figure 6.1 (a) Light rays through a lens and the corresponding focal point F. (b) The light rays are
bent by the lens and refocus as an image.
Figure 6.2 shows that a conventional (wide field and bright field) compound
microscope is composed of two lenses. The primary lens (objective) has a very short fo-
cal length and is placed close to the sample. It forms a much larger primary (intermediate)
image just below the eyepiece. The secondary lens (eyepiece) alters the primary image by
forming parallel rays that are focused on the retina by the lens of the eye. This is the
enlarged virtual image seen when looking into a microscope. Light is intensified by the
condenser lens so it more readily transmits through the sample.
Eye
Eyepiece
Objective
Sample
Slide
Condenser
Field
iris
Light
source
Figure 6.2 Compound microscope. The compound microscope contains a light source for sample
illumination, a field iris to control the light field, a condenser to focus the illuminating light, an
objective lens, and an eyepiece.
If an electron beam replaces the beam of light, the microscope becomes a trans-
mission electron microscope (Chapter 4). If light is reflected from the object instead of
passing through, the light microscope becomes a dissecting microscope. If electrons are
reflected from the object in a scanned pattern, the instrument becomes a scanning elec-
tron microscope (Chapter 4).
The microscope’s function is to allow viewing of objects that cannot be seen with the
human eye. Because we are trying to make an object look larger, the term resolution is
6-8 Chapter 6 Cellular Measurements
where λ is the wavelength of the illuminating light, n is the refractive index of the me-
dium between the objective and the specimen, and the cone angle α is 1/2 the angle of the
acceptance of light onto the lens (Figure 6.3). The value n sinα is known as the numerical
aperture (n.a.).
Objective
Specimen
Bright field
The light microscope can use several methods to view objects. Bright field microscopy is
used to view stained or naturally pigmented organelles in the cell. It involves the use of
color detection. If an object contains colored features, the eye detects those colors. The
stains or pigments absorb and alter the light that passes through the sample. Therefore,
the light that passes through the sample is seen as a color that is distinctly different from
the background or direct light that does not pass through the sample.
Most of the features in the cell are relatively colorless and difficult to see under the mi-
croscope. The easiest way to enhance the contrast of the relatively transparent cells is to
stain them. Different colored dyes are used that stain different parts of the cell. If the cells
are fixed, this is not a problem except for possible damage to the cellular integrity from
the dye.
Further image enhancement can be accomplished by using a specific mono-
chromatic light source. One way to achieve this is by using a filter. The most commonly
used filter is one that allows only blue light to pass through. Blue light, about 450 nm, is
usually the shortest wavelength used in light microscopy. It determines what cell features
you can and cannot see under the microscope. In order to see a specific structure of the
cell, it must be large enough to be able to perturb the wave motion of the light rays that
strike it. Therefore, mitochondria, about 3 µm, are one of the smallest organelles to be
seen through microscopy. Organelles that are smaller than this cannot be resolved be-
6-10 Chapter 6 Cellular Measurements
cause they are small compared to the wavelengths of visible light and are not able to per-
turb the wave motion of light traveling to our eyes.
Phase contrast
Another method of viewing objects is called phase contrast microscopy. This type of mi-
croscopy is used to view living cellular structures because they are not pigmented, thus
unable to be seen under bright field microscopy, and the structures within the cell tend to
differ in refractive indices.
Phase contrast takes advantage of the fact that when rays of light pass through
areas of differing refractive index or different lengths of optical path (thicker sections),
the light rays get out of phase with one another. This causes interference, which in turn
causes different areas to appear as differences in brightness when seen by the eye. Phase
contrast microscopy makes use of these minute shifts in the phase as the light passes
through an object. If the normal phase shift is increased by 1/4 or 1/2 wavelength, the
differences will be magnified in intensity. Phase contrast microscopes that provide 1/4
wavelength shift are called medium phase while those that provide 1/2 wavelength shift
are called dark phase.
A standard compound microscope is modified for phase contrast microscopy, as
shown in Figure 6.4, by adding annular transparent rings, one below the eyepiece and
another below the condenser. Some light transmitted from the lower annular ring passes
through the specimen and is imaged on an upper semitransparent ring. Other light trans-
mitted from the lower annular ring is refracted slightly by different parts of the specimen
and passes outside the upper semitransparent ring and is imaged on a surrounding trans-
parent diaphragm, which is one-quarter wavelength thicker. Since the rays of light travel
slightly different distances, when they meet up again, they will be slightly out of phase
with one another. If the phase differences add to each other, constructive interference, the
image will be brighter (Figure 6.5). If they cancel out, destructive interference, the image
will be darker (Figure 6.6).
Upper annular
phase
changing ring
Deflected
light
Objective
Sample
Slide
Lower
Condenser annular ring
Figure 6.4 In phase contrast microscopy, light from the lower annular ring is imaged on a semi-
transparent upper annular ring. This microscope uses the small changes in phase through the sam-
ple to enhance the contrast of the image.
Figure 6.5 The interaction of light waves that meet in phase results in constructive interference.
The amplitude of the wave is doubled by this interaction.
6-12 Chapter 6 Cellular Measurements
Figure 6.6 Interaction of light waves that meet one-half wavelength out of phase results in destruc-
tive interference. The waves cancel each other when they intersect.
If the upper annular ring is opaque instead of semitransparent, it will block all
light except that which is refracted, thus yielding a dark field phase contrast microscope.
DIC imaging is a modification of phase microscopy and also produces contrast from dif-
ferences in refractive index in the sample. Contrast is obtained by converting these phase
changes into amplitude changes. DIC does this by splitting light with a prism into two
parallel rays. These rays are routed through two different optical paths and then recom-
bined. As with the phase contrast technique, the rays interfere with one another construc-
tively or destructively depending on the phase difference. The advantage to this method is
that the phase difference can be shifted in phase by a variable amount. The most com-
monly used configuration of differential interference microscopy is called the Normarski
Interference Microscope, named for the optical theoretician George Normarski. In mi-
croscopy, the term Normarski is synonymous with the differential interference contrast
technique. This technique is used for measuring thickness of cells and observing other
items with little or no contrast. DIC imaging provides good rejection of out-of-focus in-
terference. Essentially it provides an image of the rate of change of refractive index in the
sample. The technique essentially acts as a filter that emphasizes edges and lines. Out-of-
focus refractive index changes will be blurred and have a small rate of change in the focal
plane so they will not contribute much to the contrast of the image.
Dark field
Figure 6.7 shows another method called dark field microscopy. With dark field illumina-
tion, the background field is dark because the condenser light comes from outside the
collecting angle of the objective. This is due to an opaque stop below the condenser.
Small objects, even those below the limits of resolution, can be detected easily since ob-
jects scatter light and appear to emit light on a dark field.
(a) (b)
Figure 6.7 Example of darkfield illumination. (a) shows the typical brightfield view while (b)
shows the darkfield view of the same object. Imagine the object in the darkfield view glowing like
the moon at night.
Rheinberg illumination, also called optical staining, is especially useful for look-
ing at unstained samples. This method is similar to conventional darkfield illumination
except that the central opaque light stop at the base of the substage condenser is substi-
tuted for a colored central stop. A sample viewed by this method looks white on a back-
ground that is the color of the stop. If a colored transparent annular filter is added, the
sample is colored.
Inverted microscope
If a light microscope is functionally turned upside down to the object and viewed from
under the sample, the microscope is called an inverted microscope. The light source and
condenser are above the stage pointing downward. The objective is below the stage point-
ing upward. As a result, one is looking up through the bottom of the specimen and rather
than looking at the specimen from the top, typically through a cover glass, as on a con-
ventional microscope. This is particularly useful with live cells since they can be seen
through the bottom of a container, without opening the container, and without the air in-
terface normally present between the objective and the surface of the culture. By adding
phase contrast capabilities, cells can be observed without stains.
Common biological stereo-binocular (dissecting) microscopes are also used with
reflected light to examine tissue that is too thick to allow light to pass through it.
Solid-state cameras are ideally suited for many cellular microscope applications. The
typical solid-state camera uses a technology called a charge-coupled device or CCD.
These devices consist of a large square array of tiny detectors (called pixels) each of
which produces an electric charge in response to incident photons (see Figure 6.8). Each
pixel is about 20 µm2. The charge for each pixel is collected and sent to an output ampli-
fier so that the level of light incident to each pixel is recorded. Charge accumulates over
time so the total amount of charge is proportional to the product of the light intensity and
the exposure time. The intensity of each pixel is displayed as a video matrix on the com-
puter, yielding a two-dimensional image of incident light. Tiny red, green and blue filters
placed in front of sequential pixels yield a color camera.
The CCD camera has very high signal-to-noise ratio and is available in high
resolution arrays of 2032 × 2044 pixels or greater at 1.6 frames/s. These array cameras
have a response in which the output is linearly related to the input light intensity over a
very wide range of light levels. In addition, they are not susceptible to accidental light
damage as are the high sensitivity video cameras. A lower resolution array of 1300 ×
1030 pixels at 12 frames/s is still below the 60 Hz standard video scan rate. This limits
these devices to studying either static or slow dynamic processes. Also, many times these
devices are operated at low temperatures to maximize the signal-to-noise ratio and are
therefore termed cooled.
Single pixel
CCD array
Figure 6.8 CCD Array. The CCD array is made up of many light-sensing elements called pixels.
Each pixel can independently sense light level and provide this information to a computer.
It is important to distinguish between the two types of image acquisition (see Figure 6.9)
that can be done with these microscopes. Image plane scanning, or wide field imaging,
involves the acquisition of a two dimensional image by the recording device. The device
is focused on a plane of the object and a two-dimensional picture is directly recorded.
This is the type of image that you would see if you looked into a traditional light micro-
scope. In contrast, a point-scanned device is focused on a single point of the object. The
instrumentation must have some way of scanning or focusing on different points. This is
typically done in the two horizontal (x–y) planes so the resulting picture would look the
same as the wide field image (Shotten, 1990). An example of this is the scanning electron
microscope discussed in Chapter 4.
Objective Objective
Sample Sample
Figure 6.9 Comparison of the wide field versus the point scan techniques. (a) wide field collects an
entire image while (b) point scan image must be assembled point by point.
There are two general methods for highlighting or marking different structures within a
cell or to mark different types of cells. These markers are used for a variety of purposes
from locating a certain structure within a cell, identifying which cells contain a particular
substance, how much of the substance they contain, where it is located within the cell,
how that substance may move over the cell’s cycle, if the substance moves between cells,
and even how much of the substance is present within a cell.
These markers can be put into the cells in several different ways. Some are natu-
rally taken up by the cells when introduced into their environment. Others are injected
into cells using special techniques or devices such as the tiny syringe-like microinjector.
Once put into cells, some are analyzed immediately, others are given some time to bond
to specific cellular structures, while others are activated at a later time when the re-
searcher wants to observe time specific cellular dynamics.
Fluorescent probes
The use of fluorescent markers or probes is revolutionizing the imaging of live cells and
their associated contents, structures, and dynamics. These systems all involve the excita-
tion of a fluorescent material, the subsequent fluorescence emission by the material, and
the detection of this fluorescence by a detector (Haugland, 1997).
The probes, also called fluorochromes and fluorophores, are fluorescent materi-
als that are introduced into the cellular environment. There are a number of different
ways that these probes are used with the cells:
The use of these probes is especially useful for microscopy for several reasons.
First the excitation wavelength is different from the emission wavelength. This allows for
easy detection because the excitation wavelength can easily be filtered out. Also, differ-
ent types of probes emit fluorescence at different wavelengths. This too is especially use-
ful in that several probes can be used to mark different structures within a cell and these
can be detected at the same time. In addition, the entire excitation/fluorescence cycle can
be repeated.
Two considerations are important regarding probe selection. First, probes must
be chosen based on sensitivity, or the amount of substance that is labeled within the cell.
Also the probe’s specificity or the ability to only label the substance of interest is impor-
tant. Specificity is provided by fluorochromes. These are either synthetic chemicals or
natural products that can have innate specificity, binding to certain classes of molecules,
or they can be attached to molecules that have specificity (for example, antibodies). Some
types of fluorochromes are generated by the activity of enzymes, or are sensitive to the
presence of specific ions. A final class of fluorochrome is generated by gene expression.
The advantages of using probes include the high signal levels against a low
background (noise) level, they are not prone to motion artifact, and they can be used to
monitor very small changes.
Care must be taken regarding the intensity level of the light used to excite the
probe. Photobleaching is the breakdown of the fluorescent probe within the sample as it
is exposed to light. Typically, the more intense the light, the more photobleaching that
occurs. Photobleaching causes several problems. It reduces the fluorescent signal emitted,
it generates oxygen free radicals, which can damage cells, and it may generate forms of
dye that distort the relationship between the substance being marked and other areas of
the cell.
In a flow cytometer, a laser excites fluorescence of individual cells in an interro-
gation volume. The fluorescence is then separated on the basis of color, through the use
of optical filters, and the results displayed as a histogram of cell count vs. cell type. Cells
may also be selectively isolated or sorted to high purity as they pass through this system.
Fluorescence video microscopy can capture multiple images at differing wave-
lengths of light simultaneously at video rates. The fluorescence event follows exponential
kinetics with a lifetime of approximately 350 ms.
Radioactive tracers
Radioactive tracers are used in a number of areas of cell research as an effective and safe
means of monitoring cellular interactions. This was the primary technique used for years
but is being used less and less with the development of more and better fluorescent
probes. These tracers have the advantages of high precision and specificity but suffer the
disadvantages of extensive radiation handling safety procedures, the inability to distin-
guish more than one probe at a time, and a relatively long time (hours to days) to obtain
results.
Radioactivity is caused by the spontaneous release of either a particle and/or
electromagnetic energy from the nucleus of an atom. Tritium (3H) is an important tracer
used for cellular research. Other radioactive elements used include 14C (Carbon-14), 32P
(Phosphorus-32), 125I (Iodine-125). Tritium is most prominent because replacing hydro-
gen with tritium in a molecule does not cause chemical or physiological changes in mo-
lecular activity (Heidecamp, 1996).
Radioactive elements emit radiation spontaneously, but over time a percentage
of all radioactive elements in a solution will decay. It is nearly impossible to predict
which individual element will radioactively decay, but predictions can be made about
large numbers of the elements. The radioactivity of a material is the rate of how many
decays occur each second. It is measured in becquerels.
When an alpha or beta particle, or a gamma ray passes through a substance, ions
are formed. To monitor this ionization effect, a device called an ionization chamber is
used. The most common forms of ionization chambers are the Geiger counter and the
pocket dosimeter. These chambers have two electric plates separated by an inert gas with
a voltage applied across them with a battery or power supply. When a particle enters the
6-18 Chapter 6 Cellular Measurements
chamber, ions form and are drawn to one of the electric plates as shown in Figure 6.10.
The negative ions attract to the positive plate (or anode) and the positive ions attract to
the negative plate (or cathode). This causes a small voltage change in the form of a pulse
to temporarily occur across the plates. This pulse is passed by the capacitor, which blocks
the dc voltage from the power supply. The output pulse is then read by some measuring
device. The sensitivity of the system depends on the voltage applied between the electric
plates. High-energy alpha particles are significantly easier to detect than beta particles,
and they require lower voltage. In addition, alpha particles penetrate through the metal
casing of the counter tube, whereas beta particles can only pass through a quartz window
on the tube. Consequently, ionization chambers are most useful for measuring alpha
emissions. High-energy beta emissions can be measured if the tube is equipped with a
thin quartz window and if the distance between the source of emission and the tube is
small.
Resistor Capacitor
Pulse
count
+++++++++++++++ Plates
+ −
Voltage Particle − Freq to volt
−
supply counter
+ +
+ Ions
− − − − − − − − − − −−
Meter
Figure 6.10 Basic radiation counting system. The power supply has a voltage source and series
resistor. Ionization by a radiation particle causes a voltage pulse, which passes through the capaci-
tor and is read by the meter.
A pocket dosimeter is a modification of the basic ionization chamber. It uses a
miniaturized tube that can be charged to hold a voltage without constantly recharging the
unit with a power source. Essentially it is just a capacitor, which is charged by a base unit
and carried as a portable unit. It is small enough to be easily carried in a pocket. When
exposed to ionizing radiation, the capacitor discharges slightly. Over a period of time, the
charge remaining on the dosimeter can be measured and radiation exposure can be calcu-
lated. Since dosimeters carry lower voltage potentials, they are used for the measurement
of X-ray and high-energy gamma radiation, but do not detect beta emissions.
Liquid scintillation counters are also used to measure radiation. They use a sub-
stance in the counting chamber, called a scintillant, to produce a flash of light. This flash
is produced for each radiation particle detected in the chamber. A circuit then converts
each light flash to an electric current pulse. This signal is then sent to a counting system,
which keeps track of total counts or counts per minute (cpm). If the efficiency of the sys-
tem (or the percent of actual radioactive decays that were detected) is known, the disinte-
grations per minute (dpm) can be calculated using Eq. (6.3).
Confocal laser scanning microscopes (CLSMs) have been a significant innovation for the
light microscope industry (Wright, 1993). These microscopes remove the out-of-focus
blur from the 2-D image directly through hardware rather than requiring a software de-
blurring technique (discussed in section 6.4.1). In addition, this technique yields images
from relatively thick biological tissue and successive 2-D optical sections (known as a Z-
series) can be assembled in an imaging system to construct a 3-D image. A standard light
microscope can be used for CLSM with some technically simple, yet expensive, modifi-
cations.
As opposed to a conventional wide field microscope which produces a 2-D im-
age directly, the CLSM is a point scan device, which builds a 2-D image by scanning and
assembling the image point by point with an integrated imaging system. A CLSM nor-
mally uses a laser beam as a light source because a laser provides high intensity, well-
focused, monochromatic light, which is able to easily illuminate individual points within
a specimen. Using a point light source avoids the interference of scattered light when the
entire field is illuminated. This interference is known as halation.
The scanning is typically accomplished by a series of x–y plane rotating mirrors
in the light path between the dichroic mirror and the objective that focus the laser beam
on different points of the specimen in the x–y plane. To obtain an image of the specimen,
the light from the specimen is routed to a detector. A pinhole allows only the light origi-
6-20 Chapter 6 Cellular Measurements
nating from the focal plane to reach the detector as shown in Figure 6.11. The detector
(usually a CCD camera or photomultiplier tube) measures the intensity of light and pro-
duces a signal proportional to the light at each point. This signal is then typically sent to
an imaging system where the current is converted to a voltage, digitized, processed to
reduce noise, and an image is reconstructed for display on a monitor.
CLSM has several advantages over conventional microscopy. CLSM provides
high-contrast images of specimens without the image artifacts normally present with
conventional contrasting techniques. It also has the ability to view relatively thick speci-
mens and is well suited for fluorescence applications as will be discussed later. The dis-
advantages of this technique are relatively high cost and possible damage to the specimen
from the laser illumination. In an attempt to minimize this damage, each slice may be
scanned several times at relatively low laser power and the results integrated.
Detector
Pinhole
detector
aperture
Pinhole
illumination
aperture Dichroic
or laser mirror
Objective
Specimen
Focal plane
Out of focus
In focus
Figure 6.11 Confocal laser scanning microscope. The microscope removes out-of-focus (z-plane)
blur by keeping out of focus light from reaching the detector using a pinhole.
6.2.6 Two-photon excitation microscopy
Another technique that has been developed to minimize the excitation laser’s potential
damaging effects is two-photon excitation microscopy (TPEM). This technique is based
on the principle that two photons of longer wavelength light simultaneously excite a
probe that would normally be excited by one photon of a shorter wavelength, as shown in
Figure 6.12. This method has the advantage that less photobleaching occurs because only
the focal point is being excited as compared to the entire depth of the specimen as in con-
focal and wide-field imaging. (LOCI, 2000). In addition, TPEM can obtain optical sec-
tions from deeper into a specimen. This is possible because the excitation light is not ab-
sorbed by the probe above the plane of focus and the longer wavelengths are scattered
less within the sample. Two-photon excitation has several limitations. A given probe has
lower resolution when compared to confocal imaging. This can be eliminated by the use
of a confocal aperture at the expense of a loss in signal. Also, thermal damage may occur
if the specimen contains substances that absorb the excitation wavelengths.
Detector
Objective
Specimen
Focal plane
Figure 6.12 Two photon excitation microscope. This microscope doesn’t require a pinhole but is
able to excite single points within the sample by having two photons excite the sample only at the
exact location of interest.
6-22 Chapter 6 Cellular Measurements
The advent of faster and cheaper personal computers has permitted a myriad of image
processing techniques that seem ideally suited for the microscope. Several of these tech-
niques correct the image in software as opposed to hardware. Others take the large vol-
umes of data generated by digital imaging and manipulate it into a form that is more read-
ily usable by the researcher.
Computational deblurring
One way to measure orientation is by using an orientation chamber. One type of orienta-
tion chamber tests the ability of cells to orient themselves in a gradient of chemical
attractant. The chamber is similar to a hemocytometer (see Chapter 5). The gradient is set
up by diffusion from one well to the other and the orientation of cells toward the well
containing chemical attractant is scored on the basis of their morphology or by filming
their movement.
Video enhanced contrast microscopy (VECM), as shown in Figure 6.13, uses a high-
resolution video camera and a light microscope that has differential interference contrast
(DIC) optics. By using both the analog and digital contrast enhancement as well as back-
ground image subtraction, this method is able to observe objects with dimensions an or-
der of magnitude smaller than the resolution limits of the microscope alone (Shotten,
1990). During image subtraction, for example, all identical structures in successive video
images can be erased, filtering out any interfering background images. This method in-
volves three steps (Figure 6.14). First the light microscope’s DIC optics is adjusted to just
under the saturation level of the video camera. The second step involves reducing the
background level by adjusting the black-level control. Then the gain control is adjusted to
amplify this dark-low intensity signal. The final step is required because the previous one
also captures background blemishes created by lens imperfections and unevenness of
illumination. In this step, the specimen free background that has been previously recorded
is subtracted from the live image.
6-23
6-24 Chapter 6 Cellular Measurements
Video
camera Frame grabber
Light
microscope
Monitor and
Video computer
recorder
Figure 6.13 Video-enhanced contrast microscope (VECM) system. A light microscope image is
recorded by a camera and the image is sent to a recorder and a computer for image processing.
Intensity Intensity Intensity Intensity
Figure 6.14 (a) Original low-contrast video image. (b) The threshold is set just under the saturation
level. (c) Reducing the background level. (d) Gain is adjusted to amplify dark-low intensity signal.
6.3 Cell Orientation 25
Figure 6.15 (a) An unprocessed photo of cells of the inner epidermis taken through an interference
contrast microscope. (b) The same image with digital contrast enhancement, the single structures
become apparent. The background fault remains. (c) Subtraction of the background and resulting
with further contrast enhancement.
6.4 Cell Rolling Velocity
Cell rolling velocity involves the tracking of cells, particularly blood cells, as they start to
tumble or roll as the body responds to infection or foreign matter. This is a parameter that
becomes important in the areas of healing and the body’s response to biomaterials im-
planted in the body. When an infection occurs in an area of the body, white blood cells
are responsible for attacking the infection. The way the body signals these cells is critical
to understanding the healing process. Recruitment of white blood cells from the blood to
the area of infection is required. This recruitment comes from cell adhesive molecules on
the inside of the blood vessel. These Velcro-like cell adhesion molecules (CAMs) start
the rolling process and cause the white blood cells to roll along the endothelium wall of
the blood vessel. Since this is the first step in the process of the activation of the body’s
immune system it has application in the response to tissue engineering and biomaterials.
This is typically measured using fluorescent probes, intensified fluorescence microscopy,
cell tracking techniques, and the artificial introduction of CAMs.
Intensified fluorescence microscopy (IFM) involves the use of a fluorescent
specimen, excitation of the specimen with low-level light, and detection of the fluores-
cence with a microscope and a high sensitivity video camera. High-sensitivity video
cameras use video image tubes with sensitivities about two orders of magnitude greater
than conventional video cameras. This configuration was known as video-intensified
fluorescence microscopy (VIFM). Recently, these cameras have been replaced with solid-
state slow-scan cooled CCD cameras of high resolution. These devices are ideally suited
for low-level light imaging and do not suffer from many of the problems associated with
the high-intensity video cameras. IFM images do have one major problem—out-of-focus
blur. Since the entire field of view of the specimen is illuminated, there are fluorescent
emissions from the areas above and below the focal plane causing the out-of-focus blur.
This problem is compensated for in software and is detailed in the image processing sec-
tion 6.2.7. The actual computation of rolling velocities is typically accomplished after
imaging has been completed by using cell-tracking software, which computes the rolling
velocities from relative probe locations from the prerecorded images.
Figure 6.16 shows a block diagram of an IFM system. IFM is typically used
with the lowest level of excitation light possible so that the probe can be traced for a rela-
tively long period of time. This usually gives rise to a relatively low signal-to-noise ratio,
which needs to be handled as part of the image processing procedure.
6-26
6.4 Cell Rolling Velocity 27
Recording
device
Image
Time code Frame processing
Microscope Camera generator grabber computer
Monitor
Figure 6.16 Intensified Fluorescence Microscopy system. In this system, the fluorescent signal is
received through the microscope by the camera. The video image is typically time stamped and sent
to the frame grabber. The frame grabber digitizes the signal and sends it to the computer for image
processing. The processed signal is then sent to a monitor for display by the operator and perma-
nently recorded for later playback and analysis.
6.5 Cell Pore Size Determination
Cell pore size is an important parameter for a number of reasons. When a cell secretes
hormones, neurotransmittters, or enzymes, a pore is opened in the cell, which allows
these molecules to be released (Hoffman, 1997). Also, the first indication that two cells
are going to fuse is the formation of a small pore connecting the two cells. The under-
standing and regulation of these processes are important to the formation and growth of
artificial tissue.
6.5.1 TEM
The transmission electron microscope (TEM) (discussed in Chapter 4) is a useful tool for
determining pore sizes because of its resolution in the 0.3 to 1 nm range. Electron micro-
scopes did not typically view larger objects well so it was difficult to view the cell as a
whole, locate a point of interest, and then zoom in on it. Modern electron microscope
systems, however, allow for this zooming, and the observation of whole tissues while
retaining macromolecular resolution. TEMs work in a vacuum, so living cells cannot be
observed with this type of microscope. Also, most materials in the cell are transparent to
electrons, so they must be stained with dyes. Thus it is possible that the cellular structure
being imaged may have been altered by the dye.
6.5.2 SEM
The scanning electron microscope (SEM) (discussed in Chapter 4) is also useful in cellu-
lar analysis because of its ability to show cell surface detail at resolutions in the 3 to 10
nm range and its ability to show 3-D structure. The 3-D structure is obtained through the
process of Secondary Emission Ion Scanning. As the electron beam of the scanning mi-
croscope scans the surface of an object, it can be designed to etch or wear away the out-
ermost atomic layer. The particles are analyzed with each scan of the electron beam. Thus,
the outer layer is analyzed on the first scan, and subsequently lower layers analyzed with
each additional scan. The data from each layer are then analyzed and reconstructed to
6-28
6.5 Cell Pore Size Determination 29
produce a 3-D atomic image of the object. Since electrons are relatively small, the etch-
ing is sometimes enhanced by bombarding the surface with ions rather than electrons.
Several methods can be used to deform cells (Hochmuth, 1990). The micropipet has been
an important tool for cell deformation studies. This technique involves aspirating part or
all of a cell into a glass micropipet. A pipet with a diameter of 1 µm is capable of produc-
ing a force F of 1 pN. The rigidity and other cell deformation properties can be found as a
function of the force on the cell and the displacement of the cell inside the micropipet. By
tracking the displacement of the leading edge of the cell as it moves into the pipet, cell
deformations as small as 0.1 µm can be measured.
F F
Cell Micropipet
Displacement
Figure 6.19 In the micropipet technique, force causes a displacement to determine cell deformation
properties.
6-30
6.5 Cell Pore Size Determination 31
Another technique used is called optical tweezers or a laser trap. This system develops an
optical gradient generated by a continuous laser. The optical trapping system uses an in-
verted microscope in differential interference contrast mode and a motorized x–y stage
and two associated lasers. The microscope feeds a video camera and the video is recorded
and stored on an optical memory disk recorder. The optical tweezers can be moved
throughout the viewing field and cell deformations as small as 0.1 µm can be measured.
6.7 Cell Shear Stress
When a liquid flows through a vessel of uniform diameter, the liquid adopts a parabolic
distribution of velocities from zero at the wall to a maximum at the center. Shear stresses
are generated due to regions of the liquid flowing at different speeds. A different velocity
of the liquid flowing on either side of a cell creates pressure on the cell (Figure 6.20).
This is of particular interest to bioengineers working with the circulatory system. Cell
shear stress is defined as the cellular response to this mechanical effect within the body.
Endothelial cells produce many different kinds of compounds that control vascular tone,
vessel diameter and blood flow rate. The production rate and the release of these com-
pounds are modulated by blood flow and the associated shear stress. Tissue engineers
need to understand the effect of physical parameters, such as pressure and shear stress, on
tissue formation (Lewis, 1995). For example, bone marrow cells cultivated in a petri dish
produce only two of ten types of blood cells. When these cells are cultivated in a bioreac-
tor, a device that simulates some of the physical parameters of the body, all ten types of
red blood cells are produced.
There are two problems in dealing with cell shear stress. One is knowing what is
going on in the cell when the shear stress is applied. The other is what effect the shear
stress has on the cells.
Vessel wall Blood cells
V1
V2
Blood Range of
velocities
Figure 6.20 Parabolic distribution of velocities and the shear stress it causes on a blood cell.
One of the problems with shear stress measurements involving the circulatory system is
that most methods of generating shear stress require relatively large amounts of blood.
However, the cone and plate method solves this problem. With this system, the liquid is
placed between a flat horizontal plate and a cone-shaped surface (Eylers, 1992). The cone
is rotated at a constant speed. This causes the liquid to rotate at speeds that are linearly
6-32
6.7 Cell Shear Stress 33
proportional to their distance from the axis of rotation since the slope of the cone is con-
stant. The shear stress S is the force per unit area of the wall and is defined as
dv
S =η (6.4)
dx
where η is the viscosity of the liquid (0.001 Pa⋅s for water at 20 °C), and dv/dx is the gra-
dient of the velocity between the cone and the plate. Figure 6.21 shows that with this con-
stant velocity gradient, the shear stress is proportional to the viscosity of the liquid. If the
viscosity remains relatively constant, as is the case with blood, the shear stress remains
constant. Changing shear stress values can then be accomplished by changing the speed
(v) of the rotating cone.
Cone
Liquid
dv/dx
Figure 6.21 Cone and plate system. The cone is immersed in a liquid and then rotated at a constant
rate. Since the velocity changes linearly with distance, the shear stress is a function of rotation ve-
locity.
Shear stress is reported to induce calcium increase in the cell. Fluorescent ratiometric
imaging (see section 6.13.1) is used to analyze the calcium response with respect to shear
stress. The fluorescent image is an image pair acquired at two different wavelengths and
obtained by a silicon-intensified target camera. After the signals have been sent to an
image processing system and calibrated, the images display the calcium concentration
within the cells.
6.8 Cell Adhesion
Cell adhesion, like cell rolling velocity, is an important parameter in the healing process,
in facilitating the growth of tissue, and in the body’s ability to assimilate newly implanted
biomaterials.
In some cases reduction of adhesion is the aim, in others an increase is desired.
Understanding the mechanisms that start the adhesion process, especially in the blood-
stream, are of particular interest. In addition, the strength of adhesive bonds is of prime
interest to the tissue engineer. Most studies of cell adhesion use force measurements ob-
tained from video and fluorescence microscopy imaging.
The optical tweezer, or laser trap, method is used for generating controlled forces in the
piconewton range. It uses a light microscope and the force of radiation pressure from an
infrared laser. Although piconewton sized forces are quite small, they are sufficient to
stop the motion of highly mobile cells or single enzymes. These optical tweezers allow
for measurements of subcellular components including torsional stiffness, flexibility of
the cell cytoskeleton, and forces generated by enzymes.
6.8.2 Interferometry
6.8.3 SEM
The SEM uses an intense electron source known as the field emitter, which provides
resolution to diameters less than 1 nm (Morton, 1997). This is used to study the cell adhe-
sion molecule. These molecules are found on the surface of leukocytes and are an impor-
tant mechanism in the body’s response to injury.
6-34
6.9 Cell Migration
Cell migration is the process by which cells move to a certain area, especially during
growth or in response to an abnormal condition in the body. Cell migration is an area of
great interest for tissue engineering applications such as tissue growth and biomedical
implantation.
A four-dimensional (4-D) image processing system is used to create a digital movie from
time-sequences of 3-D cellular tomographic microscope images. First, a single point im-
age is scanned over a 2-D plane. Vertical slices of the 2-D plane are acquired and a 3-D
image is reconstructed. This process must be repeated for each point in time. Then each
of these 3-D images is displayed sequentially and the result is an animation of the move-
ment within the image as a function of time. Typically, this animation is recorded on a
video recorder for ease of repetitive observation and analysis or is assembled into a com-
puter animation file. Although the animation does not improve the quality of the collected
data, it improves the qualitative interpretation of the process.
One problem with this technique is the large amount of storage space and com-
puting power required for the image processing. Cell migration observations typically
take 24 to 96 h to complete (Di Francesca and Zerbarini, 1994). Four fields of view, each
containing 5 to 20 cells, are collected at 15 to 30 min intervals and the position of each
cell is recorded in each frame at every event interval. The data are given to modeling and
analysis software to determine where cells moved and how fast they moved there.
6-35
6.10 Cell Uptake
Cell uptake is the study of the various types and amounts of biological molecules in the
cell. Cell capacitance has been used as an index of secretion in cells. The capacitance of
the cell membrane can be measured using techniques similar to the voltage clamp. Fluo-
rescent microscopy is commonly used since fluorescent probes are now available to
measure pH, sodium, magnesium, potassium, calcium and chloride ions.
To follow structures within a cell, a probe needs to be sensitive to small amounts of the
structure and only label the structure of interest. Submicron colloidal gold particles cou-
pled to antibodies are useful for this purpose (Geerts et al., 1991). These small particles
can be detected with a light microscope since the particles scatter a small amount of light,
which is detectable using video contrast enhancement microscopy. The particles appear
as black dots on a white background and the technique is called nanovid microscopy. The
biggest advantage of this technique is its ability to label specific molecules within the cell.
Another advantage is the ability for continuous observation without having to worry
about the photobleaching effects seen in fluorescent probes. A final advantage is the
small size of these particles, which allow diffusion throughout the cytoplasm.
6.10.2 Electroporation
6-36
6.11 Cell Protein Secretion
Cell protein secretion involves the movement of proteins across biological membranes
and is not well understood. Epithelial cells from certain organs produce and secrete many
important proteins that are stimulated by specific hormones and neurotransmitters. Also,
cell surface receptor proteins are secreted allowing cells to stick together and form the
extracellular matrix. As such, cellular engineers are focusing on the regulation of the ex-
ternal and internal environments of the cell that promote the secretion of specific proteins.
In the fluorescence lifetime imaging microscopy (FLIM) method, image contrast in the
fluorescence microscope is derived from the fluorescence lifetime at each point in the
image. This method is also known as Time-Resolved Fluorescence Microscopy (TRFM)
and is analogous to magnetic resonance imaging (MRI) discussed in Chapter 7. It pro-
vides chemical information about intracellular ions since the lifetimes of many probes are
altered by the presence of certain ions and/or pH. The fluorescence lifetimes of the
probes are mostly independent of their concentration and are not normally affected by
photobleaching. The most desirable feature of FLIM, however, is that it does not require
wavelength specific probes. Therefore, FLIM allows quantitative imaging using visible-
wavelength illumination. There are several methods involved in this technique (Tian and
Rodgers, 1990). Figure 6.22 shows the time correlated single photon counting method. In
this system, the time difference between the start of an excitation laser light pulse’s arri-
val at the sample and the arrival of the fluorescence photon at the detector is measured.
The detector output is recorded in a multichannel analyzer with the intensity at a given
channel being proportional to the number of data points (the number of photons) that
were collected during that time interval.
6-37
6-38 Chapter 6 Cellular Measurements
Beam Time-to-
Light amplitude
splitter Sample Timer
source
converter
Figure 6.22 Time-correlated single photon counting system block diagram. When a light source
provides an excitation pulse, some of the light is deflected to the start timer circuit while the rest
illuminates the sample. A single fluorescent photon is detected by a photomultiplier tube, which
generates a pulse that stops the timer. The time difference is then converted to an amplitude. The
various amplitudes are recorded in the multichannel analyzer and a profile of different time inter-
vals versus number of photons in that interval is displayed.
The other method using this technique simply monitors the fluorescent intensity
over time. This measurement is normally achieved with a fast-response photodetector and
a sampling oscilloscope for data acquisition. The sample is excited and the fluorescent
intensity of the sample is recorded over time.
A C
F(−)
F(∞)
A
Level of
fluorescence
B
F(+)
B
Photobleaching t
C
Figure 6.23 A sample graph of fluorescence of proteins during FRAP. (a) Before photobleaching
F(–), (b) just after protein photolysed F(+), and (c) long after protein photolysed F(∞).
If not all the proteins are mobile within the membrane, the percentage of mobile proteins
can be calculated using
F (∞ ) − F (+ )
Rm = (6.5)
F (− ) − F (+ )
where Rm is the fraction of mobile proteins, F(–), F(+) and F(∞) are the levels of fluo-
rescence (proportional to light intensity) at times immediately before, after, and long after
the area has been photolysed.
6.12 Cell Proliferation
Cell proliferation is the rate at which cells multiply and grow. Many tissue engineering
studies depend on the ability to determine the effect of various factors on cell prolifera-
tion. These include chemicals and/or conditions that may influence the rate and/or timing
of cell proliferation. These studies that help in the understanding of the regulation of cell
proliferation are key in the development of new tissue engineering methods and the bio-
materials used to support them. This is also an important parameter in developing more
successful therapies for such things as the prevention and treatment of cancer and the
screening of anticancer drugs. Cell proliferation studies are used when measuring growth
factor, drug screening, and the study of toxic, mutagenic, and carcinogenic effects of
chemical compounds on cells (Roche Molecular Biochemicals, 2000). The most basic
method involves using a microscope counting chamber, or hemocytometer, to count the
number of cells in a given area. Density of cells growing on a vessel surface is useful for
tissue engineering applications. Density is measured through the use of an ocular grid
inserted into an inverted phase contrast microscope. Cell density is recorded and then
plotted on a log scale against time in culture. However, more modern techniques are now
being used. There are two general methods of measuring cell proliferation in cell popula-
tions including measuring DNA synthesis and measuring metabolic activity.
One method of measuring DNA synthesis involves using a radioactive tracer by adding it
to a cell culture. Over time, some of the radioactive material is retained in the DNA of the
cells. The cells are washed through filters and the radioactivity of the filters containing
the radioactive cells is measured using liquid scintillation counting. Knowing the radia-
tion level of the original tracer in the cell and comparing it to the current radiation level
of the cell provides a ratio that is proportional to the number of times the cell has divided.
6-40
6.13 Cell Differentiation
Differentiation of a cell is defined as the process where cells become structurally and
functionally specialized during embryonic development. There are 200 different kinds of
cells in our bodies, yet all are developed initially from a single cell. There are many fac-
tors that affect cell differentiation, a couple of them being genetics and the environment.
Most research on cell differentiation focuses on understanding the mechanisms involved
in cell differentiation. This is especially important in tissue engineering. Learning what
cell types are involved in tissue regeneration is critical in building tissue. Understanding
the mechanisms of cell differentiation aids in the understanding of what a cell needs to
regenerate a specific part.
Staining techniques are widely used to visualize those components that are oth-
erwise too difficult to see under an ordinary light microscope because of the lack of color
contrast between the object under examination and the background or because of the lim-
ited resolving power of the light microscope. In addition, staining techniques are useful in
detecting the presence or absence of certain cell components, thus allowing a relatively
simple scheme of differentiation or identification of microorganisms. The dye used in
staining varies with the cell types.
I λ′ 1 = I λ1 − Bλ1
I λ′ 2 = I λ 2 − Bλ 2
I λ′ 1 S λ 2
R= (6.6)
I λ′ 2 S λ1
where λ1 and λ2 are the two excitation wavelengths, Iλ1 and Iλ2 are the fluorescence
intensities, Bλ1 and Bλ2 are the background light levels, and Sλ1 and Sλ2 are the shading
6-41
6-42 Chapter 6 Cellular Measurements
corrections that compensate for the small spatial distribution between the two excitation
beams. This method has probes that allow ratio imaging of Ca2+, K+, Mg2+, Cl– and pH.
Flow cytometry systems (Chapter 5) examine cells previously labeled with fluoro-
chromes one at a time and then separate the cells on the basis of their color of fluores-
cence. Fluorescence activated cell scanning and sorting uses a flow cytometer and a PC
workstation for data processing and analysis. The system uses multiple excitation wave-
lengths in the visible and UV range. Cells may be separated into a wide variety of tissue
culture containers. The PC can be used for statistical purposes providing histograms of
the various distributions of cells.
6.14 Cell Signaling and Regulation
Cell signaling and regulation involves the complex processes that occur within cells to
maintain life (regulation) and trigger new events to occur (signaling). For the biomate-
rials engineer, the dream is to manufacture material that could (1) sense its environment,
(2) integrate the sensed information and finally (3) take some action based on it. To do
this, engineers must study the processes of signaling and regulation within the cell and
develop an understanding of how and why they work as they do.
A number of techniques are used for research in this area. NMR provides nonin-
vasive images of cellular structure and chemical composition. This technique is discussed
in detail in Chapter 7. Microscope Laser Light Scattering Spectroscopy (MLLSS) is used
for motion studies of intercellular macromolecules during diffusion or flow. This method
directs laser light into a sample and subsequently detects the scattered light. Fluorescent
probes are used to mark specific membrane proteins, which can in turn be microinjected
into cells. Their paths through the cell are monitored by computer-aided video systems.
With these systems, both quantitative and qualitative observation of membrane flow in
live cells can be monitored. In addition, time lapse photography coupled with phase con-
trast microscopy allows visualization of the cell division process. However, two of the
most common methods are described below.
6.14.1 FISH
6-43
6-44 Chapter 6 Cellular Measurements
large span between emission wavelength, the intensity peaks where each probe is fluo-
rescing is readily apparent from the spectral image.
The 3-D organization of specific chromosomes within cells and tissues is also
being examined using FISH in combination with laser scanning confocal microscopy and
computer assisted 3-D image reconstruction algorithms.
Fluorescent probe
A B
Chromosome target
Figure 6.24 (a) The DNA strand is denatured with heat and put in the same culture as the fluores-
cent probe. (b) The fluorescent probe binds to the target gene.
Figure 6.25 Human chromosomes probed and counterstained to produce a FISH image. The lighter
area within each chromosome show the fluorescent probe attached to the target gene. (from Fluo-
rescence Detection: Anti-Dye Antibodies. Molecular Probes Product Literature. [Online] Available:
www.probes.com/lit/fish/part4.html)
A process known as pulse labeling is used to determine the site of molecular synthesis in
cells. A radioactive tracer is added to the cellular environment for a short period of time.
The radioactive media are then washed or diluted away. The cells are then fixed and
autoradiography is used to locate the sites of newly synthesized DNA. The site of utiliza-
6.14 Cell Signaling and Regulation 45
tion of a particular molecule is determined by pulse-chase labeling: cells are again ex-
posed to a radioactive tracer for a short period of time and then washed away. But this
time, the cells are allowed to grow for a period of time. When these cells are examined,
the radioactivity will have moved to the site of utilization in the cell.
6.15 References
Bolsover, S. R., Silver, R. A. and Whitaker, M. 1993. Ratio imaging measurements of
intracellular calcium. In D. Shotton (ed.) Electronic Light Microscopy, Techniques in
Modern Biomedical Microscopy. New York: Wiley-Liss.
Eylers, J. P. 1992. Liquids. In J. F. Vincent (ed.) Biomechanics – Materials. Oxford: Ox-
ford University Press.
Geerts, H., De Brabender, M., Nuydens, R. and Nuyens, R. 1991. The dynamic study of
cell surface organization by nanoparticle video microscopy. In R. J. Cherry (ed.)
New Techniques of Optical Microscopy and Microspectrometry. Boca Raton, FL:
CRC Press.
Heidecamp, W. H. 1996. Cell Biology Laboratory Manual. [Online] Available:
www.gac.edu/cgi-bin/user/~cellab/phpl?chpts/chpt1/intro1.html
Hochmuth, R. M. 1990. Cell biomechanics: A brief overview. Trans. ASME, J. Biomech.
Eng., 112 (3): 233-4.
Hoffman, M. 1997. Cellular Secretion: It’s in the Pits. Am. Sci., [Online] Available:
www.sigmaxi.org/amsci/Issues/Sciobs97/Sciobs97-03pits.html
Haugland, R. P. 1997. Handbook of Fluorescent Probes and Research Chemicals.
Eugene, OR: Molecular Probes Inc.
LOCI 2000. Multiple Photon Excitation Fluorescence Microscopy. [Online] Available:
www.loci.wisc.edu/multiphoton/mp.html
Lewis, R. 1995. Tissue engineering now coming into its own as a scientific field. The
Scientist. 9: 15.
Molecular Probes. 2000. Product Literature. Fluorescence Detection: Anti-Dye Antibod-
ies. [Online] www.probes.com/lit/fish/part4.html
Paesler, M. A. and Moyer, P. J. 1996. Near Field Optics: Theory, Instrumentation and
Applications. New York: John Wiley & Sons.
Pawley, J. B. 1990. Handbook of Biological Confocal Microscopy. New York: Plenum
Press.
Peters, R. and Scholz, M. 1991. Fluorescence photobleaching techniques. In R. J. Cherry
(ed.) New Techniques of Optical Microscopy and Microspectrometry. Boca Raton,
FL: CRC Press.
Photometrics 2000. CCD Grading. [Online] Available:
www.photomet.com/library_enc_grading.shtml
Richardson, M. 1997. Three Dimensional Deconvolution of Microscope Data. Vaytek,
Inc. [Online] Available: www.vaytek.com/technical.html
Roche Molecular Biochemicals 2000. Guide to Cell Proliferation and Apoptosis Methods.
[Online] Available: biochem.roche.com/prod_inf/manuals/cell_man/cell_toc.html
Soenksen, D. G., Garini, Y. and Bar-Am, I. 1996. Multicolor FISH using a novel spectral
bio-imaging system. In Asakura, T., Farkas, D. L., Leif, R. C., Priezzhev, A. V.,
Tromberg, B. J. (eds.) Optical Diagnostics of Living Cells and Biofluids. Proc. SPIE,
2678.
6-46
6.15 References 47
6-48
6.15 References 49
6.21 Calculate the shear stress for water at 20 °C in a cone and plate viscometer with
a diameter of 10 cm, a separation gap of 1 mm at the circumference, and rotation
at 1 revolution per second. Give units.
6.22 Explain the reason for and the basic principles of measuring cell adhesion.
6.23 Explain the reason for and the basic principles of measuring cell migration.
6.24 Explain the reason for and the basic principles of measuring cell uptake.
6.25 Explain the reason for and the basic principles of measuring cell protein secre-
tion.
6.26 Using fluorescence recovery after photobleaching, calculate the fraction of mo-
bile protein where F(–) = 4, F(+) = 1, F(∞) = 2.
6.27 Explain the reason for and the basic principles of measuring cell proliferation.
6.28 Explain the reason for and the basic principles of measuring cell differentiation.
6.29 Explain the reason for and the basic principles of measuring cell signaling and
regulation.
7
Nervous System
Jang-Zern Tsai
The nervous system is defined as all cells, tissues, and organs that regulate the body’s
response to both internal and external stimuli (Costello, 1994). There are many
neurological diseases and disorders that can be assessed by the use of medical equipment.
Biomedical engineers need to understand how the nervous system works in order to
design this equipment.
All cells in the body have a cell membrane surrounding them. Across this membrane
there is an electric charge referred to as the resting potential. This electric impulse is
generated by differential ion permeability of the membrane. In the cells, potassium (K+)
channels allow diffusion of K+ ions out of the cell while sodium (Na+) ions diffuse into
the cell. This Na+–K+ pump, which requires ATP to operate, pumps two K+ ions into the
interior of the cell for every three Na+ ions pumped out. K+ and Na+ ions are
continuously diffusing across the membrane from where they were just pumped, but at a
slower rate. Since there are more K+ ions inside the cell than outside, a potential exists.
The magnitude of the equilibrium membrane resting potential is given, at body
temperature, by the Nernst equation
1
2 Nervous System
Vm = (RT/nF)ln[K+]e/[K+]I (7.1)
Vm = 0.0615 log10[K+]e/[K+]i
Some cells, such as skin cells, are not excitable. Other cells, such as nerve and muscle
cells, are excitable. When a stimulating electric field acts on an excitable cell, the Na+
permeability increases, Na+ enters the cell interior, and the entering positive charge
depolarizes (reduces to approximately zero) the transmembrane potential. Later the K+
permeability increases and K+ ions flow out to counter this effect. The Na+ gates close
followed by the K+ gates. Finally, the resting potential is regenerated. See Figure 7.2.
The action potential lasts about 1 ms in nerves and about 100 ms in cardiac muscle. It
propagates in nerves at about 60 m/s and carries sensation from the periphery toward the
brain via sensory nerves. Through motor nerves, the brain commands muscles to contract.
Figure 7.1 shows an invertebrate nerve axon under stimulation of an electric potential
high enough to excite the depolarization of the axon (Stein, 1980). We can calculate the
action potential propagation velocity v = d/t, where d = distance and t = time.
7.1 Cell Potential 3
Electronic Indifferent
stimulator electrode
1m Micropipet electrode
Stimulus
Action potential 0V
Resting potential
–70 mV
Time
0 20 ms
Figure 7.1 The action potential of the nerve axon in response to electric stimulus. The
depolarization process first occurs at the stimulation site. The action potential then travels
downstream to the recording site where a penetrating micropipet is used to pick up the intracellular
potential with respect to an indifferent reference potential.
4 Nervous System
Na+ Na+
K+ K+
1 Resting phase 3 Repolarizing phase
Na+ Na+
K+ K+
2 Depolarizing phase 4 Undershoot phase
Membrane potential
2 3
+50
0 1 4
(mV)
−50
−100
t
Figure 7.2 The role of voltage-gated ion channels in the action potential. The circled numbers on
the action potential correspond to the four diagrams of voltage-gated sodium and potassium
channels in a neuron's plasma membrane (Campbell et al., 1999).
+7.2 Brain, EEG and Evoked Potentials
The brainstem is a short extension of the spinal cord and serves three major functions: a
connecting link among the cerebral cortex, the spinal cord, and the cerebellum; a center
of integration for several visceral functions; and an integration center for various motor
reflexes.
The cerebellum is a coordinator in the voluntary (somatic) muscle system and
acts in conjunction with the brainstem and cerebral cortex to maintain balance and
provide harmonious muscle movements.
Figure 7.3 shows the cerebrum, which occupies a special dominant position in
the central nervous system. Within the cerebrum are the conscious functions of the
nervous system.
Figure 7.3 The cerebrum contains the frontal, parietal, temporal and occipital lobes. From A. B.
McNaught and R. Callander, Illustrated Physiology, 3rd ed., 1975. Edinburgh: Churchill
Livingstone.
5
6 Nervous System
The electroencephalogram (EEG) provides information about the health and function of
the brain by detecting electric impulses in the brain. EEG can help diagnose conditions
such as epilepsy, brain tumors, brain injury, cerebral palsy, stroke, liver, kidney disease,
or brain death and helps physicians find the cause of problems such as headaches,
weakness, blackouts or dizziness (WFUBMC, 2000).
By removing a portion of the skull, it is possible to insert microelectrodes into
the brain and record action potentials from single cells. Because this procedure is so
invasive, the vast majority of clinical studies are made from electrodes glued onto
standard locations on the scalp as shown in Figure 7.4. These electrodes average the
action potentials from large numbers of cells and therefore do not provide action
potentials from single cells. Some localization is possible by using the monopolar
connection, which is a recording from a single electrode referenced to the average of all
other electrodes. Some localization is also possible by using the bipolar connection,
which is a recording between successive pairs of adjacent electrodes. The EEG is
typically 100 µV in amplitude with a frequency response of 0.5 to 80 Hz. Sixteen
channels are usually recorded simultaneously (Hughes, 1994).
Figure 7.4 The 10–20 electrode system for measuring the EEG. From H. H. Jasper, The Ten
Twenty Electrode System of the International Federation in Electroencephalography and Clinical
Neurophysiology, EEG Journal, 10 (Appendix), 371–375.
The EEG is typically recorded with the subject awake and resting. Figure 7.5(b)
shows that when the eyes are closed, alpha waves with a frequency of about 10 Hz appear.
As the subject becomes drowsy and falls asleep, the frequency decreases and the
7.2 Brain, EEG and Evoked Potentials 7
(c)
8 Nervous System
Figure 7.5 (a) Four types of EEG waves. (b) When the eyes are opened, alpha waves disappear. (c)
Different types of epilepsy yield abnormal waveforms. From A. C. Guyton, Structure and Function
of the Nervous System, 2nd. Ed., Philadephia: W. B. Saunders, 1972.
An evoked potential is the electric response signal of the brain to a specific stimulus. The
response is usually of low amplitude (<5 µV), especially in relation to the background
EEG signal (50 to 100 µV), so that stimulus-related cerebral evoked potentials usually
cannot be identified in routine EEG recordings. With the use of the signal averaging
technique, the background EEG and artifacts can be minimized. For example, a stimulus
occurs N = 100 times and each time the response is recorded. All 100 responses at 0 ms
are added. All 100 responses at 1 ms are added, etc. The signal averaged output
waveform is a plot of the added outputs at 0, 1, 2, etc. ms. Because these responses are
synchronized to the stimulus, they add linearly and are N times larger. The interfering
noise potentials from the EEG are random and have an amplitude equal to N . Thus
the signal-to-noise ratio (SNR) increases by a factor of N N = N (Isley et al.,
1993).
Studies of evoked potential are useful in evaluating the functional integrity of
sensory pathways. They permit the lesions of these pathways, which may not have any
visible symptoms, to be detected and localized. They may be helpful in establishing or
supporting a neurological diagnosis, following the course of certain disorders, and
assessing sensory function (such as vision or hearing) when normal behavioral or
subjective testing is not possible. The electric potentials associated with cognitive
processes can also be recorded through the scalp and are helpful in the evaluation of
patients with suspected dementia. Evoked potentials provide objectivity, because
voluntary patient response is not needed. Clinical abnormalities can be detected by an
increase in latency (the delay between the stimulus and the wave response).
During visual-evoked potential (VEP) testing, the subject stares at a central spot on a
reversing checkerboard pattern 1 m away. The pattern consists of a counterphase
alternation of dark and light squares without any change in total luminance. The reversing
rate is about twice per second. The evoked potential generated in the occipital cortex is
detected by an electrode placed on the scalp in the midoccipital position with reference to
either the linked mastoids or a midfrontal electrode. A recorded VEP is obtained by
amplifying, synchronizing, and averaging the responses to 100 such reversals. The
procedure is usually done twice to ensure reproducibility and the two eyes are tested
separately. VEP evaluates the visual nervous system from the eyes to the brain. It may
help diagnose certain visual or brain disorders (Gevins and Aminoff, 1988) (WFUBMC,
2000).
7.2 Brain, EEG and Evoked Potentials 9
During somatosensory evoked potential (SSEP) testing, peripheral nerves are stimulated
with a small electric current through bipolar electrodes on the skin of the subject’s arms
or legs. Electrodes placed on the subject’s scalp and along the spinal cord can detect a
series of positive and negative potentials following the peripheral stimulation. The
intensity of the stimulus is generally very small. About 1000 responses are averaged to
reduce the interference from endogenous cardiac, motor, and EEG activities and
exogenous electric noise. The examination is repeated at least once to ensure
reproducibility of the results. SSEP assesses the neurological pathways from the nerves in
the arms or legs through the spinal cord to the brain. It can also be used to monitor coma
patients, monitor patients during surgical procedures, and test hearing in infants and
others whose hearing cannot be tested in the standard way (Gevins and Aminoff, 1988)
(WFUBMC, 2000).
The X rays travel in all directions, but to prevent patient and operator harm, are shielded
by a collimator so only those used in the image proceed. Secondary radiation could fog
the film, but is stopped by a grid shaped like Venetian blinds. Phosphor screens emit
many light photons for each X-ray photon, thus assisting in darkening the photographic
film. To lower X-ray dose, an image intensifier may be used. The X-ray photon strikes a
phosphor layer in a vacuum tube. Many light photons stimulate a photocathode to emit
many electrons. These are accelerated to strike an output phosphor screen at +25 kV,
yielding a good image with low X-ray dose (Siedband, 1998).
Figure 7.6 The 100 kV X-ray tube generates X rays that form a shadow of the body to expose the
film. Unlike a camera, there are no lenses.
Figure 7.7 In computed tomography, a thin X-ray beam measures the attenuation along a single
path. The system is indexed to obtain attenuation in many parallel paths at one angle and then
repeated at many angles. The information is processed to yield attenuation at every element in the
2-dimensional slice. From R. A. Brooks and G. Di Chiro, Theory of image reconstruction in
computed tomography. Radiology, 117: 561–572, 1975.
12 Nervous System
Y
1/3 + 1/3 1/3 + 1 1/3 + 1/3
1 = = =
2/3 4/3 2/3
X
1 3 1
Figure 7.8 Creating images using backprojection. Each row and column signifies a certain path
angle. The values of 3 in the center row and column are uniformly backprojected with a value of 1
in each cell. These sum to the maximum value of 2, thus indicating an object in that cell. The
values of 1 are uniformly backprojected with a value of 1/3 in each cell. A sum of these two
numbers that is significantly greater relative to the others indicates an object.
One slice
in x axis
One slice
in y axis
Single line
y x
Figure 7.9 In MRI, RF coils excite one slice in the y-axis. An RF receiver measures from one slice
in the x-axis.
Figure 7.10 MRI images of the brain.(a) A sagittal view of a normal brain. (b) A sagittal view of a
patient with an arterio-venous malformation (AVM) which is an explosive growth of the capillary
bed. (c)(d) Brain images from the patient with the AVM in which the visual cortex of the brain was
activated using flashing lights while rapid MR imaging was performed. The regions of the brain
that were "activated" by the visual stimulus are displayed in white on the images. Images are from
University of Wisconsin-Madison fMRI (functional magnetic resonance imaging) laboratory.
Gamma camera
Figure 7.11 shows a gamma camera. Gamma rays from the organ travel in all directions,
but the collimator tubes ensure that only radiation from a small region reaches the NaI
detector. Light is detected by more than one photomultiplier tube, but the relative
amounts received improve spatial resolution. An electronic pulse height analyzer selects
events that have the proper gamma-ray energy. The resulting image has spatial resolution
of about 1% of the image dimension (Mettler and Guiberteau, 1991).
7.6 Brain Imaging: Nuclear Imaging 15
Figure 7.11 In a gamma camera system, radioisotopes emit gamma rays, which are collimated,
strike a NaI crystal, which emits light measured by photomultiplier tubes.
For example, CO can be made with 11C. If a portion of the brain is active, increased
blood flow carries the isotope to it, where it shows up on the image. Abnormal
functioning, tumors, seizures and other anomalies may also be mapped this way. For
example, measurement of glucose-fluorodeoxyglucose (FDG) metabolism is used to
determine tumor growth. Because small amounts of FDG can be visualized, early tumor
detection is possible before structural changes occur and would be detected by MRI or
CT.
Figure 7.12 In the PET camera, (a) the paired and (b) the hexagonal ring cameras rotate around the
patient. (c) The circular ring does not rotate but may move slightly to fill in the gaps between the
detectors.
When the retina is stimulated with light, an action potential is superimposed on the
resting potential. Measuring the electrical response of the retina to flashes of light detects
inherited retinal degenerations, congenital (existing from birth) retinal dystrophies, and
toxic retinopathies from drugs or chemicals. The electrical response of the retina to a
reversing checkerboard stimulus can help distinguish between retinal and optic nerve
dysfunction to diagnose macular disease (Vaughan and Asbury, 1995). Macular disease is
the leading cause of blindness in elderly Americans today and it is also linked to smoking.
This disease affects the central area of the retina, causing gradual loss of central vision.
The degeneration is caused by the partial breakdown of the insulating layer between the
retina and the choroids (the layer of blood vessels behind the retina) (Yahoo! Health,
2000).
The retina has electrical dipole properties that result in the corneal surface being
electropositive with respect to the outer scleral surface. Figure 7.13 shows that the record
is made by placing an active electrode on the cornea, usually one embedded in a corneal
contact lens, with saline solution bridging the gap between the electrode and the cornea,
and placing an indifferent electrode on the forehead. An alternative to the contact lens
electrode would be to use electrodes that hook over or rest on the lower eyelid. These
electrodes require a large number of responses, but are usually a better option for children
than the contact lens electrodes. The entire retina is stimulated with light and the small
voltage is amplified and recorded. The retina is stimulated with light after either dark
(scotopic) or light (photopic) adaptation (Newell, 1996).
Figure 7.13 The transparent contact lens contains one electrode. The reference electrode is placed
on the right temple.
18 Nervous System
Flash ERG
In flash ERG, the inner and outer nuclear layers of the retina produce a and b waves,
respectively (Figure 7.14). Some of the cells that make up the two layers of the retina are
the photoreceptors, horizontal cells, bipolar cells, and Muller cells. All of the cells in the
two layers give a graduated hyperpolarizing or depolarizing response to light, as
compared with the action potentials produced by the retinal ganglion cells. The a wave of
the ERG is produced by the hyperpolarizing photoreceptors. This wave has separate rod
and cone components. The b wave is generated from the activity of depolarizing bipolar
cells. The a wave represents photorecptorial activity and the b wave represents
postphotoreceptorial activity in the second-order retinal neuronal network, but the
sources are still unclear.
To differentiate between rod and cone activity, special conditions can be applied.
Since the rods of the eyes are the visual receptors used in dim light, recording from the
rod system requires long periods of dark adaptation and one very dim single flash
stimulus. Using scotopic conditions and intense white flashes produces a mixed rod and
cone response. To obtain cone response, intense light flashes accompanied by an intense
light background are required (Good, 1999). The differentiation between rod and cone
activity is useful in the diagnoses of several diseases including inherited retinal
degenerations (mentioned above), which can cause night blindness. The degeneration
starts at more peripheral parts of the retina and progressively decreases the field of vision.
This disease can be recognized in the later stages by narrowing of the retinal vessels, and
by scattered clumps of pigment throughout the retina (Encyclopaedia Britannica, 1999).
7.10 Eye, ERG, EOG and Visual Field 19
Fig 7.14 The normal standardized corneal electroretinogram. S.F. is standard flash, and responses
to 1 log unit less than and greater than standard flash intensities are shown, as well as single flash
photopic and 30 Hz responses. Scale 100 V/div. and 20 ms per div. (Good, 1999)
The resting potential of the front of the retina is electropositive with respect to the back of
the retina. This makes the cornea positive compared to the back of the eye. If the eye
looks left, this positive cornea makes an electrode to the left of the eye more positive than
an electrode to the right of the eye. We can measure the differential voltage of about 100
µV between the two electrodes to determine horizontal direction of gaze. Electrodes
above and below the eye can determine the vertical direction of gaze. This electro-
oculogram (EOG) is useful for measuring changes in direction of gaze, but cannot
measure absolute direction of gaze because variation in skin potential is larger than the
EOG signal. This test can be used to detect retinal pigment epithelial dysfunction
(Verdon, 2000).
20 Nervous System
The peripheral vision of the retina reduces rapidly away from the fovea centralis (near the
center of the retina). This can be assessed by visual field measurement, in which the
examinee detects the targets that stimulate the retina at varying distances from the fovea.
Several methods are used to measure visual field, including the confrontation test and
methods using perimeters, tangent screens, or an Amsler grid (Newell, 1996).
Confrontation test
For this test the patient fixates their vision straight ahead on the examiner at a distance of
1 m. The patient closes one eye. The examiner puts one hand midway between the patient
and the examiner with some fingers extended. The examiner then slowly moves the hand
from the periphery toward the center and the patient states the number of fingers
displayed. The temporal and nasal fields and the superior and the inferior fields are tested
in turn.
Perimeters
A perimeter consists of a half bowl with a radius of curvature of 33 cm so that the eye is
at the center of rotation of the hemisphere.
In kinetic perimetry, the patient detects a test object of fixed size and
illumination moved from a nonsensing area. In static perimetry, the patient detects a test
object of constant size and location whose light intensity is increased until it can be seen.
An isopter, the contour line that connects the points at which an object of certain
size and certain light intensity may be recognized, represents a line on the macula lutea
that passes through the points of equal visual acuity.
Tangent screen
The evaluation of the central 30° of the visual field is desirable because the
photoreceptors of the eye are concentrated in and near the fovea centralis. This can be
conducted with a test object of size from 1 to 50 mm on a tangent screen, which contains
concentric circles every 5° and radiating lines 15° or 22.5° apart stitched or marked on
black cloth 1 m from the subject. This test demonstrates the peripheral isopter, the blind
spot, and various scotomas (blind spots in the visual field).
Amsler grid
An Amsler grid is formed by intersecting 21 horizontal and 21 vertical dark lines spaced
about 5 mm with one dark dot in the middle. It encompasses the central 20° of the visual
field when held at the reading distance. The patient tests an eye by fixating it on the
7.10 Eye, ERG, EOG and Visual Field 21
central dot and indicates any irregular line section that appears blurred, wavy, curved,
discolored, or missing. This may be an indication of leakage or bleeding in the back of
the eye causing retinal swelling (Freund, 1997). Amsler grid testing is recommended for
patients with macular degeneration or other retinal disordrs that threaten central vision. It
is advantageous because most patients are able to administer the test themselves, and if a
change in condition is noticed, they can come in to see their doctors immediately.
Examiner’s view
Biprism
Rod
Housing
Adjustment knob
Figure 7.15 The Goldmann applanation tonometer. When the applanation head is moved to the
right to rest on the anesthetized cornea, the examiner rotates the adjustment knob to increase the
force.
7.10.5 Ophthalmoscopy
Ophthalmoscopy is used to inspect the interior of the eye. It permits visualization of the
optic disk, vessels, retina, choroid (layer of blood vessels behind the retina), and ocular
media. Direct ophthalmoscopy and indirect ophthalmoscopy are the two methods used.
(Newell, 1996)
Direct ophthalmoscopy
Indirect ophthalmoscopy
Mirror
Examiner’s eyes
Lamp
Inverted image of
patient’s fundus
Condensing lens
Figure 7.16 In indirect ophthalmoscopy, the examiner stands in front of the patient and holds the
condensing lens at an appropriate distance from the patient’s eye.
The ear is a sense organ specialized for hearing and balance functions. Anatomically, it is
divided into the external ear, middle ear, and inner ear. The external ear is composed of
the auricle, the external auditory canal, and the outermost layer of the tympanic
membrane. The middle ear, or tympanic cavity, is an oblong cavity bridged by an
ossicular chain consisting of three small bones between the tympanic membrane and the
inner ear. The inner ear consists of the cochlea and the labyrinth, which are the end organ
receptors for hearing and balance (Cody et al., 1981).
7.11 Ears and Audiometry 25
The auditory portion of the ear can be physiologically divided into two partsa
sound-conducting apparatus and an electromechanical transducer (Hawke, 1990). The
sound-conducting apparatus consists of the external ear, the tympanic membrane, the
ossicular chain, and the labyrinthine fluid. The electromechanical transducer transforms
the mechanical energy of sound into electric impulses to be transmitted by the auditory
nerves to the auditory cortex of the brain.
7.11.2 Audiometry
The audio signal passes through the outer ear, the middle ear, and the inner ear before
being further processed by the central auditory system. Pathology of any of these parts of
the ear can cause hearing loss. In this test, the subject responds to bursts of single-
frequency stimuli presented through calibrated earphones. The signal intensity is adjusted
and the threshold with 50% correct responses is recorded (Laszlo and Chasin, 1988).
A special vibratory transducer is placed on the mastoid process or the forehead and thus
stimulates the inner ear directly through the skull. This bypasses the outer and middle
ears. If there is a difference between the air conduction and bone conduction responses,
this indicates pathology in the outer or middle ear, which may be treatable. Inner ear
pathology is difficult to treat (Laszlo and Chasin, 1988).
In this test, the subject listens to lists of single-syllable speech discrimination words
presented through earphones and repeats what he or she hears. The result of this test is
scored from 0 to 100% based on the correctness of the subject’s answer. In contrast to the
pure tone threshold test, which addresses hearing sensitivity, this test assesses the
integrity of the entire auditory system’s ability in hearing clearly and understanding
speech communication. A low score is related to sensorineural loss, and a higher score
may be attributed to normal hearing or conductive hearing loss (Laszlo and Chasin, 1988).
This test uses words that are attenuated successively. The patient repeats back two-
syllable words that have equal stress. A threshold is determined when the patient repeats
50% of the words correctly (Laszlo and Chasin, 1988).
Impedance audiometry
In tympanometry, a probe in the ear varies air pressure from +1962 to –1962 Pa testing
middle ear function and for the presence of fluid in the middle ear. A stiff middle ear
(poor ossicular chain mobility) reduces the mobility of the eardrum, causing a 220 Hz
tone to be reflected with little attenuation. A flaccid middle ear system (possible ossicular
chain disarticulation) yields high damping and little reflection (Laszlo and Chasin, 1988).
Clicks in the ear yield electric potentials measured by an electrode at the vertex (top) of
the scalp. An average response computer separates the very small synchronized signals
from the random signals from other brain potentials. Early latency evoked potentials
(about 10 ms) are from the brain stem. Middle latency responses (about 30 ms) are
responses above the brain stem. Late latency evoked potentials (about 100 ms) evaluate
the neurological integrity of the cortical centers.
7.11.3 Otoscopy
The external auditory canal and tympanic membrane, or eardrum, can be inspected with
an otoscope, a hand-held instrument with a flashlight, a magnifying glass, and a cone-
shaped attachment called an ear speculum. Evidence of injury or congenital (present at
birth) malformation is usually obvious. One can find foreign bodies, inflammation, ulcers,
lacerations, and tumors by visually inspecting the external auditory canal with a otoscope.
This type of ear examination can also detect a rupture or puncture of the eardrum. One
may also see scarring, retraction of the eardrum, or bulging of the eardrum. A red or
swollen eardrum can indicate an ear infection. In physical diagnosis, the otologist views
normal anatomy and pathologic processes in the middle ear through the translucent
tympanic membrane (Cody et al, 1981).
7.12 Muscles
About 40% of the human body is skeletal muscles, and another 10% is smooth muscles of
internal organs and cardiac muscles from the heart. Here we are interested in
7.12 Muscles 27
characterizing the function of skeletal muscles. The primary function of skeletal muscles
is to generate force. Because of this, they are excitable. Thus skeletal muscles have two
fundamental properties: They are excitable (able to respond to stimulus) and contractible
(able to produce tension) (Biewener, 1992).
A skeletal muscle consists of numerous fibers with diameters ranging from 10 to
80 µm. Each muscle fiber contains hundreds to thousands of myofibrils (Figure 7.17).
Each myofibril has about 1500 myosin filaments and 3000 actin filaments lying side by
side (Carlson and Wilkie, 1974). Figure 7.18 shows the structure of a single myofibril.
Figure 7.17 The components of a skeletal muscle; the myofiber is the smallest complete contractile
system. Each myofiber is composed of many myofibrils. (from The Muscle Physiology Laboratory,
2000)
28 Nervous System
Myofibril
"Thin" actin
"Thick" myosin filament
filament
Sarcomere
Z disk
Figure 7.18 The structure of a single myofibril. Each myofibril contains about 1500 myosin
filaments and 3000 actin filaments in the pattern shown above. The distance shown between the
two Z disks is one sarcomere. During muscle contraction the thin (actin) filaments slide inward past
the thick (myosin) filaments pulling the two Z disks closer together, thus shortening the sarcomere.
The shortening of the thousands of sarcomeres in each myofibril causes the myofibrils, the
myofibers and then the entire muscle to contract. (from The Muscle Physiology Laboratory, 2000)
As an action potential travels along a motor nerve to muscle fibers, it initiates an
action potential along the muscle fiber membrane, which depolarizes the muscle fiber
membrane and travels within the muscle fiber. The subsequent electrochemical reaction
within the muscle fiber then initiates attractive forces between the actin and myosin
filaments and causes them to slide together. This mechanism produces muscle contraction
(Jones and Round, 1990).
Tension is developed in the muscle as it contracts. There are three types of
contraction: Isometric, or static, contraction means a muscle contracts without change in
its length. Concentric contraction occurs when a load is less than the isometric force
produced by the muscle and the load shortens the muscle. Eccentric contraction occurs
when the load is greater than the isometric force and elongates the contracting muscle
(Sigel, 1986).
7.12 Muscles 29
Muscle
Fp
Fa
Muscle length
Resting length
Figure 7.19 The overall force of a muscle, F, is the sum of the active force, Fa, and the passive
force, Fp. The active force results from voluntary contraction of the contractile elements of the
muscle. The passive force results from elongation of the connective muscle tissue beyond its
resting length. No passive force builds up when the muscle is at its resting length or less.
The amount of force generated by a stimulated muscle depends on how its ends
are restrained (Woledge et al., 1985). If neither end of a stimulated muscle is fixed, the
muscle shortens at its maximum velocity of 33 cm/s, Vmax, and no force is generated. If
we fix one end of the muscle and apply a small force to the other end, the muscle
shortens at a steady velocity less than Vmax. A force of sufficient magnitude, F0, will
prevent the muscle from shortening and isometric contraction occurs within the muscle.
The force–velocity relation is described by Hill’s equation (Woledge et al., 1985):
where a and b are constants derived experimentally, F is the muscle force during
shortening at velocity V, and F0 is the maximum isometric force that the muscle can
produce. Figure 7.20 shows a hypothetical curve for force–velocity relation according to
Hill’s equation.
30 Nervous System
Shortening velocity
Vma
Figure 7.20 The muscle force decreases with increasing shortening velocity.
The force–velocity curve can be obtained experimentally with isotonic
measurement (Lieber, 1992). The muscle is allowed to change its length against a
constant load during maximal stimulation. The muscle velocity (i.e., shortening velocity,
or lengthening velocity) is measured and plotted against the constant load.
By placing electrodes into a skeletal muscle, we can monitor the electric activity of the
muscle (Keynes and Aidley, 1991). EMG is used to detect primary muscular disorders
along with muscular abnormalities caused by other system diseases such as nerve
dysfunction (Junge, 1992). Using EMG to study muscle function is also used in various
fields such as kinesiology, psychology, and rehabilitation medicine (Basmajian and De
Luca, 1985).
The two main types of electrodes for measuring EMG signals from muscles are
surface electrodes and inserted electrodes. The inserted type further includes needle
electrodes and wire electrodes. Selection of electrode type depends on the particular
application and the convenience of use (Loeb and Gans, 1986).
The surface electrode consists of silver disks that adhere to the skin. Saline gel
or paste is placed between the electrode and the skin to improve the electric contact. The
disadvantages of surface electrodes are that they cannot effectively detect signal from
muscles deep beneath the skin and that because of poor selectivity, they cannot eliminate
cross-talk from adjacent muscles.
Needle electrodes have a relatively smaller pickup area and thus are more
suitable for detecting individual motor unit potentials. Needle electrodes can be
7.12 Muscles 31
repositioned in the muscle after they are inserted. Figure 7.21 shows various
constructions of needle electrodes (Kimura, 1989).
(a)
(b)
(c)
(d)
(e)
Figure 7.21 Examples of needle electrodes: (a) Monopolar needle with exposed pickup tip. (b)
Monopolar configuration with a pickup wire centrally located in a hypodermic needle. (c) Bipolar
configuration with two pickup wires in parallel with each other in a hypodermic needle. (d)
Monopolar configuration with a wire exposed at needle’s side hole. This can be used to detect the
activity of individual muscle fibers. (e) Bipolar configuration with two wires exposed at the side
hole of a hypodermic needle. This can be used to detect the motor unit action potential from a large
portion of the motor unit territory.
Figure 7.22 shows a wire electrode. This electrode consists of two insulated fine
wires, 25 to 100 µm in diameter, inserted through the cannula of a hypodermic needle.
The distal tips of the wires are deinsulated by about 1 to 2 mm and are bent to form
staggered hooks. One disadvantage of wire electrodes is that they tend to migrate after
insertion into the muscle.
Figure 7.22 A bipolar wire electrode uses a hypodermic needle to insert through the skin into the
muscle of interest.
We can obtain the EMG signal simply by placing a surface electrode on the skin
enveloping the muscle or by applying an inserted electrode in the muscle. A reference
electrode, usually a surface electrode, is placed on a site with minimal electric association
with the inserted site. The drawback of this monopolar configuration is that it detects not
only the signal from the muscle of interest but also unwanted signals from around the
muscle of interest.
In the bipolar configuration, two electrodes with a small distance between each
other are placed in the muscle to pick up the local signals within the muscle of interest. A
differential amplifier amplifies the signals picked up from the two electrodes with respect
to the signal picked up by a reference electrode. Because the interference signals from a
32 Nervous System
distant source are essentially equal in magnitude and phase as detected by the two
electrodes, the common mode rejection capability of the differential amplifier eliminates
the unwanted signals.
Figure 7.23 shows EMG waveforms of tongue muscle taken when the subject
was speaking some vowels.
x 10x 10
4
4
Signal vs. number of points
Si gn al v s . Nu m ber o f Po in ts
55
Amplitude
Am pl i tud e
00
-5
-5
0
0
4 50
50
100
10 0
150
150
200
200
250
25 0
x 10
4
x 10
55
Amplitude
0
Am pl itude
-5
-5
0
0
4 50
50
100
10 0
150
150
200
200
250
25 0
x 10
4
x 10
55
Amplitude
Am pl i tu de
00
-5
-5
0
0
50
50
100
10 0
150 150
200
200
250
25 0
Number of points
Num ber of Poin ts
Figure 7.23 The EMG signal measured from the hyoglossus (tongue) muscle (b), styloglossus
muscle (c), and genioglossus muscle. Wire electrodes were inserted into these tongue muscles to
take the EMG measurement. The EMG signals are associated with the speech signal (a) of the
letters /ia/iao/ian/iang/iong.
Nerve cells that control movement emerge from the spinal cord and brainstem. They
travel various distances to reach particular muscles. Neuromuscular disorder develops
when any part along the path, from the brain to the muscle, does not work properly. If the
nerve that controls the muscle is injured and the action potential from the brain cannot
reach the muscle, the muscle does not contract normally. On the other hand, if the muscle
has some disease, it may not produce normal contraction in response to the action
potential from the brain.
For adequate treatment of neuromuscular disorders, it is important to know
where the injury occurs and how serious it is. By applying a dc stimulus to the skin over a
nerve at various positions in the legs or arms, and recording the time of muscle twitch
due to the stimulation with an electrode on the skin over the muscle, one can calculate the
7.12 Muscles 33
conduction velocity of the nerve. The conduction is slowed with injury to the nerve
(Ringel, 1987).
EMG is also used for diagnosis of neuromuscular performance. Abnormal
electric signals are recorded by the EMG if there is disease in the muscle.
7.13 References
American Academy of Family Physicans (AAFP). What You Should Know About
Epilepsy. [Online] http://familydoctor.org/handouts/214.html
Basmajian, J. V. and De Luca, C. J. 1985. Muscle Alive: Their Functions Revealed by
Electromyography. 5th ed. Baltimore, MD: Williams & Wilkins.
Bickford, L. 1995. The EyeCare Connection: The EyeCare Reports. [Online]
http://www.eyecarecontacts.com/exam.html
Biewener, A. A. (ed.) 1992. Biomechanics-Structures and Systems: A Practical Approach.
New York : IRL Press at Oxford University Press.
Campbell, N. A., Reece, J. B. and Mitchell L. G. 1999. Biology. 5th ed. Menlo Park:
Benjamin/Cummings.
Carlson, F. D. and Wilkie, D. R. 1974. Muscle Physiology. Englewood Cliffs, NJ:
Prentice-Hall.
Castello, R. B. (ed.) 1994. The American Heritage Dictionary. New York: Dell.
Clark, J. W. 1998. The origin of biopotentials. In Webster, J. G. (ed.) Medical
Instrumentation: Application and Design. 3rd ed. New York: John Wiley & Sons.
Cody, T. R., Kern, E. B. and Pearson, B. W. 1981. Diseases of the Ears, Nose, and
Throat. Mayo Foundation.
Early, P. J. and Sodee, D. B. 1995. Principles and Practice of Nuclear Medicine. 2nd ed.
St. Louis : Mosby.
Encyclopaedia Britannica. 2000. Eye disease: Retinal degeneration. [Online]
http://www.eb.com:180/bol/topic?eu=117507&sctn=18#s_top
Freund, K. B. 1997. Amsler Grid Testing of Central Vision. [Online]
http://www.vrmny.com/amsler.htm
Gevins, A. S. and Aminoff, M. J. 1988. Electroencephalography: Brain Electrical
Activity. In J. G. Webster (ed.) Encyclopedia of Medical Devices and
Instrumentation. New York: John Wiley & Sons.
Good, P. 1999. The Electroretinogram. [Online] http://www.eye-
news.com/vol5_6.dir/features.dir/2feat5_6.html
Hari, R., and Lounasmaa, O. V. 2000. Neuromagnetism: tracking the dynamics of the
brain. Physics Today 33–38, May 2000.
Hine, G. J. 1967. Instrumentation in Nuclear Medicine. New York: Academic.
Hughes, J. R. 1994. EEG in Clinical Practice. 2nd ed. Boston: Butterworth-Heinemann.
International Conference on Biomagnetism. 1989. Advances in Biomagnetism. New York:
Plenum Press.
Isley, M. R., Krauss, G. L., Levin, K. H., Litt, B., Shields, R. W. Jr. and Wilbourn, A. J.
1993. Electromyography/Electroencephalography. Redmond, WA: SpaceLabs
Medical.
34 Nervous System
Jones, D. A. and Round, J. M. 1990. Skeletal Muscle in Health and Disease : A Textbook
of Muscle Physiology. Manchester, New York: Manchester University Press.
Junge, D. 1992. Nerve and Muscle Excitation. 3rd ed. Sunderland, MA: Sinauer
Associates.
Keynes, R. D. and Aidley, D. J. 1991. Nerve and Muscle. 2nd ed. Cambridge, New York:
Cambridge University Press.
Kimura, J. 1989. Electrodiagnosis in Diseases of Nerve and Muscle: Principles and
Practice. 2nd ed. Philadelphia: Davis.
Laszlo, C. A. and Chasin, D. J. 1988. Audiometry. In J. G. Webster (ed.) Encyclopedia of
Medical Devices and Instrumentation. New York: John Wiley & Sons.
Lieber, R. L. 1992. Skeletal Muscle Structure and Function: Implications for
Rehabilitation and Sports Medicine. Baltimore: Williams & Wilkins.
Loeb, G. E. and Gans, C. 1986. Electromyography for Experimentalists. Chicago:
University of Chicago Press.
Malmivuo, J. and Plonsey, R. 1995. Bioelectromagnetism: Principles and Applications of
Bioelectric and Biomagnetic Fields. New York: Oxford University Press.
Mazziota, J. C. and Gilman, S. 1992. Clinical Brain Imaging: Principles and
Applications. Philadelphia: F. A. Davis Company.
Mettler, F. A. and Guiberteau, M. J. 1991. Essentials of Nuclear Medicine Imaging. 3rd
ed. Philadelphia: W. B. Saunders.
The Muscle Physiology Laboratory. 2000. Myofilament structure. [Online]
http://muscle.ucsd.edu/musintro/fibril.shtml
The Muscle Physiology Laboratory. 2000. Fiber types. [Online]
http://muscle.ucsd.edu/musintro/fiber.shtml
NATO Advanced Study Institute on Biomagnetism. 1983. Biomagnetism: An
Interdisciplinary Approach. New York: Plenum Press.
Newell, F. W. 1996. Ophthalmology: Principles and Concepts. 8th ed. St. Louis: Mosby -
Year Book.
Ozkaya, N. and Nordin, M. 1991. Fundamentals of Biomechanics: Equilibrium, Motion,
and Deformation. New York: Van Nostrand Reinhold.
Ringel, S. P. 1987. Neuromuscular Disorders: A Guide for Patient and Family. New
York: Raven Press.
Ritch, R., Shields, M. B. and Krupin, T. (eds.) 1996. The Glaucomas. 2nd ed. St. Louis:
Mosby.
Schneck, D. J. 1992. Mechanics of Muscle. 2nd ed. New York: New York University
Press.
Siedband, M. P. 1998. Medical imaging systems. In Webster, J. G. (ed.), Medical
Instrumentation: Application and Design. 3rd ed. New York: John Wiley & Sons.
Siegel, I. M. 1986. Muscle and its Diseases: An Outline Primer of Basic Science and
Clinical Method. Chicago: Year Book Medical Publishers.
Stein, R. B. (ed.) 1980. Nerve and Muscle: Membranes, Cells, and Systems. New York:
Plenum Press.
Van Heertum, R. L. and Tikofsky, R. S. (eds.) 1995. Cerebral SPECT Imaging. 2nd ed.
New York: Raven Press.
Vaughan, D. G. and Asbury, T. 1999. General Ophthalmology. 15th ed. Appleton &
Lange.
7.13 References 35
7.14 Problems
7.1 Calculate the magnitude and sign of the cell membrane potential at body
temperature given that the intracellular potassium concentration = 150
millimoles/L and the extracellular potassium concentration = 5 millimoles/L.
7.2 Figure 7.1 shows the change of membrane potential due to an electric stimulus.
Suppose that the extracellular potassium concentration keeps constant during the
time course, sketch a figure to show the change of the intracellular potassium
concentration.
7.3 Calculate the velocity of propagation for the action potential in Figure 7.1.
7.4 Explain the origin of action potentials and how to measure them.
7.5 Explain why and how we measure the electroencephalogram.
7.6 Explain why and how we measure evoked potentials.
7.7 Draw the block diagram of a signal averager (average response computer) for
use with evoked potentials and explain how it improves the signal and by what
factor.
7.8 Calculate the improvement in signal-to-noise ratio when an average response
computer averages 200 recordings.
7.9 Explain the reason for an X-ray filter, collimator, grid, and screen.
7.10 Describe the beam paths used in a CT scanner.
7.11 Explain how magnetic fields are varied to obtain information from a single line
of tissue in an MRI scanner.
7.12 Describe the advantages of using nuclear medicine.
7.13 We wish to image the functionally active regions of the thyroid gland, which
takes up iodine when functionally active. Explain the procedure and equipment
used, including the detector type, localization method, and energy selection
method.
7.14 Explain how a gamma camera forms an image.
7.15 Explain why and how we diagnose visual system disease.
36 Nervous System
Supan Tungjitkusolmun
Cardiovascular disease is the leading cause of death in the United States. There are over
one million heart attacks per year and over 600,000 deaths, of which 300,000 die before
reaching the hospital. One important trend that has emerged is the emphasis on preven-
tion. Today, most people are aware of the things they can control that prevent heart dis-
ease (diet, not smoking, exercise, blood pressure, etc.). Attention to these areas helps re-
duce the chance of getting the disease but does not eliminate it completely. Early detec-
tion and treatment of the disease continue to be key areas of emphasis in the medical
community. The level of sophistication in dealing with cardiovascular diseases has taken
great strides in the past two decades and has been greatly aided by advances in technol-
ogy. New diagnostic techniques have been developed and old ones improved, providing
added sensitivity and specificity. It is now very common for primary care physicians to
conduct tests in their offices to detect cardiovascular diseases.
It is very important for biomedical engineers to understand the basic principles
of medical instruments. This chapter presents several important cardiovascular variables
and the most common methods by which they are measured, as well as an introduction to
the physiology of the heart and blood vessels.
The human heart and blood vessels are a transportation system that delivers es-
sential materials to all cells of the body and carries away the waste products of metabo-
lism. The heart serves as a four-chamber mechanical pump. The two chambers on the
right side of the heart send deoxygenated blood to the lungs via the pulmonary arteries
(pulmonary circulation). The other two chambers, on the left side of the heart, supply
oxygenated blood through the arteries, which branch in an orderly fashion until they
reach the capillary beds of the organs (systemic circulation). Here, there is an exchange
of nutrients, waste products and dissolved gases. The blood then flows from the capillar-
ies into the veins, which lead back to the right atrium. Hormones are also transported by
the cardiovascular system.
Figure 8.1 illustrates important features of the human heart. The heart, about the size of a
clenched fist, is located in the thoracic (chest) cavity between the sternum (breastbone)
and the vertebrae (backbone). The heart is divided into right and left halves and has four
chambers—an upper and lower chamber within each half. The upper chambers, the atria,
receive blood returning to the heart and transfer it to the lower chambers, the ventricles,
which pump the blood from the heart. Figure 8.2 shows a simplified circulatory system.
Blood returning from the systemic circulation enters the right atrium via large veins
known as the venae cavae. The blood entering the right atrium has returned from the
body tissues, where O2 has been extracted from it and CO2 has been added to it by the
cells of the body. This partially deoxygenated blood flows from the right atrium into the
right ventricle, which pumps it out through the pulmonary artery to the lungs. Thus, the
right side of the heart pumps blood into the pulmonary circulation. Within the lungs, the
blood loses its CO2 and picks up a fresh supply of O2 before being returned to the left
atrium via the pulmonary veins. This highly oxygenated blood returning to the left atrium
subsequently flows into the left ventricle, the pumping chamber that propels the blood to
all body systems except the lungs; that is the left side of the heart pumps blood into the
systemic circulation. The large artery carrying blood away from the left ventricle is the
aorta.
Figure 8.1 Basic structure of the heart. RA is the right atrium, RV is the right ventricle; LA is the
left atrium, and LV is the left ventricle. Basic pacing rates are shown.
8.2 Cardiac Biopotentials 3
Blood flows through the heart in one fixed direction—from veins to atria to ventricles to
arteries (see Figure 8.2). The presence of four one-way heart valves aides the unidirec-
tional flow of blood. The valves are positioned so that they open and close passively due
to pressure differences, similar to a one-way door. The backward gradient can force the
valve to close but cannot force it to swing open in the opposite direction.
Deoxygenated Oxygenated
blood blood
Upper body
Lung
Right Left
atrium atrium
Right Left
ventricle ventricle
Lower body
Figure 8.2 The simplified circulatory system. The blood is delivered from the right ventricle to the
lung. The oxygenated blood from the lung is then returned to the left atrium before being sent
throughout the body from the left ventricle. Deoxygenated blood from the body flows back to the
right atrium and the cycle repeats.
Two of the heart valves, the right and left atrioventricular (AV) valves, are posi-
tioned between the atrium and the ventricle on the right and left sides, respectively (see
Figure 8.1). These valves allow blood to flow from the atria into the ventricles during
ventricular filling (when atrial pressure exceeds ventricular pressure), but prevent the
backflow of blood from the ventricles into the atria during ventricular emptying (when
ventricular pressure greatly exceeds atrial pressure). If the rising ventricular pressure did
not force the AV valves to close as the ventricles contracted to empty, much of the blood
would inefficiently be forced back into the atria and veins instead of being pumped into
the arteries. The right AV valve is also called the tricuspid valve because it consists of
three cusps, or leaflets. Likewise, the left AV valve, which consists of two leaflets, is
often called the bicuspid valve or, alternatively, the mitral valve.
4 Heart and Circulation
The other two heart valves, the aortic and pulmonary valves, are located at the
junction where the major arteries leave the ventricles. They are also known as semilunar
valves because they are composed of three leaflets, each resembling a shallow half-moon
shaped pocket. These valves are forced open when the left and right ventricular pressures
exceed the pressure in the aorta and pulmonary arteries (i.e. during ventricular contrac-
tion and emptying). The valves close when the ventricles relax and ventricular pressures
fall below the aortic and pulmonary artery pressures. The closed valves prevent blood
from flowing from the arteries back into the ventricles from which it has just been
pumped. The semilunar valves are prevented from swinging back into the ventricles by
the anatomical structure and positioning of the cusps (Sherwood, 2001).
A cardiac cycle contains a sequence of events during one heart beat. The ventricles con-
tract during systole. The ventricles fill during diastole. In an average human at rest, the
entire cardiac cycle takes about 0.8 s (giving a pulse of about 75 beats/min). Heart rate is
the number of heartbeats per minute. We can easily measure our own heart rate by count-
ing the pulsations of arteries in the wrist (radial artery) or neck (carotid artery). Figure 8.3
shows the synchronization of the electrocardiogram (see section 8.2.2) and the pressures
in the left ventricle.
8.2 Cardiac Biopotentials 5
Figure 8.3 In the top figure, the electrocardiogram (ECG) initiates the cardiac cycle. The cardiac
sounds (section 8.5) are also shown. The bottom figure shows that ejection occurs when the pres-
sure in the left ventricle exceeds that in the arteries.
During diastole, the pressure in the left ventricle is low and less than that of the
arteries. Thus the aortic valve is closed. The blood from the left atrium flows into the left
ventricle. The left atrium subsequently contracts due to the excitation from the pacemaker
cells in the heart (see section 8.1.3) and further fills the left ventricle. The ventricles con-
tract, which increases the pressure and closes the mitral valve. The pressure increases
until it exceeds the aortic pressure, the aortic valve opens, and blood flows into the aorta.
The blood continues to flow from the left ventricle into the aorta as long as the ventricu-
lar pressure is greater than the arterial pressure, as shown in Figure 8.3. However, if the
venticular pressure is much greater than the aortic pressure, there is a problem with the
valve and it may be stenotic (narrow). The dicrotic notch results when the aortic valve
slams shut and the AV valve is still closed. Table 8.1 summarizes the events of the car-
diac cycle.
6 Heart and Circulation
Table 8.1 Duration and characteristics of each major event in the cardiac cycle.
The aortic pulse contour changes with aging due to changes in aortic compliance
(distensibility). Reflection of the systolic pulse from the distal aorta distorts the aortic
contour and creates two systolic shoulders. A pulse from a younger person with a more
compliant aorta is rounded. That from an older person with decreased aortic compliance
has a reflected wave, which causes a more peaked pulse and higher pressure (hyperten-
sion) (O’Rourke, 1995).
Currently, there are numerous electric heart rate monitoring systems that con-
tinuously measure heart rate during rest and exercise. They rely on sensing electrodes
placed on the skin surface, which measure the electric frequency of the heart, or the elec-
trocardiogram (section 8.2.2). Some models, like the ear clip, use a photosensor to meas-
ure change in tissue light absorption caused by the pulsatile increase in arteriole volume,
which is then converted into a beat-per-minute readout.
Figure 8.1 shows that the sinoatrial node (SA node), which serves as the heart’s pace-
maker, is located in the wall of the right atrium, near the point where the inferior vena
cava enters the heart. It is composed of specialized muscle tissue that combines character-
istics of both muscle and nerve (see Chapter 7). Nodal tissue contracts like muscle, thus it
generates electric impulses similar to those found in nervous tissue. This tissue is self-
excitable, and can contract without any signal from the nervous system. Each time the SA
node contracts, it initiates a wave of excitation that travels through the wall of the heart at
a rate of approximately 1 m/s (Berne, 1981). The impulse spreads rapidly, and the two
atria contract simultaneously. Cardiac muscle cells are electrically coupled by the interca-
lated disks between adjacent cells. At the bottom of the wall separating the two atria is
another patch of nodal tissue, the atrioventricular (AV) node. The atria and ventricles are
separated by nonconductive tissue except for the AV node. When the wave of excitation
reaches the AV node, it is delayed for about 0.1 s (AV-nodal delay), which ensures that
the atria will contract first before the ventricles contract. After this delay, the signal to
contract is conducted to the apex of the ventricles along the bundle of His, and the wave
of excitation then spreads upward through the ventricular walls via the Purkinje fibers
(Davies et al., 1983).
8.2 Cardiac Biopotentials 7
Simple tests of the functions of the heart include monitoring the heart rate, the
blood pressure, the cardiac rhythm, and the heart sounds. More sophisticated diagnoses
can be performed by studies of the electrocardiogram, the echocardiogram, the stress test,
Holter monitoring, and angiography. These techniques will be discussed in the subse-
quent sections.
8.2.1 Electrodes
Surface recording electrodes convert the ionic current within the body to electron current
in the metal wires, which is then sent to a recorder. Electrodes are generally made from
metal. The silver/silver chloride (Ag/AgCl) electrode is electrochemically stable and is
thus most widely used (Brown et al., 1999). An electrode is usually constructed by elec-
trolyzing a silver plate as an anode in an aqueous solution of sodium chloride to form a
film of AgCl on the surface of the silver. The reaction is Ag + Cl– → AgCl + e–. Often a
transparent electrolyte gel containing chloride ions as the principle anion is used. The gel
also allows for good contact between the skin–electrode interface. Figure 8.4 shows a
disposable surface electrode. Tobuchi et al. (1981) discuss several techniques used to
manufacture Ag/AgCl electrodes. Other metals such as nickel-plated brass can be used as
reusable ECG electrodes but they have larger motion artifacts.
Figure 8.4 A disposable surface electrode. A typical surface electrode used for ECG recording is
made of Ag/AgCl. The electrodes are attached to the patients’ skin and can be easily removed.
8.2.2 Electrocardiogram
Contraction of cardiac muscle during the heart cycle produces electric currents within the
thorax. Voltage drop across the resistive tissue can be detected by electrodes placed on
the skin and recorded as an electrocardiogram (ECG, sometimes called EKG). Figure 8.3
shows components of an electrocardiogram (P, Q, R, S, and T waves).
The electrocardiogram contains information on the electric rhythm, electric con-
duction, muscle mass, presence of arrhythmia (irregular heart beat), ischemia (lack of
blood flow) or infarction, even electric disturbance and drug effects on the heart (see Fig-
ure 8.5). The P wave in the ECG represents the contraction of the atria. From Figure 8.3,
the R wavealso called the QRS complexof the ECG corresponds to the contraction
of the ventricles. The T wave is due to repolarization (re-establishing the electric poten-
tial) of the ventricles. The atria repolarize also, but this event is covered up by the QRS
complex. The downslope of the T wave is referred to as the vulnerable period. It is during
this time of the cardiac cycle that a premature impulse may arrive and fibrillation can
occur. Fibrillation is an irregular contraction of cardiac muscle, which results in an inef-
fective propulsion of blood. When ventricular fibrillation occurs a normal rhythm can be
re-established by passing a strong electric current. This puts the entire myocardium in a
refractory state in which no impulse can occur. Thus, “resetting” the heart. Direct current
defibrillation shock has been found to be more effective than alternating current shock
(Berne, 1981).
I aVR V1 V4
(a)
I aVR V1 V4
(b)
I aVR V1 V4
(c)
8.2 Cardiac Biopotentials 9
I aVR V1 V4
(d)
I aVR V1 V4
(e)
I aVR V1 V4
(f)
Figure 8.5 The electrocardiogram. (a) The normal ECG. (b) 1st degree AV block in which the
delay from the P wave to the Q wave is lengthened. (c) acute inferior myocardial infarction (lack of
blood flow to heart muscle, which causes tissue to die), in which the S–T sement is depressed. (d)
right atrial hypertrophy (increase in muscle mass of the atria), in which V4 is large. (e) ventricular
tachycardia (faster than normal heart rate) with clear AV dissociation. (f) Wolff–Parkinson–White
syndrome with atrial fibrillation.
RA
Figure 8.6 Block diagram of an electrocardiograph. The normal locations for surface electrodes are
right arm (RA), right leg (RL), left arm (LA), and left leg (LL). Physicians usually attach several
electrodes on the chest of the patients as well.
Figure 8.7 A circuit of an ECG amplifier. The instrumentation amplifier, located on the left of the
circuit provides a high input impedance and has a gain of 25 in the dc-coupled stages. (From Web-
ster, J. G. (ed.) 1998. Medical instrumentation: application and design, 3rd ed. New York: John
Wiley & Sons.)
8.2 Cardiac Biopotentials 11
Connecting each lead of the ECG to a buffer amplifier prevents loading error
caused by the added resistance of the skin. The buffer amplifier offers a very high input
impedance and unity gain. If a differential amplifier or inverting amplifier were used
without a buffer, the gain would be affected (see section 2.2.5).
12-lead ECG
The electric potential generated by the heart appears throughout the surface of the body.
We can measure the potential differences between surface electrodes on the body. Differ-
ent pairs of electrodes at different locations generally yield different results because of
the spatial dependence of the electric field of the heart (Neuman, 1998b).
Physicians attach several surface electrodes and record a few beats of the 12-
lead diagnostic ECG (see Figure 8.6). The 12-lead ECG monitors the heart’s electrical
activity from 12 different angles. By selecting different pairs of electrodes, we can obtain
ECGs that have different shapes. Common locations for the electrodes are the
limbsright arm, left arm, right leg, and left leg. A pair of electrodes, or combination of
several electrodes through a resistive network that gives an equivalent pair, is referred to
as a lead. The most widely used bedside ECG diagnostic system records 12 different po-
tential differences, or ECG leads (Tompkins, 1993).
Figure 8.8 shows the Einthoven triangle. The three leads shown are lead I, right
arm (RA) to left arm (LA); lead II, right arm (RA) to left leg (LL); and lead III, LA to LL.
The scalar signal on each lead of Einthoven’s triangle can be represented as a voltage
source, thus we can write Kirchhoff’s voltage law for three leads as
I − II + III = 0 (8.1)
12 Heart and Circulation
RA LA
LL
Figure 8.8 Einthoven’s triangle. Lead I is from RA to LA, lead II is from RA to LL, and lead III is
from LA to LL.
Three additional leads, called the unipolar leads, or augmented limb leads, are
routinely used in taking diagnostic ECGs. These leads are based on signals obtained from
more than one pair of electrodes. Unipolar leads consist of the potential appearing on one
electrode taken with respect to an equivalent reference electrode, which is the average of
the signals from two or more electrodes. The three different unipolar leads are positive
electrodes placed on the right arm, augmented right (aVR), the left arm, augmented left
(aVL), and the left foot, augmented foot (aVF). The other leads that are frequently meas-
ured in clinical ECGs are the precordial leads. Physicians place an electrode at various
anatomically defined positions on the chest wall (Neuman, 1998b). Required frequency
response is 0.05 to 150 Hz.
ECG monitoring
The ECG may be monitored continuously when the patient is in emergency care, in a
coronary care unit, an intensive care unit, or during stress tests, by a recorder or telemetry.
In these cases, only one lead, usually lead II, is monitored on a display. The system
automatically monitors rhythm disturbances and gives an alarm when the heart rate is too
slow or too fast. To avoid motion artifacts and muscle noise the frequency response is
reduced to 0.5 to 40 Hz.
Medical personnel usually monitor one lead of the patient’s ECG using a bedside monitor
in the critical care unit (CCU) of a hospital. As the patient recovers, the doctor may feel
comfortable with the patient moving around but may want to maintain ECG monitoring.
A bedside monitor limits the movement of the patient to the length of the ECG leads.
With a telemetry system, a nurse attaches the patient leads to the patient in the same way
as for the bedside monitor. Instead of the patient leads feeding back directly to the bed-
side monitor, they feed a portable transmitter. The transmitter, which resembles a port-
able radio, is small enough for the patient to carry. The transmitter sends the patient’s
ECG by modulating it with a radio-frequency (RF) carrier, much in the same way that a
cordless telephone works. Manufacturers may transmit the information using frequency
modulation (FM), amplitude modulation (AM) or even pulse code modulation (PCM)
depending upon design. Antenna systems pick up the RF signal from the transmitter and
send it to the receiver. The receiver then removes the RF signal from the patient’s ECG
and sends the resulting signal to the monitor for analysis. The signal strength of the
transmitter and tuning of the system limit the distance a patient can move away from the
antenna system. In the ECG telemetry system, the receiver alarms on high and low ECG
limits the user can adjust, or possibly some arrhythmia conditions (Reinhard and Fowler,
1987).
Exercise stress testing is an evaluation of the patient’s cardiovascular system using one
lead of an ECG, monitor, treadmill, and blood pressure unit. It is done to screen patients
for heart disease and help predict or unmask potential coronary problems.
The procedures for stress testing are as follows. The technician first abrades the
skin and applies stress electrodes to the patient’s chest. Then, a baseline ECG and blood
pressure are taken. The exercise protocol (there are numerous protocols, which gradually
increase both speed and elevation of the treadmill) is started. During exercise, the physi-
cian monitors the patient’s ECG, heart rate, blood pressure, and physical symptoms. S–T
segment depression may indicate cardiac ischemia due to coronary arterial occlusion.
When the test ends, ECG and blood pressure measurements are done for documentation
of the heart rate at maximum effort or submaximum effort. It is important to continue
monitoring the patient’s ECG, heart rate, and blood pressure during recovery, because
some problems are uncovered during that phase.
8.2.3 Electrograms
Electrograms are electric signals recorded from within the heart. Electrograms are meas-
ured by introducing a catheter electrode into the veins. The catheter is then placed into
the heart under X-ray guidance, which is usually a biplane cardiac imaging system with a
C-arm, allowing multidirectional biplane projections with extreme angulation capabilities.
We can use electrograms to study abnormal heart rhythms under controlled situations to
diagnose the specific problem with the heart’s electric system. Electrophysiologists can
8.2 Cardiac Biopotentials 15
obtain information on the propagation of excitation from the SA node to the ventricles
with the catheter.
There are certain groups of patients who have problems that can be corrected by
radio-frequency ablation—the use of RF energy via a catheter for lesion making to mod-
ify the conduction pathway of the heart (Huang and Wilber, 2000). One example is pa-
tients with refractory arrhythmias of the upper chamber who need complete interruption
of the AV conduction pathway. Electrophysiologists usually monitor the abnormal con-
duction of the heart by applying three ablation catheters, each having three recording
electrodes and one ablation electrode at the tip. The catheters are placed at different loca-
tions, such as the SA node, the AV node, and in the ventricle. By observing the timing of
the excitation activity at various locations, the electrophysiologist can determine the loca-
tions of unwanted conduction pathways. They are then heated and destroyed with RF
energy.
Any invasive procedure that requires catheters to be inserted in the body in-
cludes some risk. However, the risk is quite small and the electrophysiology study is rela-
tively safe. One possible side effect is that some patients may develop bleeding at the site
of catheter insertion into the vessels.
Pressure port
Catheter
Figure 8.9. A system for cardiac pressure and flow measurement. The pressure is usually measured
with a catheter placed in the right side of the heart. An external strain gage pressure sensor is also
shown. (Adapted from Orth, J. L. 1995. System for calculating compliance and cardiac hemody-
namic parameters, US Patent, 5,423,323.)
Pressure measurement systems in use today are essentially all electric strain gages and
apply the principle of the Wheatstone bridge, as described in sections 2.1.6 and 9.11.3.
Older pressure sensors had unbonded strain gage wires mounted under stress between a
frame and movable armature so that preload was greater than any expected external com-
pressive load. This was necessary to avoid putting the wires in compression. Thus blood
pressure caused diaphragm movement, which caused resistance change, which caused a
voltage change.
Modern pressure sensors use an etched silicon diaphragm embedded with strain
gages formed by diffusing impurities into it. The strain gage electric resistance changes
with diaphragm deformation, which responds to pressure applied. Unfortunately the resis-
tance of the silicon also changes with temperature, but the effect of this change may be
minimized by diffusing four strain gages into the diaphragm. These then form the four
arms of a resistance bridge. Two of the strain gages are placed close to the center of the
diaphragm, where their resistance increases with applied pressure, and the other two close
to the periphery, where their resistance decreases with pressure. However, if the tempera-
ture changes, all four resistances change by the same percentage and this does not change
the output from the resistance bridge.
8.3 Cardiac Pressures 17
8.3.2 Catheter
In the cardiac catheterization laboratory, catheters are inserted into the chambers of the
heart to measure pressures, flows, and oxygen saturation to determine if valve replace-
ment is required. Catheters can also inject radiopaque dye for X-ray fluoroscopy, which
can image vessel narrowing to determine if vessel replacement is required.
A catheter is a flexible tube for insertion into a narrow opening such as the
blood vessels, so that fluids may be introduced or removed. Figure 8.9 shows a Swan–
Ganz catheter. The first person who passed a catheter into the human heart was Werner
Forsmann, a medical student at Eberswalde, Germany. In 1929, at the age of 25, he
passed a 65 cm catheter through one of his left antecubital veins, guiding it by fluoros-
copy until it entered his right atrium. Forsmann then climbed up the stairs to the Radiol-
ogy Department where the catheter position was then documented.
Catheters are inserted through sheaths (the layers of connective tissue that en-
velop structures such as nerves, arteries, tendon, and muscle) into the arteries and the
veins of the body then pass up to the heart. The pressures within different areas of the
heart are measured. Two approaches may be used to apply a catheter to the heart.
The brachial (in the arm) approach usually utilizes cutdown on the brachial
artery (an artery that extends from the auxillary artery, down the side and inner surface of
the upper arm to the elbow, where it divides into the radial and ulnar arteries) and vasilic
vein at the elbow. The direct brachial approach may have advantages in a very obese pa-
tient, in whom the percutaneous femoral technique may be technically difficult and
bleeding hard to control after catheter removal.
The indirect femoral approach. The percutaneous femoral approach has its own
set of advantages. Arteriotomy and arterial repair are not required; it can be performed
repeatedly in the same patient at intervals, whereas the brachial approach can rarely be
repeated more than two or three times with safety.
Cardiac pressures are usually measured by one of several ways. In the catheter-
type system, the blood pressure communicates with the pressure sensor via a fluid-filled
catheter. The sensor is outside the body. With an appropriate length and diameter, accu-
rate pressure readings can be obtained.
With the catheter-tip sensor, the elastic unit is placed in a blood vessel, thus the
elastic member is in direct contact with the blood. This arrangement avoids the damping
and resonance sometimes encountered with catheter-type systems and is more accurate
than the catheter-type sensors (Grossman and Baim, 1991).
culation per unit time is termed the cardiac output, generally expressed in liters/minute
(L/min).
Cardiac output of a normal patient is directly correlated with the size of the body.
Most investigators use total body surface area as the standardizing variable. The ratio of
the cardiac output to area is called the cardiac index. Table 8.2 lists several variables re-
lated to cardiac output.
Table 8.2 Some physiological variables. The data presented in this table are the average values of a
group of subjects.
Variables ±SD)
Mean (±
Weight (kg) 70
Cardiac output (mL/s) 110
Heart rate (min–1) 76
Mean velocity, ascending aorta (mm/s) 16
LV end-diastolic volume (mL) 125 (±31)
LV end-systolic volume (mL) 42 (±17)
LV stroke volume (mL) 82 (±20)
LV ejection fraction 0.67 (±0.11)
LV mean wall thickness (mm) 10.9 (±2.0)
Cardiac output (CO) = heart rate (HR) × stroke volume (SV) (8.2)
End-diastolic volume is the amount of blood in the ventricle at the end of dias-
tole. End-systolic volume is the volume of blood remaining in the ventricle at the end of
systole when ejection is complete.
The stroke volume, the volume of ejected blood from the ventricles, is about 80
mL/beat. The average resting heart rate of human is about 70 beats/min. Thus cardiac
output is about 80 × 70 = 5,600 mL/min = 5.6 L/min. Cardiac output is regulated by
changes in both HR and SV. Heavy exercise increases both HR and SV, and CO can in-
crease to as high as 25 L/min.
In 1870, Fick stated that if the concentration of O2 in arterial and mixed venous blood is
known, and the rate of inhalation of O2 is also known, the cardiac output, CO, can be
computed from the formula (Webster, 1998)
dm dt
CO = (8.4)
Ca − Cv
8.4 Cardiac Output 19
Example 8.1 A patient’s O2 concentration, meaused in the pulmonary artery, is 0.12 L/L.
The O2 concentration measured in the patient’s aorta is 0.19 L/L. A spirometer is used to
obtain the patient’s O2 consumption rate 0.250 L/min. Calculate the patient’s cardiac
output.
The patient’s O2 concentrations and consumtion are known: Ca = 0.19 L/L, Cv
= 0.12 L/L and dm/dt = 0.250 L/min. From Eq. (8.4)
dm dt 0.250 L/min
CO = = = 3.57 L/min
Ca − C v 0.19 L/L − 0.12 L/L
8.4.2 Thermodilution
Figure 8.10 The relationship of the temperature gradient and time. (Adapted from Baker, P. D., Orr,
J., Westenskow, D. R. and Johnson, R. W. Method and apparatus for producing thermodilution
cardiac output measurements utilizing a neural network, US Patent, 5,579,778.)
The cardiac output can be calculated from (Webster, 1998)
|Q|
CO = t1
(m3/s) (8.5)
ρbcb ∫ | ∆Tb (t ) | dt
0
where Q is the heat injected in joules, ρb is the density of the blood in kg/m3, cb is the
specific heat of the blood in J/(kg⋅K), and ∆Tb is the temperature gradient function as
shown in Figure 8.10. The area under the temperature versus time curve is inversely pro-
portional to the blood flow or cardiac output. Therefore, healthy patients with normal
cardiac output will have less area under temperature versus time curve, than unhealthy
patients with some type of coronary insufficiency. If cardiac output is reduced, the bolus
takes longer to pass through the heart.
Example 8.2 Given the following data: blood density = 1060 kg/m3, heat capacity of the
blood = 3640 J/(kg⋅K), amount of heat injected = 1.25 J, and the area under the curve in
Figure 8.10 is 1.0 s⋅°C, calculate the cardiac output.
From Eq. (8.5), the cardiac output can be computed as
8.4 Cardiac Output 21
1.25 J
CO = 3
= 3 × 10 −7 m 3 /s. (8.6)
1060 kg/m × 3640 J/(kg ⋅ °C) × (1.0 s ⋅ °C)
The disadvantage of this method is its invasiveness. Also, the lower temperature
bolus can be rapidly warmed to body temperature. Repeat measurements are easier to
perform than for the dye dilution technique (see below). It is more difficult to calculate
the size of the temperature perturbation than it is to measure the quantity of injected dye,
however. Another disadvantage of the thermodilution method is that the injecting catheter
acts as a heat exchanger between the blood and the cold saline. The cold saline left at the
end of the injection can further cool the blood. However, the error is not too significant.
External detectors are used to record activity from short-lived radiopharmaceuticals. This
is a procedure that yields a high degree of contrast between the blood and the surrounding
structures. There are two classes of radionuclides:
1. Positron emission tomography (PET), which uses substances that release posi-
trons, which travel a short distance and then react with electrons to emit pairs of photons
on paths 180° apart. This technique is excellent for measuring the extent of myocardial
ischemia and is highly specific for the diagnosis of coronary artery disease (Marwick
1995).
2. Single photon emission computerized tomography (SPECT), which is a
method based on radionuclides such as 99mTc. The imaging device in SPECT is usually a
gamma-ray camera. Diagnostic accuracy and adjunctive assessment of ventricular func-
tion make SPECT the currently favored stress imaging radioisotope technique ( Merz
1997).
Chapter 7 discusses the principles of radionuclides.
22 Heart and Circulation
8.4.5 Echocardiography
Sound of frequencies greater than 20 kHz, the highest frequency that a human ear is able
to detect, are called ultrasound. The application of the ultrasonic imaging technique to the
heart is called echocardiography. This technique is noninvasive. The transducer contains
a piezoelectric crystal, which transmits and receives ultrasound resulting in images.
Echocardiography can measure the size of the heart, functions of the heart, the blood flow
through the heart (when combined with the Doppler technique), and the movement of the
heart wall. It can help identify tumors and clots in the heart. Structural abnormalities of
the heart wall, valves, and the blood vessels going in and out of the heart can be detected
by this technique as well.
Cardiac diagnosis uses high-frequency ultrasound (1.6 to 2.25 MHz). The sound
waves are both emitted and received by the same sensor. Sound waves are reflected at
interfaces between media of different acoustic impedances. Motion mode, or M-mode,
echocardiograms are generated by means of a single ultrasonic beam traversing cardiac
structures with definable anatomic shapes (chamber walls, interventricular septum) or
characteristic motions (cardiac valves). The echocardiograms are then recorded on an
oscilloscope, where the time-of-flight of reflected sound occupies the ordinate, and
slower time the abscissa. Advances in echocardiographic equipment including real-time
cross-sectional scanning (two-dimensional echocardiography), Doppler echocardiography,
and contrast echocardiography have broadened the role of echocardiography in the diag-
nosis of a variety of cardiac diseases.
M-mode echocardiography
With this method, the ultrasonic sensor is placed on the anterior chest. The beam passes
through the right ventricle, interventricular septum, left ventricular cavity, posterior left
ventricular wall, pericardium and lung. Figure 8.11 shows that these structures are por-
trayed on the ordinate and time is displayed on the abscissa. Additional information is
obtained by sweeping the sensor in an arc between the base and apex of the heart. By
proper positioning of the sensor all four cardiac valves, both ventricles, and the left
atrium can be visualized (Bom, 1977). M-mode echocardiography provides good axial
resolution and is therefore superior in qualitative diagnoses such as pericardial thickening,
small effusion, and constrictive hemodynamics (Cikes and Ernst, 1983). However, be-
cause it lacks spatial geometry, most potential information from using many cardiac
cross-sections is unavailable (Cikes and Ernst, 1983).
8.4 Cardiac Output 23
Figure 8.11 Simultaneous recording of motion mode (M-mode) and two dimensional echocardio-
grams. The arrows on the right image indicates the position of the ultrasound beam from which the
M-mode recording was made. LVW = left ventricular wall, LV = left ventricle, LA = left atrium,
RV = right ventricle (From College of Veterinary Medicine, Univ. of Tennessee. 2000. M-Mode
Echocardiography [Online] http://www.vet.utk.edu/m-mode.html)
Figure 8.12 2-D echocardiography of the long axis view of the right ventrical (RV): (a) the ultra-
sonic beam angle through the heart, (b) the cross-sectional diagram of the image and (c) the actual
2-D image. TV = tricuspid valve, AML = anterior mitral leaflet. (Adapted from Rafferty, T. 1992.
Transesophageal two-dimensional echocardiography
http://www.gasnet.org/reference/echomanual/html/2-d_echo..html)
8.4 Cardiac Output 25
Table 8.2 Relative advantages of echocardiographic examination techniques. (Adapted from Roe-
landt, 1983)
Doppler echocardiography
Contrast echocardiography
Contrast echocardiography is based on the fact that the injection of almost any liquid
through a small catheter produces a cloud of echoes on the M-mode echocardiogram. The
most commonly used dye for contrast echocardiography is indocyanine green. The mi-
crobubbles are too large to pass through the capillaries and yield a cloud of echos. Thus,
an injection of indocyanine green on the right side of the heart remains in the right side of
the heart unless there is a right-to-left shunt. We measure the cardiac output by tracing
the intravascular bubbles with a high degree of sensitivity by ultrasound (Meltzer and
Roelandt, 1982).
MRI provides higher spatial resolution than ultrasound and the capability of measuring
the velocities in the three spatial directions for any plane through the heart (Ingels et al.,
1996). In contrast to radionuclide angiography, MRI utilizes no ionizing radiation. It can
also diagnose the characteristics of the cardiac muscle. Also, the stroke volume can be
calculated by integrating the flow curve in the main pulmonary artery during a cardiac
cycle. MRI offers excellent anatomical detail of the cardiovascular system, particularly in
congenial heart disease and in diseases of the aorta (Underwood, 1992).
Cardiac sounds shown in Figure 8.3 are associated with the movement of blood during
the cardiac cycle. Table 8.3 gives descriptions of the origins of various heart sounds. The
first heart sound is low-pitched, soft, and relatively long. It is associated with the close of
the AV valves. The opening of the valves does not produce any sound. The sounds are
caused by vibrations within the walls of the ventricles and major arteries during valve
closure, not by the valves snapping shut. The first heart sound signals the onset of systole.
The second heart sound has a higher pitch, is shorter and sharper. It is associated with the
closure of the semilunar valves that occurs at the onset of ventricular relaxation as the left
and right ventricular pressures fall below the aortic and pulmonary artery pressures, re-
spectively. Thus the second heart sound signals the onset of ventricular diastole. Mur-
murs are vibrations caused by turbulence in the blood moving rapidly through the heart.
Table 8.4 lists the types of murmurs and what they represent.
Table 8.3 The heart sounds. The 1st and 2nd heart sounds are most prominent.
Sound Origin
1st sound Closure of mitral and tricuspid valves
2nd sound Closure of aortic and pulmonary valves
3rd sound Rapid ventricular filling in early diastole
4th sound Ventricular filling due to atrial contraction
Table 8.4 Timing of murmurs. For example, if the physician hears the 1st heart sound, a swishing
sound, and then the 2nd heart sound, the patient likely suffers from AV valve insufficiency.
A stenotic valve is a stiff, narrowed valve that does not open completely. Thus,
blood must be forced through the constricted opening at tremendous velocity, resulting in
turbulence that produces an abnormal whistling sound similar to the sound produced
when you force air rapidly through narrowed lips to whistle. An insufficient valve is one
that cannot close completely, usually because the valve edges are scarred and do not fit
together properly. Turbulence is produced when blood flows backward through the insuf-
ficient valve and collides with blood moving in the opposite direction, creating a swish-
ing or gurgling murmur (Sherwood, 2001). Improved diagnosis can be obtained from
heart sound audiograms, where the magnitude of each frequency component is displayed.
8.5.1 Stethoscopes
8.5 Cardiac Sounds 27
A stethoscope is used to transmit heart sounds from the chest wall to the human ear.
High-frequency sounds, or murmurs, are easier to hear with the diaphragm. The bell,
which should be applied lightly to the chest, transmits low-frequency sounds more effec-
tively—for example, the diastolic murmur of mitral stenosis and third and fourth heart
sounds. Murmurs arise from turbulent blood flow and are characterized by their timing,
quality, and intensity. Intensity is graded from 1 (just audible) to 6 (audible without a
stethoscope).
Figure 8.13 A stethoscope with bell and diaphragm modes. Adapted from Mohrin, C. M., 1995.
Stethoscope. US Patent, 5,389,747.
Figure 8.13 shows a diagram of a stethoscope which incorporates both the bell
and diaphragm modes into the same chestpiece. Using this particular model (Mohrin,
1995), we can interchange between the bell and diaphragm modes by simply pressing the
chestpiece against a patient’s body and twisting the bell housing so that the two dia-
28 Heart and Circulation
phragms (shown in Figure 8.13) adjust their openings. If the two openings of the dia-
phragms coincide with each other, the stethoscope is operating in the bell mode. If there
is no through opening, it is in the diaphragm mode. Thus, physicians are able to change
between the bell and diaphragm modes without having to remove the bell housing from
the patient’s body.
8.5.2 Microphones
Microphones in use today are either crystal microphones (piezoelectric effect) or dy-
namic microphones (Faraday’s principle) (Peura and Webster, 1998). Piezoelectric sen-
sors are used as microphones since they are in the form of a thin film. They are very use-
ful if one is interested in detecting surface vibrations of an object, such as the heart sound.
When a force, f, is applied to a polarized crystal, the resulting mechanical deformation
gives rise to an electric charge Q. Microphones turn an acoustical pressure into a voltage.
Q = kf = Kx (8.7)
where k is the piezoelectric constant in C/N, K is the proportionality constant in C/m, and
x is the deflecting distance of the piezoelectric sensor.
Figure 8.14 illustrates a charge amplifier. Chapter 2 gives the voltage across an
initially uncharged capacitor, C, by
1 t1 1 t1 dx x
v=
C ∫
0
idt =
C ∫
0
K
dt
dt = K
C
(8.8)
where t1 is the integrating time and i is the current flowing through the capacitor, C. A
charge amplifier acts as a high-pass filter, and Figure 8.15 shows the time constant τ is
τ = RC (8.9)
Hence it only passes frequencies higher than the corner frequency fc = 1/(2πRC).
8.5 Cardiac Sounds 29
Figure 8.14 The piezoelectric sensor generates charge, which is transferred to the capacitor, C, by
the charge amplifier. Feedback resistor R causes the capacitor voltage to decay to zero.
Figure 8.15 The charge amplifier responds to a step input with an output that decays to zero with a
time constant τ = RC.
30 Heart and Circulation
8.6.1 PET
Positron emission tomography uses isotopes that emit photons 180° apart (i.e. in opposite
directions). It has been successfully used to assess myocardial viability. This technique
has the distinct advantage of allowing for the assessment of both myocardial perfusion
and metabolism. This is particularly important since severely ischemic but viable myo-
cardium might show almost no flow, despite the presence of myocardial metabolism.
Under normal circumstances, the myocardium uses fatty acids for its metabolism. How-
ever, when blood flow slows and oxygen supply decreases, anaerobic metabolism devel-
ops, and consequently, the myocardium begins using glucose instead of fatty acids as a
source of energy. A high ratio of regional glucose utilization to myocardial flow has been
found to be a reliable sign of severely ischemic but viable muscle.
Several studies have demonstrated that PET is a better predictor of myocardial
viability than thallium-201 scintigraphy. This probably relates to the improved imaging
resolution of PET and to its added benefit of also assessing metabolism. The advantage of
PET over thallium is lost when sophisticated quantitative analysis of perfusion by thal-
lium is undertaken. The disadvantages of PET are its high cost and nonportability.
Thallium, a member of the heavy metal group of elements, had little biologic use other
than as a rodenticide until its application in myocardial imaging in the early 1970s. Al-
though a heavy metal capable of causing typical heavy metal poisoning at high doses,
doses of 5,000 to 10,000 times less than that associated with clinical poisoning can be
employed for myocardial imaging. The biologic properties of thallium at these low con-
centrations closely resemble those of potassium.
Thallium-201 imaging of the myocardium is one of the more difficult nuclear
medicine procedures from the standpoint of technique. The relatively low photon abun-
dance and low energy of the imaged photons require excellent equipment and technique
8.6 Myocardial Viability 31
to obtain high-quality images. The cost of Thallium-201 and the time required for exer-
cise testing and imaging makes the repetition of technically inadequate studies difficult
(Ritchie et al., 1978).
8.7 Circulation
The three kinds of blood vessels are arteries, veins, and capillaries, which in the human
body have been estimated to extend a total distance of 100,000 km. Arteries carry blood
away from the heart to organs throughout the body. Within these organs, arteries branch
into arterioles, tiny vessels that give rise to the capillaries. Capillaries form networks of
microscopic vessels that infiltrate each tissue. It is across the thin walls of capillaries that
chemicals are exchanged between the blood and the interstitial fluid surrounding the cells.
At the downstream end, capillaries rejoin to form venules, and these small vessels con-
verge into veins. Veins return blood to the heart. Notice that arteries and veins are distin-
guished by the direction in which they carry blood, not by the quality of the blood they
contain. Not all arteries carry oxygenated blood, and not all veins carry blood depleted of
oxygen. However, all arteries do carry blood from the heart to capillaries, and only veins
return blood to the heart from capillaries. The arterial walls are muscular, thus the arteries
are able to control the blood flow through constriction and dilation. The size of the capil-
lary is 10 µm in diameter. The capillaries are where the gaseous exchange of CO2 and O2
occurs.
Peripheral vascular diseases are a group of distinct diseases or syndromes in-
volving the arterial, venous, or lymphatic system (Strandness et al., 1986). The major
vascular disease is coronary artery disease, which kills more than a million people every
year, and afflicts more than 5 million others with disabling symptoms. The development
of various noninvasive tests for diagnosis of arterial disease continues to improve rapidly.
These tests enable us to diagnose and follow the course of arteriosclerosis (abnormal
condition of an artery) much more accurately than was possible in the past. The Doppler
ultrasonic method is the most reliable method of evaluating the severity of lower extrem-
ity ischemia and monitoring postoperative results. A promising new development is Du-
plex scanning (section 8.11.1) of the peripheral arteries, combining Doppler spectral
analysis of blood flow with B-mode ultrasound imaging. Using the Duplex scan to local-
ize the stenosis and to accurately estimate its severity may allow selected vascular recon-
structions to be done in the future without preoperative arteriography.
Venography remains the standard for the diagnosis of deep venous thrombosis
(blood clots), but does have significant limitations. In addition to the cost and discomfort,
some venograms are not diagnostic because of technical reasons. Likewise, invasive
techniques have become useful in the long-term evaluation of treated and untreated pe-
ripheral vascular disease. The increasing sophistication of instrumentation has provided
32 Heart and Circulation
not only physiological information but also anatomic depiction of intraluminal lesions by
ultrasonic imaging. Direct arteriography has also seen new developments with the advent
of digital subtraction angiography. While venous digital subtraction angiography has not
proven to be as clinically useful as originally hoped, application of digital subtraction
techniques to direct arteriography has permitted significant reduction in amounts of con-
trast agent required. By combining digital subtraction techniques with direct arterial an-
giography, patients at high risk of renal failure can be studied safely while combined arte-
rial systems such as abdominal and carotid vessels can be simultaneously studied in pa-
tients with normal renal function. As for the future, magnetic resonance imaging (MRI)
holds promise as an arterial diagnostic tool, particularly with large vessels. However, it
appears unlikely that it will replace ultrasound and contrast arteriography.
R= 8Lη/πr4, (8.11)
∆P is the difference between the mean arterial blood pressure (MABP) and right atrial
pressure. Since the right atrial pressure is about zero, ∆P = MABP. R is the resistance
through all vascular beds in parallel. Fung (1997) provides further information on the
biomechanics of the blood vessels.
For detecting or measuring blood flow, ultrasonic Doppler flowmetry has be-
come the main technique of noninvasive investigation, both in the clinical and research
laboratory. The continuous wave (CW) and pulsed varieties each have their advocates,
the former being easier to use while the latter affords more precise flow interrogation.
Although the audible signal suffices for many clinical applications, recordings are re-
quired for others and for most research endeavors. Recordings made with the zero-
crossing detector are adequate to some extent for real-time frequency spectrum analyzers.
These devices not only depict the velocity flow envelope with precision but also allow
analysis of the power spectrum at each frequency. Increased velocity and disturbances in
8.8 Blood Flow 33
the pattern of blood flow, revealed by these processing methods, are good indicators of
vascular disease. Measurement of volume flow requires knowledge of the cross-sectional
area of the vessel (which can be obtained from the B-mode image) and the mean of the
velocity profile. Since velocity profiles are seldom completely axisymmetric, precise
flow measurement has been difficult to achieve with either the CW or pulsed-Doppler.
Laser Doppler methods are applicable only to the cutaneous tissues and are dif-
ficult to quantify. At present their role remains uncertain. The use of magnetic resonance
imaging (MRI) to measure blood flow, though promising, is in its infancy. Positron emis-
sion tomography (PET) has proved useful in the evaluation of local blood flow, espe-
cially in the brain, but is not widely available and remains prohibitively expensive. The
transcutaneous electromagetic flowmeter seems to have little merit. Monoitoring the
clearance of radionuclides, a semi-invasive technique for evaluting blood flow in isolated
areas, was initially greeted with enthusiasm; however, the washout curves proved to be
difficult to interpret, and the results were inconsistent. At present isotope clearance is
employed only for special clinical studies or for research work.
8.8.1 Dilution
This technique was discussed in section 8.4.3. Applying a similar approach, we can de-
termine the blood flow.
This technique utilizes Faraday’s law of magnetic induction, which states that a voltage is
induced in a conductor that is moving in a magnetic field. It can measure pulsatile flow.
The blood is an electrical conductor, therefore voltage will be developed when the blood
flows through a magnetic field. Figure 8.16 illustrates the principle of an electromagnetic
flowmeter. If the magnetic field, the direction of motion and the induced voltage are mu-
tually at right angles (orthogonal), then
e = Blu (8.13)
B
B
l
N S
e
−
u
Dc excitation
If the generation of the magnetic field is done with dc, the induced voltage at the elec-
trodes is also dc. The induced voltage is small compared to surface potentials generated
at the electrodes. It is difficult to distinguish changes in blood flow from changes in the
surface potentials due to polarization at the electrodes and electrodes may detect other
signals such as electric impulses from the heart. Thus dc excitation is not used commer-
cially.
Ac excitation
If the generation of the magnetic field is done with ac then the induced voltage at the
electrodes is alternating. The surface potentials at the electrodes are removed by capaci-
tive coupling. The ac excitation causes a different problem than dc—the change in mag-
netic field intensity causes an induced current in the conductor, which is proportional to
the rate of change of the magnetic field intensity. The induced current cannot be sepa-
rated from the desired signal at the electrodes. One of the solutions is to measure the
voltage from the electrodes when the magnetic field intensity is not changing.
8.8 Blood Flow 35
Ultrasonic flowmetry is a commonly used technique for measuring blood velocity in the
peripheral arteries. It can measure pulsatile flow. A piezoelectric sensor is used to convert
from electric to acoustic signals (see section 8.5.2). Figure 8.17 shows a system for blood
flow measurement using an ultrasonic flowmeter. The Doppler effect shifts the frequency
of the ultrasonic beam from the oscillator when it intercepts the moving blood cells by an
amount ∆f, which can be computed from
2uf 0 cos θ
∆f = , (8.14)
c
where f0 is the fundamental frequency of an ultrasonic wave from the source, traveling at
velocity c through the blood. The ultrasonic wave intercepts a stream of moving blood
with velocity u crossing the ultrasound beam at an angle θ to produce the frequency shift
∆f. The factor of 2 occurs because the Doppler shift arises both on absorption of the
sound by the moving blood particles and between the transmitting blood cell and the re-
ceiving transducer.
36 Heart and Circulation
Oscillator
Scan head
Blood vessel
Doppler
angle
Computer
RF AF F/V
Detector amplifier
amplifier converter
Figure 8.17 Ultrasonic flowmeter. The sensor at the scan head transmits the signal from the oscilla-
tor and receives the reflected wave from the blood cells. The RF (radio frequency) amplifier ampli-
fies the received signal and the carrier frequency, then AF (audio frequency) signal is produced by
a detector. Adapted from Picot, P. A. and Fenster, A. 1996. Ultrasonic blood volume flow rate me-
ter. US Patent, 5,505,204.
The Doppler probe has a piezoelectric ultrasound transducer integrated into the
tip. The crystal on the tip sends and receives ultrasound waves. The timing of the sending
and receiving velocity allows the ultrasonic flowmeter to measure blood flow velocities.
The returning signal is transmitted, in real time, to the display device. The display shows
the average velocity of all the red cells within the sample volume. When the sample area
is constant, the peak velocity is accurately tracked in the center of the artery and the key
parameters remain relatively positionally insensitive and reliable.
Computerized parameters of intracoronary flow velocity, including peak and
mean diastolic and systolic velocities, diastolic and systolic velocity integrals, mean total
velocities, and the total velocity integral are automatically analyzed.
The Doppler ultrasonic flowmeter can assess blood flow from the Doppler shift that
sound waves experience when they travel through a large blood vessel. However, the
ultrasonic flowmeter is not able to measure the blood flow in small blood vessels, or
microcirculation. In contrast, this is possible if laser light is used instead of sound (Shep-
8.8 Blood Flow 37
herd and Öberg, 1990). Light is capable of measuring the velocities of red blood cells
even at the relatively slow speeds with which they move through capillaries. Laser-
Doppler blood flowmetry thus ulilizes the short wavelengths (extremely high frequencies)
of visible and infrared light to measure the blood flow velocity in a similar manner as the
Doppler ultrasound blood flowmeter. The spectral purity of the laser makes it practical to
detect the slight frequency shifts produced by the interactions between photons and mov-
ing red blood cells.
Figure 8.18 shows the principle of Laser-Doppler measurement. Laser light is
transmitted to the blood vessels and reflected to a detector. Light that intercepts the mov-
ing red blood cells is Doppler-shifted. Light in tissue is scattered by stationary tissue and
it reaches the detector without experienceing a Doppler shift (Shepherd and Öberg, 1990).
A laser-Doppler catheter velocimeter can be used to measure intravascular blood
flow velocity. This can be achieved easily because optical fibers have minute diameters
and permit high resolution. However, catheters disturb the flow of blood in their
neighborhood, and because light is multiply scattered in passing through the blood, com-
bining the high spatial and temporal resolution of laser-Doppler velocimetry with catheter
delivery is not straightforward. At the present time, the development of instruments for
the practical measurement of blood flow velocity through a catheter is an area of active
research.
The major advantage of the laser-Doppler flowmeter is the ability to measure
blood flow in regions other than the limbs. Another advantage is its frequency response.
Not only can the laser-Doppler flowmeter be used to examine variations in blood flow
through the cardiac cycle, but also, on a somewhat longer time scale, slower rhythms can
be observed.
Figure 8.18 Laser-Doppler flowmetry. Light that intercepts the red blood cells experiences a Dop-
pler frequency shift.
Arterial pressure is an essential check on the performance of the heart. The systolic and
diastolic pressures not only provide those maximum and minimum values that are an im-
portant aspect of the pulsations, but also permit a rough estimate of mean pressure in
large arteries. Normal blood pressure is considered to be 120/80 mmHg. Mean blood
flow is essentially the same at all cross sections of the arterial tree, in the total population
of peripheral arterioles as in the ascending aorta. Mean pressure, however, falls continu-
ally from the ascending aorta to the terminus of the vena cava, an indication that the en-
ergy imparted to the blood by ventricular ejection is gradually dissipated as it travels
through the vascular system. The decrement in pressure per unit length of the system, or
vascular resistance, is greatest in the microcirculation.
Blood pressures measured in the distal portions of the extremities are especially
useful for identifying the presence of hemodynamically significant disease. The degree of
circulatory impairment is usually reflected by the relative or absolute reduction in resting
pressure. Stress testing increases the sensitivity of pressure measurement, permitting the
recognition of clinically significant disease in limbs in which resting pressures are
borderline normal. Unfortunately, pressure measurements are not always reliable.
Figure 8.19 The sphygmomanometer detects arterial opening and closing that occurs between sys-
tolic and diastolic pressures.
Cuff
Occluded
blood vessel
Figure 8.20 The pressure of the cuff occludes the blood vessel. When the arterial pressure is
greater than the pressure applied by the cuff, Korotkoff sounds are created and blood pressure can
be measured.
40 Heart and Circulation
Figure 8.21 Top: Cuff pressure with superimposed Korotkoff sounds, which appear between sys-
tolic and diastolic pressures. Bottom: the oscillometric method detects when the amplified cuff
pressure pulsations exceed about 30% of maximal pulsations. From Geddes, L. A. Cardiovascular
devices and their applications. New York: Wiley. 1984.
Arterial and venous blood pressure can be measured by inserting a catheter into the blood
vessel and maneuvering it until the end is at the site at which the blood pressure is to be
measured. In this case the diaphragm of the sensor is mounted at the end of the catheter
outside the body (similar to section 8.2). Catheters for use in the artery are usually thin,
short, and highly flexible. A catheter can be inserted into the artery inside a needle as
well. The alternative method is to use a catheter-tip sensor where the miniature dia-
phragm is at the tip inside the vessel. Although accurate, these methods are very invasive.
Arterial tonometry is a noninvasive technique for monitoring the arterial blood pressure
in a continuous manner. A linear array of pressure sensors is pressed against the radial
artery so that at least one sensor is directly over the lumen. Pressure is increased from
low to high so that measurements are made when the artery is half collapsed. By selecting
the maximal peak-to-peak pressure reading from all sensors, the arterial pressure is de-
termined in the same way as for the applanation tomometer (section 7.10.4). Zorn et al.
(1997) compared the Colin Pilot 9200 tonometric blood pressure measurements with in-
tra-arterial blood pressure measurements. Tonometric values were slightly less than the
intra-arterial pressure measurements; the mean difference for systolic blood pressure was
2.24 ± 8.7 mmHg and for diastolic pressure was 0.26 ± 8.88 mmHg.
∆d
Distension (%) = × 100 (8.15)
dD
where ∆d is the diameter change and dD is the diastolic diameter (Nakatani et al., 1995).
42 Heart and Circulation
Arterial
Blood
wall
Figure 8.22 A catheter is inserted through the blood vessels. A rotating ultrasonic transducer is
attached at its tip and illuminates the walls.
8.10.2 Angiography
Peripheral angiograms are most commonly done to test the arteries that supply blood to
the head and neck or the abdomen and legs. Because arteries do not show up well on or-
dinary X rays, angiograms utilize a contrast agent containing iodine, which is injected
into the arteries to make them visible on radiographs. A catheter is placed into the artery,
via the groin, and is manipulated by the physician into the artery requiring study. Once
the catheter is in place, contrast agent, or dye is injected through the catheter into the ar-
teries and a series of X rays taken. The dye allows the doctor to see the inside of the pa-
tient’s artery and determine how well blood is moving through the vessel.
In digital subtraction angiography (DSA), a first image is taken without contrast
agent. Then a second image is taken with contrast agent. The two images are digitized,
then subtracted, yielding no image in all regions without contrast agent. Vessels contain-
ing constrast agent show very well, so less contrast agent is required.
One application of vessel volume flow measurement is the prediction of stenosis of both
common and internal carotid arteries by monitoring common carotid blood flow. Gener-
ally, an adequate degree of stenosis (usually accepted as 50% diameter reduction), results
in a measurably decreased volume flow. Vessel volume flow measurements may also be
applied in the diagnosis and treatment of vascular malformations. Specifically, the meas-
urement of volume flow may help to distinguish arteriovenous malformations and fistulae,
which are high-flow lesions, from venous malformations, which are low-flow lesions.
Moreover, volume flow measurements provide a quantitative way both of assessing blood
steals (deviation of normal flow) and of evaluating the effectiveness of embolization
(mass of foreign matter in the blood vessel) therapy. Renal dialysis patients may also
benefit from these measurements.
To estimate the volume flow rate of blood through an artery, pulsed Doppler ultrasound
measures the velocity of the blood at 16 or more discrete locations across the vessel di-
ameter. From the velocity measurement, and a measurement of the diameter of the vessel,
we estimate the volume of blood flowing through the vessel. Other techniques, such as
color M-mode, directly measure the one-dimensional velocity profile and are also used.
A clinical color Doppler ultrasound instrument is used with a position and orien-
tation sensing device, and a computer with video digitizer to acquire blood velocity
measurements in two dimensions across an entire blood vessel lumen. Subsequently, the
44 Heart and Circulation
volume flow rate can be determined by integrating the two-dimensional velocity profile
over the vessel lumen area. The volume flow rate measurements are then displayed to an
operator in real time as a scrolling graph.
With the developments in video and in computer technology over the past two decades,
we are now able to see structures that are very fine, movements that were too fast too be
seen, and images that were too dim or too noisy to be seen. Generally, the components
used in video microscopy are a camera, mixer, recorder, processor, analyzer, and monitor.
It also utilizes advances in image-processing (Inoué, 1986). Chapter 6 discusses the prin-
ciples of video microscopy.
8.12 References
Berne, R. and Levy, M. 1981. Cardiovascular Physiology. 4th ed. St. Louis: Mosby.
Bom, N. (ed.) 1977. Echocardiography: With Doppler Applications and Real Time Imag-
ing. The Hague: Martin Nijhoff Medical Division.
Brown, B. H., Smallwood, R. H., Barber, D. C., Lawford, P. V., and Hose D. R. 1999.
Medical Physics and Biomedcial Engineering. Bristol, UK: IOP Publishing.
Burch, G. and Winsor, T. 1972. A Primer of Electrocardiography. 6th ed. Philadelphia:
Lea and Febiger.
Cikes, I. and Ernst, A. 1983. New aspects of echocardiography for the diagnosis and
treatment of pericardial disease. In J. Roelandt (ed.) The Practice of M-mode and
Two-dimensional Echocardiography. Boston: Martinus Nijhof.
Cohn, P. F and Wayne, J. 1982. Diagnostic Methods in Clinical Cardiology. Boston: Lit-
tle, Brown.
College of Veterinary Medicine, University of Tennessee. 2000. M-Mode Echocardi-
ography [Online] http://www.vet.utk.edu/m-mode.html
Davies, M. J., Anderson, R. H. and Becker, A. E. 1983. The Conduction System of the
Heart. London: Butterworth.
Forestieri, S. F. and Spratt, R. S. 1995. Angiography using ultrasound. US Patent,
5,394,874.
Fung, Y. C. 1997. Biomechanics: Circulation. 2nd ed. New York: Springer-Verlag.
Grossman, W. and Baim, D. S. 1991. Cardiac Catheterization. Angiography and Inter-
vention, 4th ed. Philadelphia: Lea & Febiger.
Huang, S. K. S., and Wilber, D. J. (eds.) 2000. Radiofrequency Catheter Ablation of Car-
diac Arrhythmias: Basic Concepts and Clinical Applications, 2nd ed. Armonk, NY:
Futura.
Ingels, N. B. Jr., Daughters, G. T., Baan, J., Covell, J. W., Reneman, R. S. and Yin, F. C.
1996. Systolic and Diastolic Function of the Heart. Amsterdam: IOS Press.
Inoué, S. 1986. Video Microscopy. New York: Plenum Press.
8.12 References 45
Marwick, T. H., Shan, K., Go, R. T., MacIntyre, W. J. and Lauer, M. S. 1995. Use of
positron emission tomography for prediction of perioperative and late cardiac events
before vascular surgery. Amer. Heart J. 130 (6): 1196-202.
Meltzer, R. S. and Roelandt, J. (eds.) 1982. Contrast Echocardiography. Hague: Mar-
tinus Nijhoff.
Merz, C. N. and Berman, P. S. 1997. Imaging techniques for coronary artery disease:
current status and future directions. Clinical Cardiology. 20 (6): 526-32.
Mohrin, C. M. 1995. Stethoscope. US Patent, 5,389,747.
Moran, P. R. 1982. A flow zeugmatographic interlace for NMR imaging in humans.
Magn. Reson. Imaging 1: 197–203.
Nakatani, S., Yamagishi, M., Tamai, J., Goto, Y., Umeno, T., Kawaguchi, A., Yutani, C.
and Miyatake, K. 1995. Assessment of coronary artery distensibility by intravascular
ultrasound: application of simultaneous measurements of luminal area and pressure.
Circulation 91: 2904–10.
Neuman, M. R. 1998a. Biopotential electrodes. In J. G. Webster (ed.) Medical Instrumen-
tation: Application and Design. 3rd ed. New York: John Wiley & Sons.
Neuman, M. R. 1998b. Biopotential amplifiers. In J. G. Webster (ed.) Medical Instrumen-
tation: Application and Design. 3rd ed. New York: John Wiley & Sons.
Onoda, M. 1994. Electrocardiograph system. US Patent, 5,284,151.
O’Rourke, M. 1995. Mechanical principles in arterial disease. Hypertension 26: 2–9
Orth, J. L. 1995. System for calculating compliance and cardiac hemodynamic parame-
ters. US Patent, 5,423,323.
Peura, R. A. and Webster, J. G. 1998. Basic sensors and principles. In J. G. Webster (ed.)
Medical Instrumentation: Application and Design. 3rd ed. New York: John Wiley &
Sons.
Picot, P. A. and Fenster, A. 1996. Ultrasonic blood volume flow rate meter. US Patent,
5,505,204.
Rafferty, T. 1992. Transesophageal Two-Dimensional Echocardiography [Online]
http://www.gasnet.org/reference/echomanual/html/2-d_echo..html
Reinhard, C. J. and Fowler, K. A. 1987. Telemetry system and method for transmission
of ECG signals with heart pacer signals and loose lead detection. US patent,
4,658,831.
Ritchie, J. L., Hamilton, G. W. and Wackers, F. J. T. (eds.) 1978. Thallium−201 Myocar-
dial Imaging. New York: Raven Press.
Shepherd, A. P. and Öberg, P. Å. (eds.), 1990. Laser-Doppler Blood Flowmetry. Boston:
Kluwer Academic Publishers.
Sherwood, L. 2001. Human Physiology: From Cells to Systems. 4th ed. Pacific Grove,
CA: Brooks/Cole.
Strandness, D. E., Didisheim, P., Clowes, A. W. and Watson, J. T. (eds.) 1986. Vascular
Diseases: Current Research and Clinical Applications. Orlando: Grune & Stratton.
Tobuchi, K., Kato, T. and Namba, K., 1981. Silver-silver chloride electrode. US Patent,
4,270,543.
Togawa, T., Tamura, T. and Öberg, P. Å. 1997. Biomedical Transducers and Instruments.
Boca Raton FL: CRC Press.
Tompkins, W. J. (ed.) 1993. Biomedical Digital Signal Processing. Englewood Cliffs NJ:
Prentice Hall.
46 Heart and Circulation
8.13 Problems
8.1 Sketch the anatomy of the heart chambers and valves and show the direction of
blood flow.
8.2 Explain why the wall of the left ventricle is thicker and more muscular then the
wall of the right ventricle.
8.3 Describe the relation between the arterial pressure and the left ventricular pres-
sure during a cardiac cycle.
8.4 Explain why the AV node delays the wave of excitation.
8.5 Draw a typical ECG waveform, label all waves, and explain what is happening
within the heart muscle during each wave.
8.6 Explain why different pairs of electrodes are used on specific locations on the
body for an ECG.
8.7 An ECG has a scalar magnitude of 0.8 mV on lead I and a scalar magnitude of
0.2 mV onlead III, calculate the scalar magnitude of lead II.
8.8 Given the stroke volume = 80 mL, LV end-diastolic volume = 51 mL , and the
heart rate = 80 bpm, calculate the LV ejection fraction and the cardiac output.
8.9 Use the Fick method to calculate the cardiac output (in mL/min) of a patient
whose arterial concentration of O2 is 0.20 L/L, venous concentration of O2 is
0.13 L/L, and the O2 consumption is 0.3 L/min.
8.10 Calculate the cardiac output using the thermodilution method from the following
t1
data: ∫ | ∆T
0
b | dt = 3 s ⋅ °C , Q = 1250 J, ρb = 1060 kg/m3, and cb = 3640
J/(kg·K).
8.11 Explain why and how we measure cardiac output by the thermal and dye dilu-
tion techniques and compare the two methods.
8.12 A Doppler-ultrasonic blood flowmeter has a carrier frequency of 7.5 MHz, with
a transducer angle of 30°. The velocity of the sound is 1500 m/s and the audio
frequency is 12.6 kHz. Calculate the blood velocity.
8.13 Describe the cardiac sounds, their origins, and the time they occur in the cardiac
cycle.
8.14 An artery has a diameter of 3 mm. The viscosity of the blood is 0.0028 Pa⋅s at
37 °C, and the mean arterial blood pressure is 90 mmHg at one end and 0 mmHg
8.12 References 47
at the other end. Calculate the blood flow rate in mL/min, assuming that the
length of the artery equals 0.5 m.
8.15 Sketch an electromagnetic flowmeter and explain its principle of operation.
8.16 Sketch an ultrasonic flowmeter and explain its principle of operation.
8.17 Draw the block diagram for and describe the automatic indirect auscultatory
peripheral blood pressure measurement system.
8.18 Draw the block diagram for the automatic indirect oscillometric (not ausculta-
tory) peripheral blood pressure measurement system. Show all pneumatic con-
nections between parts. Show location of all sensors and explain how they work.
Describe the measurement cycle. Sketch the resulting waveforms. Explain how
the signal processing identifies significant pressures.
8.19 Explain why and how we measure direct cardiac pressure using a catheter
(catheter-type and catheter tip).
8.20 Explain why and how we measure vessel pressure using arterial tonometry.
8.21 Explain why and how we measure vessel pressure distensibility.
8.22 Explain why and how we measure vessel volume flow.
9
Lung, Kidney, Bone and Skin
Shilpa Sawale
This chapter describes four major organs in the body: lung, kidney, bone, and skin. Each
of these organs has its own properties and functions: the lungs help us respire, the
kidneys help clean the blood, bone supports the body, and skin protects the body.
Measurements are performed to confirm whether each of these organs is functioning
properly, and also to measure some of their properties.
9.1 Lung
The exchange of gases in any biological system is called respiration. To sustain life, the
human body needs oxygen, which is utilized in cells with other essential nutrients during
the metabolic oxidation process. Carbon dioxide is a by-product of cellular metabolism.
The hemoglobin in the blood is the dominant transport mechanism by which oxygen is
brought to cells. Carbon dioxide produced by cells is dissolved in the blood plasma and
carried off for disposal via the lungs. The most important function of the lungs is to
supply tissue with adequate oxygen and to remove excess carbon dioxide.
To properly exchange gases during inspiration, alveolar pressure must be less
than the atmospheric pressure. There are two ways of producing the pressure difference
necessary for inspiratory flow: (1) the alveolar pressure can be reduced below
atmospheric. This is called natural or negative pressure breathing; (2) the atmospheric
pressure can be raised above normal and therefore above the resting alveolar pressure.
This is called positive pressure breathing.
Normal breathing is accomplished by active contraction of inspiratory muscles,
which enlarges the thorax. This further lowers intrathoracic pressure, which normally is
less than atmospheric pressure and hence, the air at atmospheric pressure flows through
the nose, mouth and trachea to the lungs.
The process of bulk gas transport into and out of the lungs and the diffusion of
gases across the alveolar membrane is known as pulmonary function. Tests performed to
determine parameters of system efficiency are called Pulmonary Function Tests (PFTs)
(Feinberg, 1986). PFTs do not tell us the ultimate answer. For example, finding a
1
2 Lung, Kidney, Bone and Skin
decreased lung volume reveals restriction, but not the cause of it. However, in several
circumstances PFTs are a useful and powerful tool: diagnosing a lung disease and finding
the extent of the abnormality, following a patient during the course of a disease to
determine the efficiency of the treatment or the need for supplemental oxygen and
mechanical ventilation, determining if preoperative patients can withstand the surgery,
assessing disability, and deciding whether an individual can perform an occupational task
requiring a certain work load (Petrini, 1988).
9.2 Pulmonary Volume
The most commonly used measures of the mechanical status of the ventilatory system are
the absolute volume and changes of volume of gas space in the lungs achieved during
various breathing maneuvers. Pulmonary volumes have been divided according to
physiological limits (Petrini, 1988). The common terms used to describe various
capacities (capacities refer to quantities that include more than one volume) are as
follows:
1. Total Lung Capacity (TLC) is the largest volume to which the subject’s lung can
be voluntarily expanded.
2. Residual Volume (RV) is the smallest volume to which the subject can slowly
deflate his or her lung.
3. Functional Residual Capacity (FRC) is the resting lung volume achieved at the
end of normal expiration. It is the point where inward pulling forces of the lung
and outward pulling forces of the chest wall are equal in magnitude.
4. Tidal Volume (TV) is the volume inspired and expired during normal breathing.
All of the lung capacities are based on these four parameters (Eqs. (9.1) to (9.4)).
IRV is the inspiratory reserve volume. ERV is the expiratory reserve volume.
In dealing with volumes of gases, the conditions under which the values are
reported must be well defined and carefully controlled because gases undergo large
changes under different thermodynamic conditions. These volumes and specific
capacities, represented in Figure 9.1, have led to the development of specific tests to
quantify the status of the pulmonary system.
3
4 Lung, Kidney, Bone and Skin
TLC
IRV
IC
TV VC
ERV
FRC
RV
Figure 9.1 The spirometer measures lung capacities and lung volumes. Because the subject cannot
make the lung volume equal to zero, the spirometer cannot measure RV and FRC.
The measurement of changes in the lung volume have been approached in two different
ways. The first involves measurement of the gas through the airway opening. This is
referred to as spirometry. The second method is to measure changes in the volume of the
gas space within the body using plethysmographic techniques.
Spirometer
The spirometer was one of the early instruments used for estimating respiratory volumes.
It collects the gas passing through the airway opening and computes the volume it had
occupied within the lungs. The tests performed by the spirometer can be classified into
two major groups: single-breath tests and multiple-breath tests. There are three types of
tests under the category of single-breath tests: tests that measure expired volume only,
tests that measure expired volume in a unit time, and tests that measure expired
volume/time. Maximal Voluntary Ventilation (MVV) is measured in the category of
multiple-breath tests (Feinberg, 1986).
A spirometer is an expandable compartment consisting of a movable, statically
counterbalanced, rigid chamber or bell, a stationary base, and a dynamic seal between
them. This seal is often water. Changes in internal volume of the spirometer are
proportional to the displacement of the bell. This motion can be observed on a calibrated
scale or a rotating drum (kymograph) as shown in Figure 9.2.
9.2 Pulmonary Volume 5
Counterweight
Pulley
Bell
Kymograph
Water seal
One-way
Valves Soda-lime
cannister
Mouthpiece
Tube to the
patient
Figure 9.2 In the water sealed spirometer, expired CO2 is removed in the soda-lime cannister.
The mouthpiece of the spirometer is placed in the mouth of the subject whose
nose is blocked. As the gas moves into and out of the spirometer, the pressure of the gas
in the spirometer changes, causing the bell to move.
The system can be modeled as two gas compartments connected such that the
number of moles of gas lost by the lungs through the airway opening is equal and
opposite to the number gained by the spirometer. For rebreathing experiments, most
spirometer systems have a chemical absorber (soda lime) to prevent build up of carbon
dioxide. With the exception of water vapor in a saturated mixture, all gases encountered
during routine respiratory experiments obey the ideal-gas law during changes of state:
N
P= RT = ρRT (9.5)
V
where P is the pressure of the ideal gas, R is the universal gas constant, T is the absolute
temperature, ρ is the mole density, which for a well-mixed compartment equals the ratio
of moles of gas N in the compartment to the compartment volume V.
6 Lung, Kidney, Bone and Skin
Body plethysmograph
Because alveolar pressure cannot be measured directly, an indirect method has been
developed using a device known as a total body plethysmograph. This device is a
volumetric displacement box into which the patient is placed. The box is closed and
sealed so that no gas can enter or escape except through the connection leading to a
patient mouthpiece or mask. When the outlet connection is attached to a spirometer,
pneumotachometer, or respiratory function computer, a wide variety of measurements
can be easily made. It is used clinically not only to evaluate the absolute volume and
volume changes of the lungs, but also to provide a continuous estimate of alveolar
pressure, from which airway resistance can be found. Figure 9.3 shows a total body
plethysmograph.
9.2 Pulmonary Volume 7
Body plethysmograph
PL
Flow Q
PB
Lung
PL
VL
Pneumotachometer
PB
Shutter VB
Figure 9.3 The total body plethysmograph measures lung volume with the shutter closed and the
airway resistance via a pneumotachometer with the shutter open.
In order to compute the alveolar pressure using plethysmograph pressure
response to thoracic volume changes, the Volume of Thoracic Gas (VTG) must be known.
A patient sits inside the plethysmograph with a pneumotachometer flow meter in his or
her mouth. Both mouth and box pressures are displayed on the x and y axes of the
oscilloscope for one breath cycle against a total mechanical obstruction to airflow. This is
accomplished using a solenoid operated shutter to block the pneumotachometer. When a
patient breathes against a closed shutter, the air flow is zero, and thus the pressure
measured at the mouth is the same as the alveolar pressure (Feinberg, 1986).
The tissues of the body are composed of liquids and solids and can be
considered incompressible when compared with the gas in the lungs. During breathing
movements, the tissues change shape, not volume. Also because the volume of the
plethysmograph is constant, changes in the volume of the plethysmograph are zero.
Consequently the change in the volume of the gas in the lungs ∆VL is equal and opposite
to the change in the volume of the gas space in the box ∆VB.
During compression, ∆VL = –∆VB. There is a large change in pressure of the
lung ∆PL, while there is a small change in pressure in the box ∆PB. As a first order
approximation, VL/VB = –∆PB/∆PL. The partial pressure of water vapor in the lung must
be subtracted from the total pressure to yield compressible dry gas. The lung is
isothermal whereas the box is adiabatic. All this can be calibrated by plotting ∆PB/∆PL
during breathing with the shutter closed, from which we learn VL (Primiano, 1998). After
calibration, we can measure ∆PB and use it to predict ∆PL. Then we can open the shutter
to obtain the airway resistance
Impedance plethysmography
ρbL
Zb = (9.7)
∆A
Electrodes
Meter
Figure 9.4 A model for two electrode impedance plethysmography for cylindrical vessels.
∆V can be calculated using the following equation:
− ρ b L2 ∆Z
∆V = (9.8)
Z2
If the assumptions are valid, the above equation shows that we can calculate ∆V
from ρb and other quantities that are easily measured.
However, the lung does not have a regular shape, so we cannot use equations to
predict volume changes. In respiratory impedance plethysmography, respiration is
electrically monitored by measuring changes in chest impedance. The electric impedance
9.2 Pulmonary Volume 9
of the thoracic cavity changes with breathing movements and can be sensed in order to
monitor ventilatory activity. Ventilation produces changes in the shape and content of the
chest that are electrically observable as small impedance variations. A 100 kHz
impedance measurement between electrodes on each side of the chest yields a waveform
that follows ventilation. Electrodes can be easily attached to a segment of the chest and
the resulting impedance of the chest measured. As resistivity increases in response to
increased air in the chest, the impedance of the chest increases. It is useful for monitoring
breathing of infants as in infant apnea monitors.
Inductance plethysmography
Pneumotachometer
The most commonly used instrument for measurement of flow is the pneumotachometer.
This instrument is designed to measure respiratory gas flow during inspiration and
expiration. It is based on the pneumatic equivalent of Ohm’s law:
10 Lung, Kidney, Bone and Skin
∆P (difference in pressure)
Q(flow) = (9.9)
R (fixed resistance of pneumotachometer)
∆P ∆P
(a) (b)
Figure 9.5 A pneumotachometer measures flow from a pressure drop ∆P across resistance
elements such as (a) a fine mesh screen or (b) capillary tubes or channels.
The amount of gas transferred from the lung to blood is a function of the pressure
gradient and the diffusion capacity (DL) if the gas is diffusion limited. The equation is
where Vg is the rate of transfer of gas (g) from alveoli (A) to the capillaries (c), and P is
the pressure. DL is the diffusing capacity in mL/(min⋅mmHg) and includes the
parameters of the surface area of gas transfer, the thickness of the membrane across
which the gas is transferred, and the properties of the gas such as the molecular weight
and solubility. Since DL is measured from the gas side, Vg is negative since the gas is
moving away from the lungs (Primiano, 1998).
Carbon dioxide diffuses across the alveolar membrane much more easily than
oxygen, so a diffusion defect affects oxygen transfer first. Because of this and because of
the important role oxygen plays in sustaining life, it is important to evaluate a diffusion
capacity for oxygen. But obtaining the PaO2 requires a sample of arterial blood. To avoid
this, clinicians use carbon monoxide as the tracer gas as its properties are quite close to
those of oxygen, and so its diffusion capacity provides a reasonable estimate of the
diffusion capacity of oxygen. In addition, it also has affinity for hemoglobin at low
concentrations; hence all carbon monoxide that enters the blood chemically combines
with the hemoglobin in the red blood cells. One of the many methods that have been used
to find out the diffusion capacity of carbon monoxide involves a single breath technique.
The subject inspires a mixture of air, 0.3% (or less) carbon monoxide and helium
(approximately 10%) from RV to TLC. The subject holds his breath at TLC for about
10 s and then forcefully exhales down to RV. Although it requires subject cooperation, it
can be performed quickly and can be repeated easily. Also, it does not require samples of
arterial blood flow. The computation of DLCO from measurements made during this
single breath technique are based on a one compartment model of the lung. If a well-
mixed alveolar compartment is filled with a mixture of gases containing some initial
alveolar fraction FACO, then during breath holding with the airway open, the CO
diffuses into the blood in the pulmonary capillaries, and the alveolar FACO decreases
exponentially with time (Primiano, 1998).
D CO ( Patm − PA H 2 O )(t 2 − t1 )
FˆA CO(t 2 ) = FˆA CO(t1 ) × exp − L
VA (9.11)
9.6 Kidney
Kidneys are paired bean-shaped organs lying on either side of the spine in the upper part
of the abdomen. Each kidney is connected to the aorta and vena cava by a single renal
artery and vein. By means of these blood vessels, the kidneys filter about 1200 mL/min
of blood. Each kidney consists of approximately a million microscopic units called
nephrons, which are made up of two components: a glomerulus and a tubule. The key
separation functions of the kidney are: elimination of water-soluble nitrogenous end-
products of protein metabolism, maintainence of electrolyte balance in body fluids and
elimination of excess electrolytes, contribution to the obligatory water loss and discharge
excess water in the urine, and maintenance of the acid–base balance in body fluids and
tissues. Normal renal function is to remove water, electrolyte, and soluble waste products
from the blood stream. The kidneys also provide regulatory mechanisms for the control
of volume, osmolality, electrolyte and nonelectrolyte composition, and pH of the body
fluids and tissues. Figure 9.6 shows the general anatomy of the kidney.
9.5 Pulmonary Airway Resistance 13
Ureter
Kidneys
Bladder
Urethra
Figure 9.6 The kidneys excrete urine through the ureters to the bladder, where it is voided through
the urethra.
9.7.1 Creatinine
The glomerulus acts as a filter and separates water and crystalloids from the blood plasma,
while retaining blood cells and proteins. Nonselective glomerular filtration is followed by
the selective re-absorption in the renal tubes. Of the 180 L of water filtered each day,
perhaps 1 to 2 L are excreted. Likewise, all glucose, amino acids, and the small amount
of protein filtered together with most of the sodium, chloride, bicarbonate and calcium is
re-absorbed. The re-absorption of these ions is precisely regulated to return to the body
the exact quantity required to maintain the composition of the extracellular fluid.
Creatinine, a byproduct of phosphate metabolism in muscle, is frequently used as a
measure of GFR (Glomerular Filtration Rate) because it is freely filtered and not
reabsorbed: the quantity crossing the glomerular filter is thus same as the steady state
excretion of urine. For creatinine, the clearance (or volume of plasma that contains the
quantity excreted in unit time) is the same as GFR and is expressed in L/day (Gregory,
1988).
14 Lung, Kidney, Bone and Skin
UV
GFR ≈ C = (9.13)
P
where U is the urinary concentration, P is the plasma concentration, V is the urine flow
rate and C is the creatine clearance. The normal glomerular filtration rate averages 120
mL/min, about 116 mL/min is re-absorbed, and about 4 mL/min is excreted as urine.
9.8.1 Pyelogram
The main components of the urinary tract have either a tubular or cavitational structure
and their delineation with radio-opaque medium is readily possible. In this method the
radioactive dye is injected into the urethral passage. After a fixed amount of time has
passed, the X ray is taken. A satisfactory pyelogram demonstrates the kidney’s position,
size, shape, its cup-shaped cavity, and its functional ability to concentrate the opaque
medium.
The one shot Intravenous Pyelogram (IVP) is performed when there is suspected
or potential injury to the kidney or the ureter. If one kidney is injured it is important to
know if the other kidney is functioning normally. The one shot intravenous pyelogram is
a very good way to determine this. The presence of foreign bodies, retroperitonial
hematomas, and spine fractures can also be determined by the one shot IVP. In this
technique 1 mL/kg of nonionic radiocontrast agent is usually injected as a rapid
intravenous bolus. Five minutes after the injection of contrast agent, an anteroposterior
radiograph of the abdomen is obtained. In a hypotensive patient, the one shot IVP may
not be helpful as there may be no secretion of the contrast due to decreased renal blood
flow.
9.9 Hemodialysis
Dialysis is necessary when the GFR falls below 5 to 10% of normal. There are two types
of renal failure: acute (days or weeks) and chronic (months or years). Chronic renal
failure is a group of conditions in which renal function is slowly eroded and from which
recovery is very unusual, while in acute renal failure recovery is within a few days or
weeks. Thus dialysis is a temporary measure in acute renal failure but in chronic renal
failure it is needed for life or until a kidney is successfully transplanted. Hemodialysis is
a process in which blood is continuously circulated in contact with a permeable
membrane, while a large volume of balanced electrolyte solution circulates on the other
side of the membrane. Diffusion of dissolved substances from one stream to the other
removes molecules that are in excess in the blood and replaces those for which there is a
deficiency. Three processes occur simultaneously during dialysis—osmosis, diffusion
and ultrafiltration. Osmosis is movement of fluid across membranes from a lower
9.9 Hemodialysis 15
where TMP is the Transmembrane Pressure (mmHg) and KUF is the Coefficient of
Ultrafiltration of the dialyser ((mL/h)/mmHg). Transmembrane pressure is the arithmetic
difference between the pressure on the blood side of the dialyser (usually positive) and
the dialysate pressure (usually negative). The negative pressure is calculated using the
formula
AP + VP
NP = TMP − (9.15)
2
where NP is the negative pressure, AP is the arterial pressure, and VP is the venous
pressure. The transmembrane pressure can be calculated using the formula:
Dialysate output
Dialyser
Hollow fibers
Dialysate input
Adjustable constriction
Hollow fibers
(a) (b)
Figure 9.7 Typical dialyser (a) indicating the counter current flow of the blood and the dialysate (b)
Cross-sectional view of the dialyser.
Hemodialysis membranes vary in chemical composition, transport properties,
and biocompatibity. These membranes are fabricated from three classes of biomaterials:
regenerated cellulose, modified cellulose, and synthetics. Chapter 4 describes some
details of biomaterials. Cuprophan (cupramonium cellulose), cellulose acetate,
polyacrylonitrile, polymethyl methacrylate polycarbonate, and polysulphate are some
other commonly used materials to prepare dialysis membranes.
Figure 9.8 shows a schematic of a hemodialysis machine (Gregory, 1988). The
blood from the patient is passed into the dialyser with the help of a roller blood pump.
Blood clots rapidly when it comes in contact with most foreign substances. To avoid
blood clots an anticoagulation agent is added. Heparin is the most commonly used
anticoagulation agent. The blood passes through the dialyser where the diffusion takes
place between the dialysate and the blood. Clean blood that comes out of the dialyser is
then injected into the patient. The blood is taken out from the arterial side of the patient
and is injected into the venous side. A venous bubble trap is inserted in the circuit to
make sure that the blood going back to the patient does not have an air bubble in it. An
ultrasonic sensor or photocell is used for bubble detection. The ultrasonic sensor monitors
any change in sound transmission and the photocell monitors any change in light
transmission. Whenever there is air on the venous side, the signals being transmitted are
interrupted and trigger an alarm. The pressure on the venous side is also monitored using
a pressure gage.
9.9 Hemodialysis 17
Heparin infusion
Pump
Air/foam detector
Dialysate
pump
Concentrate
Spent dialysate Heater
to waste Adjustable
constriction
Proportioning
Conductivity and pump
pH cell
Figure 9.8 Blood from an artery flows past a membrane in the hemodialysis machine, then returns
to a vein.
Proportioning pumps first mix the dialysate concentrate with water until a
desired concentration is reached. The dialysate temperature, conductivity, pH, and flow
are monitored and controlled. The dialysate is warmed by a heater to near normal body
temperature. The high temperatures may cause hemolysis (rupture of RBCs). The
negative pressure of the dialysate is usually generated by having the dialysate pump draw
the dialysate through the dialyser and by placing an adjustable constriction upstream of
the dialyser as shown in Figure 9.7. The pH of the dialysate must match the acid/base
balance of the patient’s blood. The dialysate is then passed through the dialyser in the
opposite direction to blood flow. The dialysate coming out of the dialyser is then passed
through a blood leak detector to check if there has been a rupture of the membrane and if
any blood has leaked into the dialysate circuit side.
The blood leak detector uses a beam of light that shines through the used
dialysate and into a photocell. Normally the dialysate is clear and lets light pass through.
Even a tiny amount of blood causes a break between the light beam and the photocell.
Any break in the transmission of the light beam triggers an alarm. The spent dialysate is
then drained off.
Heating element
Pump
Valve
Check valve
Pump
Dialysate Catheter
supply Peritoneal Body
cavity fluids
Spent
dialysate
To provide a mobile system, the control hardware, dialysate supply and spent
dialysate are commonly mounted on a wheeled stand. The dialysate supply is normally
positioned high on the stand, above the abdomen, to provide a gravity flow. The spent
dialysate collection is positioned at the bottom of the stand for same reason. Most
systems are equipped with peristaltic pumps to control flow. Traveling or portable
systems may exclusively use gravity to provide a smaller, lighter system. The dialysate is
warmed by the control electronics prior to flowing through the catheter and into the
peritoneal cavity to prevent thermal shock. After approximately 30 min, the dialysate is
pumped out of the cavity and into a collection bag. The weight of the dialysate supply
and spent dialysate is monitored to determine the amount of fluid and waste removed
from the body.
The key measurement in the peritoneal dialysis process is the weight of the
fluids. It is not practical to attempt to measure the change in concentration of the
dialysate. Therefore, the amount of fluid pumped into the peritoneal cavity and the
amount of fluid removed are measured as the means of monitoring the amount of water
and waste diffused from the body fluids. The circuit in Figure 9.10 shows a possible
means of measuring the fluid weight.
15 V
Rf
Ri
13 kΩ −
vo
+
2 kΩ vi
refreshed. If peritoneal dialysis can accurately be simplified into the model shown in
Figure 9.9 (two volumes of fluid separated by a membrane), a relationship can be
developed to determine the rate at which the body fluid wastes or solutes diffuse across
the membrane.
The concentrations of the solutes are defined as
N
C= (9.17)
V
J s = ωRT (C b − C d )
N N (9.18)
J s = ωRT b − d
V b Vd
where R is the gas constant (J/(mol·K)), T is the absolute temperature, 310 K, and ω is the
solute permeability of the peritoneal membrane (1 × 10–5 moles/(N⋅s)). The designations
of b and d stand for the body fluid and dialysate, respectively.
Through some algebra and knowing Js, the surface area of the peritoneum and
the initial concentrations, the concentration in the body determined as a function of time
is
Vd Vb
C (t ) = C o e −αt + (9.19)
V b + V d V b + Vd
This relationship can be plotted to determine at what point the dialysate should be
refreshed. Normally, 30 min cycles are done. After 30 min, the body fluid concentration
is approximately 0.9524Co using the values given above. The concentration after n cycles
is simply computed as
C (t ) = (0.9524) n Co (9.20)
and may be used to estimate the reduction in body fluid solute concentration following
the dialysis. As mentioned previously, the rate of dialysis is monitored by weighing the
dialysate and spent dialysate.
where TBW is the total body water in m3, Ht is the height of the person in m, R is the
resistance in Ω, and Wt is the weight in N. a, b, and c are constants determined by
calibrating against isotopic dilution techniques (see section 10.4.1) or using other
equations estimating lean body mass, from which body fat can be estimated.
This method is controversial. Patterson (1989) points out that wrist-to-ankle
measurement is influenced mostly by the impedance of the arm and leg and less than 5%
of the total impedance is contributed by the trunk, which contains half the body mass.
The National Institutes of Health (1994) concludes that there are many conditions in
critical illness for which conventional BIA is a poor measure of TBW.
Another method that is used to estimate extracellular water is by deduction.
There are three basic components of total body mass: water, lean body mass (muscle) and
body fat. Section 10.4 gives many methods of estimating body fat. Other methods
provide us with equations that estimate lean body mass. If one of each of these methods
are used to obtain an approximation for lean body mass and body fat, these can be
subtracted from the total body mass to obtain an estimate of the extracellular water.
two different types of materials plus water. They are collagen, the major organic fraction,
which is about 40% of the weight of solid bone and 60% of its volume and bone mineral,
the so called inorganic component of the bone, which is about 60% of the weight of the
bone and 40% of its volume. Either of the components may be removed from the bone,
and in each case the remainder, composed of only collagen or bone mineral, looks like
the original bone. The collagen remainder is quite flexible, like a chunk of rubber, and
can even be bent into a loop. When collagen is removed from the bone, the bone mineral
remainder is very fragile and can be crushed with the fingers. Collagen is produced by the
osteoblastic cells and mineral is then formed on the collagen to produce bone (Cameron,
1978).
The strength of the bone depends to a large extent on the mass of bone mineral present. In
diseases like osteoporosis the bone mineral mass is considerably reduced. Up to a few
years ago osteoporosis was difficult to detect until a patient appeared with a broken hip or
a crushed vertebra. By that time it was too late to use preventive therapy. Thus bone
mineral is very important and commonly measured to detect bone diseases such as
osteoporosis. The bone mineral content of specimens can be measured by ashing the
specimen in a furnace or demineralizing it in a decalcifying solution (Ashman, 1989).
The most commonly used technique for noninvasively measuring bone mineral content in
the bone is dichromatic or dual photon absorptiometry (DPA).
In the early part of the 20th century, X rays were used to measure the amount of
bone mineral present in the bone. But there are some major problems with using an
ordinary X ray: the usual X-ray beam has many different bands of energy, and the
absorption of X rays by calcium varies rapidly with energy in this range of energies; the
relatively large beam contains much scattered radiation when it reaches the film; the film
is a poor tool for making quantitative measurements since it is nonlinear with respect to
both the amount and the energy of X rays. The net result of these problems is that a large
change in the bone mineral mass (30 to 50%) must occur between the taking of two X
rays of the same patient before a radiologist can be sure that there has been a change.
Figure 9.10 shows that in dual photon absorptiometry, three problems with the X-ray
technique are largely eliminated by using an X-ray source filtered to yield two
monoenergetic X-ray beams at about 30 keV and 70 keV, a narrow beam to minimize
scatter, and a scintillation detector that detects all photons and permits them to be sorted
by energy and counted individually. Tests are frequently made in the spine, hip, and
forearm but can be done on the entire body.
9.12 Bones and Joints 23
Scintillation
detector
Dual beam
Soft tissue
Bone
Figure 9.10 In a dual photon absorptiometer, an X-ray source is filtered to emit at two discrete
energies.
Hologic (2000) improves clinical bone densitometry by integrating bone mineral
density (BMD) with Instant Vertebral Assessment (IVA). IVA, possible only with fan-
beam technology, generates a high-resolution image of the entire lateral spine in just 10 s,
enabling physicians to visually assess vertebral status for a more accurate determination
of fracture risk than just BMD alone. A different instrument measures the speed of sound
(SOS, in m/s) and broadband ultrasonic attenuation (BUA, in dB/MHz) of an ultrasound
beam passed through the heel, and combines these results to obtain the Quantitative
Ultrasound Index (QUI). The output is also expressed as a T-score and as an estimate of
the Bone Mineral Density (BMD, in g/cm2) of the heel.
Tensile loads
When forces are applied to any solid object, the object is deformed from its original
dimensions. At the same time, internal forces are produced within the object. The relative
deformations created at any point are referred to as strains at that point. The internal force
intensities (force/area) are referred to as stresses at that point. When bone is subjected to
forces, these stresses and strains are introduced throughout the structure and can vary in a
very complex manner (Ashman, 1991).
24 Lung, Kidney, Bone and Skin
F = kx (9.22)
where F is the applied force, x is the change in length or elongation of the spring, and k is
the spring constant or stiffness of the spring.
Force σc = Ultimate
F Fult = Failure tensile strength
load
∆L Fy = Yield
Force, F
Stress, σ σy = Yield load strength
L
AE/L = Stiffness
E = Elastic
modulus
A, E
Deflection, ∆L
Strain, ε
(a) (b)
FL
∆L = (9.23)
AE
where ∆L is the elongation of the cylinder, L is the original unstretched length, A is the
cross-sectional area, F is the force and E is the elastic (Young’s) modulus (18 GPa for
compact bone).
Tensile, or uniaxial, strain, ε, can be calculated using the formula
∆L
ε= (9.24)
L
F
σ = (9.25)
A
9.12 Bones and Joints 25
These tests can be performed on specimens with different lengths, cross-sectional areas
and under forces of varying magnitudes (Black, 1988).
Figure 9.11(b) shows that a cylinder of bone tested in tension yields a linear
region (also known as the elastic region) where the atoms of the bone are displaced only
slightly by reversible stretching of the interatomic bonds. This is followed by a nonlinear
region where yielding and internal damage occurs, often involving irreversible
rearrangement of the structure. After yielding, nonelastic deformation occurs until finally
failure or fracture results. The load at which yielding occurs is referred to as the yield
load, Fy. The load at which failure occurs is called the ultimate or failure load, Fult. This
curve describes the behavior of the structure since the curve differs for a different cross-
sectional area or different length. It also represents the mechanical behavior of the
material as opposed to the behavior of the structure. In the stress–strain curve, the
material yields at a stress level known as the yield strength and fractures at a stress level
known as the fracture strength or the ultimate tensile strength. Toughness is the ability to
absorb energy before failure and is calculated from the total area under the stress–strain
curve, expressed as energy per unit volume in J/m3 (Black, 1988).
The above quantities can be measured using any of the displacement type
sensors (e.g. strain gage, LVDT (Linear Variable Differential Transformer), capacitive,
and piezoelectric sensors). The most commonly used sensors are the strain gage and the
LVDT.
Shear Loads
When forces are applied parallel to a surface or along the edge of an object, the object
deforms in a way shown in Figure 9.12. The sides of the object perpendicular to the
forces stretch and shear stresses and strains result.
L ∆L
F
W
A = LW
∆L
γ = (9.26)
L
F
τ = τ = F/A (9.27)
A
where ∆L is the distance of shear deformation, L is the original length, A is the cross-
sectional area and F is the acting force.
The shear modulus, G, can be calculated using τ and γ using the relationship:
τ
G= (9.28)
γ
See more information about measuring shear stress and strain in Chapter 6.
ρl
R= (9.29)
A
∆R / R ∆ρ / ρ
G= = (1 + 2µ ) + (9.30)
∆L / L ∆L / L
∆D D
µ=− (9.31)
∆L L
where D is the diameter of the cylindrical specimen. Poisson’s ratio is the ratio between
the lateral strain and axial strain. When a uniaxial tensile load stretches a structure, it
increases the length and decreases the diameter.
Figure 9.13 shows four strain gage resistances that are connected to form a
Wheatstone bridge. As long as the strain remains well below the elastic limit of the strain
gage resistance, there is a wide range within which the increase in resistance is linearly
proportional to the increase in length.
9.12 Bones and Joints 27
R2 R1
vi
Rx
R3 R4
vo
Figure 9.13 Four strain gage resistances R1, R2, R3, and R4 are connected as a Wheatstone bridge.
vi is the applied voltage with the bottom terminal grounded. vo is the output voltage, which must
remain ungrounded and feeds a differential amplifier. Potentiometer Rx balances the bridge. See
section 2.1.6 for more information on Wheatstone bridges and potentiometers.
The types of strain gages are dictated by their construction. Unbonded strain
gages may be formed from fine wires that are stretched when strained, but the resulting
sensor is delicate. The wire’s resistance changes because of changes in the diameter,
length and resistivity. Bonded strain gages may be formed from wires with a plastic
backing, which are glued onto a structural element, such as carefully dried bone.
Integrated strain gages may be formed from impurities diffused into a silicon diaphragm,
which forms a rugged pressure sensor (Peura and Webster, 1998). A strain gage on a
metal spring is useful for measuring force within a uniaxial tensile test machine.
9.12.4 LVDT
A transformer is a device used to transfer electric energy from one circuit to another. It
usually consists of a pair of multiply wound, inductively coupled wire coils (inductors)
that facilitate the transfer with a change in voltage, current or phase. A linear variable
differential transformer (LVDT) is composed of a primary coil and two secondary coils
connected in series opposition, as shown in Figure 9.14. The ac excitation is typically 5 V
at 3 kHz. The coupling between these two coils is changed by the motion of the high
permeability magnetic alloy between them. When the alloy is symmetrically placed, the
two secondary voltages are equal and the output signal is zero. When the alloy moves up,
a greater voltage is transformed to the top secondary coil and the output voltage is
linearly proportional to the displacement. The LVDT is useful in determining the strain
on tendons and ligaments (Woo and Young, 1991).
28 Lung, Kidney, Bone and Skin
vo
x
vi Secondary coil
Primary
coil
Secondary
coil
High permeability
alloy core
Figure 9.14 In a linear variable differential transformer, displacement of the high permeability
magnetic alloy changes the output voltage.
The most commonly used technique for measurement of stress and strain on
bone specimens is the Uniaxial tension test using the LVDT. Figure 9.15 shows the set up
for this test. It consists of one fixed and one moving head with attachments to grip the test
specimen. A specimen is placed and firmly fixed in the equipment, a tensile force of
known magnitude is applied through the moving head, and the corresponding elongation
is measured. Then using Eqs. (9.24) and (9.25), the uniaxial stress and strain can be
calculated.
9.12 Bones and Joints 29
Moving head
Grip
Specimen
LVDT Bonded strain
gage
Grip
Fixed head
Figure 9.15 The uniaxial tension test measures force versus elongation.
A direct method of measuring ligament strain is by mounting a strain gage load cell
within cadaveric knees (Markolf et al., 1990). A noncontact method for measuring
ligament strain is the video dimension analyzer (VDA) (Woo et al., 1990). Reference
lines are drawn on the specimen using stain and the test videotaped. The VDA system
tracks the reference lines and yields strain versus time. Tendon and ligament tissues
contain low-modulus elastin, which bears the majority of the load for strains up to about
0.07. At higher strains, collagen fibers, which possess a zigzag crimp, take an increasing
portion of the load, resulting in an upward curvature of the stress–strain curve (Black,
1988).
Proteoglycans are important components of the extracellular matrix of articular
cartilage and other soft tissues. Viscosity of proteoglycans extracted from cartilage can be
measured using a cone-on-plate viscometer described in section 6.7.1 (Zhu and Mow,
1990). The angle of the cone α is 0.04 and the diameter of the plate 2a is 50 mm. The
shear rate γ = ω/α, where ω is the rotational speed of the plate (rad/s). Apparent viscosity
ηapp = 3T/(2ωa3γ), where T = torque. Another viscometer is the Ostwald capillary
viscometer, which calculates the coefficient of viscosity η from the pressure gradient
dp/dL, the volume rate of flow Q, and the tube radius R using the equation η =
[πR4/(8Q)](dp/dL) (Fung, 1981). Another viscometer is the Couette coaxial viscometer in
which an outer rotating cylinder transmits torque through the test fluid to an inner coaxial
cylinder.
30 Lung, Kidney, Bone and Skin
Diarthrodial (synovial) joints have a large motion between the opposing bones. To
diagnose disease and to design artificial joints, we desire to measure friction between
them. The resistance to motion between two bodies in contact is given by frictional force
F = µW, where µ is the coefficient of friction and W is the applied load (Black, 1988).
Surface friction comes either from adhesion of one surface to another due to roughness
on the two surfaces or from the viscosity of the sheared lubricant film between the two
surfaces. Lubrication of bone joints is an important factor in determining coefficient of
friction. Rheumatoid arthritis results in overproduction of synovial fluid in the joint and
commonly causes swollen joints. The synovial fluid is the lubricating fluid that is used by
the joints. The lubricating properties of the fluid depend on its viscosity; thin oil is less
viscous and a better lubricant than thick oil. The viscosity of synovial fluid decreases
under the large shear stresses found in the joint.
The coefficient of friction is measured in the laboratory using arthrotripsometers
(pendulum devices). Here a normal hip joint from a fresh cadaver is mounted upside
down with heavy weights pressing the head of femur into the socket. The weight on the
joint is varied to study the effect of different loads. The whole unit acts like a pendulum
with the joint serving as the pivot. From the rate of decrease of the amplitude with time,
the coefficient of friction is calculated. It can be concluded that fat in the cartilage helps
to reduce the coefficient of friction. When synovial fluid is removed, the coefficient of
friction is increased considerably. Figure 9.16 shows an arrangement for measuring the
coefficient of friction (Mow and Soslowsky, 1991).
9.12 Bones and Joints 31
Bone joint
(pivot)
Weights
Rocking movement
Figure 9.16 Decay of oscillation amplitude in the pendulum device permits calculation of the
coefficient of friction of a joint.
We wish to measure wear in joints of artificial materials such as ultra-high-
molecular-weight polyethylene (UHMWPE) (Dowson, 1990). For polymeric materials,
the volume V = kPX produced by wear during sliding against metallic or ceramic
countersurfaces is proportional to the applied load P and the total sliding distance X. k is
a wear factor indicative of the wear resistance of a material.
Bone position is important for calculating the loading forces that act on it. The most
complicated bones in the body on which these forces are calculated are in the spinal cord.
Calculating forces for other bones, such as the femur, is relatively easy. These loading
forces are static and dynamic in nature.
A goniometer is an electric potentiometer that can be attached to a joint to
measure its angle of rotation. For working details of the goniometer refer to Chapter 10.
Human joint and gross body motions can be measured by simple protractor type
goniometers, electrogoniometers, exoskeletal linkage devices, interrupted light or normal
and high speed photography, television-computer, X-ray or cineradiographic techniques,
sonic digitizers, photo-optical technique, and accelerometers (Engin, 1990).
32 Lung, Kidney, Bone and Skin
Bending a slab of cortical or whole living or dead bone yields piezoelectric potentials of
about 10 mV, which we can measure using electrodes (Black, 1988). If the strain is
maintained, the potentials rapidly decay to a very low value, called the offset potential.
Some workers hypothesize that these potentials have a role in directing growth, healing,
and remodeling.
9.13 Skin
The primary function of the skin is to act as a buffer between the body and the outside
world. It absorbs sunlight, softens blows, insulates from heat and cold, retains fluid
within the body and keeps foreign organisms and extraneous chemicals out of it. It plays
an important role in body’s defense against infection. It also acts as the body’s radiator
and as such is the most significant element in the thermoregulation system.
Transepidermal Water Loss (TWL or TEWL) refers to the rate at which water migrates
from the viable dermal tissues through the layers of the stratum corneum to the external
environment. In the absence of profuse sweating, the TWL is predominantly controlled
by the diffusion of water vapor in the stratum corneum caused by the difference in vapor
concentration between the inside and outside surfaces.
TWL measurements have been used in studying the restoration of barrier
function in wounded skin, as well as in evaluating occlusive properties of topical
preparations. TWL can be measured using the direct measurement technique or the
indirect measurement technique. Flow hygrometry is a commonly used technique for
direct measurement. The indirect measurement technique relies on establishing a
boundary air layer over skin of known geometry. The two commonly used techniques of
indirect measurement are the closed cup method and the open cup method (Marks, 1981).
Flow hygrometry
In flow hygrometry, the flux of water vapor out of a fixed area of skin is determined by
measuring the increase in water vapor concentration in the flowing gas stream. The
increase in humidity is read at a sensor output once steady-state conditions have been
reached. TWL can be calculated using the formula:
K ×V × R
TWL = (9.34)
A
9.13 Skin 33
where K is the instrument constant, V is the increase in the sensor output, R is the gas
flow rate and A the skin area isolated by the measuring chamber. TWL is usually
measured in mg/(cm2⋅h). There are numerous possible sources of error in the flow
hygrometry system, which include uncertainty in the gas flow rate, uncertainty in the
actual area of skin exposed to flowing gas, absorption of water vapor in tubing
connecting the skin chamber with the sensor, and leaks in the seal of the chamber to the
skin site. These errors can be minimized by careful design of the system.
In this method, transfer of water vapor to the sensor is by convection, which
normally requires gas tanks or pumps, valves, and tubing in addition to the sensor and the
skin chamber. In contrast, indirect measurement techniques, which are described later,
require little auxiliary equipment. Figure 9.17 shows a flow hygrometer (Smallwood and
Thomas, 1985).
Meter
Gas
source
Humidity sensor
Flow rate = R
Skin
Figure 9.17 The flow hygrometer measures the increase in humidity of gas flowing over the skin.
In the closed cup method, the humidity sensor is sealed into one end of a cylinder of
known length. The cylinder is placed on the skin, trapping a volume of unstirred air
between the skin and the detector as shown in Figure 9.18(a). Within a few seconds, the
sensor voltage begins to rise steadily followed by a gradual decrease in rate of change.
The application of diffusion principles to the water vapor in the volume of trapped air
predicts this behavior and shows that the TWL is directly proportional to the slope of the
transient linear portion of the detector output curve shown in Figure 9.18(b). The TWL is
found from:
dv
TWL = K × l × (9.35)
dt
where K is an instrument calibration constant, l is the distance between the detector and
the skin surface, v is the detector voltage and t is the time. The closed chamber method
34 Lung, Kidney, Bone and Skin
does not permit recordings of continuous TWL because when the air inside the chamber
is saturated, skin evaporation ceases.
Sensor
v
Transient linear
portion
l
Time
Skin
(a) (b)
Figure 9.18 The closed cup hygrometer: (a) configuration of measuring cup, (b) typical sensor
output curve.
Figure 9.19(a) shows the open cup hygrometer. The sensor is mounted in the wall of the
cylindrical chamber with the top end of the chamber open and the bottom end placed
against the skin. As in the closed cup method, the cylindrical chamber defines a volume
of undisturbed air in which the transfer of water vapor is controlled by diffusion in air. A
few minutes after applying the chamber to the skin, the sensor reaches a new equilibrium
in which the rate of flow of vapor out of the open end of the cylinder is equal to the rate
of flow of vapor out of the skin, the TWL. In this condition the rate of flow can be
calculated, according to the principles of diffusion, from the product of diffusion
coefficient of water vapor in the air and the concentration gradient. For the configuration
shown in Figure 9.19(a) the sensor is mounted at a distance l from the end of the cylinder
where the humidity is determined by the room condition. Thus the increase in sensor
response divided by the length determines the humidity gradient. The equation for TWL
measurement with the open cup method is then:
TWL =
(V − V0 ) × K × D (9.36)
l
where V and V0 are equilibrium and initial sensor voltages, K is an instrument constant,
D is the diffusion coefficient of water in air and l is the distance between the sensor and
the end of the cylindrical chamber open to the room air. Air movement and humidity are
the greatest drawbacks of this method. This method is currently used in commercially
available devices.
9.13 Skin 35
Sensor
l voltage
Sensor
Vo
Skin
Time
(a) (b)
Figure 9.19 The open cup hygrometer: (a) configuration of measurement cylinder, (b) typical
sensor output curve
In the flow hygrometry method, the investigator has the option of selecting the
water content of the sweep gas within the limits of the humidity sensor. For indirect
methods, unless measurements are made in a humidity-controlled room, the variable
ambient condition can have a profound effect on the resulting measurements.
9.13.2 Color
Among the physical skin parameters, color is important in clinical dermatology. The
experienced dermatologist uses color information in several ways. First, he frequently
bases his first-view diagnosis or elaboration of differential diagnosis to a considerable
amount on specific colors of the skin lesions themselves. Second, the color differences
often allow the clinician to judge the distribution of lesions on the human body, even
without close inspection and palpation. Finally, the intensity of color may provide useful
information about the severity of a pathological process, and changes in color to normal
tell the dermatologist if the treatment works.
The color under which we perceive the skin depends on a number of variables,
including pigmentation, blood perfusion, and desquamation (shedding of skin). Because
color perception is a subjective sensory and neurophysiological process, the evaluation of
color is highly observer dependent. This has been a concern not only in dermatology, but
also in industries such as dye production and printing in which highly consistent colors
are necessary. In order to measure color objectively, color-measuring devices have been
developed instead of having it judged by subjective observers.
These devices are useful in dermatology, skin pharmacology, toxicology and
cosmetology because they allow a quantitative measurement instead of a historically
subjective grading of reaction.
Colors may be described by their hue (color position in the color wheel),
lightness (called value) and saturation (called chroma). Any color is expressed as a
combination of (H/V/C) (hue, value, chroma) (Berardesca, 1995).
36 Lung, Kidney, Bone and Skin
The instruments used to measure the variations in the color of skin are the
Dermaspectrometer and the Erythema meter. Their working principle can be explained as
follows: Inflammatory skin erythema is the result of an increased presence of
erythrocytes in skin structures. In inflammation, cutaneous vessels are dilated and blood
flow is increased. Because oxyhemoglobin has absorption in the spectral range from 520
to 580 nm, it absorbs green light, while red light is reflected to the absorber. Changes in
skin redness affect the absorption of green light, but affect less that of red light. An
erythema index of the skin can therefore be based on the ratio between the reflection of
red and green light.
Similarly melanin pigmentation of the skin lead to an increased absorption both in the
green and the red parts of the spectrum, and a melanin index may be defined as follows:
The DermaSpectrometer emits green and red light (568 and 655 nm) from an
LED source. The Erythema Meter emits green and red light (546 and 671 nm) from a
tungsten lamp. Figure 9.20 shows that the light reflected from the skin is detected with a
photosensor. A microprocessor calculates the erythema and the melanin index, which is
then displayed.
9.13 Skin 37
Green
Red Green
Red
Epidermis
Superficial vascular
Dermis plexus
Subcutis
The Chromameter and Micro Color are another pair of instruments based on the same
working principle as the DermaSpectrometer and Erythema meter. The perceived color of
an object as the spectral distribution of reflected light depends on the light with which the
object is illuminated. This refers both to the color of lighting and the angle under which
the light hits the object. As the perceived color of the object may be expressed as the
proportion of the remitted to the absorbed light, equal lighting over the whole visible
spectrum is essential. Selective lighting results in limited color information; e.g., a red
object would appear black when perceived under a green lamp.
Lighting is achieved using a Xe flash lamp that emits an intensive white light
covering the whole visible spectrum. In order to compensate for variation in illumination,
part of the emitted Xe light is sent to a set of color sensors, whereas the rest illuminates
the object (dual beam system, Figure 9.21(a)). The color sensors analyze the illuminating
light, and through a microprocessor, the light reflected from the object is adjusted for
variations in the illuminating light. Because the measuring unit is a hollow chamber, the
light emitted from the flash lamp hits the surface of the object from all angles. In the
Chromameter, only the light remitted at 0° to the axis of the instrument (90° to the object
surface) is collected for color measurement (d/0 measurement principle, Figure 9.21(b)),
whereas in the Micro Color, the light reflected at 8° to the axis (d/8 measurement) is
collected. The reflected light is transferred to three photodiodes. Before each diode, a
38 Lung, Kidney, Bone and Skin
color filter ensures a specific spectral sensitivity with peaks at 450 nm (blue), 550 nm
(green), and 610 nm (red). These photodiodes simulate the human eye with its three blue,
green, and red sensitive cones in the central fovea. The light reaching the sensors is
transformed into electric signals that are used by the microprocessor for the calculation of
the color values.
Reflected light
Sensor
Sensors
Light
Emitted light
Object Object
(a) (b)
Figure 9.21 (a) The spectrum of object light is compared with the spectrum of the light source
alone to allow for differences in lighting of the object (b) Illuminating light is diffuse whereas
measured light is perpendicular to the object (d/0).
9.14 References
Ashman, R. B. 1989. Experimental techniques. In S. C. Cowin (ed.) Bone Mechanics.
Boca Raton FL: CRC Press.
Berardesca, E., Elsner, P., Wilhem, K. -P. and Maibach, H. I. 1995. Bioengineering of the
Skin: Methods and Instrumentation. Boca Raton FL: CRC Press.
Bhattacharya, A. and McGlothlin, J. D. 1996. Occupational Ergonomics: Theory and
Applications. New York: Marcel Dekker.
Black, J. 1988. Orthopaedic Biomaterials in Research and Practice, New York:
Churchill Livingstone.
Bronzino, J. D. (ed.) 2000. The Biomedical Engineering Handbook. 2nd ed. Boca Raton
FL: CRC Press.
Cameron, J. R. and Skofronick, J. G. 1978. Medical Physics. New York: John Wiley &
Sons.
Cohen, S. H. 1981. Non-invasive Measurements of Bone Mass and Their Clinical
Applications. Boca Raton FL: CRC Press.
Dowson, D. 1990. Bio-tribology of natural and replacement synovial joints. In V. C.
Mow, A. Ratcliffe and S. L.-Y. Woo (eds.) Biomechanics of Diarthrodial Joints. Vol.
II. New York: Springer-Verlag.
Ducheyne, P. and Hastings, G. W. 1984. Functional Behavior of Orthopedic
Biomaterials Vol. 1: Fundamentals, Vol. 2: Applications. Boca Raton FL: CRC Press.
9.14 References 39
9.15 Problems
9.1 Sketch a spirometer output. Label and describe lung volumes that it can measure.
Label and describe lung volumes that it cannot measure.
9.2 You attach a 10 kΩ rotary potentiometer with 5 V across it to a spirometer
pulley so the potentiometer wiper voltage varies from 0 to 5 V. You notice that a
10 kΩ from the wiper to ground causes loading error. Draw an op amp circuit
that prevents this loading error.
9.3 You wish to measure lung volume using a body plethysmograph. Sketch the
equipment. Describe (1) sensor type and location, (2) system and method for
calibration, (3) procedure for measurement, (4) results of measurement, (5)
equation, and (6) any corrections required.
9.4 In a 3000 L body plethysmograph with closed shutter, the subject blows to
achieve a lung pressure of 10000 Pa. The box pressure changes by 10 Pa.
Calculate the lung volume. State any approximations.
9.5 Give the equations and describe how to measure airway resistance.
9.6 Given resistivity of the thorax is 300 Ω·cm. Assume electrode area and thoracic
dimensions and estimate the impedance of the thorax for use in impedance
plethysmography for measuring pulmonary volume changes.
9.7 Describe the importance of measuring pulmonary air-flow. List the different
methods used for the measurement. Explain any one of them.
9.8 Sketch a graph of the results obtained when measuring diffusing capacity of the
lung. Label axes and significant quantities.
9.9 The measured plasma concentration of creatinine is 1 mg/100 mL. The urine
flow is 4 mL/min. The urinary concentration is 30 mg/100 mL. Calculate the
glomerular filtration rate.
9.10 Describe a pyelogram. Explain its importance in diagnosing kidney disease.
Explain how this test is performed clinically.
9.11 Describe safety devices required for hemodialysis.
9.15 Problems 41
9.12 Find the transmembrane pressure given KUF is 4 (mL/h)/mmHg. Total fluid to
be removed is 2810 mL and hours on the dialysis is 4 h.
9.13 Find the negative pressure across the membrane given transmembrane pressure
is 315 mmHg. The arterial pressure and venous pressure are 130 mmHg and 100
mmHg respectively.
9.14 Explain why we measure bone stress. List and describe different sensors that can
be used for this measurement.
9.15 A bone specimen with a radius of 1.26 cm is tested in a uniaxial tension test.
The original length of the specimen is 30 cm. After applying a force of 1000 N
the length is changed to 31.5 cm. Calculate the stress and strain generated in the
specimen. Calculate the elastic (Young’s) modulus.
9.16 Describe arthrotripsometers, what they measure and their operating principle.
Also explain how there is a exponential decay of the oscillations of the
pendulum with respect to time.
9.17 Explain the importance of measuring water loss. Explain how water loss is
measured using flow hygrometry.
9.18 Explain how skin color is important to a dermatologist. List different methods of
skin color measurement and explain any one of them.
10.5 Measurement of Body Movement 1
10
Body Temperature, Heat, Fat, and Movement
Chao-Min Wu
This chapter deals with measurements of physical parameters from the total body, such as
temperature, heat, fat, and movement. These four parameters are somewhat related
because they all deal with energy. Heat is the thermal energy content of a body. It is
defined as the vibratory motion of its component particles. Body heat is the result of a
balance between heat production and heat loss, which occurs through conduction,
convection, radiation and evaporation. Temperature is a measure of the tendency of a
body to transfer heat from or to other bodies. Balancing heat production within the body
against heat loss to the surroundings determines the body temperature. Body fat is energy
related in that it is part of the heat storage process. We discuss how the human body
regulates body temperature and how this regulation is related to heat production, heat loss,
and heat storage in the form of body fat. We further discuss clinical temperature
measurement techniques and common devices. Then we move our discussion to various
methods of body heat and body fat measurements. We compare various techniques
among these measurements as well. Finally, we cover the measurement of body
movement. In each section, we explain why and how to measure these physical
parameters.
1
2 Body Temperature, Heat, Fat, and Movement
–
Anterior Sweating
hypothalamus Skin vasodilation
set-point Decrease metabolism Core
temperature
Posterior Skin
Muscle shivering
hypothalmus temperature
Skin vasoconstriction
Increase metabolism
+
Skin thermal
receptors
Hypothalamic thermal
receptors
Figure 10.1 The human temperature regulation system can increase or decrease body temperature.
The heat maintenance region is located in the posterior hypothalamus where it
combines sensory signals transmitted from the anterior hypothalamus and peripheral
temperature sensors to control the temperature-increasing or temperature-decreasing
mechanisms of the body. This part of the hypothalamus acts like a thermostat controller
in room temperature regulation. When the body is too hot, the heat maintenance region
triggers temperature-decreasing procedures to reduce body heat by sweating, skin
vasodilation, and decreasing tissue metabolism. Sweating is the main mechanism by
which the body causes heat loss via water evaporation. The effect of skin vasodilation is
an increase in heat loss by increasing blood flow to the skin surface. Vasodilation is
handled through the autonomic nervous system. Excessive tissue metabolism is inhibited
to decrease heat production. In turn, it decreases body temperature.
On the other hand, temperature-increasing procedures are instituted when the
body temperature is too cold. They are: skin vasoconstriction, shivering, and increased
thyroxine secretion. The reaction of skin vasoconstriction, in contrast to skin vasodilation,
is caused by stimulation of the posterior hypothalamus. Shivering is a reaction of skeletal
muscles to call for an increase in heat production of the body. It is excited by the cold
signals from the skin receptors (Guyton and Hall, 1996). The primary motor center for
shivering is located at the posterior hypothalamus (dorsomedial portion). The response
time of skin vasomotor control and shivering is within minutes. However, the increase of
thyroxine secretion is a long-term process of heat production increase through the
4 Body Temperature, Heat, Fat, and Movement
endocrine system. It usually takes several weeks to complete. Thyroxine is the hormone
secreted by the thyroid gland to increase the rate of tissue metabolism.
In addition to the physiological control of body temperature, behavioral
response to the change of environment plays an important role in body temperature
control, especially in a very cold environment. It is usually an efficient way to maintain
our body temperature. For example, we put on more clothes when we expect severe cold
weather.
Radiation from the skin surface is the main mechanism of heat loss from the body. It
accounts for 60% of the body heat loss under normal conditions (Guyton and Hall, 1996).
The relationship between the energy emitted from the skin surface and the skin surface
temperature forms the basis of surface temperature measurement. The surface
temperature is determined by the skin blood supply and the condition of subcutaneous
tissues. It is a good indicator for bone fractures and inflammation. Abdominal placement
of a thermal sensor (thermistor) for skin temperature measurement is important in
neonatal monitoring against cold stress. Medical thermography is a technique to scan the
skin surface and map its thermal distribution. Physicians have used this technique to
10.2 Clinical Temperature Measurement 5
diagnosis tumors and breast cancers. However, modern imaging techniques, such as MRI
and mammography, are more popular diagnostic tools for these carcinomas.
Liquid crystal thermometers are made of chemical compounds with the property of
reflecting light within a range of temperature (26 to 40 °C). For example, the crystals can
be formulated to a change in color through red, yellow, green, and blue as the substance
is heated from 38 °C through to 40 °C. They are constructed by impregnating Mylar
strips with encapsulated liquid crystals. With attached adhesive, they are often used for
surface temperature measurements, especially when monitoring the temperature trend in
an operation room. Liquid crystal thermometers are easy to use, inexpensive, and
disposable but less precise than other thermometers (see next section). In general, their
resolution is around 0.2 °C to 0.5 °C.
Mercury thermometer
Caution should be taken in handling and storing of the mercury thermometer. The
accuracy of the mercury thermometer is usually manufacturer-dependent.
Electronic thermometer
3
10
10
1
← β = 3500 K
← β = 4000 K
R/R25°C
← β = 3000 K
−1
10
−3
10
−50 0 50 100 150 200
Temperature (°C)
Figure 10.2 Examples of resistance–temperature curve for three NTC thermistors, β = 3000 Κ,
3500 K, and 4000 K.
1 1
RT = R0 exp β −
T T0 (10.1)
(
V = C1 (T1 − T2 ) + C 2 T12 − T22 ) (10.3)
where C1 and C2 are constants that depend on the thermocouple pair with T in kelvins. If
the temperature difference T1 – T2 is small, the second term of Eq. (10.3) can be dropped.
The sensitivity or thermoelectric power of the thermocouple is the derivative of Eq. (10.3)
with respect to T1 is
dV
α=
dT1 . (10.4)
Voltmeter
Copper Copper
Constantan
Figure 10.3 The J type thermocouple is formed from copper and constantan.
Digital
display
The major advantages of electronic thermometers are the small size of sensors,
ease of use, and fast response time. Their disadvantages are similar to those of mercury
thermometers, such as mucosal damage and cross–infection if not carefully handled. A
10.2 Clinical Temperature Measurement 9
disposable probe cover is used to prevent cross-infection. Sliding the probe from the floor
of the mouth to the sublingual pocket warms up the probe and decreases oral temperature
measurement error and response time.
Infrared thermometer
A basic infrared thermometer (IRT) design includes four parts: a waveguide (tube) to
collect the energy emitted by the target, a sensor to convert the energy to an electric
signal, an emissivity adjustment to match IRT calibration to the emitting characteristics
of the object being measured, and a sensor temperature compensation circuit to ensure
that temperature variations within the IRT are not transferred to the final output. Figure
10.5 shows a block diagram of an infrared thermometer. The tympanic membrane emits
infrared energy. The collection of infrared radiation is enhanced by a waveguide, which
improves reflection of the infrared radiation onto the sensor when the shutter opens. The
pyroelectric sensor converts infrared radiation into electric signals, which are fed through
the amplifier, multiplexer (MUX), and analog-to-digital converter (ADC) to the
microprocessor for processing and display. The microprocessor handles emissivity
adjustment and temperature compensation, and calculates the patient (target) temperature
according to Eq. (10.5):
Tb =
[χ (N T − N T0 ) + T04 ] (10.5)
4
(
N T = Aσε a Tb4 − Ta4 , ) (10.6)
where A is the effective body (target) area, σ the Stefan–Boltzmann constant (= 5.67 ×
10–8 W/m2·K4), εa the emissivity of surroundings (sensor), Tb the body temperature,
and Ta the sensor temperature (Ta and Tb are both in kelvins).
10 Body Temperature, Heat, Fat, and Movement
Shutter
Ambient sensor
Ear
IR Micro
Ta processor
Tb MUX A/D
Amp.
Sensor
Digital
Waveguide Window display
Shutter
switch
Figure 10.5 The infrared thermometer opens a shutter to expose the sensor to radiation from the
ear.
The ambient sensor that monitors the sensor temperature (Ta), is a thermistor.
The infrared sensor is a pyroelectric sensor followed by a current-to-voltage converter
that is used for the pyroelectric sensor (Fraden, 1991). The pyroelectric sensor, like the
piezoelectric sensor does not respond to steady inputs and in response to a step input
yields a high-pass first order output of the form e-t/τ. The whole measurement cycle,
from shutter opening to display, is 1 s. Its accuracy and resolution are about ±0.1 °C. The
main advantages of the infrared thermometer are that it only contacts the ear canal and
not the eardrum (i.e. less infection), ease of use, and fast response. It is suitable for the
temperature measurement of young or high-risk patients. However, inaccurate
measurements occur when patients have curved ear canals or the cerumen (ear wax)
obstructs the tympanic membrane.
Direct calorimetry has been used on animals and in humans for energy and thermal
regulation studies. Since the beginning of the century, many direct calorimeters have
been built. They are usually classified under one of four principles of operation: gradient
layer calorimeter, air–flow calorimeter, water flow calorimeter, and compensating heater
calorimeter. The gradient layer calorimeter is usually used for studies in temperature
regulation of human subjects while they are performing different work. The air–flow
calorimeter is popular in studies of energy expenditure for humans and animals. The
water flow calorimeter can be used for both applications without social isolation. The
compensating heater calorimeter is commonly used in experiments for small animals.
Webb (1985) provides detailed information on the historical development of these direct
calorimeters and a list of human calorimeters.
Gradient layer calorimeters are isothermal calorimeters. They are chambers lined on all
surfaces with a layer of insulating material and surrounded by a heat sink metal wall or a
constant temperature water jacket. As sensible (i.e. nonevaporative) heat from the subject
passes through the thin layer of insulating material to the heat sink (metal wall or water
jacket), the temperature gradient across the layer of insulation is sensed by a network of
thermocouples (e.g. copper–constantan). Benzinger and Kitzinger (1949) first introduced
this type of calorimeter. These thermocouples are used to detect the temperatures of the
inner and outer surfaces. The evaporative heat loss from the subject can be estimated
from the increased moisture in the ventilating air that exits the chamber. The main
objective of the gradient layer calorimeter design is to prevent any sensible heat loss from
convecting out of the chamber by heating up the ventilating air. The design and
construction of a gradient layer calorimeter is a complicated engineering task. It includes
the control of the ventilation and cooling system and instrumentation for the temperature
measurement for a huge chamber. The performance of the calorimeter depends on initial
calibration and routine calibration with known electric energy and is controlled by a
computer. McLean and Tobin (1987) provide a detailed description of how to design and
construct a gradient layer calorimeter. They reported that the accuracy of total heat loss
measurement is around ±1% or ±3 W within the operation range (10 to 40 °C). The
response time of the measurement to a unit change of heat output (i.e. when a person is
first placed inside) is 80% of the change is recorded within 3 min and 90% within 10 min
but the full response takes nearly an hour. Their calibration equations are listed as follows:
VC =
(QD N − QD C − QD L ) (10.7)
kc
QDC
VT = (10.8)
VDk t
12 Body Temperature, Heat, Fat, and Movement
C vC C vC
QE C + QC C + ∆p
V V
VE = (10.9)
ke
QC = (V + V )k + QC
N C T L (10.10)
VC
QC E = (VE − VT )k − ∆p (10.11)
vC
where VC, VT, and VE are measured voltages from the chamber, thermocouples, and
evaporative heat, respectively; QD N and QD E are the rates of nonevaporative and
evaporative heat production in the chamber; QC C is the rate of nonevaporative heat
convection out of the chamber; QD L is the rate of heat leakage through any gaps in the
heat-sensitive layer; VD and νD are chamber and evaporative heat measuring air flow rates;
kc, ke, and kt are constants (sensitivities); ∆p is a term due to unequal sensitivities or
unequal sampling rates; during the calibration, as performed with the chamber controlled
at 25 °C and the evaporative heat measuring unit at 10 °C, the desired value k = kc = ke
VD /νD = kt VD = 46.5 W/mV (40 kcal/h⋅mV) is obtained by adjusting the output circuit of
the calorimeter when the ventilation system is in operation and with no load in the
chamber ( QC N = QD E = 0).
The advantages of the gradient layer calorimeter are accuracy and fast response;
however, they are expensive and not transportable. Figure 10.6 shows that they are
complicated in design and construction.
10.3 Measurement of Body Heat: Calorimetry 13
Insulation layer
Constantan
Metal wall
Cooling tubes
Copper
Thermocouple
connection
Figure 10.6 In a gradient layer calorimeter, thermocouples measure the difference in temperature
across the wall. Ventilating system and measurements not shown.
An air–flow calorimeter is a convection calorimeter where heat loss through the surface
is prevented by insulation. Heat loss from the subject is carried away from the chamber
by air–flow. The rate of heat removal by the air–flow is estimated from the mass flow
rate and temperature rise of the air. An important objective for the design of an air–flow
calorimeter is to prevent any water condensation on the surface of the chamber. Figure
10.7 shows the principle of an air–flow calorimeter. The seated subject is confined in a
chamber with polyurethane insulation applied to the outsides. The heat loss from the
subject is transferred to the ventilating air, whose temperature is measured at the inlet and
the outlet of the chamber. The heat loss ( Q a ) from the subject is the product of air mass
flow rate ( m ), specific heat (ca), and temperature change (T2 – T1 ) of the ventilating air:
Q a = m c a (T2 − T1 ) (10.12)
14 Body Temperature, Heat, Fat, and Movement
T2
Thermometer
Air in
Insulated chamber
(polyurethane)
T1
Figure 10.7 An air flow calorimeter measures inlet and outlet temperatures, flows, and humidity.
The advantages of the air–flow calorimeter are simplicity of construction and
fast response to heat change (7 to 20 min). However, they need very accurate humidity
and temperature measurements.
T1
Water in
Inlet thermometer
Figure 10.8 The water flow calorimeter measures the inlet and outlet water temperature.
AC Power
meter
Power
source
Heater
Figure 10.9 The compensating heater calorimeter requires less heater power when the subject
supplies heat.
16 Body Temperature, Heat, Fat, and Movement
Open-circuit system
The open-circuit system determines VDO 2 from the minute ventilation rate ( VDE ) and the
difference ( ∆F ) between inspired ( FI ) and expired ( FE ) gas concentrations. This
calculation of VDO 2 includes a correction for change in barometric pressure, temperature,
and relative humidity of the inspired air for adjustment of the volume to standard
temperature, pressure, and dry air condition (STPD: 0 °C, 760 mmHg, dry). Energy
expenditure (EE, kcal/day) is estimated by the standard (Weir, 1949) equation:
[( ) ( ) ]
EE = VDO 2 (3.941) + VDCO 2 (1.11) × 1440 (10.13)
where VDO 2 and VDCO 2 are expressed in L/min, and 1440 = the number of minutes in a
day. It assumes that 12.3% of the total calories arise from protein metabolism. When
urinary nitrogen excretion ( ND ) is measured, a third term (– ND (2.17)) in g/min is added to
the bracket of Eq. (10.13) to account for protein correction of energy expenditure.
Open-circuit calorimeters use either the mixing chamber or the dilution
technique to measure inspired and expired concentrations of oxygen ( FIO 2 , FEO 2 ) and
10.3 Measurement of Body Heat: Calorimetry 17
carbon dioxide ( FICO 2 , FECO 2 ), and inspired and expired minute ventilation ( VDI ,
VDE ). In an ideal situation, VDO 2 is determined by
However, it is technically difficult to measure the difference between VDI and VDE . The
relationship between VDI and VDE is commonly calculated by equating the quantity of
nitrogen in inlet and outlet air stream, i.e. the Haldane transformation. Then VDI is
obtained by the equation
FEN 2 D
VDI = VE (10.16)
FIN 2
where FEN2 = expired nitrogen concentration and FIN2 = inspired nitrogen concentration.
If only N2, O2, and CO2 are analyzed, VI can be obtained by
1 − FEO 2 − FECO 2 D
VDI = VE (10.17)
1 − FIO 2 − FICO 2
where FEN2 = 1 – FEO2 – FECO2 and FIN2 = 1 – FIO2 – FICO2. Substituting Eq.
(10.17) into Eq. (10.14), the formula for VO 2 becomes
or
when FIN2 = 1 – FIO2, because FICO2 = 0.03% in most situations. This equation is used
by most of commercially available systems. In general VDI and VDE are of similar
magnitude but not equal; the same is true for FIO2 and FEO2. The problem with using
the Haldane transformation is that it becomes increasingly inaccurate as the FIO2
increases above 0.6 (see below). Consequently, extreme precision in estimates of these
four variables would be necessary if VDO 2 were to be calculated directly from Eq. (10.14).
If we assume VDI is equal to VDE , there will be significant errors in VDO 2 measurement, for
example in a patient whose respiratory rate is 10 breaths per minute at a 1000 mL tidal
18 Body Temperature, Heat, Fat, and Movement
volume ( VDT ) and whose VDO 2 is 250 mL/min. The respiratory quotient (RQ, given by the
ratio VDCO 2 / VDO 2 ) of this patient is 0.7. This means that 25 mL per breath of O2
produces 17.5 mL per breath of CO2. In this case, the difference between VDI and VDE is
75 mL. If VDO 2 is measured at an O2 concentration of 60% (FIO2 = 0.6) assuming that
VDI is equal to VDE , an error of 18% would be obtained for VDO 2 measurement (205
mL/min). When RQ reaches 1.0, this error becomes small. Measurement of VDCO 2 is
much simpler. When inspired air is room air, it contains only very small amount of CO2
(< 0.03%). VDCO 2 may be calculated accurately from a simpler form
This simpler form is also applied to measurement of VDCO 2 for patients with mechanical
ventilation where FICO2 is 0.0.
Most commercially available open-circuit systems use mixing chambers as in
Figure 10.10. Expired air from the patient passes through a facemask or mouthpiece and
is directed to the mixing chamber. At the end of the mixing chamber, a small sample is
drawn by a vacuum pump through O2 and CO2 sensors that measure FEO2 and FECO2.
A pressure transducer and thermistor are necessary for pressure- and temperature-
compensated gas measurements. A volume transducer is used to measure minute
ventilation ( VDE ). A microprocessor controls the instrument and provides minute-by-
minute data on O2 consumption and CO2 production, RQ, and energy expenditure of the
subject.
From
Inspired
gas source Computer
O2 CO2 control
Pressure
Switch transducer
Vacuum
pump
Thermistor
Exhaust
Patient
Expired gas Flow
from patient transducer
Mixing chamber
Figure 10.10 A microcomputer-based open-circuit system includes a mixing chamber, O2 and CO2
analyzers, and the various variables (pressure, flow, temperature) used to calculate VDO 2 .
When measuring respiratory exchange gas, there are two broadly used methods:
the air may be directly expired through a mask or mouthpiece as mentioned above or it
may be drawn from a hood or canopy enclosing the subject's head. The latter is
sometimes called a flow-by system. It needs a pump to draw constant air over a subject's
10.3 Measurement of Body Heat: Calorimetry 19
head and this air–flow is mixed with the patient's expired air. The process of VDO 2 and
VCO 2 measurements is the same as before. However, the patient's expired air is diluted
by the constant air flow, the difference in O2 and CO2 concentrations are smaller than
those measured by the former method. Consequently this demands that the analyzers be
more precise and free of drift. The Deltatrac metabolic monitor (Datex–Engstrom
Division, Instrumentarium Corp., PO Box 900, FIN-00031, Helsinki, Finland) and
Vmax29n (SensorMedics Corp., Yorba Linda, CA) represent this type of system. It is
suitable for use with a mechanical ventilator or a canopy for patients of all ages. The
Deltatrac consists of an infrared CO2 analyzer, a paramagnetic O2 analyzer, a 4 L mixing
chamber, and a CRT display (Weissman et al., 1990). When the Deltatrac is used with a
mechanical ventilator (Figure 10.11), the expired gases enter the mixing chamber, from
which gas is sampled and analyzed to determine FEO2 and FECO2. After the expired
gases pass through the mixing chamber, the gases are diluted with room air driven by a
flow generator (Q, 40 L/min). VCO 2 is then calculated as
(
VDCO 2 = Q F * CO 2 ) (10.20)
where F*CO2 is the concentration of CO2 in the diluted gas sample. The respiratory
quotient (RQ) is then calculated using the Haldane transformation.
1 − FIO 2
RQ = (10.21)
FIO 2 − FEO 2
− FIO 2
FECO 2
where FIO2 is the O2 concentration from the inspired limb of the ventilator. VDO 2 is then
determined by the formula
VDCO 2
VDO 2 = . (10.22)
RQ
20 Body Temperature, Heat, Fat, and Movement
Metabolic
monitor
Humidifier
Inspiratory
limb
CO2 and O2
FIO2 analyzers
Patient FECO2,
FEO2 F*CO2
Expiratory limb
Constant
Exhalation flow
valve generator
Mechanical
ventilator
Room air
Figure 10.11 A dilution system used by Deltatrac uses mechanical ventilation. FIO2 is the inspired
oxygen concentration, FECO2 is the true expired carbon dioxide concentration, FEO2 is the
expired oxygen concentration, and F*CO2 is the diluted carbon dioxide concentration.
A microprocessor controls the sampling of gases. CO2 measurements are made
by alternately sampling the diluted CO2 (F*CO2) and the true CO2 concentration
(FECO2). O2 measurements occur by alternately sampling the difference between
inspired and expired O2 (FIO2 – FEO2) and the measurement of FIO2. Weissman and
colleagues (1990) reported that the accuracy of the Deltatrac for VDO 2 and
VDCO 2 measurements is within ±7%. They also showed a different calculation procedure
for canopy measurement of VDO 2 and VDCO 2 in spontaneously breathing patients.
In addition to the mixing chamber and the dilution techniques mentioned above,
some mixing chamber systems use a breath-by-breath technique for VDO 2 and
VDCO 2 measurements. The breath-by-breath technique was originally designed for
exercise testing and has been adapted for use with a mechanical ventilator. It makes
measurements for every breath of the patient's expiration and differs from the mixing
chamber system in that the mixing chamber is absent. Gas samples are collected on a
breath-by-breath basis and are averaged over a preset time to obtain mixed expired values
(Branson et al., 1995). Furthermore, bedside mixing chamber systems can be modified
and used for ambulatory and field studies by dispensing with the CO2 analyzer to
simplify the procedure. However, this results in lower accuracy.
Closed-circuit system
10.3 Measurement of Body Heat: Calorimetry 21
Closed-circuit systems calculate VDO 2 by measuring over time the volumetric change
from a source of oxygen. They use the volumetric loss and replenishing techniques to
measure VDO 2 . The former observes how long it takes to consume all the oxygen that is
within a spirometer. The latter measures the amount of oxygen needed to keep the
spirometer full. A generic volumetric loss closed-circuit system consists of a CO2
absorber, a breathing circuit, a spirometer (see Figure 9.2) full of 100% O2, and a CO2
analyzer (Figure 10.12). Closed-circuit systems allow VDO 2 and VDCO 2 measurements
during mechanical ventilation and during spontaneously breathing ventilation with a
mouthpiece plus nose-clip or facemask. During measurements, the patient inspires O2
from the spirometer and the expired gases enter the mixing chamber. Samples of the
expired gas are analyzed to obtain FECO2 before passing through the soda lime scrubber.
Movements of the spirometer are recorded as VT and VDE . The volume change in the
spirometer as a function of time is used to calculate the rate of O2 consumption. In a
similar way as previously described, the rate of CO2 production is determined from
measurements of VDE and FECO2. During mechanical ventilation, the bellows in Figure
10.12 is driven by the ventilator to deliver gas to the patient.
From ventilator
Bellows
Spirometer
Patient
CO2
Scrubber As oxygen
Mixing supply and
chamber analyzer
Sample
gas
CO2 analyzer
Figure 10.12 A closed-circuit system uses the volumetric loss principle. The spirometer is used as
an oxygen supply and the volume change as a function of time in the spirometer is used to calculate
the rate of oxygen consumption.
The whole body chamber is an early design of a volume-replenishing closed-
circuit system. In the whole body chamber the subject is kept in a sealed chamber that is
ventilated with a constant, measured supply of fresh air. The respiratory gas exchanges
produced by the subject are measured in terms of their effects on the composition of
chamber air. Samples of well-mixed air are drawn from the chamber for continuous
analysis. The differences in oxygen and carbon dioxide concentrations in the in-going
and out-going air are measured and used to calculate energy expenditure. McLean and
Tobin (1987) stated that the Atwater and Benedict respiration chamber represents one of
the earliest designs of this type of calorimeter. The fundamental design of a whole body
22 Body Temperature, Heat, Fat, and Movement
chamber is similar to and functions as the mixing chamber in Figure 10.12. During the
experiment, air is drawn from the chamber by an air pump and passes through an external
circuit. This external circuit includes one CO2 absorber, two moisture absorbers, a
spirometer, and an oxygen supply. Moisture in the air is absorbed by passage through a
container of sulfuric acid and this is followed by absorption of CO2 in soda lime; the
moisture given off from the soda lime is then absorbed by a second container of sulfuric
acid. The pressure drop in the system, caused by the uptake of oxygen by the subject and
the absorption of any expired CO2, is detected by the spirometer. As pressure in the
chamber decreases, the volume of spirometer decreases. This then releases a valve on the
oxygen inlet line. As the oxygen supply replenishes the chamber with oxygen, the
spirometer bell rises and the valve is closed again. A meter is used to measure the
delivery of oxygen. The amount of CO2 production is obtained from a weight change of
the soda lime container plus the moisture collected in the second sulfuric acid container.
The whole body chamber offers fast response, accuracy and precision and
allows comparative measurements during sleep, at rest, before and after meals, and
during controlled periods of exercise. Ravussin and Rising (1992) reported that the
response time for their respiration chamber (a total volume of 19500 L) measured during
CO2 dilution was 3 min for 93% of the total response and 4 min for 99% of the total
response. In their review of techniques for the measurement of human energy expenditure,
Murgatroyd et al. (1993) reported that an accuracy of ±1% in gas exchange measurement
by the whole body chamber is achievable. They further reported that an accuracy of 1.5
to 2% of energy expenditure was obtained if the precision of the gas exchange rate is
better than 4 mL/min over a period of 30 min in a chamber with a volume of 10000 L.
Although the whole body indirect calorimeters provide fast response, accuracy and
precision of measurement, satisfactory results are only obtained with careful chamber
design and construction, and by the choice of the best possible instrumentation.
Nevertheless, expensive cost and artificial environment (the subject is not under free-
living conditions) are still its disadvantages. Current design of the closed-circuit
volumetric replenishment indirect calorimeter employs an ultrasonic transducer to
monitor the volume level of a bellows that is used as an oxygen reservoir. As oxygen is
depleted, a microcomputer-controlled calibrated volumetric pulse of oxygen is delivered
to maintain the bellows at the preset reference level. The VDO 2 is calculated by
multiplication of the pulse volume by the number of pulses per minute. In addition, the
ultrasonic transducer is also used to monitor VT. Measurement of VDO 2 made by this
technique has been found to be quite accurate. A good correlation (r > 0.99) between
measured and actual VDO 2 has been reported by several authors (Branson et al., 1988;
Keppler et al., 1989). Advantages of the closed-circuit indirect calorimeters are that they
are not affected by fluctuations of FIO2 and by an FIO2 > 0.6. Details of technical
considerations have been discussed in Branson et al. (1995).
The doubly labeled water (DLW) method was first introduced by Lifson, Gordon, and
McClintock in the 1950s (Lifson et al., 1955). They used the principles of indirect
calorimetry to measure total energy expenditure from the disappearance rates of two
10.3 Measurement of Body Heat: Calorimetry 23
stable isotopes: deuterium (2H) and oxygen 18 (18O). Since the early 1980s (Schoeller and
van Santen, 1982), the use of doubly labeled water has provided a technique where the
total energy expenditure can be measured in a free-living individual over 1 to 2 weeks. It
also has considerable appeal for field studies to estimate the energy cost of activities and
thermogenesis by subtracting the basal metabolic rate from the total energy expenditure.
The principle of the method is based on six assumptions listed as follows:
1. The body is in steady state; i.e. the volume of total body water pool remains
constant over time.
2. The turnover rates of water and carbon dioxide are constants over time.
3. The volume of labeled water distribution is only equal to total body water.
4. The isotopes are lost only as water and carbon dioxide.
5. Losses of labeled water and carbon dioxide in urine and exhaled carbon dioxide
have the same level of enrichment as in body water.
6. The background levels of the isotopes remain constant.
k18 – k2 = rCO2
Figure 10.13 Principle of the doubly labeled water method. r is the production rate, k represents
rate constants determined from the experiment.
Figure 10.13 shows the principle of the method. After a subject drinks a dose of
DLW, the deuterium mixes with the body water pool and the 18O mixes with both water
24 Body Temperature, Heat, Fat, and Movement
and bicarbonate pools. The production rate of CO2 is calculated as the difference
between 18O and 2H disappearance rates. The formula for this calculation is:
W
rCO 2 = VCO 2 = × (k18 − k 2 )× 22.4 (10.23)
2
where r (in L/h) is the production rate, W is the size of total body water in mol and k18
and k2 are fractional disappearance rates for 18O and 2H, respectively. They are
computed from measurements of the isotopic concentrations c18 and c2 using k = (1/c) ×
(∆c / ∆t). The factor 2 in the denominator of Eq. (10.23) is the molar proportion of
oxygen in carbon dioxide in relation to water. In practice, Lifson et al. (1955) modified
Eq. (10.23) by including the fractional factor: f1 (0.93) for 2H2O gas/2H2O liquid, f2
(0.99) for H2 gas/H218O liquid, and f3 (1.04) for C18O2 gas/ H218O liquid to take a
natural fraction of the isotopes between different body fluids into account. Eq. (10.23)
becomes
W f −f D
rCO 2 = VDCO 2 = (k18 − k 2 ) − 2 1 m × 22.4
2 f 3 2 f 3 18
or
W D
rCO 2 = VDCO 2 = (k18 − k 2 ) − m × 22.4 (10.24)
2.08 624
where m D is the rate of water loss as vapor in g/h. A minimum of two urine samples (one
a few hours after the dose and another at the very end of the experiment) are collected to
compute the rate constants (k18 and k2) and the total body water (W). A number of cross-
validation studies have been performed by comparison with gas exchange analysis and
the whole body chamber under resting condition as well as during sustained heavy
exercise (see reviews in Ritz and Coward, 1995; Mclean and Tobin, 1987; Prentice et al.,
1991; Ravussin and Rising, 1992). Results showed that the doubly labeled water method
is a suitable technique for assessing energy expenditure in free-living circumstances with
an accuracy of ±5%. However, large underestimates were found in obese subjects
(Ravussin and Rising, 1992). Advantages of the doubly labeled water method are that it is
noninvasive (oral dosing), nonintrusive (urine sampling), and suitable for noncompliant
subjects and field studies. In addition, deuterium distribution space can be used to
estimate total body water for an estimate of body composition (see section 10.4.1).
Nevertheless, the high cost for the expense of 18O-enriched water and technical
complexity are disadvantages of the doubly labeled water method. The amount of dosage
and its resulting cost depend on how precise the mass spectrometer is. In-depth
knowledge of the assumptions of the doubly labeled water technique is required to apply
it to various physiological and pathological situations.
10.5 Measurement of Body Movement 25
The use of the deuterium oxide dilution technique to measure body water to predict FFM
uses the fact that fat is mainly anhydrous and water is a relatively constant component
25
26 Body Temperature, Heat, Fat, and Movement
(73.2%) of FFM. It is based on the assumption that the isotope has the same distribution
volume as water, is exchanged by the body in a manner similar to water, and is nontoxic
in the amount used. During the experiment, subjects are asked to drink or are injected
with a known amount of the labeled water (deuterium oxide water). Then the amount of
the labeled water in saliva, urine, or plasma is determined after an equilibrium period (3
to 4 h). The calculation of total body water (TBW) volume is based on the simplified
relationship:
where C1V1 is the amount of the isotope given, and C2 is the final concentration of the
isotope in saliva, urine, or plasma and V2 is the volume of TBW. Available techniques
for the assay of deuterium in aqueous solution are gas chromatography (see section
3.15.2), mass spectrometry (see section 3.12), and filter-fixed infrared absorptiometry.
The quantity of the isotope given depends on the analytical system used and the objective
of the research. Lukaski (1987) reported that a 10 g dose of deuterium oxide (99.7%
purity) mixed with ~300 mL of distilled-deionized water is used routinely in his
laboratory for the filter-fixed infrared absorptiometry. Some investigators have given an
oral dose of deuterium oxide of 1 g/kg body weight (Mendez et al., 1970) for gas
chromatography and an oral dose of 2 g D2O has been used for mass spectrometry
(Halliday and Miller, 1977).
For adult white individuals (Wellen et al., 1994), the calculation of fat free mass
from TBW volume is done by dividing V2 by 1.04 and 0.732 to account for the estimated
4% nonaqueous H+ exchange and water fraction of the FFM respectively.
V2
FFMTBW (kg ) = (10.26)
(1.04 × 0.732 )
(weight − FFMTBW )
%BFTBW = × 100 (10.27)
weight
where FFMTBW is FFM derived from TBW (V2), weight is the subject's body weight,
and %BFTBW is the derived percent body fat from FFMTBW. For populations of low
(children) or high (obese) TBW, water fraction of the FFM (0.732) in Eq. (10.26) should
be adjusted to maintain an accurate estimation of %BFTBW. This change of TBW is
attributed to the fact that 15% of adipose tissue is water. In general, the accuracy
associated with measuring body fat by the TBW method is about 3% to 4% (Lukaski,
1987).
The most accurate method for estimating body fat is the direct measurement of body
density from underwater weighing. This method is often used as the reference method for
evaluating new indirect methods such as bioelectrical impedance analysis. Densitometry
10.4 Measurement of Body Fat 27
assumes that the human body has two distinct compartments: fat mass and fat free mass
(Siri, 1961; Brozek et al., 1963). The measurement of body density is accomplished by
measuring body volume (according to Archimedes' principle) and body weight (by the
difference between the subject's weight in air and weight in water). Archimedes' principle
states that the volume of a submerged object is equal to the amount of water the object
displaces. In a human subject, the apparent volume of the subject is determined by
correcting for the volume in the lung and the density of the water using the water
temperature at the time of the measurement.
The weighing device consists of a plastic chair suspended from a scale into a
huge tank containing water maintained at an average temperature of 35 °C. The tare
weight for the chair and water temperature are measured before the experiment. During
the measurement, the subjects sit on the chair with minimal clothes and are submerged
under water. While remaining under water, the subjects exhale completely and hold their
breath. Underwater weight is recorded after bubbles stop coming from the subject's
mouth and oscillation of the scale due to the waves has subsided. Underwater weight is
recorded 10 times for each subject and the mean of the three highest weights is used.
Residual lung volumes are measured on land to account for its contribution to the
subject's buoyancy. The body density (BD) is expressed in g/cm3 and is calculated from
Eq. (10.28) (Guo et al., 1987):
Wa
BD = (10.28)
Wa − (W w − T )
− RV
Dw
4.95
%BF = – 4.50 (10.29)
BD
4.57
%BF = – 4.412 (10.30)
BD
Lukaski (1987) reported that Eqs. (10.29) and (10.30) give results within 1%
body fat if the body density is within 1.10 to 1.03 and that Siri's equation yields higher
values than the Brozek’s equation for subjects with more than 30% body fat. Due to the
errors associated with variations of water and bone mineral contents, an error of 1 to 3%
in prediction of percent body fat with this method is expected (Lohman, 1990). In general,
the error is greater in populations with lower bone mineral content (children, women, and
old people) and populations with high bone mineral content (athletes). Noninvasive and
28 Body Temperature, Heat, Fat, and Movement
accurate measurement of body density, and good estimation of percent body fat have
made the underwater weighing a standard method for evaluation of other new and
portable techniques such as bioelectrical impedance analysis.
Another technique for direct measurement of body fat is dual photon absorptiometry or
dual energy X-ray absorptiometry. DPA/DEXA has been used in clinical settings to
diagnose osteoporosis and also to evaluate types of therapy for osteoporosis by directly
measuring bone mineral contents of the patients. Photon aborptiometry was first
introduced by Cameron and his colleagues (1962, 1963) to measure bone mineral content
(BMC) in the limbs with a single isotope (Iodine-125 at 27.3 keV or Americium-241 at
59.6 keV). This technique is called single photon absorptiometry (SPA) and is based on
the principle that bone mineral content is directly proportional to the amount of absorbed
photon energy by the target bone (see Figure 9.9). The necessity of constant thickness of
absorber and uncontrollable variations in transmission due to the presence of fat tissue
limit the measurement of BMC to bones only in the extremities. Mazess and his
colleagues (Mazess et al., 1970; Judy, 1970) developed dual photon absorptiometry to
measure contents of bone mineral and soft tissue in the appendicular parts of the skeleton
(e.g. femur) with the isotope Gadolinium-153 (153Gd, 44 and 100 keV). This method has
been applied to measure BMC in the axial skeleton (e.g. vertebrae) with the isotopes
Americium-241 (241Am, 59,6 keV) and Cesium-137 (137Cs, 662keV) by Roos and
Skoldborn (1974). The underlying principle of DPA/DEXA for the measurement of body
fat is that the ratio of absorbance of the two different energy level photons (R = µL/µH) is
linearly related to the percent of fat in the soft tissue of the body. The mass attenuation
coefficients for bone mineral and soft tissues are constant and are not affected by the
amount of bone and soft tissues along the path of the photon beam. They depend on the
energy level of the beam. The basic equations for DPA are as follows (Sorenson et al.,
1989):
N
ln 0 L = µ sL M s + µ bL M b, (10.31)
NL
N
ln 0 H = µ sH M s + µ bH M b, (10.32)
NH
where N0L and NL are the number of incident and transmitted low energy photons, N0H
and NH are the number of incident and transmitted high energy photons, µsL and µsH are
the mass attenuation coefficients of soft tissue at the low and high energies, µbL and µbH
are the same quantities for bone mineral, and Ms and Mb are the area densities in g/cm2
of soft tissue and bone mineral in the beam path. By solving Eqs. (10.31) and (10.32)
simultaneously, Mb can be derived as:
10.4 Measurement of Body Fat 29
N 0H N
R ln − ln 0 L
NH NL
Mb = (10.33)
(Rµ bH − µ bL )
where R = µsL/µsH. Based on the same basic equations, DEXA uses a constant voltage
X-ray generator and k-edge filters (made of cerium) to separate the X-ray beam into two
energy levels, or a switching-pulse system that rapidly alternates the constant voltage X-
ray generator to produce two beams with high and low energy simultaneously. Current
examples of the former type of densitometer include Norland XR-36 (Norland Medical
System, Inc., Fort Atkinson, WI) and Lunar DPX series (Lunar Radiation Corp., Madison,
WI). An example of the latter type of densitonometer is the Hologic QDR-1000 series
(Hologic, Inc., Waltham, MA). In recent years, DEXA has become the main stream in the
densitometer industry. DEXA scans faster and has better image quality over DPA.
Most of the current commercially available densitometers usually include a
scanning and examination table, X-ray system (scanner and scintillation detector), a
computer system with additional software to control and calculate BMC and body
composition, and positioning or calibration accessories. During the measurement, the
patient needs to lie supine on the table and a series of transverse scans are made from
head to toe at 1 cm intervals (Mazess et al., 1990). Measurements take 10 to 20 min for a
total body measurement or 1 to 5 min for a regional scan. Typical transverse scan image
area is around 60 cm × 200 cm. The value of R (weighted for mass of tissue in each pixel)
is determined for the number of pixels containing the soft tissue alone. The %BF can be
derived from the R value through application of a regression equation of %BF and the R
value resulting from the calibration phantom provided by the manufacturer. Total body
bone mineral content is calculated from the bone-containing pixels. These calculations
are usually processed by the software provided by the manufacturer. Technology of
automatic exposure control is also applied to DEXA to assure that each pixel receives the
proper exposure regardless of tissue thickness. In general, the precision error for the
percent body fat is within approximately 2% (Lukaski, 1987; Mazess et al., 1990; Jensen,
1992; Prentice, 1995). The advantages of DPA/DEXA include portability, capability of
measuring total body or regional composition of bone mineral and body fat, high
reproducibility (3 to 5%), and less dependence on the technician's skills and experience.
Narrow width table and size limitation prohibit using the DPA/DEXA technique to assess
very obese subjects. Radiation exposure of DPA/DEXA is between 2 and 10 mrem.
Indirect measurement techniques of body fat such as anthropometry and BIA require
empirically derived regression equations to estimate body fat or total body water (TBW).
These techniques are suitable for population studies where individual distinction is less
critical. They often need to be validated by reference methods such as densitometry and
the isotope dilution method. Anthropometry is a rapid and inexpensive way to evaluate
nutritional status for a population in a field study, but it requires a skilled technician and
30 Body Temperature, Heat, Fat, and Movement
Shoulder
Triceps
skinfold
Lange
caliper
a b
Figure 10.14 (a) Lange skinfold caliper used for assessing thickness of subcutaneous fat. (b)
Illustration of an example of skinfold measurement, triceps skinfold taken on the midline posterior
surface of the arm over the triceps muscle.
During the measurement of skinfold thickness, a double layer of skin and
subcutaneous fat tissue is grasped between the thumb and forefinger and pulled forward
from the site by 1 cm to allow the jaw of the caliper, maintaining a constant jaw pressure
of 10 g/mm2, to impinge on the skin. A distance of at least 1 cm should be maintained
between the fingers and the caliper arms to prevent any interference from the pressure of
the fingers. The calipers should be applied perpendicular to the body surface and
10.4 Measurement of Body Fat 31
perpendicular to the longitudinal axis of the skinfold (see Fig. 10.14(b)). The release of
the pressure of the caliper arms requires the measurement to be made within 4 s to
standardize differences among subjects in compressibility of skinfolds. The sequence of
measurements is repeated three times to reduce intrameasurer bias. For a more complete
review of specific anthropometric measurement techniques, see the Anthropometric
Standardization Reference Manual (Lohman et al., 1988).
Many regression equations are available for the prediction of body density and
thus body fat from skinfold thickness measurements. These equations are generally valid
for adult Caucasian populations, but overestimate body fat in elderly individuals. For
example, Jackson and Pollack (1978) and Jackson et al. (1980) published equations for
estimating body density in adult men, Eq. (10.34), and women, Eq. (10.35), aged 20 to 50
years.
where BD is body density, X1 is the sum of chest, abdomen, and thigh skinfolds (mm),
X2 is the sum of triceps, thigh, and suprailium skinfolds in mm, and X3 is the age.
Correlation coefficients (r) for Eqs. (10.34) and (10.35) are 0.905 and 0.842, respectively.
These body densities can subsequently be used for percent body fat estimation with Eq.
(10.27).
The advantages of body fat prediction from measurement of skinfold thickness
are that it is simple, inexpensive, and noninvasive. However, the accuracy and reliability
of skinfold thickness measurement depends on the skill of the measurer, interaction
between the measurer and the subject, and the characteristic of the subject. The potential
error associated with the use of skinfolds, assuming the correct equation is used, the
measurer is well trained, and correct skinfold sites chosen, is 3 to 4% (Lohman, 1990). In
addition, body fat of edematous (excess fluid in the tissues) patients is often
overestimated and very obese subjects cannot fit in the standard calipers.
Bioelectrical impedance analysis (BIA) for body fat measurement is used for
measurement of fat free mass of the body. It is also based on the two-compartment model
(in parallel) as mentioned earlier. Fat free mass is comprised mainly of electrolyte-
containing water (73.2%) in the body; it represents the main path of conduction for the
applied current. In contrast, fat mass behaves as an insulator. Therefore, the impedance of
the body is controlled by fat free mass. The method of BIA is based on the principle that
the impedance of an isotropic conductor is related to the length and cross-sectional area
of the conductor for a constant signal frequency. Prediction equations are often generated
using BIA values to estimate fat free mass. With a constant signal frequency and a
relatively constant conductor configuration, the impedance to an ac current can be related
to the volume of the conductor:
32 Body Temperature, Heat, Fat, and Movement
ρL2
Z= (10.36)
V
where Z is the impedance in ohms (Ω), ρ is the resistivity in Ω⋅cm, L is the conductor
length in cm, and V is the volume in cm3. A general procedure to measure the whole
body impedance is to let the subject, 2 h after eating and 30 min after voiding, lie supine
on a table with arms and legs sufficiently apart. Current injector electrodes are placed in
the middle of the dorsal surfaces of the right metacarpal–phalangeal (hand) and
metatarsal–phalangeal (foot) joints, respectively. Detector electrodes are positioned in the
midline of the dorsal surfaces of right wrist and ankle. Measurements of body resistance
and reactance are made on ipsilateral and contralateral sides of the body (as shown in
Figure 10.15). The lowest value of resistance for each subject is used to predict fat free
mass (FFM). Hoffer et al. (1969) first used four surface electrodes (tetrapolar) for the
study of TBW. The tetrapolar technique is used to minimize contact impedance and skin–
electrode interaction. They introduced 100 µA of ac current at 100 kHz in 20 healthy
subjects and 34 patients whose TBW were measured with the isotope dilution method
(see section 10.4.1) and showed that (stature)2/Z was the best predictor of TBW. The
correlation coefficients (r) of (stature)2/Z against TBW for 20 healthy subjects and 34
patients were 0.92 and 0.93, respectively. Lukaski et al. (1985) demonstrated in his total
body impedance study in 37 men that the resistance, R, is a better predictor than the
impedance (square root of R2 + Xc2) when the reactance, Xc, is small. They used a
current source of 800 µA at 50 kHz (RJL Systems, Detroit, MI). Consequently, the best
predictor of TBW (r = 0.95) and FFM (r = 0.98) was (stature)2/R. Kushner and Schoeller
(1986) cross-validated the prediction of TBW with BIA by generating a prediction
equation from a sample of 40 nonobese adults and applying it to 18 obese patients. They
found that (stature)2/R was a good predictor of TBW (r = 0.96). Segal et al. (1985) also
confirmed the significant role of (stature)2/Z as a predictor of FFM (r = 0.912) with 75
male and female subjects ranging from 4.9 to 54.9% body fat, but reported
overestimation of FFM for obese subjects. Because fat free mass includes virtually all
water in the body, the formula that is used to predict TBW via BIA is often used to
estimate FFM. A general model that is often used to predict FFM or TBW in kg is:
Stature 2
FFM or TBW = C1 + C 2 (Weight ) + C 3 (10.37)
R
FFM
% BF = 100 − (10.38)
Weight
10.4 Measurement of Body Fat 33
has been used to estimate percent body fat. Segal et al. (1988) and Gray et al. (1989) have
suggested that prediction of FFM can be enhanced by the use of gender and fatness
specific equations.
Excitation ac current source (I)
Detected voltage drop (E)
Figure 10.15 In resistance measurement on the ipsilateral side of the body current flows through
one arm, the trunk, and one leg.
Current commercially available BIA instruments (RJL Systems, Detroit, MI;
Dieto-System, Milan, Italy; etc.) are often equipped with multiple frequency function and
specialized software. They provide empirically derived regression equations and are
capable of estimating TBW, FFM, %BF, and extracellular water (ECW). Theoretically
TBW is modeled as a parallel R–C circuit with the cell membrane as a capacitor to
separate intracellular and extracellular fluids. At low frequency (~1 to 5 kHz), the current
only reaches the extracellular fluid while at high frequency (~50 kHz and higher) it
penetrates both fluids. Therefore, at low frequency the impedance can be used to estimate
ECW. Deurenburg (1995) reported in his study of 103 healthy males and females, aged
19–51 years that ECW was best predicted by resistance measured at 1 kHz (r = 0.93) and
TBW was best estimated by resistance measured at 100 kHz (r = 0.95). Regression
coefficients (C1, C2, and C3) estimated from his study for TBW and ECW were 0.368,
0.133, 2.1 and 0.177, 0.073, 2.0, respectively. Segal et al. (1991) also confirmed the
significant role of resistance measured at 5 kHz and 100 kHz for prediction of ECW (r =
0.93) and TBW (r = 0.95) in 36 healthy males. Regression coefficients (C1, C2, and C3)
estimated from their study for TBW and ECW were 0.455, 0.14, 3.43 and 0.284, 0.112, –
6.115, respectively. They cross-validated on two randomly selected subsets (18 each) and
showed that the prediction equations were reproducible and valid.
BIA offers the advantage of rapid, safe, noninvasive measurements of percent
body fat. In addition, multifrequency BIA can differentiate extracellular water (ECW)
and total body water (TBW). Problems with the BIA method include that there is no
generally accepted regression equation as a reference for all sample groups and BIA
instruments and that accurate prediction of a change of ECW is quite difficult.
Body movement is studied to learn how to analyze the movement of the body and to
understand its underlying principles. It is useful for improving performance of motor
34 Body Temperature, Heat, Fat, and Movement
skills (such as skating) and providing efficient methods of using the body in daily life
skills (e.g., walking and running). Analysis of body movement requires observation and
measurement. Many methods are available, ranging from simple visual observation to
complicated computerized systems. Typical research or clinical gait analysis laboratory
setups include goniometers, accelerometers, electromyography (EMG, see Chapter 7),
force plates, and kinematic (motion) analysis systems. A kinematic analysis system varies
from simple and subjective visual observation to more expensive and complicated video
or optoelectronic systems that can provide objective, quantitative data of three-
dimensional motion of selected points on the body. With additional force plate outputs,
the kinematic analysis system can be used as a kinetic analysis system to provide force
and moment information for human limb movement. In this section, we discuss working
principles and applications for goniometers, accelerometers, and video and optoelectronic
kinematic analysis systems. For additional information, refer to texts by Allard et al.
(1995), Whittle (1996), Winter (1990), and Medved (2001).
Goniometers
A goniometer is a device that measures relative angle of a joint that connects two body
segments. The goniometer (see Figure 10.16(a)) consists of two attachment arms, which
are fixed to two limb segments of the subject on either side of a joint to measure the joint
angle without interfering with natural movement. Traditionally, a goniometer uses a
resistance potentiometer to convert changes in rotation to an electric output proportional
to the positions of two attachment arms. The potentiometer is aligned with the joint axis
to establish the zero position and is calibrated to measure the joint angle in degrees.
Multiaxial measurements of the joint are possible with two or three potentiometers
mounted orthogonally in different planes. The output of the goniometer provides
continuous analog data of joint motion at relatively low cost. However, to ensure accurate
measurements, fixation to the body with cuffs around the soft tissues must be performed
carefully to prevent off-axis problems (Chao, 1980). The measurement error caused by
the translation of the joint center during movement can be improved by replacing the
rigid arms of the traditional goniometer with a flexible parallelogram linkage designed to
accommodate the motion change outside the measurement plane (Thomas and Long,
1964). A triaxial parallelogram electrogoniometer (see Figure 10.16(b)) has been used to
record motions in more than one plane for the lower limb (Isacson et al., 1986). Strain
gages have also been used as an alternative to the potentiometer. They deform when they
are stretched or compressed and the electric output is proportional to the change in joint
rotation. The working principle of strain gage and its related circuitry is discussed in
section 8.3.1.
10.5 Measurement of Body Movement 35
Plastic ribbons
Parallelogram
Connecting rods linkage
Vi Knee joint
Vout = k θ Triaxial
goniometer
a b
Figure 10.16 (a) A goniometer attached to the shank and thigh to measure knee rotation. Vi is the
input voltage. Vout is the output voltage that is proportional to the angle of knee rotation. (b)
Subject wearing a triaxial goniometer on knee joint.
Accelerometers
Accelerometers are instruments that measure acceleration. Basically, they all contain a
small mass and make use of Newton's law (F = ma) to measure the force that is required
to accelerate the known mass, as shown in Figure 10.17. There are two classes of
accelerometers used in the analysis of body movement: strain gage (usually piezoresistive)
and piezoelectric accelerometers. A strain gage accelerometer consists of strain gages
(wires) bonded to a cantilevered mass and a base to which the cantilever beam is attached.
When the base is accelerated, the cantilever beam deforms due to the inertia of the mass,
thus changing the strain in the wires. The strain in the wires changes their resistances. We
measure the change of resistance with a Wheatstone bridge circuit (see section 9.11.2),
which requires a differential amplifier. The resulting electric output is proportional to the
acceleration of the mass. A piezoresistive accelerometer works under the same principle
as a strain gage accelerometer. Instead of using strain sensitive wire, it uses piezoresistive
strain elements as sensors. Piezoelectric accelerometers measure the force directly. The
piezoelectric sensor converts the force produced by the acceleration to an electric charge
or voltage (see Eq. (8.7) and section 8.5.2 for the working principle and its charge
amplifier). For the analysis of body movement, the accelerometer is attached to a body
segment at a specific point to measure acceleration in one direction. Two- or three-
dimensional measurements are possible with several accelerometers grouped together
orthogonally. The use of accelerometers for the analysis of body movement provides an
alternative way to measure the velocity and displacement of the limb segment when the
initial values of the limb’s velocity and displacement are known (Morris, 1973). Some
practical limitations prevent widespread use of the accelerometer in body motion analysis:
necessity to exclude the effect of the field of gravity from true kinematic acceleration,
difficulty in extracting the rotational acceleration, and low frequency noise caused by the
baseline drift in the measurement output.
36 Body Temperature, Heat, Fat, and Movement
Strain
E
a
V
Seismic Cantilever
mass
Figure 10.17 Vertical acceleration of the accelerometer frame bends the cantilever beam because
the seismic mass remains at rest. Voltage output (V) is proportional to the acceleration (a). E is the
supply voltage.
Gait can be measured on a single film using sequential strobe lights and a camera. Gait
can be measured by video recorder. Figure 10.18 shows that a video or optoelectronic
kinematic analysis system consists of active or passive markers placed at selected skin
surface locations on bony landmarks, sensors that track markers' positions, and a
computer system to control the timing for active marker display and signal processing of
markers' tracings. The sensors, either optoelectronic circuitry or a video system, collect
kinematic data from active (infrared LED) or passive markers and low-pass filter the data
before an analog-to-digital converter (ADC), which feeds to a computer for later
computation of the position and velocity of the rigid body in three-dimensional (3-D)
space. A computerized analysis system often yields graphic displays that can provide
information on joint positions, joint angles, segment angles, velocity, and acceleration as
a function of time.
Camera 1 Camera 4
Markers
EMG
transmitter Camera
(option) Walk way control Display
Computer system
and
Interface
Force
platforms Analog interface
(option)
Camera 2 Camera 3
A/D
converter
EMG
receiver
Figure 10.18 An example of a gait analysis setup includes a four-camera kinematic system, two
force platforms, and an electromyogram (EMG) telemetry system.
10.5 Measurement of Body Movement 37
A j xi + B j yi + C j z i + D j
x ij = (10.39)
I j xi + J j yi + K j z i + 1
E j xi + F j yi + G j z i + H j
y ij = (10.40)
I j xi + J j yi + K j z i + 1
10.6 References
Allard, P., Stokes, I. A., and Blanchi, J.-P. 1995. Three-Dimensional Analysis of Human
Movement. Champaign IL: Human Kinetics.
Abdel-Aziz, Y. I. and Karara, H. M. 1971. Direct linear transformation from comparator
coordinates into object space coordinates in close-range photogrammetry. Proc.
Close-range Photogrammetry. Am. Soc. of Photogrammetry. 1–18.
Bair, D. N. and Davies, A. 1995. Temperature probes and principles of thermometry. In
R. L. Levine and R. E. Fromm, Jr. (eds.) Critical Care Monitoring: From Pre-
Hospital to the ICU. St. Louis: Mosby.
Behnke, A. R., Feen, B. G., and Welham, W. C. 1942. Specific gravity of healthy men.
JAMA, 118: 495–98.
Benzinger, T. H. and Kitzinger, C. 1949. Direct calorimetry by means of the gradient
principle. Rev. Sci. Instrum., 20: 849−60.
Branson, R. D., Lacy, J., and Berry, S. 1995. Indirect calorimetry and nutritional
monitoring. In R. L. Levine and R. E. Fromm, Jr. (eds.) Critical Care Monitoring:
From Pre-Hospital to the ICU. St. Louis: Mosby.
Branson, R. D., Hurst, J. M., Davis, K. Jr., and Pulsfort, R. 1988. A laboratory evaluation
of the Biergy VVR calorimeter. Resp. Care, 33: 341–7.
Brozek, J., Grande, I., Anderson, J. T., and Keys, A. 1963. Densitometric analysis of
body composition: revision of some quantitative assumption. Ann. NY Acad. Sci., 40:
110–3.
Cameron, J. R., Grant, R., and MacGregor, R. 1962. An improved technique for the
measurement of bone mineral content in vivo. Radiology, 78: 117.
Cameron, J. R. and Sorenson, J. 1963. Measurement of bone mineral in vivo: an
improved method. Science, 142: 230–2.
Chao, E. Y. 1980. Justification of triaxial goniometer for the measurement of joint
rotation. J. Biomechanics, 13: 989–1006.
Cobbold, R. S. 1974. Transducers for Biomedical Measurements: Principles and
Applications. New York: John Wiley & Sons.
Deurenburg, P. 1995. Multi-frequency impedance as a measure of body water
compartments. In P. S. W. Davies and T. J. Cole (eds.) Body Composition
Techniques in Health and Disease. Cambridge: Cambridge University Press.
Fraden, J. 1991. Noncontact temperature measurements in medicine. In D. L. Wise (ed.)
Bioinstrumentation and Biosensors. New York: Marcel Dekker.
Gray, D. S., Bray, G. A., Gemayel, N., and Kaplan, K. 1989. Effect of obesity on
bioelectrical impedance. Am. J. Clin. Nutr., 50: 255–60.
10.6 References 39
Guyton, A. C. and Hall, J. E. 1996 Textbook of Medical Physiology. 9th ed. Philadelphia:
W. B. Saunders.
Halliday, D. and Miller, A. G. 1977. Precise measurement of total body water using trace
quantities of deuterium oxide. Biomed. Mass Spectrom, 4: 82–7.
Hoffer, E. C., Meador, C. K., and Simpson, D. C. 1969. Correlation of whole body
impedance with total body water. J. Appl. Physiol., 27: 531–4.
Isacson, J., Cransberg, L., and Knuttsson, Z. 1986. Three-dimensional electrogoniometric
gait recording. J. Biomech., 19: 627–35.
Jackson, A. S. and Pollock, M. L. 1978. Generalized equations for predicting body
density of men. Br. J. Nutr., 40: 497–504.
Jackson, A. S., Pollock, M. L., and Ward, A. 1980. Generalized equations for predicting
body density of women. Med. Sci. Sports Exerc., 12: 175–82.
Jensen, M. D. 1992. Research techniques for body composition assessment. J. Am.
Dietetic Assoc., 92: 454–60.
Judy, P. E. 1970. A Dicromatics Attenuation Technique for the in vivo Determination of
Bone Mineral Content. Ph.D. Thesis, University of Wisconsin, Madison, WI.
Keppler, T., Dechert R., Arnoldi, D. K., Filius R., and Bartlett, R. H. 1989. Evaluations
of Waters MRM-6000 and Biergy VVR closed-circuit indirect calorimeters. Resp.
Care, 34: 28–35.
Kushner, R. F. and Schoeller, D. A. 1986. Estimation of total body water by bioelectrical
impedance analysis. Am. J. Clin. Nutr., 44: 417–24.
Lange, K. O. and Brozek, J. 1961. A new model of skinfold caliper. Am. J. Phys.
Anthropol., 19: 98–9.
Lifson, N., Gordon, G. B., and McClintock, R. 1955. Measurement of total carbon
dioxide production by means of D218O. J. Appl. Physiol., 7: 704–10.
Lohman, T. G. 1990. Body composition assessment in sports medicine. Sports Med.
Digest., 12: 1–2.
Lohman, T. G., Roche, A. F., and Martorell, R. (eds.) 1988. Anthropometric
Standardization Reference Manual. Champaign, IL: Human Kinetics Books.
Lukaski, H. C., Johnson, P. E., Bolonchuk, W. W., and Lykken, G. I. 1985. Assessment
of fat-free mass using bioelectrical impedance measurements of the human body. Am.
J. Clin. Nutr., 41: 410–17.
Lukaski, H. C. 1987. Methods for the assessment of human body composition: traditional
and new. Am. J. Clin. Nutr., 46: 537–56.
Levine, R. L. and Fromm, R. E., Jr. (eds.) 1995. Critical Care Monitoring: From Pre-
Hospital to the ICU. St. Louis: Mosby.
Mazess, R. B., Ort, M., Judy, P., and Mather, W. 1970. Absorptiometric bone mineral
determination using 153Gd. Proc. CONF-700515, Bone Measurement Conf.,
Chicago, IL, 308.
Mazess, R. B., Barden, H. S., Bisek, J. P., and Hanson, J. 1990. Dual-energy X-ray
absorptiometry for total-body and regional bone-mineral and soft-tissue composition.
Am. J. Clin. Nutr., 51: 1106–12.
McLean, J. A. and Tobin, G. 1987. Animal and Human Calorimetry. Cambridge, UK:
Cambridge University Press.
Medved V. 2001. Measurement of Human Locomotion. Boca Raton, FL: CRC Press.
40 Body Temperature, Heat, Fat, and Movement
Mendez, J., Keys, A., Anderson, J. T., Grande, F. 1960. Density of fat and bone mineral
of mammalian body. Metabolism, 9: 472–7.
Mendez, J., Procop, E., Picon-Reategui, F., Akers, R., and Buskirk, E. R. 1970. Total
body water by D2O dilution using saliva samples and gas chromatography. J. Appl.
Physiol., 28: 354–7.
Milsum, J. H. 1966. Biological Control System Analysis. New York: McGraw-Hill.
Morris, J. R. W. 1973. Accelerometry—a technique for the measurement of human body
movements. J. Biomechanics, 6: 729–36.
Murgatroyd, P. R., Shetty, P. S., and Prentice, A. M. 1993. Techniques for the
measurement of human energy expenditure: a practical guide. Int. J. Obesity, 17:
549–68.
Pace, N. and Rathburn, E. N. 1945. Studies on body composition. III The body water and
chemically combined nitrogen content in relation to fat content. J. Biol. Chem., 158:
685–91.
Prentice, A. M. 1995. Application of dual-energy X-ray absorptiometry and related
techniques to the assessment of bone and body composition. In P. S. W. Davies and
T. J. Cole (eds.) Body Composition Techniques in Health and Disease. Cambridge
UK: Cambridge University Press.
Prentice, A. M., Diaz, E. O., Murgatroyd, P. R., Goldberg, G. R., Sonko, B. J., Black, A.
E., and Coward, W. A. 1991. Doubly labeled water measurements and calorimetry in
practice. In R. G. Whitehead and A. Prentice (eds.) New Techniques in Nutritional
Research. New York: Academic Press.
Ravussin, E. and Rising, R. 1992. Daily energy expenditure in humans: measurements in
a respiratory chamber and by doubly labeled water. In J. M. Kinney and H. N.
Tucker (eds.) Energy Metabolism: Tissue Determinants and Cellular Corollaries.
New York: Raven Press.
Ritz, P. and Coward, W. A. 1995. Doubly labeled water measurement of total energy
expenditure. Diabetes & Metabolism, 21: 241–51.
Roos, B. O. and Skoldborn, H. 1974. Dual photon absorptiometry in lumbar vertebrae. I.
theory and method. Acta Radiologica: Therapeutics, 13: 266–80.
Schoeller, D. A. and van Santen, E. 1982. Measurement of energy expenditure in humans
by doubly-labeled water method. J. Appl. Physiol., 53: 955–9.
Segal, K. R., Gultin, B., Presta, E., Wang, J., and van Itallie, T. B. 1985. Estimation of
human body composition by electrical impedance methods: a comparative study. J.
Appl. Physiol., 58: 1565–71.
Segal, K. R., Van Loan, M., Fitzgerald, P. I., Hodgdon, J. A., Van Itallie, T. B. 1988.
Lean body mass estimated by bioelectrical impedance analysis: a four-site cross-
validation study. Am. J. Clin. Nutr., 47: 7–14.
Segal, K. R., Burastero, S., Chun, A., Coronel, P., Pierson, R. N., and Wang, J. 1991.
Estimation of extracellular and total body water by multiple-frequency bioelectrical-
impedance measurement. Am. J. Clin. Nutr., 54: 26–9.
Siri, W. E. 1961. Body composition from fluid spaces and spaces, analysis, of methods.
In J. Brozek and A. Henschel (eds.) Measuring Body Composition. Washington DC:
National Academy Sciences, 223–44.
Tanner, J. M. and Whitehouse, R. H. 1955. The Harpenden skinfold caliper. Am. J. Phys.
Anthropol., 13: 743–6.
10.6 References 41
10.7 Problems
10.1 Calculate the value of the thermistor temperature coefficient α for T = 298 K
and β = 3000 K.
10.2 Design a circuit that converts the resistance change of a thermistor to a voltage,
and then amplifies it to measure temperature.
10.3 Explain the Seebeck effect of a thermocouple and why and how to use it for
temperature measurement.
10.4 Design a circuit that uses an electronic cold junction and thermocouple, then
amplifies it to measure temperature.
10.5 Explain the operation of a pyroelectric infrared thermometer, sketch its block
diagram, and sketch the response after the shutter opens.
10.6 Compare direct and indirect calorimetry.
10.7 Explain how to use the gradient layer calorimeter for heat measurement.
10.8 Figure 10.6 shows a simple gradient layer calorimeter. Sketch an expanded
gradient layer calorimeter that includes ventilation. Show location of all sensors
and explain how they work. Give a symbol for the output of each sensor.
Explain by means of equations, block diagrams and/or words how to obtain the
quantities given in equations (10.7) through (10.11) from your sensor outputs.
10.9 The oxygen consumption is 0.25 L/min. The carbon dioxide production is 0.06
L/min. Calculate the energy expenditure. Give units.
10.10 Define %BF and explain its purpose.
10.11 Explain how to measure %BF using densitometry.
42 Body Temperature, Heat, Fat, and Movement
10.12 Describe how to use the bioelectrical impedance method for body fat
measurement.
10.13 Draw the circuit for a resistance potentiometer goniometer with an amplifier
with a gain of 2
10.14 Draw the circuit for a 4-arm piezoresistive strain gage accelerometer and
amplifier.
10.15 Draw the circuit for a piezoelectric accelerometer and amplifier to pass
frequencies higher than 1 Hz.
10.16 You want to analyze human locomotion; list the devices you need and explain
why.