4 6003574465387038522 PDF
4 6003574465387038522 PDF
4 6003574465387038522 PDF
Analytical Chemistry is the branch of chemistry concerned with determining the qualitative and
quantitative composition of substances. The process of determining the property of a material
system is called analysis. The determination of the identity of chemical species is known as
qualitative analysis. If the identity of a chemical species is known, the concentration of that
species in the sample can be determined by quantitative analysis. The substances that are the
objects of analysis is called analyte while the material in which the analyte is found is called the
matrix of the analyte.
The methods followed while doing a chemical analysis of the sample can be classified as either a
classical (wet) method or an instrumental method. Classical method includes gravimetric
analysis and volumetric analysis, while instrumental methods include spectroscopic,
chromatographic and electroanalytical methods. It is interesting to contrast classical methods
with instrumental methods of analysis. Instrumental methods are usually faster and more
sensitive after the necessary calibration have been established. However, in many cases the
classical methods are more accurate even though much slower than instrumental methods.
Instrumental methods offer improved detection limits over classical methods.
Steps in Chemical Analysis
Selecting chemical analysis method (classical and instrumental) is only one of the several steps
in the total chemical analysis. Before analysis, sample must be collected in a process called
sampling. The obtained sample must undergo chemical or physical pretreatment. After
chemical analysis, the data from the analysis must be mathematically manipulated to provide a
meaningful result and the result must be clearly communicated to interested persons. If any of
the five steps (sampling, sample pretreatment, chemical analysis, data handling and
communication of results) in the total chemical analysis is not performed properly, the analytical
results cannot be expected to be correct.
1
Analytical chemistry helps diagnose parathyroid diseases in human’s measurements of
ionized calcium in blood serum.
Determination of nitrogen in foods establishes their protein content and thus their
nutritional value.
Analysis of steel during its production permits adjustment in the concentrations of such
elements as carbon, nickel, and chromium to achieve a desired strength, hardness,
corrosion resistance, and ductility.
Research in all science, biochemistry, biology, chemistry, geology, physics, agriculture,
health and other science.
Errors in Chemical Analysis
Errors in the analytical laboratory are basically of two types: determinate errors and
indeterminate errors. Determinate errors, also called systematic errors, are errors that
were known to have occurred, or at least were determined later to have occurred, in the
course of the lab work. They may arise from avoidable sources, such as contamination,
wrongly calibrated instruments, reagent impurities, instrumental malfunctions, poor sampling
techniques, errors in calculations, etc. Results from laboratory work in which avoidable
determinate errors are known to have occurred must be rejected or, if the error was a
calculation error, recalculated. Determinate errors may also arise from unavoidable sources.
An error that is known to have occurred but was unavoidable is called a bias. Such an error
occurs each time a procedure is executed, and thus its effect is usually known and a
correction factor can be applied. Determinate errors may be divided into four categories:
sampling errors, method errors, measurement errors, and personal errors.
Indeterminate errors, also called random errors, on the other hand, are errors that are not
specifically identified and are therefore impossible to avoid. Since the errors cannot be
specifically identified, results arising from such errors cannot be immediately rejected or
compensated for as in the case of determinate errors. Rather, a statistical analysis must be
performed to determine whether the results are far enough “off-track” to merit rejection.
Examples might include fluctuations in room temperature, fluctuations in line voltage,
mechanical vibrations, cosmic rays, etc.
Statistics establish quality limits for the answers derived from a given method. A given
laboratory result, or a sample giving rise to a given result, is considered “good” if it is within
these limits. In order to understand how these limits are established, and therefore how it is
known if a given result is unacceptable, some basic knowledge of statistics is needed.
2
Statistical Analysis of Data
The procedure used to determine whether a given result is unacceptable involves running a series
of identical tests on the same sample, using the same instrument or other piece of equipment,
over and over. In such a scenario, the indeterminate errors manifest themselves in values that
deviate, positively and negatively, from the mean (average) of all the values obtained. Given this
brief background, let us proceed to define some terms related to elementary statistics.
1. Mean
In the case in which a given measurement on a sample is repeated a number of times, the average
of all measurements is called the mean. It is calculated by adding together the numerical values
of all measurements (Xi) and dividing this sum by the number of measurements (n). In this text,
we give the mean the symbol m. The true mean, or the mean of an infinite number of
measurements (the entire population of measurements), is given the symbol µ, the Greek letter
mu.
Xi ,
m
n
2. Deviation
How much each measurement differs from the mean is an important number and is called the
deviation. A deviation is associated with each measurement, and if a given deviation is large
compared to others in a series of identical measurements, there is problem with the analysis.
Such a measurement is called an outlier. Mathematically, the deviation is calculated as follows:
d = |m – e|
in which d is the deviation, m is the mean, and e represents the individual experimental
measurement. (The bars refer to absolute value, which means the value of d is calculated without
regard to sign.)
3. Standard deviation
The most common measure of the dispersion of data around the mean is the standard deviation:
The term n is the number of measurements, and n – 1 is referred to as the number of degrees of
freedom. The term s represents the standard deviation. The significance of s is that the smaller it
is numerically, the more precise the data (the more the measurements are “bunched” around the
3
mean). For an infinite number of measurements (where the mean is µ), the standard deviation is
symbolized as σ (Greek letter sigma) and is known as the population standard deviation. An
infinite number of measurements is approximated by 30 or more measurements.
4. Relative standard deviation
One final deviation parameter is the relative standard deviation (RSD). It is calculated by
dividing the standard deviation by the mean and then multiplying by 100 or 1000:
4
5. Normal Distribution
For an infinite data set (in which the symbols µ and σ apply), a plot of frequency of occurrence
vs. the measurement value yields a smooth bell-shaped curve. It is referred to as bell-shaped
because there is equal drop-off on both sides of a peak value, resulting in a shape that resembles
a bell. The peak value corresponds to µ, the population means. This curve is called the normal
distribution curve because it represents a normal distribution of values for any infinitely
repeated measurement. The normal distribution curve is a picture of the precision of a given data
set. The more points there are bunched around the mean and the sharper the drop-off away from
the mean, the smaller the standard deviation and the more precise the data.
6. Variance (V)
The term that is sometimes useful in statistics is the variance (V). This is the square of the
standard deviation. The sample variance is given by
V = s2
Precision refers to the repeatability of a measurement. If you repeat a given measurement over
and over and these measurements deviate only slightly from one another, within the limits of the
number of significant figures obtainable, then we say that the data are precise, or that the results
exhibit a high degree of precision. The mean of such data may or may not represent the real
value of that parameter. In other words, it may not be accurate.
Accuracy refers to the correctness of a measurement, or how close it comes to the correct value
of a parameter. For example, if an analyst has an object that he or she knows weighs exactly
5
1.0000 g, the accuracy of a laboratory balance (weight measuring device) can be determined.
The object can be weighed on the balance to see if the balance will read 1.0000 g. If a series of
repeated weight measurements using this balance are all between 0.9998 and 1.0002 g, we say
the balance is both precise and accurate. If, on the other hand, a series of repeated weight
measurements using this balance are all between 0.9983 and 0.9987 g, we say that the balance is
precise, but not accurate. If repeated weight measurements using this balance are all between
0.9956 and 0.9991 g, the data are neither precise nor accurate. Finally, if repeated weight
measurements using this balance are all between 0.9956 and 1.0042 g, such that the mean is
1.0000 g, then the balance is not precise but it appears to be accurate. These facts on accuracy
and precision are illustrated further in the following Figure.
If it is established that a measuring device provides a value for a known sample that is in
agreement with the known value to within established limits of precision, that device is said to be
calibrated. Thus, calibration refers to a procedure that checks the device to confirm that it
provides the known value.
Confidence interval is range of results around a mean value that could be explained by random
error. In general, we can write
6
Table: Confidence Intervals for Normal Distribution Curves between the Limits m ± zs
Example: What is the 95% confidence interval for the amount of aspirin in a single analgesic
tablet drawn from a population where m is 250 mg and s2 is 25?
SOLUTION
According to the above Table, the 95% confidence interval for a single member of a normally
distributed population is
Thus, we expect that 95% of the tablets in the population contain between 240 and 260 mg of
aspirin.
Among the most commonly used statistical significance tests applied to small data sets
(population samples) is the series of Student's tests. One of these tests is used for the comparison
of two means, which is commonly applied to many cases. Typical examples are:
Example 1: Comparison of analytical results obtained with the same method on samples A and
B, in order to confirm whether both samples contain the same percentage of the measured
analyte or not.
Example 2: Comparison of analytical results obtained with two different methods A and B on the
same sample, in order to confirm whether both methods provide similar analytical results or not.
The outcome of these tests is the acceptance or rejection of the null hypothesis (H0). The null
hypothesis generally states that: "Any differences, discrepancies, or suspiciously outlying results
are purely due to random and not systematic errors". The alternative hypothesis (Ha) states
exactly the opposite.
7
The null hypothesis for the aforementioned examples is: The means are the same, i.e. in Example
1: both samples contain the same percentage of the analyte; in Example 2: both methods provide
the same analytical results. The differences observed (if any) are purely due to random errors.
The means are significantly different, i.e. in Example 1: each sample contains a different
percentage of the analyte; in Example 2: the methods provide different analytical results (so at
least one method yields systematic analytical errors).
To compare two means using student’s t-test, the two means and the corresponding standard
deviations are calculated by using the following equations (nA and nB are the number of
measurements in data set A and data set B, respectively):
nA nB
Xi Xi
XA
i 1 n A
and X B n
i 1 B
nA nB
( X A X i )2 (X B X i )2
SA i 1
and S B i 1
nA 1 nB 1
(n A 1) S A (nB 1) S B
2 2
S AB
n A nB 2
XA XB
texp
1 1
S AB
n A nB
texp value is compared with the critical (theoretical) tth value corresponding to the given degree
of freedom N (in the present case N = nA + nB - 2) and the confidence level chosen. Tables of
critical t values can be found in any book of statistical analysis, as well as in many quantitative
analysis textbooks. If texp > tth then H0 is rejected else H0 is retained.
An F-test is a simple calculation to compare the precision of two sets of measurement. An F-test
can provide insights into two main areas: 1) Is method A more precise than method B? 2) Is there
a difference in the precision of the two methods? To calculate an F-test, the standard deviation of
the method which is assumed to be more precise is placed in the denominator, while the standard
deviation of the method which is assumed to be least precise is placed in the numerator.
8
2
S1
F 2
S2
Comparing the calculated F with tabulated F, if the F-table value is smaller than the calculated
value, you can reject the null hypothesis.
(1) The N values comprising the set of observations under examination are arranged in ascending
(2) The statistic experimental Q-value (Qexp) is calculated. This is a ratio defined as the
difference of the suspect value from its nearest one divided by the range of the values (Q:
rejection quotient). Thus, for testing x1 or xN (as possible outliers) we use the following Qexp
values:
X 2 X1 X X N 1
Qexp or Qexp N
X N X1 X N X1
9
(3) The obtained Qexp value is compared to a critical Q-value (Qcrit) found in tables. This critical
value should correspond to the confidence level (CL) we have decided to run the test (usually:
CL=95%).
(4) If Qexp > Qcrit, then the suspect value can be characterized as an outlier and it can be rejected,
if not, the suspect value must be retained and used in all subsequent calculations.
Example: The following replicate observations were obtained during a measurement and they are
arranged in ascending order: 4.85, 6.18, 6.28, 6.49, 6.69. Can we reject observation 4.85 as an
outlier at a 95% confidence level?
Answer: Qexp = (6.18 - 4.85) / (6.69 - 4.85) = 0.722. Qexp > Qcrit at CL: 95% for N=5). Therefore
we can reject 4.85.
Aqueous solutions are solutions in which the solvent is water. For every solution, there is a limit
to the amount of solute that can be dissolved in the solvent. For example, if we add salt to water,
it will dissolve. If we add more salt to the water, it too will dissolve. However, at a certain point,
no more salt will dissolve and any additional salt will simply fall to the bottom of the container.
At this point, the solution is said to be saturated. A saturated solution is one which has the
maximum amount of solute dissolved in the solvent. A solution that has not reached the
maximum limit of dissolved solute is said to be unsaturated. The amount of solute dissolved in
the solvent can be referred qualitatively as concentrated if it has a large amount of solute. A
solution which has only a relatively small amount of dissolved solute is termed dilute.
10
Three common ways of expressing the relative amounts of solute to solvent are as follow:
Example: Suppose that 5.00 g calcium chloride is dissolved in 40.0 g water. Calculate the mole
and mass fractions of calcium chloride in the solution.
Solution:
11
Example: Calculate the mole and mass fractions of a solution made by dissolving 9.0 g ethanol
in enough water to make 200.0 g solution.
Solution:
Example: Calculate the molar concentration of a solution made by dissolving 42.3 g glucose in
250.00 mL water.
Solution: By defination, M n
V
Mole of glucose: Molar mass of glucose (C6H12O6): 6.(12.011) + 12.(1.008) + 6.(16.00) = 180.16
Example: What mass of sodium hydroxide (NaOH) is needed to make up 0.5 L of a 0.5M
solution? [Atomic mass: Na = 23, O = 16, H = 1].
12
Dilution
Dilution is the addition of solvent (or other substances) to a solution, thereby reducing the solute
concentration. Dilution is important in the chemical laboratory because of:
The first step involved when two or more solutions are mixed is to recalculate the
concentration(s) in the mixture. The following generic example will serve to illustrate the
procedure: Suppose that two solutions, A and B, are mixed. The volumes and concentrations of
A and B are (VA, [A]0), and (VB, [B]0) respectively. Note that the subscript “0” means here that
this is the concentration of the original solution.
13
Normality of Solution
C A xVA
VB
CB
If concentration given in percentage and specific density of the solute given, to calculate
10 xsdx%
molarity, M
M .mass
C A xVA
VB
CB
14