Fourier Optics in Image Processing - 24!09!28!16!48 - 07
Fourier Optics in Image Processing - 24!09!28!16!48 - 07
Fourier Optics in Image Processing - 24!09!28!16!48 - 07
Image Processing
SERIES IN OPTICS AND OPTOELECTRONICS
Series Editors: Robert G W Brown, University of
California, Irvine, USA
E Roy Pike, Kings College, London, UK
This international series includes books on all aspects of theoretical and applied
optics and optoelectronics. Titles in this series range in level from advanced text-
books for graduate students to research monographs on specialized subjects includ-
ing: Nonlinear Optics and Laser Radiation, Interferometry, Waves, Crystals and
Optical Materials. These titles will also meet the requirements of scientists and
engineers working in optics, optoelectronics and related fields in the industry.
Handbook of Solid-State Lighting and LEDs
Zhe Chuan Feng (Ed.)
Nanophotonics and Plasmonics: An Integrated View
Dr. Ching Eng (Jason) Png and Dr. Yuriy Akimov
Handbook of Optoelectronic Device Modeling and Simulation: Lasers,
Modulators, Photodetectors, Solar Cells, and Numerical Methods – Volume Two
Joachim Piprek (Ed.)
Handbook of Optoelectronic Device Modeling and Simulation:
Fundamentals, Materials, Nanostructures, LEDs, and Amplifiers – Volume One
Joachim Piprek (Ed.)
Handbook of GaN Semiconductor Materials and Devices
Wengang (Wayne) Bi, Hao-chung (Henry) Kuo, Pei-Cheng Ku,
and Bo Shen (Eds.)
Handbook of Optoelectronics, Second Edition: Applied Optical Electronics –
Volume Three
John P. Dakin and Robert Brown (Eds.)
Handbook of Optoelectronics, Second Edition: Enabling Technologies –
Volume Two
John P. Dakin and Robert Brown (Eds.)
Handbook of Optoelectronics, Second Edition: Concepts, Devices, and
Techniques – Volume One
John P. Dakin and Robert Brown (Eds.)
Optical Microring Resonators: Theory, Techniques, and Applications Vien Van
Thin-Film Optical Filters, Fifth Edition
H. Angus Macleod
Optical MEMS, Nanophotonics, and Their Applications
Guangya Zhou and Chengkuo Lee (Eds.)
By
Neil Collings
MATLAB® is a trademark of The MathWorks, Inc. and is used with permission. The MathWorks
does not warrant the accuracy of the text or exercises in this book. This book’s use or discussion
of MATLAB® software or related products does not constitute endorsement or sponsorship by The
MathWorks of a particular pedagogical approach or particular use of the MATLAB® software.
CRC Press
Taylor & Francis Group
6000 Broken Sound Parkway NW, Suite 300
Boca Raton, FL 33487-2742
© 2019 by Taylor & Francis Group, LLC
CRC Press is an imprint of Taylor & Francis Group, an Informa business
No claim to original U.S. Government works
Printed on acid-free paper
Version Date: 20180514
International Standard Book Number-13: 978-1-4987-6068-3 (Hardback)
This book contains information obtained from authentic and highly regarded sources. Reasonable
efforts have been made to publish reliable data and information, but the author and publisher cannot
assume responsibility for the validity of all materials or the consequences of their use. The authors and
publishers have attempted to trace the copyright holders of all material reproduced in this publication
and apologize to copyright holders if permission to publish in this form has not been obtained. If any
copyright material has not been acknowledged please write and let us know so we may rectify in any
future reprint.
Except as permitted under U.S. Copyright Law, no part of this book may be reprinted, reproduced,
transmitted, or utilized in any form by any electronic, mechanical, or other means, now known or
hereafter invented, including photocopying, microfilming, and recording, or in any information
storage or retrieval system, without written permission from the publishers.
For permission to photocopy or use material electronically from this work, please access
www.copyright.com (http://www.copyright.com/) or contact the Copyright Clearance Center, Inc.
(CCC), 222 Rosewood Drive, Danvers, MA 01923, 978-750-8400. CCC is a not-for-profit organization
that provides licenses and registration for a variety of users. For organizations that have been granted
a photocopy license by the CCC, a separate system of payment has been arranged.
Trademark Notice: Product or corporate names may be trademarks or registered trademarks, and
are used only for identification and explanation without intent to infringe.
Visit the Taylor & Francis Web site at
http://www.taylorandfrancis.com
and the CRC Press Web site at
http://www.crcpress.com
To my wife, Lydia, and my children. Ines, Ernest, and
Emma.
Contents
Preface xi
Acronyms xiii
Chapter 1 Introduction 1
1.1 COMPUTER VISION 1
1.2 TEMPLATE MATCHING AND FEATURE EXTRACTION 3
1.3 FOURIER OPTICS 4
1.3.1 Diffraction 5
1.3.2 Spatial frequency of a grating 8
1.3.3 Convolution and correlation 8
1.3.4 Fourier shift theorem 10
1.4 DIGITAL TECHNIQUES 10
1.5 TRADE-OFF WITH DIGITAL (NUMERICAL) APPROACHES 11
vii
viii Contents
6.8 3D CORRELATORS 95
Chapter 7 Filtering 99
7.1 SYNOPSIS 99
7.2 VECTORS AND MATRICES 101
7.3 RANDOM VARIABLES AND FUNCTIONS 104
7.4 HYPOTHESIS TESTING 108
7.5 FIGURES OF MERIT 111
7.5.1 Peak sharpness measures 112
7.5.2 Signal to noise ratio 112
7.5.3 Discrimination capability 113
7.6 FILTER CODING TECHNIQUES 113
7.6.1 Phase-only filtering 114
7.6.2 Binary phase-only filters 114
7.6.3 Wiener filter 115
7.6.4 Constrained filters 115
7.7 FILTERS BASED ON TRAINING IMAGES 117
7.7.1 SPR filters 118
7.7.2 SDF filters 119
7.7.3 Trade-off and unconstrained filter design 121
7.8 PHASE-ONLY CORRELATION 121
Bibliography 145
Index 173
Preface
xi
xii Preface
xiii
xiv Acronyms
Introduction
CONTENTS
1.1 Computer vision . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1
1.2 Template matching and feature extraction . . . . . . . . . . . . . . . . 3
1.3 Fourier optics . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4
1.3.1 Diffraction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5
1.3.2 Spatial frequency of a grating . . . . . . . . . . . . . . . . . . . . . 8
1.3.3 Convolution and correlation . . . . . . . . . . . . . . . . . . . . . . . 8
1.3.4 Fourier shift theorem . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10
1.4 Digital techniques . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10
1.5 Trade-off with digital (numerical) approaches . . . . . . . . . . . . . 11
1
2 Fourier Optics in Image Processing
1.3.1 Diffraction
An intuitive mathematical treatment of optical diffraction is that based
on the Huygens-Fresnel principle [90]. According to this principle, light
propagates as a wavefront and, in order to calculate successive positions
of the wavefront, it is sufficient to propagate wavelets from secondary
sources on the wavefront. Wavelets are spherical waves which are so-
lutions of the Helmholtz equation which governs the complex ampli-
tude of monochromatic optical disturbances propagating through free
space. The resulting wavefront in the output plane is the envelope of
the wavelets from these secondary sources:
1 eikR
Z
′
U (r ) = U (r) cos [θ(r)] ds, (1.2)
iλ R
S
12
z02 + (x′ − x)2 + (y ′ − y)2
R =
(x′ − x)2 + (y ′ − y)2
≃ z0 + (1.3)
2z0
(x + y ) (x2 + y 2 ) x′ x + y ′ y
′2 ′2
≃ z0 + + − (1.4)
2z0 2z0 z0
x′ x + y ′ y
≃ z0 − (1.5)
z0
Equation (1.4) is the approximation used for Fresnel diffraction where
6 Fourier Optics in Image Processing
R
y
θ
x
z
r’
r
the fourth power terms in the coordinate variables are ignored. Equa-
tion (1.5) is the approximation used for Fraunhofer diffraction where
the quadratic terms of the coordinate variables are ignored. Under the
Fresnel approximation, Equation (1.4),
(x′2 + y ′2 )
′ ′ 1
U (x , y , z0 ) = exp ik z0 + .
iλz0 2z0
2 (1.6)
(x + y 2 ) x′ x + y ′ y
Z Z
U (x, y, 0) exp ik − dxdy
2z0 z0
Fourier Transform
The formula for the propagation from the aperture plane to the output plane,
Equation (1.6), is the FT of the aperture function multiplied by a constant.
The two-dimensional FT is a transform between the spatial domain, x, y,
and the frequency domain, fx , fy . The frequency plane appears at the
output plane or the focal plane of the lens, where the spatial coordinates
are related to fx , fy by x′ = fx λz0 , y ′ = fy λz0 , and x′ = fx λf , y ′ = fy λf ,
respectively. In the case of the lens, the complex amplitude of the electric
field can be expressed as a function of two spatial coordinates, and the
transform, Û (fx , fy ), is a function of the two frequency coordinates
ZZ ∞
Û (fx , fy ) = U (x, y)exp[−2πi(fx x + fy y)] dxdy (1.8)
−∞
When propagating from the output plane back to the aperture plane, the
inverse transform is used. This is defined as follows
ZZ ∞
U (x, y) = Û (fx , fy )exp[2πi(fx x + fy y)] dfx dfy (1.9)
−∞
eikz0 (x − x′ )2 + (y − y ′ )2
ZZ
′ ′
U (x , y , z0 ) = U (x, y, 0) exp ik .
iλz0 2z0
dxdy
(1.12)
GLOSSARY
Feature extraction: Searching for pixel neighbourhood features
within the scene.
Fourier optics: The field of optics where the spatial Fourier trans-
forming aspect of optical diffraction is exploited.
Spatial frequency: The natural unit of the spatial Fourier series and
transform.
15
16 Fourier Optics in Image Processing
by the physical extension of the pixel. The total extent of the input,
U (x, y, 0), is then bounded by the aperture of the EASLM. This limi-
tation is known as apodisation. For a unit width rectangular aperture
on the x-axis, the notation rect(x) is used, with corresponding spectral
function, sinc(fx), where sinc(fx) = sin(πfx )/(πfx) (Figure 2.1).
Figure 2.2Labelling of the pixel sizes, pixel repeats, and total dimen-
sions of the EASLM
as follows Z ∞
δ(x − x0 )g(x) dx = g(x0 )
−∞
Z Z L−1
XM −1
′ ′ eikz0 X
U (x , y , z0 ) = U (x, y, 0)rect(x/L∆x).
iλz0
l=0 m=0
rect(y/M ∆y)rect(x/a)rect(y/b)δ(x − l∆x)δ(y − m∆y)) . (2.3)
′
x x + y′ y
exp ik − dxdy
z0
Due to the windowing effect of the full aperture rect functions,
rect(x/L∆x) and rect(y/M ∆y), the limited summation over the δ
functions in Equation (2.3) can be replaced by comb functions. The
Fourier transform of the spatial comb function is a spatial frequency
comb function, Equation (2.1).
The sampling converts the analogue Fourier transform into a dis-
crete Fourier transform (Section 1.4). The input image, U (x, y, 0), is a
LxM matrix of values, with individual elements U (l∆x, m∆y, 0). Its
Fourier transform is given by
L−1
XM −1
X
Û (fx , fy ) = U (l∆x, m∆y)exp [−2πi (fx l∆x + fy m∆y))]
l=0 m=0
(2.4)
Therefore,
eikz0
U (x′ , y ′ , z0 ) = [Û (fx , fy ) ⊗ comb(fx)comb(fy ))].
iλz0 (2.5)
sinc(fx a)sinc(fy b)
plane of the lens will increase. The quality of the lens is also critical for
maintaining an accurate representation of the spectral function in the
Fourier plane. Aberrations, in particular field curvature, will increase
the spot size in the Fourier plane, if they are not corrected. Correct
lens design, following the references cited in Section 1.3.1, is important.
In order to assess the quality of a given lens, the modulation transfer
function (MTF) is used. The MTF will be discussed in Chapter 4.
Glass
ITO layer
Alignment
layers
Liquid Crystal
Mirror electrode
Light blocking
metal layer
Interconnect
G metal layer
D S
Pixel transistor Storage capacitor
Silicon
as the planar aligned nematic (PAN) cell, which is no longer used for
microdisplay because of poor speed and uniformity, is attractive for
both amplitude and phase modulation. An alternative name for this
effect is the zero degree twist or electrically controlled birefringence
(ECB) cell. The birefringence modulation consequent on applying a
voltage to the cell generates a phase modulation effect when the polar-
isation of the light is aligned with the liquid crystal director, as used
in the Hamamatsu X13138 and JDC JD8714 devices. A phase modu-
lation of 2π can be achieved in a cell where the optical path difference
between the light reflected from a pixel in the zero voltage state and
one in the fully driven state is one wavelength. If it is desired to use
this effect for amplitude modulation, the polarisation of the light is
aligned at 45◦ to the liquid crystal director, and the thickness of the
LC layer is halved because the optical path difference required is only
one half wavelength for full amplitude modulation.
A second liquid crystal effect called the vertically aligned nematic
(VAN)is used predominantly in microdisplay because of the excellent
contrast between the black zero voltage state and the driven state.
This is also known as the tilted homeotropic (TH) mode. It is used in
the JVC LCOS for projector applications. The liquid crystal used in
such devices has a negative dielectric anisotropy where the long axis of
the molecule aligns perpendicular to the applied field. The availability
of good liquid crystal mixtures for this mode is more limited than
the more common positive dielectric anisotropy liquid crystals used
in the other modes. However, it has potential use in EASLMs if high
birefringence liquid crystals can be developed.
Spatial Light Modulators 27
The liquid crystal configuration of the two modes which have been
discussed does not twist within the cell. Therefore, when the electric
field driving the pixel is removed, the liquid crystal is slow to return
to the equilbrium state due to hydrodynamic flow in the liquid crystal
medium. Twisted liquid crystal modes benefit from a reduced flow and
the liquid crystal returns to its equilibrium state more quickly. Three
modes which employ twist are the 45◦ twist mode (also known as the
hybrid-field-effect (HFE) mode), the twisted nematic (TN) mode, and
the mixed twisted nematic (MTN) mode. The HFE mode has a 45◦
twist and an excellent tolerance to cell gap non-uniformity. This was
important in the early days of optically addressed spatial light modula-
tors (OASLMs) (Section 2.6) where the backplane had poor planarity
[23]. The TN mode is a 90◦ twist cell which is used in transmissive
devices. The polarisation vector of the incident light is aligned with
the director of the liquid crystal cell on the front surface of the cell in
both the HFE and TN modes. The MTN mode mixes the polarisation
rotation property of a TN mode with the birefringence modulation ef-
fect of the PAN mode. In common with the TN mode, there is a 90◦
twist, but the incident polarisation for reflective devices is at 20◦ to
the alignment of the front surface. The 90◦ twist produces a fast switch
speed in both the TN and MTN devices. Further speed improvement
can be achieved with the optically compensated bend (OCB) mode.
This can be viewed as a 180◦ twist cell because the alignment pre-tilt
on the opposing substrates of the cell is in the opposite sense (Figure
2.4). The idea for an OCB LCOS was presented in [279].
(a) (b)
Cross-section of (a) PAN cell; and (b) OCB cell, illustrating
Figure 2.4
opposite sense of alignment pre-tilt in the latter
tem by means of the liquid crystal layer. The speed of devices which
employ nematic liquid crystals is inferior to the electrically addressed
nematic devices. Higher speed reflective devices based on ferroelectric
liquid crystals have been extensively researched [181]. CRL Smectic
Technology marketed a ferroelectric liquid crystal OASLM called the
BOASLM (Binary OASLM) in the early 1990s. It was based on a hy-
drogenated amorphous silicon photoconductor and had a write time of
between 100 and 500 µs, with a resolution of greater than 50 lp/mm
(10% MTF). The write sensitivity was less than 30 µW/cm2 and the
active area was a 30 mm diameter circle.
The OASLMs used in the systems which will be described in Chap-
ter 6 are the LCLVs. Therefore, EASLMs will be referred to as SLMs
in the rest of the book, retaining LCLV for the OASLMs which have
been used.
GLOSSARY
Apodisation: The word literally means “removal of the feet”. It is
used to refer to the removal of the light amplitude by a hard aper-
ture, such as that outside of the reflecting pixellated surface of an
EASLM, or a soft aperture, such as, for example, that of the sinc
function.
Diffractive optical
elements
CONTENTS
3.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 35
3.2 Design of Diffractive Optical Elements . . . . . . . . . . . . . . . . . . . . 37
3.3 Fabrication . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 41
3.1 INTRODUCTION
In order to transform and filter the images within an optical process-
ing system, a useful tool is the diffractive optical element (DOE). The
DOE is a beam shaping optical element for a coherent light beam,
where beam shaping is achieved by significant modulation of the re-
fractive index of the element on the microscale. In contrast to refractive
optical elements such as a lens, or a prism, the modulation of refractive
index occurs over shorter spatial distances. Due to this high spatial fre-
quency modulation, individual elements of the image can be processed
differently to other elements, so that geometrical transformation of the
image can be accomplished [34]. This is known as space variant inter-
connection in contrast to the space invariant interconnection provided
by a lens or a bulk prism. DOEs are widely used in optical systems,
which require space variant connections, such as optical interconnects
and multichannel Fourier optical systems. Feature extraction and im-
age transforms with DOEs have been described in [222, 224, 225, 226].
The first hand-drawn diffraction element was a matched filter for
an optical correlator [33]. This was called a detour phase hologram be-
cause the phase information was encoded in the position of slits in each
amplitude modulated pixel. The encoding was effected by decentring
the slit by an amount corresponding to the magnitude of the phase.
35
36 Fourier Optics in Image Processing
The iterative algorithm has been used for a number of design prob-
lems, including beam shaping [266]. In particular design problems, the
original algorithm described by Gerchberg and Saxton [89] stagnates.
Stagnation is the inability to reduce the error below a floor which ex-
ceeds the desired error criterion. A large number of techniques have
been developed for avoiding stagnation in particular instances, but no
standardised procedure has emerged that is suitable for a wide range
of problems. A significant improvement was the development of three
more refined algorithms, the input-output [77], the output-output, and
the hybrid input-output algorithms [78]. The choice of one or other
of these algorithms is the prerogative of the designer. The totality of
these with other algorithms developed since are known as IFTAs (It-
erative Fourier Transform Algorithms). A significant improvement has
been the introduction of a noise dump area surrounding the DOE [88].
A further improvement is the introduction of additional terms into
the MSE criterion with adjustable parameters. These allow the DOE
designer to control the minimisation rate of the new criterion and par-
tially eliminate stagnation [143]. Improvements in this methodology
allow the optimisation of the uniformity and efficiency of the DOE at
the same time [134].
The result of the optimization algorithm is a design for the DOE
which is not unique. If the algorithm is repeated with a different ran-
dom phase seed, then a different design will result. The performance
(efficiency and MSE) of different designs will be comparable but dis-
tinct. A more computationally expensive design method which does
not stagnate is simulated annealing (SA) [136]. Starting from a ran-
dom phase distribution, the phase hologram is progressively refined
by making changes to each pixel, checking the MSE, and accepting
the change or not, depending on the “throw of a dice”, made at each
change. This algorithm is considered to give the best solution to the
design problem. The MSE is known as the cost function which can be
further elaborated according to any additional constraints required in
the optimisation, for example the efficiency.
The Gerchberg-Saxton algorithm generates a phase distribution
which, under illumination with a collimated coherent light beam, will
reproduce the target pattern in the far field. If it is desired that the
pattern is generated in the near field, then the transforms applied in
the iteration loop must be adapted accordingly. Fractional Fourier and
Fresnel transforms were applied to this problem in [287]. Neither of
these transforms possesses an inverse, so the propagation from the tar-
40 Fourier Optics in Image Processing
Evanescent waves
The plane wave waves which propagate to the output plane satisfy (fx2 +
fy2 )λ2 ≤ 1. If (fx2 + fy2 )λ2 > 1, then the wave amplitude diminishes as it
propagates along the z axis. This is an evanescent wave. These waves do
not contribute to the output plane, and therefore represent energy loss in
the transfer from the DOE to the output plane. They can be neglected in
scalar diffraction designs, but are more important in rigorous design where
the feature size is smaller and spatial frequencies are higher.
The rigorous design of DOEs is required when the feature size of the
DOE is reduced to the order of magnitude of the wavelength of the illu-
minating light beam. This is currently a research field, although there
Diffractive Optical Elements 41
3.3 FABRICATION
In respect of the construction of optical systems, the optical designer
is the poor cousin of the electronics design engineer. The availability
42 Fourier Optics in Image Processing
TIC GmbH offer RIE of the DOEs that they have produced by either
laser or e-beam lithography. Raith GmbH sells equipment for e-beam
lithography.
In semiconductor lithography systems, a considerable effort is
spent on eliminating the errors which arise during fabrication of high-
resolution structures over relatively large areas. Potential error sources
arise due to pattern distortion, misalignment, and substrate distortion.
Pattern distortion will be minimal when the address grid is finer than
the feature size. Misalignment is a significant problem at the multi-
ple exposure step for multilevel phase elements. Substrate distortion is
usually accommodated by replication onto higher optical quality sub-
strates. In order to reduce potential errors during the design phase, the
SA algorithm can be constrained to favour contiguous areas of similar
phase level in the DOE.
In order to realise low-cost optical systems, it is important to be
able to replicate the DOEs. A variety of embossing, moulding and cast-
ing techniques have been developed for optical elements. Embossing is
the process of pushing a seal into hot wax. A silicon or quartz seal
is called a stamp. Copies of silicon or quartz stamps can be made in
nickel, and the resulting elements are called nickel shims. Companies,
such as NIL Technology, and EV Group make stamps and shims for
nanoimprint lithography. CDA GmbH have teamed up with Holoeye
Photonics AG to offer a design and fabrication competence for emboss-
ing DOEs into either Polymethyl Methacrylate (PMMA), Polycarbon-
ate (PC), Cyclic Olefin Copolymer (COC), or Cyclic Olefin Polymer
(COP). Moulding describes the process of using fluid materials to fill
the space between the mould and a reference surface, so that, when
cool, the injected material takes the form of the mould. Nickel shims,
fabricated from the lithographically mastered DOE, are used for in-
jected moulding [192]. Companies, such as NIL Technology, Temicon,
and Rothtec will fabricate shims to customer specifications. Casting is
a higher quality process than either injection moulding or embossing.
The UV-curable or sol-gel materials are introduced between the mould
and a substrate, and subsequently cured by either UV light or thermal
curing. The replica formed by one of the above techniques is now a
low-cost component for the optical system, provided that the quantity
required is sufficient to justify the expense of replication.
Diffractive Optical Elements 45
GLOSSARY
Binary optics: The use of VLSI fabrication techniques such as pho-
tolithography and etching, to create high-resolution surface-relief
structures in optical substrates.
4.1 INTRODUCTION
A camera is composed of an image sensor with an associated optical
element. In consumer cameras, the optical element is an imaging lens
system, which images an object onto the image plane where the image
sensor is located. Cameras are used for image capture for the input to
Fourier optical systems. Image sensors are used at the output plane and
on intermediate image planes within the Fourier optical system. Signif-
icant data reduction within the system suggests that high-complexity
image sensors may not be optimal at the output of the system. In or-
der to understand the limitations of the imaging process, the optical
transfer function will be described in Section 4.2. This is followed by an
overview of image sensor technology, with a list of exemplar products
in the marketplace.
47
48 Fourier Optics in Image Processing
aI1 (xi , yi ) + bI2 (xi , yi ) 7−→ aO1 (xo , yo ) + bO2 (xo , yo ) (4.1)
where O1 , O2 (xo , yo ) are the outputs corresponding to inputs
I1 , I2 (xi , yi ), and a, b are constants.
An important consequence of this linearity is that the output can
be written as a summation of the inputs each multiplied by a constant
which is a function of the coordinates of the input, output and the
wavelength of the light, λ,
+∞
ZZ
O(xo , yo ) = W (xo , yo ; xi , yi ; λ)I(xi , yi )dxi dyi , (4.2)
−∞
W = WA (xo − xi , yo − yi ; λ) (4.3)
where the subscript A refers to the area or patch over which the system
is spatially invariant. The condition of spatial invariance is known as
the isoplanatic condition, and the patch is known as the isoplanatic
patch. Although the isoplanatic condition applies over a limited area
only, the integral equation is used with the proviso that W varies ac-
cording to location in the input/output planes.
When the input is a monochromatic point source, the spread func-
tion is known as the point spread function (PSF), h(xi − xo , yi − yo )
[90] and
+∞
ZZ
O(xo , yo ) = h(xo − xi , yo − yi )I(xi , yi )dxi dyi (4.4)
−∞
ˆ x , fy ) is the FT of I(xi , yi ),
where Ô(fx, fy ) is the FT of O(xo , yo ), I(f
and H(fx , fy ) is the FT of h(xo − xi , yo − yi ). H(fx , fy ) is called the
optical transfer function (OTF).
The transfer function approach is valid for an ideal isoplanatic opti-
cal system. This ideal system is one where the translation of an object
point in the object plane produces a proportional translation of the im-
age point in the image plane. The diffraction-limited OTF is the trans-
fer function for an ideal imaging system which is free from aberrations.
The shape of the diffraction-limited OTF as a function of frequency
is determined by the limiting aperture of the system, for example, the
circular frame of the lens. This has been calculated for both square
and circular limiting apertures [90]. It decreases monotonically from
a maximum at zero spatial frequency to zero at a cut-off frequency,
which is equal to the reciprocal of the product of the f-number of the
lens and the wavelength of the light. The f-number is the focal length
of the lens divided by the lens aperture. Higher f-number lenses have
lower cut-off frequencies.
The diffraction-limited OTF is an important metric for evaluating
the performance of camera lenses. The OTF of camera lenses is, in
general, a complex quantity, since it includes the aberration of the lens.
Therefore, in order to compare it with the diffraction-limited OTF of
the limiting aperture of the imaging system, the modulus of the OTF of
the lens is used. This is called the modulation transfer function (MTF).
The MTF is measured using groups of Ronchi gratings, each grating
of a different spatial frequency. When the lens is used to make a unity
magnification image of these gratings, the contrast in the image of each
grating is reduced in comparison with the object grating. The contrast
is the ratio of the local maximum to minimum intensity (Imax /Imin ).
The MTF is the ratio of the difference between Imax and Imin in the
image plane, normalised with respect to the difference between Imax
and Imin in the object plane. The orientation of the lines for lens testing
is usually perpendicular to two orthogonal axes, as in the 1951 U.S. Air
Force (USAF) test chart. The resolution of the lens is described by its
50% MTF which is the spatial frequency at which the MTF is 50%
of its low frequency value. Companies such as Image Science Ltd offer
equipment for measuring the MTF of lenses.
Point source inputs are an alternative to Ronchi gratings. The re-
sults of point source testing are of interest in the case of lenses for
astronomical telescopes because distant stars behave as point sources.
In this case, the squared magnitude of the PSF is used for the assess-
50 Fourier Optics in Image Processing
and lines for image sensor testing. EASLMs have a larger pixel repeat
than image sensors. Therefore, the MTF of an amplitude modulating
EASLM can be measured directly by a good quality camera employing
a high quality lens. Phase modulating EASLMs are measured on the
basis of diffraction efficiency. The diffraction efficiency of a grating is
the power of light diffracted into the first diffraction order normalised
by the power of light in the incident beam. The diffraction from a
Ronchi grating of four pixel period on an EASLM illustrates the first
order diffraction (Figure 4.1). The most significant spot is the zero or-
der corresponding to light which has not been diffracted. The diffracted
spot in the centre of the top of the image is due to the inherent pixel-
lation of the EASLM. There are corresponding spots on the left and
right extremes of the image. The first diffraction orders lie on either
side of the zero order. They are situated at one quarter of the distance
to the diffracted spots due to the pixellation, since there are four pixels
per period of the grating. They are known as the +1 and -1 orders. The
second order diffraction spots are missing since the grating is symmet-
ric. The third orders show faintly because their intensity is 1/9th of the
first order intensity. Finally, the intensity of the pixellated diffraction
in the horizontal direction is reduced because it coincides with an even
harmonic of the grating period, which is suppressed due to the symme-
try of the grating. In order to measure the line response function of an
image sensor, a rectangular slit is imaged onto the device. Due to the
small pixel repeat of image sensors, the diffraction spreading of the slit
imaging system is included in the response function. This is removed
by deconvolution (Section 1.3.3).
When an image is sampled by an EASLM or an image sensor,
there is an upper limit to the spatial frequency representation, which is
(2∆x)−1 , where ∆x is the pixel repeat in the x-direction. If the spatial
frequency content of the image exceeds this, then aliasing results. Alias-
ing is the confusion between the actual pattern and a similar pattern
of lower resolution. In the time domain, this is the familiar example
of sampling a spinning wheel with a lower frequency strobe light. The
wheel appears to move slowly and may be observed rotating in the op-
posite sense. In Fourier optics, a comparable example is the use of an
SLM to represent a lens function in an optical system. For short focal
length lenses where the optical path changes rapidly at the periphery
of the lens, the pixel repeat of the SLM can be insufficient. The effect of
this is that the SLM samples the lens structure and this has the effect
of replicating the lens in the image plane of the encoded lens [40]. In or-
52 Fourier Optics in Image Processing
U(fx,0)
(a)
fx
-B B
U(fx,0)
(b)
fx
(c)
fx
-fs2 fs2
Figure 4.2 The real part of the x-component of the spatial frequency
spectrum of an image:(a) unsampled; (b) sampled at spatial frequency
fs1 ; (c) sampled at spatial frequency fs2
Sampling
The rule-of-thumb that the sampling frequency of an image should be at
least twice the highest spatial frequency of the image is sometimes referred
to as the Nyquist criterion [84] or the Whittaker-Shannon theorem [90]. A
1D spatial frequency spectrum of the real values of the Fourier transform
of an image is shown in Figure 4.2(a). When this spectrum is sampled,
this results in the appearance of higher orders, which are mentioned in
Section 2.1.1 and illustrated as the diffracted spots due to the pixellation
Transfer Functions and Cameras 53
in Figure 4.1. When the sampling frequency, fs1 , is greater than 2B, where
B is the maximum frequency of the spectrum of Figure 4.2(a), the higher
orders are separate from the original spatial frequency spectrum (Figure
4.2(b)). The original object can be reconstructed accurately by low pass
filtering this replicated spectrum. However, when the sampling frequency,
fs2 , is less than 2B, the higher orders are not clearly separated from the
original spatial frequency spectrum (Figure 4.2(c)). In the border region
between the original spectrum and the higher order, it is uncertain whether
the frequency value corresponds to a frequency in the original spectrum
or in the higher order. This is the aliasing due to an insufficient sampling
frequency.
FTF7046M are full frame devices and the KAI-29050 is interline trans-
fer (Table 4.1). The full frame CCD clocks the photocharge line by
line to a shift register. Consequently, there is charge smearing dur-
ing the transfer from the light-sensitive to the masked regions of the
CCD. It is preferable to use a light shutter with the full frame de-
vice. These devices offer 100% Fill Factor but use less silicon so that
they are less expensive. The frame transfer CCD has the sensitivity
of the full frame device but it is typically more expensive due to the
larger sensor size needed to accommodate the frame storage region.
Although the frame storage region is protected from incident light,
there is still smearing during the readout time. However, the readout
time is shorter than for the full frame CCD, because the full sensor
is readout at once rather than line by line. Interline devices have a
charge storage area next to every pixel. This effectively reduces the
light sensitive area of the sensor, so that the Fill Factor is reduced.
This can be partially compensated by the use of microlens arrays to
focus the received light on the photosensor area. The compensation
usually works best for parallel light illumination but for some applica-
tions which need wide angle illumination the sensitivity is significantly
compromised. The interline-transfer CCD incorporates charge trans-
fer channels called Interline Masks. These are immediately adjacent to
each photodiode so that the accumulated charge can be rapidly shifted
into the channels after image acquisition has been completed. The very
rapid image acquisition virtually eliminates image smear.
Despite its proud history, several disadvantages of the CCD have
been appreciated. In order to display images on analogue monitors,
the early CCDs were equipped with interlaced scanning. Progressive
scanning is the technique of choice for coupling into EASLMs and PCs.
Frame transfer and full frame CCDs are inherently progressive scan,
but interline transfer CCDs can be either progressive or interlaced. A
further disadvantage is the limited readout speed, due to the serial
output data stream. In Table 6.1, the speeds of the CCD sensors are
quoted in terms of the serial output data rate. Further disadvantages
of CCD sensors are blooming and smearing. Blooming occurs when the
photogenerated charges exceed the capacity of the pixel and spill over
into adjacent pixels. Smearing is the generation of additional charges
during the readout process, which is an issue in the full frame devices.
The mainstream image sensor nowadays is the CMOS array.
The invention of MOS image sensors preceded that of the CCD
[185], but the lower fixed-pattern noise of CCDs gained the day. Ad-
Transfer Functions and Cameras 55
CMOSIS CMV50000 CMOS 36.4 x 27.6 mm 7920x6004 4.6 8.5e 14,000e 30 fps
CMOSIS Nanoeye GS CMOS 3.4 x 3.4 mm 640x640 3.6 16e 16,000e 100 fps
Dalsa FTF7046M CCD 36 x 24 mm 6936x4616 5.2 11e 40,000e 1.2 fps
e2v EV75C776 CMOS 11.5 x 11.5 mm 4096x4096 2.8 2e 7,000e 45 fps
Fairchild CIS1910A sCMOS2.0 12.5 x 7 mm 1920x1080 6.5 4e 30,000e 50 fps
Fairchild LTN4625A sCMOS2.0 25.3 x 14.3 mm 4608x2592 5.5 5e 40,000e 120 fps
Nikon DS-Qi2 CMOS 36 x 23.9 mm 4908x3264 7.3 2.2e 60,000e 6 fps
ON semi. KAF-1603 CCD 13.8 x 9.2 mm 1536x1024 9 15e 100,000e 10 MHz
ON semi. KAI-29050 CCD 36.2 x 24.1 mm 6644x4408 5.5 12e 20,000e 40 MHz
ON semi. LUPA1300-2 CMOS 1.4 ′′ diag 1280x1024 14 37e 30,000e 500 fps
Sony IMX253 CMOS 17.6 mm diag 4112x3008 3.45 3e 10,000e 46.4 fps
Transfer Functions and Cameras 57
58 Fourier Optics in Image Processing
high quantum efficiency, high resolution, and a large field of view si-
multaneously in one image. This makes them particularly suitable for
high fidelity, quantitative scientific measurement in low-light-level con-
ditions. With regard to potential applications in Fourier optical sys-
tems, the high sensitivity of the CCD camera would make it appropri-
ate for scene capture. The flexibility of the CMOS camera is important
for processing within and at the output of the system, where, in addi-
tion, the provision of an anti-reflection (AR) coating, which prevents
reflection within the system, is important. The application can also be
influenced by the readout mode for the camera: either rolling or global
shuttering. The rolling shutter allows the readout of columns of pixels
to be staggered, which offers lower noise and higher frame rate than
global shutter. However, this restricts the image capture application
to static or slow moving objects. Therefore, the performance data for
CMOS sensors in Table 4.1 refers to global shuttering, where the whole
frame is captured at the same time.
An alternative to the CCD and CMOS sensors is the position sen-
sitive detector (PSD) and the profile sensor (PS). The PSD locates the
beam position on a 2D photodiode surface by locating the electrodes
at the periphery of the surface [250]. An example is the S1880 sensor
from Hamamatsu. The PS is a photodiode array which is specialised for
peak detection [174]. Each pixel is divided into two sub-pixels, called
X and Y. All the X-pixels in a row are connected together, and all the
Y-pixels in a column are connected together. A scan of the rows gives
the X location of the peak intensity, and a scan of the columns gives
the Y location of the peak intensity. The speed of the PS sensor is high,
800 fps for the 512x512 pixel version based on a 10 bit ADC.
4.4 CAMERAS
Modern cameras deliver a very homogeneous and stable dark image
which is beneficial especially in low light applications. The speed of
simple camera processing operations has been improved by the use
of FPGA based frame grabber cards. The low-voltage differential sig-
nalling (LVDS) interface, which is supported by many image sensor
manufacturers in their products, specifies the electrical characteristics
of a differential, serial communications protocol, to enable camera de-
velopers to route the data output from a CMOS image sensor directly
into an FPGA. Each pair of such signals enables data to be transferred
from the image sensor to the FPGA at rates at about 600 Mbit/sec.
Transfer Functions and Cameras 59
Separate clock sources are used to enable the FPGA to accurately re-
cover synchronized data from the imager. LVDS channels have a low
susceptibility to noise because sources of noise add the same amount of
common-mode voltage to both lines in a signal pair. The use of LVDS
interface on sensors, in general, reduces power consumption, which is
important when the power budget is critical.
Sony have developed a new standard called Scalable Low Voltage
Signalling with an Embedded Clock (SLVS-EC) interface, which allows
a high-speed interface with a lower pin count. As a result, the cost of
building cameras with high resolution and high data rates will be lower
than with the LVDS interface. The SLVS technique is based on a point-
to-point signalling method that has evolved from the traditional LVDS
standard. SLVS also has a low susceptibility to noise, but because the
specification calls for smaller voltage swings and a lower common-mode
voltage, lower power drive circuitry can be used to drive the interface.
In the SLVS-EC design, the clock signal is embedded in the data from
the imager and recovered by dedicated circuitry on the FPGA. Hence
the data can be transmitted at much higher data rates and over much
further distances. Specifically, each of the channels on the SLVS-EC
interface can support effective data rates of over 2 Gbit/sec.
In order to realise low-cost optical systems, it will be important
to fabricate cameras at low cost. Wafer scale camera technology is
aimed at fabricating the image sensor and associated lens in volume
by processing 8 inch diameter silicon wafers and multilayer wafer-scale
lens assemblies in mask aligner systems. The complete cameras can
be subsequently diced and separated. The Omnivision OVM7690 is a
single chip image sensor, embedded processor and wafer-level optics in
one compact, small-profile package of 2.5 x 2.9 x 2.5 mm. The CMOS
image sensor (CMOSIS) NanEye module has a camera volume of only
1.5 mm3 and a low power dissipation, which is important for remote
systems. Details of the larger NanEye GS are included in Table 4.1. For
most measurement applications, it is very convenient to have a linear
response of the output signal to changes in light input. Deviations from
such a linear behavior are addressed as non-linearity. Cameras employ
the necessary electronic components and firmware algorithms to correct
the non-linear behavior in real time providing for linearity better than
99%.
60 Fourier Optics in Image Processing
GLOSSARY
Airy pattern or disc: This is the diffraction pattern of a circular
aperture, consisting of a central lobe surrounded by a sequence of
rings of lower intensity.
Read noise: Noise associated with all circuitry that measures and
converts the voltage on a pixel into an output signal.
CHAPTER 5
Light sources
CONTENTS
5.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 61
5.2 Coherence . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 63
5.2.1 Double-slit experiment . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 65
5.3 Laser diode optics . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 68
5.4 Vertical cavity surface emitting laser . . . . . . . . . . . . . . . . . . . . . . 68
5.5 Light emitting diode . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 69
5.1 INTRODUCTION
The principal Fourier optical system architectures have been demon-
strated using high coherence length gas lasers. In the interest of build-
ing compact, low-cost systems, the sources of choice are the semicon-
ductor diode laser (DL) and light emitting diodes (LEDs). The DL was
invented in 1962 by Robert Hall of General Electric Corporation. His
team assembled a device from pieces of p- and n-type gallium arsenide.
When these are polished and contacted, and one face of the sandwich
is polished, the p-n junction is ready for current injection. If the p-type
material is the anode and the n-type the cathode, then current can be
injected into the p-n junction diode with low resistance. Coherent laser
radiation is emitted from the polished face of the p-n junction when
the injected current exceeds a threshold level. Hall’s device required a
10 amp current at liquid nitrogen temperatures. The current was de-
livered in a 1 to 10 microsecond pulse, in order to avoid over-heating
the sandwich. Important advances in both the material and architec-
ture of the laser have been made in the last 50 years. The threshold
current has been reduced to sub-milliamp levels in laboratory devices,
although tens of milliamps is more customary for commercial devices.
In order to access applications spanning blu-ray disc readers and au-
tomobile headlamps to fibre optic communication, a wavelength range
61
62 Fourier Optics in Image Processing
2x2 2y 2
I(x, y, z) = I(0, 0, z)exp − + , (5.1)
ωx (z) ωy (z)
where ωx (z), ωy (z) are the Gaussian beam radii in the x- and y-
directions at a distance z from the beam waist. The beam waist of the
diode laser, where the Gaussian beam radii take the values ωx0 and
ωy0 , is located at the diode. The Gaussian beam radius is the radius
at which the intensity of the beam has decreased to e12 of its on-axis
value. Due to the small aperture of emission, the radiation is emitted
in a diverging beam of, typically, 7 degrees parallel to the junction, and
12 degrees perpendicular to the junction. The divergence angle here is
the angle subtended between the axis of the beam and the Gaussian
beam radius. The divergence angles can be expressed in terms of the
Gaussian beam radii at the waist
Light Sources 63
λ λ
θk = , θ⊥ = . (5.2)
πωx0 πωy0
Therefore, the brightness of a laser of power, P , and wavelength,
λ, with a diffraction limited output beam is given by
P P
B= = 2. (5.3)
πωx0 ωy0 πθk θ⊥ λ
where (5.2) has been used. The units for the brightness are W sr −1 m−2 .
The centre wavelength of the laser is an important parameter be-
cause this scales the size of the FT. There is a ±0.5% to 1.5% vari-
ability of the centre wavelength of high power laser diodes, due to
fabrication tolerances. If the temperature of the laser mounting is al-
lowed to increase to around 60 deg. C, this can result in an additional
+1% increase in the centre wavelength of the emission. Moreover, the
light emission is composed of a range of wavelengths around the centre
wavelength, and this is the spectral width of the emission, or linewidth.
This is quantified by the full width half maximum (FWHM), which is
the width of the spectral distribution at the intensity which is one half
the maximum intensity. The FWHM of diode lasers is, typically, a few
nanometres.
The LED was the pre-cursor of the DL. The story of the LED ex-
tends back to 1907, when Round discovered electroluminescence [292].
The first commercial red LED was made at Ferranti’s Photon Devices
Group [202]. Improvements in the LED have led to applications in do-
mestic and automobile lighting. Although the brightness of an LED
is orders of magnitude less than that of the DL and the linewidth is
about an order of magnitude larger, this disadvantage is offset by the
much lower price and the longer lifetime. Future application in Fourier
optical systems will require careful design in order to comply with the
coherence requirements of the system.
5.2 COHERENCE
In order to describe light propagation using wavefronts, the vibrations
of the electric field at two neighbouring points in space and time must
be correlated. Coherence is a measure of this correlation in both the
spatial and temporal dimensions. A fully coherent beam is one where
correlation extends over space and time, for example a laser beam. A
wavefront links points in space where the electric field vectors vibrate
64 Fourier Optics in Image Processing
devices and elements used in the optical system. The degree of coher-
ence of a diode laser source depends on both the size of the diode and
the quality of the collimation optics.
In order to represent the correlation between the amplitude and
phase of the light field at two locations in space, the mutual coherence
function, Γmn (τ ) is used, where
Γmn (τ )
γmn (τ ) = p p , (5.5)
Γmm (0) Γnn (0)
and the absolute value of γmn (0) is the value which was measured in
the fringe visibility experiment.
Finally, the mutual intensity, Jmn is equal to Γmn (0).
′
eikz0 xx
Z
′
U (x , z0 ) = U (x, 0)exp ik − dx. (5.7)
iλz0 z0
Using the Fourier shift theorem (Section 1.3.4)
eikz0
U (x′ , z0 ) = (sinc(fx)e−iπfx D + sinc(fx )eiπfx D ) (5.8)
iλz0
eikz0
U (x′ , z0 ) = 2sinc(fx)cos(πfxD), (5.9)
iλz0
where x′ = fx λz0 .
Bright fringe
U(x,z)
Bright fringe
Bright fringe
Bright fringe
Bright fringe
Double slit Lens Screen (U(x’,f))
3D reflectors [91, 123, 216], aplanatic designs [283], and many others.
The main emphasis has been on maximising flux transfer rather than
the coherence of the light output. Superluminescent LEDs (SLEDs)
have a small source size, but the power is low and the cost is high, so
they are currently not suitable for the systems discussed here.
GLOSSARY
Coherence length: The distance along a ray over which a defined
phase relationship between the complex field amplitudes is main-
tained.
Optical correlators
CONTENTS
6.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 71
6.2 System analysis . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 72
6.2.1 Geometric optics . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 72
6.2.2 Diffraction integrals . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 74
6.3 Holographic recording materials . . . . . . . . . . . . . . . . . . . . . . . . . . . 78
6.4 The Vander Lugt correlator . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 83
6.4.1 Background . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 83
6.4.2 Original VLC . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 84
6.5 The joint transform correlator . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 86
6.6 Hybrid correlators . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 89
6.7 Correlator systems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 93
6.8 3D Correlators . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 95
6.1 INTRODUCTION
The primary application of Fourier optics in image processing is the
optical correlator system. The system design involves the calculations
of both ray optics and diffraction integrals. This is the subject of Sec-
tion 6.2. In order to construct a correlator system, an essential element
was the complex spatial filter. Vander Lugt developed a technique for
recording a complex spatial filter in 1966 and constructed the first cor-
relator based on this in the same year [264]. The spatial filter in this
correlator was recorded on a holographic recording material, which
had to be processed and replaced in the system after processing. Holo-
graphic recording materials are discussed in Section 6.3. The Vander
Lugt correlator (VLC) architecture and its variants are discussed in
Section 6.4. In order to avoid the high precision required in the align-
ment of the spatial filter in the VLC, a novel correlator architecture
was proposed and demonstrated [280]. This became known as the joint
71
72 Fourier Optics in Image Processing
" #
1 − fd1 d
A B 1 0 1 d 1 0
= 1 =
C D − f2
1 0 1 − f11 1 − f12 + f1df2 − 1
f1 1 − fd2
f1 f2
f= (6.3)
f1 + f2 − d
This formula allows the construction of correlators employing two
SLMs, where the focal length required does not correspond to that
available from stock optics. These systems are discussed in Section 6.6.
When neither an achromat nor a combination of lenses suffice, the
Fourier lens should be designed from scratch. There are a number of
papers which indicate the approach to be adopted, some of which were
mentioned in Section 1.3.1. The papers presented provide the expe-
rience of past workers who were concerned with the design of rela-
tively long focal length Fourier lenses. The clearest exposition of the
design problem is given in [22], where three-lens and four-lens designs
are presented. These designs give a small amount of negative (barrel)
distortion. This is not important in systems where the Fourier plane
74 Fourier Optics in Image Processing
Q[−c]; the scaling by a constant, V [b], and its inverse, V [1/b]; the
Fourier transformation, F [U (x)], and its inverse, F −1 [Û (fx )]; and free-
space propagation, R[d], and its inverse, R[−d] [90]. They are defined
as follows:
ikcx2
Q[c](U (x)) = e 2 U (x); (6.4)
p
V [b](U (x)) = |b|U (bx); (6.5)
Z ∞
F [U (x)] = U (x)exp[−2πifx x]dx; (6.6)
−∞
and ∞
1
Z
ik ′ 2
R[d][U (x)] = √ U (x)e 2d (x−x ) dx. (6.7)
iλd −∞
U(x,y) U(x,y)
(a) (b)
Therefore, the Fourier transform of U (x) appears in the bfp of the lens,
multiplied by a phase factor which is quadratic in the x′ coordinate.
This phase factor is not important if the focal plane is the last plane
in the optical system, e.g., where an image sensor is placed. However,
if it is an intermediate plane, then this phase factor can influence the
quality of the final stage of the system. When the input is placed in
the ffp of the lens, then this phase factor disappears. The latter is the
recommended arrangement for the first and intermediate stages of a
Fourier optical system.
Vander Lugt based his operational notation on Fresnel functions
which he termed ψ functions. Their use can be illustrated in the case
of the converging beam arrangement (Figure 6.1b). The ψ function is
defined as
ik(x2 +y 2 )
ψ(x, y; 1/d) = e 2d . (6.9)
Therefore, a lens of focal length, f , is represented by
ik(x2 +y 2 )
ψ(x, y; −1/f ) = e− 2f . (6.10)
and
ik(xx′ +yy ′ )
ψ(x − x′ , y − y ′ ; 1/d) = ψ(x, y; 1/d)ψ(x′ , y ′ ; 1/d)e− d . (6.12)
Optical Correlators 77
Recording Interrogation
(write) (read)
BAR*
R
q
A B (a)
BA*R
A RA*B
q
R (b)
B
RAB*
This consists of three beams, the first two terms form an on-axis beam,
and the third and fourth terms give off-axis beams at angles θ and −θ
to the normal to the HRM (Figure 6.2a).
When two signal beams (A) and (B) are incident at angles θ and
−θ to the normal to the HRM (Figure 6.2b), the resulting interference
pattern is
|U (x, y)|2 = |A(x, y)eirx + B(x, y)e−irx|2 = |A(x, y)|2 + |B(x, y)|2
+ A(x, y)B ∗ (x, y)e2irx + A∗ (x, y)B(x, y)e−2irx
(6.18)
R|U (x, y)|2 = R|A(x, y)|2 +R|B(x, y)|2 + RA(x, y)B ∗ (x, y)e2irx
+ RA∗ (x, y)B(x, y)e−2irx
(6.19)
This consists of three beams, the first two terms form an on-axis beam,
and the third and fourth terms give off-axis beams at angles 2θ and
−2θ to the normal to the HRM (Figure 6.2b).
The development of HRMs has been similar to SLMs in a major
respect, they have been developed for a different application area, in
this case holographic memory. The resolution of each of the preced-
ing HRMs is higher than the wavelength of visible light. This allows
high density recording and a good range of angles for the recording
beams. Moreover, an enhanced diffraction efficiency is achieved when
the depth of the recording is greater than approximately λ/sin2 θ. Un-
der this condition, the grating is considered “thick” [242]. In addition to
high diffraction efficiency, the angular sensitivity is high, so that small
deviations of the angle of the illumination beam reduce the diffraction
efficiency considerably. This makes the HRM suitable for high density,
volume holographic recording, because a large number of holograms
can be recorded with a very small crosstalk. However, when the same
materials are used for Fourier plane filtering in an optical correlator,
this angular sensitivity results in a decrease in intensity of the corre-
lation beam as the object is displaced to different spatial positions in
the input scene.
It was demonstrated in Section 1.3.3 that the correlation signal is
displaced in proportion to the displacement of the input image. This
mathematical feature of the correlation integral is known as shift in-
variance. The ideal Fourier plane filter would maintain the intensity
of the correlation signal as the target moves across the input plane.
This is a property of a space invariant interconnection, as discussed in
Chapter 3. One of the factors which can limit the space invariance is
the depth of the recording. When the depth of the recording is greater
than approximately λ/sin2 θ, the grating is considered “thick” [242]. In
this case, the angular sensitivity is high, so that small deviations of the
illumination beam reduce the diffraction efficiency considerably. Thick
gratings produce a correlation signal intensity which is dependent on
Optical Correlators 83
the object position [66]. The interconnection between the input plane
and the correlation plane now becomes spatially variant in that the
transfer function changes as the object is displaced across the input
plane. Some workers, however, prefer to distinguish between this effect
and spatial variance due to system level limitations [31].
BS L2 HSF
SLM1 L3
L1
D
HSF
SLM1
BS L2
L1
Laser diode
BS L2 Image sensor
SLM
L1
Laser diode
SLM
L1 L4
The displacements of the input image and filter from the optical axis
produce linear phase ramps on the FTs of these distributions in the
Fourier plane. This is a consequence of the Fourier shift theorem (Sec-
tion 1.3.4), where a phase ramp of r in the input plane produced a
displacement of −rz0 /k in the FT. Here, a displacement of −H∆x/2
in the input plane produces a phase ramp of H∆xk/2z0 in the FT.
Similarly, a displacement of H∆x/2 produces the opposite phase ramp
of −H∆xk/2z0 in the FT. The intensity distribution in the plane of
the image sensor or LCLV will be
1 ˆ 2 + |F̂ |2 + IˆF̂ ∗ exp(−2πifx H∆x)
|U (x′ , y ′ , z0 )|2 = [|I|
(λz0 )2 (6.23)
+Iˆ∗ F̂ exp(2πifx H∆x)]
generally of larger SBWP than the filter image, the zero order in the
JTC is twice the extent of the zero order in the VLC. This entails an
adequate separation of the input and filter images in the input plane,
which impacts the SBWP available for the input image.
The original advantage of the JTC was that it avoided the use of a
recording medium to store the Fourier transform of the filter. This ad-
vantage has been eroded by the use of SLMs in the filter plane (Section
6.6). However, the capture of the JPS on an image sensor or LCLV al-
lows Fourier plane processing which can improve the correlator perfor-
mance. Binarisation of the JPS can be performed with a binary LCLV
[120], and with an image sensor [122]. The nonlinear transfer charac-
teristic of an LCLV can also be tailored to provide more information
about the correlation spot [117].
BS L2 BS
SLM1 SLM2
L3
L1
Laser diode
Image sensor
adjusting the focal length of lens, L2. In order to calculate the focal
length of L2, the size of the FT of the image is calculated and equated
to the size of SLM2. For square SLMs, the highest spatial frequency
generated by SLM1 should be located at the corner pixel of SLM2.
When this condition holds, and SLM1 has the same resolution as SLM2,
then every element of the FT in Equation (2.3) coincides with a pixel
of SLM2. For SLMs of rectangular aspect ratio, the highest spatial
frequency should match the minor axis of SLM2, and the pixels along
this axis will be in coincidence with the elements of the digital FT along
this dimension. For an SLM with a pixel repeat of ∆x1 , the largest
spatial frequency generated is ηmax = 1/2∆x1 . In the focal plane of
the Fourier lens, this is located at x′max = f λ/2∆x1 . Therefore, if
the number of pixels and the pixel repeat of SLM2 are L and ∆x2
respectively, then
fλ L∆x2
= (6.25)
2∆x1 2
This equation can be used to calculate the required focal length of the
Fourier lens, according to
L∆x2 ∆x1
f= (6.26)
λ
Two lenses, f1 and f2 , separated by a distance, d, are used to synthesize
Optical Correlators 91
M
SLM
(a)
BS L2 HM
(b)
L3
Image sensor
Figure 6.8Inverted VLC: (a) storing the templates in an HM; (b) re-
play of the templates from the HM. BS-Beamsplitter; HM-Holographic
memory; L-Lens; SLM-Spatial light modulator
a number of filters for each input scene, the number of correlations per
second (cps), is limited by the speed of the template SLM. The template
SLM is replaced by an angle multiplexed HM, the access speed of which
is much faster. The different templates in a filter bank can be accessed
at a template rate of kHz when the addressing beam is modulated by
an acoustooptic deflector. In the system illustrated in Figure 6.8, the
templates are stored by interfering the template beam with a plane ref-
erence wave (Figure 6.8a). The SLM displays the phase-only FT of the
template so that the template is recorded in an edge-enhanced format
in the HM. The HM used was a PR crystal called Lithium Niobate.
The phase-only FT employs the phase modulation discussed in Sec-
tion 2.3.1. It provides a high efficiency light throughput, which will be
quantified in the Chapter 7. In order to perform the correlation of mul-
tiple templates with a given scene image, the templates are accessed by
angle-multiplexing a plane reference wave which is counterpropagating
to the reference wave which recorded the edge-enhanced template (Fig-
ure 6.8b) [68, 69]. The FT of the template beam counterpropagates to
the SLM which now displays an FT of the scene image. The product of
the conjugate template and the scene image is Fourier transformed by
lens L3 to give a correlation signal on the image sensor. This system
was constructed under European funding (Brite) and the performance
details are provided in Table 6.1 [235].
The inverted (or transposed) VLC correlator is capable of search-
ing through a significant number of templates for every input image.
Therefore, it solves the problem of orientation and scale inflexibility in
the standard correlator by allowing the storage of all the necessary ori-
entational and scale changes in the template. A transposed correlator
was the basis of a Small Business Innovation Research (SBIR) grant in
the mid-1990s to Advanced Optical Systems, Inc. to develop a missile
guidance system [35]. If, in addition, the input scene is of low SBWP,
then the electronic processing can be significantly extended. The FFT
of the input scene and the multiplication with the FT of the template
can be performed electronically. This is most efficiently performed with
binary versions of the two FTs. In which case the multiplication is an
exclusive-OR (XOR) logic operation. The final FT is performed at high
speed using a binary SLM [20, 178].
The systems shown in Figures 6.7 and 6.8 employ on-axis filtering,
as opposed to the off-axis filtering used in the original VLC. The use
of the off-axis reference beam provided the advantage of separating
the three beams in the correlation output Equation (6.17). However, it
Optical Correlators 93
is low. Therefore, the filter SLM contains the template in phase mod-
ulated form, in order to promote light efficiency (Section 7.6.1). The
JPL greyscale optical correlator (GOC) employed a fast ferroelectric
greyscale SLM for the template images, which allowed a fast speed
when searching through a large database of templates for each input
image. Martin Marietta funded the construction of a fast correlator at
Boulder Nonlinear Systems (BNS) [233]. Ferroelectric LCOS was used
for both the input and template SLMs and this design was patented
[239]. The work was pursued at BNS under funding from the U.S. Army
Missile Command and resulted in an optical correlator equipped with
256 x 256 FLCOS operating at 4 kHz [74]. The overall speed of the
system was limited by the camera head to 979 cps. The FLCOS were
tested in analogue modulation mode, probably using ancillary retarda-
tion plates in front of the device [16]. The analogue mode results from
intermediate orientations of the liquid crystal director between the two
stable in-plane orientations discussed in Section 2.3.1. The retardation
plates convert the partial rotation of the director into analogue phase
values. Finally, the National Optics Institute of Canada (INO) pro-
duced an optical processor based on TFT LCDs which could be either
configured as an HC with a phase modulated filter, or configured for
spatial filtering (Section 8.5).
In Europe there was a complementary effort to build correlator sys-
tems funded mainly by the European Commission. The lead contrac-
tor in a European consortium, Thomson-CSF, constructed and suc-
cessfully tested a compact correlator, which employed a PR crystal,
bismuth silicon oxide (BSO), to record the JPS [210]. The PR crystal
was complemented by a mini-YAG (yttrium aluminium garnet) laser
and a TFT LCD input device. In the first version of this correlator,
the speed was limited by the input device. In a later version, this was
replaced by a 256x256 pixel FLC SLM, and gave in excess of 1000 cps
[59]. The performance of this correlator for fingerprint recognition was
documented in [218]. The European Space Agency funded TU Dresden
to construct a number of correlators. The entry in Table 6.1 provides
one example of these, based on a JTC with image sensor output which
can be accommodated in a linear system of 22 cm length. The SBWP
recorded is that of the input plane SLM. Where a JTC is implemented,
the SBWP of the input image will be less than one half of the SBWP
in the table, due to the need to accommodate both the input and
template images, with appropriate spacing between them, on the same
input device. The Institute of Optical Research (IOF) in Stockholm
Optical Correlators 95
6.8 3D CORRELATORS
Spatial light modulators are limited with regard to real world scenes.
It was shown in Chapter 1 that the digital picture function is an ap-
proximation to the picture function of an object. However, when the
object is embedded within a 3D scene, a number of complications arise.
Scale, aspect, illumination, and occlusion of the object all have to be
accommodated. In order to capture the 3D scene electronically for ob-
ject recognition, four approaches will be mentioned: digital holography
[121]; capturing a sequence of 2-D projections [219]; integral imaging
[83]; and the construction of sub-images [197]. Digital holography is a
technique for electronic processing of holograms in digital form. Typ-
ically, the hologram of a 3D scene is captured on an image sensor.
This hologram is an interference pattern between the scene and a ref-
erence wave. Numerical reconstruction and cross-correlation with the
template is performed electronically. Optical processing is beneficial
when the reference template can be placed in the input scene [219]. It
was demonstrated, by simulation of a JTC, that the 3D position of the
object can be recovered. This involves the capture of 2D projections of
the 3D scene by translating the image sensor in a plane orthogonal to
the line connecting the viewer to the scene. In general, the imaging of
3D scenes is preceded by a camera calibration in order to calculate the
elements of the homography matrix, which relates the coordinates of
the image sensor to those of the scene, but, with suitable approxima-
tions in this case, the location of the template within the scene can be
ascertained.
Integral imaging is more than 100 years old. It was invented in the
form of integral photography by Gabriel Lippmann in 1908. Professor
Lippmann was a prolific inventor who won the Nobel Prize in the same
year for his invention of a method for colour photography based on
recording interference fringes in a high-resolution silver halide emul-
sion. Integral photography involved the recording of a scene using an
TABLE 6.1 Compact correlator systems
array of small lenses in place of the single lens in a camera. When the
photographic plate was developed and viewed with a similar array of
lenses, the original scene exhibited parallax when viewed from differ-
ent positions within the lens array aperture. These cues give a realistic
rendition of the original 3D scene. Integral imaging is the name given
to the first aspect of integral photography. It is the basis of light field
cameras such as those marketed by Raytrix and Lytro. It is also known
as plenoptic imaging. 3D object recognition based on integral imaging
involves digital computation and is known as computational integral
imaging [139]. The advantage of integral imaging compared with dig-
ital holography is that incoherent light can be used. An interesting
use of lenslet arrays was the creation of sub-images from the elemen-
tal images formed by lenslet arrays [197]. The advantage of creating
these sub-images is that they can be used for the measurement of out-
of-plane rotations and shifts using sequences of 2D correlations. The
disadvantage is that this involves the complexity of electronic wiring.
GLOSSARY
Holographic recording material (HRM): A medium which records
high spatial frequencies by modification of its optical properties, es-
pecially the refractive index.
Filtering
CONTENTS
7.1 Synopsis . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 99
7.2 Vectors and matrices . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 101
7.3 Random variables and functions . . . . . . . . . . . . . . . . . . . . . . . . . . . 104
7.4 Hypothesis testing . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 108
7.5 Figures of Merit . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 111
7.5.1 Peak sharpness measures . . . . . . . . . . . . . . . . . . . . . . . . . . 112
7.5.2 Signal to noise ratio . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 112
7.5.3 Discrimination capability . . . . . . . . . . . . . . . . . . . . . . . . . . 113
7.6 Filter coding techniques . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 113
7.6.1 Phase-only filtering . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 114
7.6.2 Binary phase-only filters . . . . . . . . . . . . . . . . . . . . . . . . . . . 114
7.6.3 Wiener filter . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 115
7.6.4 Constrained filters . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 115
7.7 Filters based on training images . . . . . . . . . . . . . . . . . . . . . . . . . . . 117
7.7.1 SPR filters . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 118
7.7.2 SDF filters . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 119
7.7.3 Trade-off and unconstrained filter design . . . . . . . . . . 121
7.8 Phase-only correlation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 121
7.1 SYNOPSIS
A fundamental task in image processing is that of detecting a relevant
object in a scene, such as face recognition. If the position of the recog-
nised face is required, then the task is called localisation. If a particular
face is sought, then the task is called identification. In Chapter 1, the
capability of digital cameras to recognise and localise faces was men-
tioned. Identification is a more complex task. This chapter is concerned
with the engineering of filters in order to improve the performance of
the correlator system in particular directions, such as: discrimination,
in order to facilitate identification; throughput, in order to facilitate
99
100 Fourier Optics in Image Processing
and
2
j + k 2 jl + mk jn + ok
j k
j l m
AAT = l m = lj + mk l2 + m2 ln + om (7.5)
k n o
n o nj + ok nl + om n2 + o2
j 2 + l2 + m2 jk + lm + no
= (j 2 + l2 + m2 )(k 2 + m2 + o2 )
kj + ml + on k 2 + m2 + o2
−(jk + lm + no)2
Filtering 103
AA T ) is
and det(A
j 2 + k 2 jl + mk jn + ok
l2 + m2 ln + mo
lj + mk l2 + m2 ln + om = (j 2 + k 2 )
nl + om n2 + o2
nj + ok nl + om n2 + o2
lj + mk ln + mo lj + mk l2 + m2
− (jl + mk) + (jn + ok)
nj + ok n2 + o2 nj + ok nl + om
vector which is ordered from low to high spatial frequencies. The zero
order component is the first component of the transform. The 2D DFT
is ordered similarly, with the low spatial frequencies at the beginning
of the rows and columns. In order to design filters for optical systems,
where the zero order is at the centre of the matrix, a shift operation
must be performed on the results calculated digitally. The zero order
must be shifted from the top corner of the 2nd quadrant to the top
corner of the 4th quadrant. In order to maintain the 1D transform
nature of the rows and columns, the 1st and 3rd quadrants must be
interchanged. This shift operation is shown for the 2D transform in
Figure 7.1.
2 1
3 4
y=pdf(x) z=cdf(y)
0.4
0.5
-3 3
-3 3
Figure 7.3 Image with no noise and with added noise of variance 0.1, 1.0
and 4.0
where p(x, y) is the joint probability of variable, n(t1 ), taking the value
x, and variable, n(t2 ), taking the value y. For a stationary process, the
covariance is a function of τ = t1 − t2 . Additive white Gaussian noise
is a random signal with zero mean, a variance of σ 2 , and a covariance
of σ 2 if τ = 0, and zero if τ 6= 0.
The same statistical functions can be computed for random vari-
ables which are not infinite, using summations in place of integrations.
If X and Y are discrete distributions of M components, they are rep-
resented as a column vector and
M X
X M
Cov(X, Y ) = (xl − mX )(yk − mY )T (7.10)
l=1 k=1
108 Fourier Optics in Image Processing
where
M M
1 X 1 X
mX = xl , and mY = yk . (7.11)
M M
l=1 k=1
The same functions can be computed when there are more than two
discrete distributions; for example, when a number of different mani-
festations of the template image are available. The N template images
are expressed in raster scanned vector format, s 1 , s 2 ,..ssn ,...ss N , where
each s n is a column vector of M components, M being the number of
pixels in the image. Then the mean, s̄s, of a set of N samples is the
average over all samples
N
1 X
s̄s = sn (7.12)
N n=1
s′ = s − s̄s (7.13)
True False
Positives Positives
False True
Negatives Negatives
0.9
0.8
0.7
Detection rate
0.6
0.5
0.4
0.3
0.2
0.1
0
0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1
False alarm rate
balloon object which has been defined by the training set, the classifier
must have a capacity for generalisation. Patterns within a class have
to be recognised as belonging to the same class. This is made more
difficult when there are objects such as faces which form a class on their
own. The situation can be expressed graphically by forming a feature
space, where the axes correspond to common features of the patterns
which can be quantified, e.g., size and ellipticity of the face/balloon.
The patterns can then be represented by points in this feature space
(Figure 7.6).
Decision
boundary
Faces
Ellipticity x x
x x
o o Balloons
o o
Size
Figure 7.6 Feature space representation of patterns
some filters the different criteria are antagonistic, in the sense that the
decrease of one criterion will lead to the increase of another.
filter (BPOF) (Section 7.6.2). When the additive noise is not “white”
but “coloured”, the optimal filter is a Wiener filter which is presented
in Section 7.6.3. The MF and Wiener filter demand a complex coding
of the filter. If the complex modulation values of the filter are fitted to
the operating curve of the SLM, this is known as constrained filtering
(Section 7.6.4).
F̂ ∗ (fx, fy )
F̂P OF = (7.15)
|F̂ (fx , fy )|
This filter yields a high Horner efficiency, but the SNR is compromised
because it has no noise suppression capability, although some ameliora-
tion of this has been devised [146]. The discrimination capability of the
POF can be accentuated by attenuating a selected number of spatial
frequencies in the filter. This has been employed in the discrimination
of an “F” from an “E” in [1]. The high discrimination of this filter is
a handicap when there is variability in the template. The design of a
POF for this eventuality has also been described [129]. The POF was
used in one of the compact correlators listed in Table 6.1 [240]. Al-
though the POF produces a sharper correlation peak for a “hit” than
the MF, the filter which optimises the PCE is the inverse filter [148]
1
F̂IF = (7.16)
F̂ (fx, fy )
A further drawback of the POF is that it is only relevant for images
which have a FT with a phase function. In Chapter 1, it was shown
that the FT of real symmetric inputs is real (Equation 1.10). If the FT
is, furthermore, positive, then the POF would be the null matrix.
F̂ (fx , fy )
F̂W = (7.19)
F̂ (fx , fy ) + Pn (fx , fy )
Im Im
x x x
x x
x x
x
x x Re x
x x x
Re
x
x x
(c) x (d)
for the other class are small, and vice versa. The filter corresponding
to each eigenvector was correlated with the input image to assess the
performance of this technique in [154]. When there are more than two
classes, this methodology can be extended [93]. Alternatively, the fea-
ture space data can be mapped into a decision space, as in the example
of the support vector machine (Section 7.3). The output plane of the
correlator is segmented into a number of areas corresponding to the
number of classes, so that a spot in one area identifies the class of the
object. Since a significant SBWP is required of the filter in this case,
CGHs are preferred [94]. Both dimensionality reduction and feature
space mapping were used in [44]. The techniques of SPR are valuable
when there is significant inter- and intra-class variation. All the above
techniques have been applied to the samples of the input object. Two-
class discrimination has also been applied in the Fourier domain [71].
GLOSSARY
Binary phase-only filter (BPOF): A filter which is derived from
thresholding the phase-only filter.
123
124 Fourier Optics in Image Processing
tive lens can be engineered to have the precise focal length required for
creating a correctly sized Fourier transform of the input on the filter
SLM, there is no requirement for creating a synthetic focal length us-
ing two refractive lenses, Equation (6.3). An alternative approach is to
ensure that the devices are orientationally aligned but their positions
may not be precisely known. In this case, the lenses can be constructed
holographically to comply with the geometry of the folding [215]. An
additional advantage is that the aberrations of the lenses can also be
controlled. The capability to represent lens functions on an SLM (Sec-
tion 4.2) has allowed systems to be constructed where the focal length
and aberrations can be corrected under software control. The compact-
ness of the arrangement is limited by the focal length of the SLM lens,
which is, in turn, limited by the pixel repeat, due to the requirement to
avoid aliasing (Section 4.2). Nevertheless, a folded system based on one
LCOS device has been realised [288]. One half of the device displays
the input scene multiplexed with a lens, and the second half displays
the FT of the template. Due to the phase-only modulation employed
in the LCOS device, the multiplexing of a lens function with the in-
put scene is achieved by simply adding the two functions. If the total
phase exceeds 2π, then the value is reduced by an integer multiple of
2π. In addition to the lens function added to the input scene, a grating
was added to the template half of the device. The grating displaces
the correlation signal from the high intensity zero order spot by a dis-
tance proportional to the spatial frequency of the grating. The use of
one SLM for both the input scene and the template images reduces
the SBWP of the system. The adjustability of the arrangement can
be maintained with a two SLM system where lens functions are multi-
plexed on both the scene and the template SLMs [289]. The improved
SBWP was utilized for spatially multiplexing both the input scene and
the filter function. Spatial multiplexing is practicable when the SBWP
of the image is much less than the SBWP of the SLM. In this case, the
SBWP of the image was 256 x 256 so that it could be replicated 2 x 2
times across the 792 x 600 pixel device.
The idea of systematically integrating complex optical systems us-
ing planar technologies such as reactive ion etching (RIE) lithography
or ion-beam lithography was introduced at Bell Technical Labs when
there was a concentrated effort on optical computing [115]. This is now
known as PIFSO (planar integration of free-space optics). A Fourier
transform system has been constructed with PIFSO [183]. The power
of DOE design is illustrated by the achromatic nature of the Fourier
126 Fourier Optics in Image Processing
lens, which is not required in the systems which have been described
here.
Mirror
DOEs Devices
the filter, such as the input and frequency plane SLMs in an HC sys-
tem, the alignment required is more exacting. Translational alignment
along the 3 axes and rotational alignment around the same axes (roll,
yaw, and pitch) requires a systematic procedure. For example, the roll
misalignment around the optical axis can be nulled by comparing the
far-field diffraction patterns of the 2 SLMs [189]. Here, the alignment of
the SLMs is a precursor to the construction of a high-value solid block
module. Failure of one of the SLMs entails the replacement of the com-
plete module. An alternative is to retain the principle advocated in
[25], namely, to restrict the module to one device array and to build
the alignment into the mating of the modules. In order to accomplish
this, the mating surfaces should be either flat or round, and precisely
machined with respect to datum surfaces [271].
where (∆x, ∆y) are the distances between cells in the x- and y- direc-
2
tions, σGL is the variance of the grey levels (GL), and λh , λv are the
reciprocals of the correlation length in the horizontal and vertical axes
of the picture.
The location of target images in remote sensing images is an ap-
130 Fourier Optics in Image Processing
Im Im
xa
x x x x Re x x Re
a cgt c g
t
x
(a) (b)
Figure 8.2Coding of the four bases of DNA for the input plane of an
optical correlator: (a) amplitude coding along the real axis; (b) phase
coding on the unit circle
Improved contrast of the viewed image resulted from the phase con-
trast imposed by the spatial filter. When an SLM is used in the spatial
filtering plane, more sophisticated spatial filters can be realised [175].
The microscope is preferably equipped with a fibre-coupled laser illu-
mination together with a spinning diffuser for speckle reduction [200].
However, partially coherent illumination can be used in the form of fil-
tered white light which is apertured down so that the spatial coherence
is improved sufficiently [176].
Cytopathology involves the visual inspection of cell tissue. It can be
performed using spatial filtering in the microscope [199] or on CCD im-
ages [95] of tissue. It is an example of a repetitive task where abnormal
tissue is a rare occurrence and the false negative rates are, understand-
ably, high. Hence the interest in automating the visual inspection pro-
cess. An example is cervical cancer screening, where pap-smear slides
are prepared from cellular samples obtained from the cervix. Where fa-
cilities exist, such slides are sent to a cytology laboratory for staining.
This is followed by high-resolution microscopy and automated feature
extraction and classification of the cytoplasm and the nucleus. An op-
tical processing approach, consisting of two stages, can be employed to
isolate regions of interest (ROIs) for closer (possibly visual) inspection
[180]. The first, automatic stage is the recognition of large nuclei in
a hybrid correlator, using a filter which is the Fourier transform of a
circular dot. Abnormal cells are identified on the basis of their enlarged
nuclei. A normal cell’s nuclear diameter is between 6 and 8 µm, while
an abnormal cell’s nuclear diameter ranges between 10 and 22 µm. The
circular dot has a diameter which is the lower limit for the diameter of
the nucleus in an abnormal cell. The correlation peak width or height
can be thresholded to separate the cells with nuclear diameters less
than this diameter (normal cells) from those with diameters greater
than the filter diameter (abnormal cells). In information processing,
simple masks, such as the FT of the circular dot, are called kernels.
The convolution of an image with such a kernel is a basic technique in
morphological image processing, where ‘hits’ correspond to those parts
of the image that are larger than the dot. Complementary filtering can
be performed on the complementary image, where ‘misses’ correspond
to those parts of the complementary image that are smaller than the
dot. This involves a complementary kernel to the normal image, and
the combination of the two convolutions (‘hit’ and ‘miss’) is called
the hit/miss transform (HMT). The HMT detects the shape/size of
the cell nucleus. If the nucleus is circular with a diameter larger than
134 Fourier Optics in Image Processing
the weft spatial frequency. These were used to calculate four bandpass
filters which were displayed on the filter SLM. The bandpass filtering
removes the zero order and the warp and weave spatial frequencies so
that the output image has less structure than the original image and
defects can be found relatively easily by thresholding the digital image.
Optical implementation of spatial filtering saves computing time, mem-
ory, and storage for high-resolution imaging of wide areas of fabric. A
speed on the production line of several metres per second was cited
as the justification for optical detection of weave defects in [21]. Both
wavelet filtering and Wiener filtering (Section 7.6.3) were employed in
the INO optical correlator (Table 6.1), used in spatial filtering mode. In
this mode the filter LCTV is operated in amplitude modulation mode.
The wavelet selected was the Mexican hat wavelet which is close to the
difference of Gaussians (DOG) spatial filter (Figure 8.4), represented
by
fx2 fx2
0.9
0.8
Normalised filter transfer function
0.7
0.6
0.5
0.4
0.3
0.2
0.1
0
-10 -8 -6 -4 -2 0 2 4 6 8 10
Spatial frequency
memory was ideal for the storage of the templates of a machine part
at all possible orientations. These could be read out at high speed in
order to establish the orientation of the part in a scene.
Spectral filtering of the scene can be advantageous in product in-
spection. Boulder Nonlinear Systems worked on an HC for the inspec-
tion of produce such as apples [16]. An electronically switchable liquid-
crystal tunable filter, placed in front of a CCD camera, is employed to
select the frequency of imagery that is brought into the correlator. The
filter could be switched between different wavelengths in the visible
and near-infrared ranges within 25 µs. Certain wavelengths enhance
defects in the produce. Two examples given in the paper were scabs on
a red apple which are enhanced with red filtered light and sub-surface
bruising which is detected in the infrared. Subsequent detection of the
defect in the correlator signals removal of that apple from the line of
produce which is being inspected. The paper also discusses the two
types of LCOS which were used in the correlator. The configurations
of the liquid crystal layer in these two types of LCOS were PAN and
FLC. Moreover, the paper catalogues how the modulation of ampli-
tude and phase generated by each of these configurations can be varied
according to the orientation of retardation plates placed in front of the
LCOS device in the paths of the incident and reflected beams.
8.7 BIOMETRICS
The market for biometric systems is expected to exceed $20 billion by
2020. Biometric identification is used to confirm the identity of a person
for security purposes. Three prominent types of identification systems
are based on the uniqueness of the fingerprint, the iris, or the face. Fin-
gerprints are probably the most widely used. They have been used for
identification purposes for over 100 years. A database of fingerprints,
FVC2004, is included in the handbook [172]. For a limited access sys-
tem, such as the workers in a laboratory, a straightforward correlation
of the print with a limited database can be successful. I personally ver-
ified the accuracy of a system employed to identify the fingerprints of
workers in the Hamamatsu research company. This system was based
on a joint transform correlator (JTC) [140]. For more demanding recog-
nition tasks, a digital simulation of the phase-only correlation (POC)
technique (Section 7.8) has been used [114]. The POC is particularly
useful in those cases where the captured fingerprint is of poor quality
[113]. This was successfully tested on a larger factory of 700 employees.
140 Fourier Optics in Image Processing
The POC digital simulation has been used to coarsely assign two fin-
gerprints, prior to detailed matching in a convolutional neural network
in [208]. An alternative to fingerprints is to use the finger vein. The
network pattern of the finger vein is distinct for each individual and
is internal. An improved finger-vein identification algorithm based on
template matching was proposed and confirmed experimentally [237].
A digital simulation of POC has also been used for iris recognition with
a view to a digital signal processing implementation [184].
Face recognition software is commonly used on digital cameras to
locate faces in scenes. Face authentification by digital approaches is an
emerging biometric, which has been accelerated by a recent trend in 3-D
face mapping. Distinctive features such as the eye socket contour, and
the chin and nose shape are not subject to variation with illumination
conditions. They can also be measured under a variety of angles of the
face, rather than a straight-ahead shot. The face identification system
on the Apple iPhone projects more than 30,000 infrared dots onto the
user’s face. The projected grid method yields an improved recognition
accuracy of 1 in a million false positives, where the wrong person would
be allowed to unlock the phone. Qualcomm has a computer vision kit
that is capable of active depth sensing, using an infrared illuminator,
infra-red camera and a 16-megapixel RGB camera. The illuminator
fires a light that creates a dot pattern (using a filter), and the infra-
red camera searches for and reads the pattern. By calculating how
the dots warp over a subject and the distance between points, the
system can estimate the 3D profile of faces. Identification of faces from
a limited database has been claimed by NEC, based on their NeoFace
software. Interestingly, a multichannel approach to face localization
based on correlating 10 landmark features across the face was outlined
in [86]. The approach was extended to car detection based on the MIT
Streetscene database, and to pedestrian detection based on the Daimler
pedestrian dataset. The interest of the approach is, firstly, that the
HOG low level descriptors (Chapter 1) are used to pre-process the
images for correlation, and, secondly, the memory and computational
cost of performing correlation in the spatial frequency domain scales
favourably compared with alternative approaches when the number of
training images increases.
Face identification using the 2D correlation approach follows the
capture by a standard camera and segmentation and excision of the
faces from the scene. The identification task, in this case, is com-
pounded by varying pose, illumination, and expression. In order to
Systems: Packaging and Applications 141
peak but the cryptographic key. In order to increase the security of the
system, an extra feature is incorporated. The extra security feature is
the bonding of a phase mask to the primary identification amplitude
pattern such as the fingerprint, picture of a face, or signature [118].
The phase mask and the primary pattern are separately identifiable in
an optical correlator.
using a MACH filter which is based on the silhouette of the head and
shoulders in all orientations and aspects [214].
Unmanned autonomous vehicles (UAVs) operate at low altitudes,
for example down to 10 m above the terrain, and are equipped with
a 2D visible wavelength image sensor complemented by a fish-eye lens
with a wide field of view. In order to navigate the UAV, one option is
to generate a digital elevation map (DEM), which is a running 2.5D
model of the terrain [255]. An efficient means for generating the DEM
using a JTC is also described. It consists of dividing the scene into
a large number of sub-images, which can be tracked from frame-to-
frame in order to generate an optical flow field. The optimal size of
the sub-images was 24 x 24 pixels and the JTC was capable of 50,000
correlations per second [254].
145
146 Bibliography
[14] H.R. Bagley, J.A. Sloan, and D.W. Small. Vander Lugt optical
correlator on a printed circuit board, August 19 1997. US Patent
5,659,637.
[42] D.T. Carrott and T.M. Burke. Optical correlator based auto-
mated pathologic region of interest selector for integrated 3D
ultrasound and digital mammography, May 28 2002. US Patent
6,396,940.
[53] Heang Ping Chan, Jun Wei, Yiheng Zhang, Mark A. Helvie,
Richard H. Moore, Berkman Sahiner, Lubomir Hadjiiski, and
Daniel B. Kopans. Computer-aided detection of masses in digital
tomosynthesis mammography: Comparison of three approaches.
Medical Physics, 35(9):4087–4095, 2008.
[82] L. E. Franks. A model for the random video process. The Bell
System Technical Journal, 45(4):609–630, April 1966.
[91] J. M. Gordon and Ari Rabl. Reflectors for uniform far-field ir-
radiance: fundamental limits and example of an axisymmetric
solution. Appl. Opt., 37(1):44–47, Jan 1998.
[92] Jan Grinberg, Alex Jacobson, William Bleha, Leroy Miller, Lewis
Fraas, Donald Boswell, and Gary Myer. A new real-time non-
coherent to coherent light image converter the hybrid field effect
liquid crystal light valve. Optical Engineering, 14(3):217–225,
1975.
154 Bibliography
[94] Zu-Han Gu, James R. Leger, and Sing H. Lee. Optical implemen-
tation of the least-squares linear mapping technique for image
classification. J. Opt. Soc. Am., 72(6):787–793, Jun 1982.
[99] Tsutomu Hara, Yoshiharu Ooi, Yoshiji Suzuki, and Ming H. Wu.
Transfer characteristics of the microchannel spatial light modu-
lator. Appl. Opt., 28(22):4781–4786, Nov 1989.
[119] Bahram Javidi and Francis Chan. Image Recognition and Clas-
sification, chapter 6, pages 189–216. CRC Press, 2002.
[152] Yann LeCun, Yoshua Bengio, and Geoffrey Hinton. Deep learn-
ing. Nature, 521(7553):436–444, May 2015.
[153] John N. Lee and Arthur D. Fisher. Device developments for opti-
cal information processing. volume 69 of Advances in Electronics
and Electron Physics, pages 115–173. Academic Press, 1987.
[159] Scott D. Lindell and William B. Hahn, Jr. Overview of the Martin
Marietta transfer of optical processing to systems (TOPS) optical
correlation program. Proc. SPIE, 1701:21–30, 1992.
[161] Tomasz Lis, Jessica Gaskin, John Jasper, and Don A. Gregory.
Digital optical correlator x-ray telescope alignment monitoring
system. Optical Engineering, 57, 2018.
[166] J.R. Lucas, A.M. Pollack, and S.A. Mills. Reflective optical cor-
relator with a folded asymmetrical optical axis, May 10 1994. US
Patent 5,311,359.
[172] Davide Maltoni, Dario Maio, Anil K. Jain, and Salil Prabhakar.
Handbook of Fingerprint Recognition. Springer Publishing Com-
pany, Incorporated, 2nd edition, 2009.
[173] Tariq Manzur, John Zeller, and Steve Serati. Optical correlator
based target detection, recognition, classification, and tracking.
Appl. Opt., 51(21):4976–4983, Jul 2012.
[228] J.A. Sachs, H.V. Goetz, D. Keith, L. Li, S.H. Linn, A. Parfenov,
S.E. Brice, T.J. Scheffer, J.A. Van Vechten, and J. Xue. Trans-
missive, optically addressed, photosensitive spatial light modula-
tors and color display systems incorporating same, August 2011.
US Patent 7,990,600.
[230] Marija Scholl and James W. Scholl. Optical processing for range
and attitude determination. Proc. SPIE, 1694:23–28, 1992.
[257] Jiangmin Tian, Guoyou Wang, Jianguo Liu, and Yuanchun Xia.
Chinese license plate character segmentation using multiscale
template matching. Journal of Electronic Imaging, 25, 2016.
[269] Paul Viola and Michael Jones. Robust real-time object detection.
International Journal of Computer Vision, 57:137–154, 2004.
[270] K. von Bieren. Lens design for Optical Fourier Transform sys-
tems. Appl. Opt., 10(12):2739–2742, Dec 1971.
[272] Li Wang and Garret Moddel. Resolution limits from charge trans-
port in optically addressed spatial light modulators. Journal of
Applied Physics, 78(12):6923–6935, 1995.
[274] Shunli Wang, Liangcai Cao, Huarong Gu, Qingsheng He, Claire
Gu, and Guofan Jin. Channel analysis of the volume holographic
correlator for scene matching. Opt. Express, 19(5):3870–3880,
Feb 2011.
Bibliography 171
[275] Shunli Wang, Qiaofeng Tan, Liangcai Cao, Qingsheng He, and
Guofan Jin. Multi-sample parallel estimation in volume holo-
graphic correlator for remote sensing image recognition. Opt.
Express, 17(24):21738–21747, Nov 2009.
[283] Roland Winston and Weiya Zhang. Novel aplanatic designs. Opt.
Lett., 34(19):3018–3019, Oct 2009.
[284] C.G. Wynne. Simple Fourier transform lenses II. Optics Com-
munications, 12(3):270–274, 1974.
[285] Jingqun Xi. High efficiency directional light source using lens
optics, June 2 2015. US Patent 9,046,241.
[288] Xu Zeng, Jian Bai, Changlun Hou, and Guoguang Yang. Com-
pact optical correlator based on one phase-only spatial light mod-
ulator. Opt. Lett., 36(8):1383–1385, Apr 2011.
[289] Xu Zeng, Takashi Inoue, Norihiro Fukuchi, and Jian Bai. Parallel
lensless optical correlator based on two phase-only spatial light
modulators. Opt. Express, 19(13):12594–12604, Jun 2011.
[291] Jie Zhang, Binjie Xin, and Xiangji Wu. Density measurement of
yarn dyed woven fabrics based on dual-side scanning and the FFT
technique. Measurement Science and Technology, 25(11):115007,
2014.
[292] Nikolay Zheludev. The life and times of the LED, a 100-year
history. Nature Photonics, 1(4):189–192, 2007.
Index
173
174 Index