Electro-Optical System Analysis and Design (Cornelius J)
Electro-Optical System Analysis and Design (Cornelius J)
Electro-Optical System Analysis and Design (Cornelius J)
COrnellus.J. Willers
-----
Library of Congress Cataloging‐in‐Publication Data
Willers, Cornelius J.
Electro‐optical system analysis and design: a radiometry perspective / Cornelius
J Willers.
pages cm
Includes bibliographical references and index.
ISBN 978‐0‐8194‐9569‐3
1. Electrooptics. 2. Optical measurements. 3. Electrooptical devices‐‐Design and
construction. I. Title.
TA1750.W55 2013
621.381ʹ045‐‐dc23
2013002619
Published by
SPIE—The International Society for Optical Engineering
P.O. Box 10
Bellingham, Washington 98227‐0010 USA
Phone: +1 360 676 3290
Fax: +1 360 647 1445
Email: [email protected]
Web: http://spie.org
Copyright © 2013 Society of Photo‐Optical Instrumentation Engineers (SPIE)
All rights reserved. No part of this publication may be reproduced or distributed
in any form or by any means without written permission of the publisher.
The content of this book reflects the work and thought of the author(s).
Every effort has been made to publish reliable and accurate information herein,
but the publisher is not responsible for the validity of the information or for any
outcomes resulting from reliance thereon.
Cover image “Karoo Summer,” by Fiona Ewan Rowett (www.fionarowett.co.za),
used with permission.
Printed in the United States of America.
First printing
Preface
On Sharing
Teachers cross our paths in life. Some teachers have names, others leave
their marks anonymously. Among my teachers at the Optical Sciences
Center at the University of Arizona were James Palmer, Eustace Dereniak,
and Jack Gaskill. They freely shared their knowledge with their students.
Some teachers teach through the pages of their books, and here I have
to thank Bill Wolfe, George Zissis, and many more. Many years ago,
R. Barry Johnson presented a short course which influenced my career
most decisively.
The intent with this book is to now share some of my experience, ac-
cumulated through years of practical radiometry: design, measurements,
modeling, and simulation of electro-optical systems. The material pre-
sented here builds upon the foundation laid at the Optical Sciences Center.
I had the opportunity to share this material in an academic environment
at graduate level in an engineering school, thereby clarifying key concepts.
Beyond the mathematics and dry theory lies a rich world full of subtle in-
sights, which I try to elucidate. May this book help you, the reader, grow
in insight and share with others.
The reductionist approach holds the view that an arbitrarily complex sys-
tem can be understood by reducing the system to many, smaller systems
xxiii
xxiv Preface
that can be understood. This view is based on the premise that the com-
plex system is considered to be the sum of its parts, and that by under-
standing the parts, the sum can be understood. While the reductionist
approach certainly has weaknesses, this approach works well for the class
of problems considered in this book. The methodology followed here is to
develop the theory concisely for simple cases, developing a toolset and a
clear understanding of the fundamentals.
The real world does not comprise loose parts and simple systems.
Once the preliminaries are out the way, we proceed to consider more com-
plex concepts such as sensors, signatures, and simple systems comprising
sources, a medium, and a receiver. Using these concepts and the tools de-
veloped in this book, the reader should be able to design a system of any
complexity. Two concurrent themes appear throughout the book: frag-
menting a complex problem into simple building blocks, and synthesizing
(designing) complex systems from smaller elements. In any design pro-
cess, these two actions take place interactively, mutually supporting each
other. In this whirlpool of analysis and synthesis, uncontrolled external
factors (e.g., the atmosphere, noise) influence the final outcome. This is
where the academic theory finds engineering application in the real world.
This book aims to demonstrate how to proceed along this road.
Toward the end of the book, the focus shifts from a component-level
view to an integrated-system view, where the ‘system’ comprises a (sim-
ple or composite) source, an intervening medium, and a sensor. Many
real-world electro-optical applications require analysis and design at this
integrated-system level. Analysis and design, as a creative synthesis of
something new, cannot be easily taught other than by example. For this
purpose several case studies are presented. The case studies are brief and
only focus on single aspects of the various designs. Any real design pro-
cess would require a much more detailed process, beyond the scope of this
book.
General Comments
The purpose with this book is to enable the reader to find solutions to real-
world problems. The focus is on the application of radiometry in various
analysis and design scenarios. It is essential, however, to build on the foun-
dation of solid theoretical understanding, and gain insight beyond graphs,
tables and equations. Therefore, this book does not attempt to provide an
extensive set of ready-to-use equations and data, but rather strives to pro-
vide insight into hidden subtleties in the field. The atmosphere provides
opportunity for a particularly rich set of intriguing observations.
Preface xxv
Riana Willers for patience and support, as co-worker on our many projects
— her light footprints fall densely on every single page in this book: advis-
ing, scrutinizing every detail, debating symbols and sentences, editing text
and graphics, compiling the nomenclature and index, and finally, acting as
chapter contributor. Riana is indeed the ghost writer of this book! Fiona
Ewan Rowett for permission to use her exquisite “Karoo Summer” on the
front cover. The painting beautifully expresses not only the hot, semi-arid
Karoo plateau in South Africa, but also expresses radiated light and vi-
brant thermal energy, the subject of this book. My teachers at the Optical
Sciences Center who laid the early foundation for this work. Ricardo San-
tos and Fábio Alves for contributing to the chapter on infrared detector
theory and modeling. The pyradi team for contributing their time toward
building a toolkit of immense value to readers of this book. Derek Griffith
for the visual and near-infrared reflectance measurements. Hannes Calitz
for the spectral measurements, and Azwitamisi Mudau for the imaging
infrared measurements. Dr Munir Eldesouki from KACST for permis-
sion to use the Bunsen flame measured data in the book. The many col-
leagues, co-workers, and students at Kentron (now Denel Dynamics), the
CSIR, KACST, and the University of Pretoria for influencing some aspect
of the book. Scott McNeill and Tim Lamkins for patience and guiding me
through the publication process. Scott’s untiring patience in detailed cor-
rection deserves special mention. Eustace Dereniak for encouraging me
to submit the book for publication. Barbara Grant, Eustace Dereniak and
an anonymous reviewer for greatly influencing the book in its final form.
Finally, Dirk Bezuidenhout, and the CSIR for supporting the project so
generously in the final crucial months before publication.
Mark Twain wrote that he did not allow his schooling to get in the
way of his education. It is my wish that you, my esteemed reader, will
delve beyond these written words into the deeper insights. Someone else
said that the art of teaching is the art of assisting in discovery. May you
discover many rich insights through these pages.
Nelis Willers
Hartenbos
March 2013
Contents
Nomenclature xvii
Preface xxiii
2 Introduction to Radiometry 19
2.1 Notation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 19
2.2 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . 20
2.3 Radiometry Nomenclature . . . . . . . . . . . . . . . . . . . 23
2.3.1 Definition of quantities . . . . . . . . . . . . . . . . . 23
2.3.2 Nature of radiometric quantities . . . . . . . . . . . . 25
2.3.3 Spectral quantities . . . . . . . . . . . . . . . . . . . . 25
2.3.4 Material properties . . . . . . . . . . . . . . . . . . . . 27
2.4 Linear Angle . . . . . . . . . . . . . . . . . . . . . . . . . . . 27
vii
viii Contents
3 Sources 57
3.1 Planck Radiators . . . . . . . . . . . . . . . . . . . . . . . . . 57
3.1.1 Planck’s radiation law . . . . . . . . . . . . . . . . . . 60
3.1.2 Wien’s displacement law . . . . . . . . . . . . . . . . 62
3.1.3 Stefan–Boltzmann law . . . . . . . . . . . . . . . . . . 63
3.1.4 Summation approximation of Planck’s law . . . . . . 64
3.1.5 Summary of Planck’s law . . . . . . . . . . . . . . . . 65
3.1.6 Thermal radiation from common objects . . . . . . . 65
3.2 Emissivity . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 65
3.2.1 Kirchhoff’s law . . . . . . . . . . . . . . . . . . . . . . 69
3.2.2 Flux transfer between a source and receiver . . . . . 70
3.2.3 Grey bodies and selective radiators . . . . . . . . . . 71
3.2.4 Radiation from low-emissivity surfaces . . . . . . . . 73
3.2.5 Emissivity of cavities . . . . . . . . . . . . . . . . . . . 74
Contents ix
4 Optical Media 97
4.1 Overview . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 97
4.2 Optical Mediums . . . . . . . . . . . . . . . . . . . . . . . . 98
4.2.1 Lossy mediums . . . . . . . . . . . . . . . . . . . . . . 98
4.2.2 Path radiance . . . . . . . . . . . . . . . . . . . . . . . 99
4.2.3 General law of contrast reduction . . . . . . . . . . . 102
4.2.4 Optical thickness . . . . . . . . . . . . . . . . . . . . . 103
4.2.5 Gas radiator sources . . . . . . . . . . . . . . . . . . . 103
4.3 Inhomogeneous Media and Discrete Ordinates . . . . . . . 104
4.4 Effective Transmittance . . . . . . . . . . . . . . . . . . . . . 105
4.5 Transmittance as Function of Range . . . . . . . . . . . . . . 108
4.6 The Atmosphere as Medium . . . . . . . . . . . . . . . . . . 108
4.6.1 Atmospheric composition and attenuation . . . . . . 108
4.6.2 Atmospheric molecular absorption . . . . . . . . . . 111
4.6.3 Atmospheric aerosols and scattering . . . . . . . . . . 112
4.6.4 Atmospheric transmittance windows . . . . . . . . . 116
4.6.5 Atmospheric path radiance . . . . . . . . . . . . . . . 118
4.6.6 Practical consequences of path radiance . . . . . . . . 120
4.6.7 Looking up at and looking down on the earth . . . . 121
4.6.8 Atmospheric water-vapor content . . . . . . . . . . . 121
4.6.9 Contrast transmittance in the atmosphere . . . . . . 124
4.6.10 Meteorological range and aerosol scattering . . . . . 127
4.7 Atmospheric Radiative Transfer Codes . . . . . . . . . . . . 129
4.7.1 Overview . . . . . . . . . . . . . . . . . . . . . . . . . 129
4.7.2 Modtran™ . . . . . . . . . . . . . . . . . . . . . . . . 129
Bibliography . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 130
Problems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 133
6 Sensors 221
6.1 Overview . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 221
6.2 Anatomy of a Sensor . . . . . . . . . . . . . . . . . . . . . . 221
6.3 Introduction to Optics . . . . . . . . . . . . . . . . . . . . . . 223
6.3.1 Optical elements . . . . . . . . . . . . . . . . . . . . . 223
6.3.2 First-order ray tracing . . . . . . . . . . . . . . . . . . 225
6.3.3 Pupils, apertures, stops, and f -number . . . . . . . . 226
6.3.4 Optical sensor spatial angles . . . . . . . . . . . . . . 230
6.3.5 Extended and point target objects . . . . . . . . . . . 232
6.3.6 Optical aberrations . . . . . . . . . . . . . . . . . . . . 232
6.3.7 Optical point spread function . . . . . . . . . . . . . . 235
6.3.8 Optical systems . . . . . . . . . . . . . . . . . . . . . . 236
6.3.9 Aspheric lenses . . . . . . . . . . . . . . . . . . . . . . 237
6.3.10 Radiometry of a collimator . . . . . . . . . . . . . . . 238
6.4 Spectral Filters . . . . . . . . . . . . . . . . . . . . . . . . . . 240
6.5 A Simple Sensor Model . . . . . . . . . . . . . . . . . . . . . 240
6.6 Sensor Signal Calculations . . . . . . . . . . . . . . . . . . . 242
6.6.1 Detector signal . . . . . . . . . . . . . . . . . . . . . . 242
6.6.2 Source area variations . . . . . . . . . . . . . . . . . . 244
6.6.3 Complex sources . . . . . . . . . . . . . . . . . . . . . 245
6.7 Signal Noise Reference Planes . . . . . . . . . . . . . . . . . 245
6.8 Sensor Optical Throughput . . . . . . . . . . . . . . . . . . . 248
Bibliography . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 250
Problems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 250
xii Contents
Index 477
Chapter 1
Electro-Optical System Design
1.1 Introduction
Optical flux has a source and, for the applications considered in this book,
also a destination (sometimes called a receiver or absorber). Having a
source and a destination, it must also have a channel, path, or medium.
The approach in this book is to consider all three components interacting
with the flux. The presence of more than one component implies that the
flux can be seen to operate in a system context, with elements of the system
including at least a source, a medium, and a receiver. Accepting the notion
of an electro-optical system, the system can be subjected to actions such as
analysis, design, and testing.
The fundamental approach taken in this book is that an electro-optical
system should be considered as a system with cause-and-effect implica-
tions. Although the components in the system may not interact in a physi-
cal or causal manner, the performance of the system can be expressed as a
set of relationships. In these relationships the system’s performance leads
to interdependencies between parameters of the various components. For
example, the maximum range performance of a laser rangefinder depends
on laser power, atmospheric transmittance, and sensor noise all of which
require tradeoff analysis to optimize the system. Hence, notwithstanding
the autonomy of each component, from a system perspective, the design
process induces a synthetic parameter interdependence between the vari-
ous components in the system.
The premise for this approach asserts that electro-optical systems can
be optimized by trading off parameters between different components of
1
2 Chapter 1
the system. Such a capability provides freedom and power to optimize the
system by appropriate design choices. Returning to the laser rangefinder,
the cost and complexity of laser power can be traded against the cost and
complexity of noise in the receiver: the selection of the appropriate low-
noise design may ease the burden of higher-power laser technology.
This chapter provides an overview of the generic design process, there-
by providing a basis for electro-optical system design. The basic philoso-
phy outlined here has application in practically all design situations. The
ideas presented come from system engineering principles, practical expe-
rience, and common sense. This chapter structures the activities that most
designers use intuitively. Although it is true that simple systems can be
handled intuitively, the design of complex systems requires stricter disci-
pline. The thoughtless execution of a procedure can be dangerous, but a
well-structured approach provides a better basis for a sound design.
1.2.1 Definitions
cannot capture the subtleties inherent in the design process or the require-
ments for a good design. However, it may be enlightening to attempt a
definition. The definition will be given in two phases: a short version in
this section, and a more-complete version in Section 1.2.11.
Any design should start with a clear goal to be achieved and an initial
design or design concept. The design goal can be translated into perfor-
mance measures (see Section 1.2.8) by predicting performance in calcu-
lation, simulation, or by building the item. Once a system is built, the
measured performance is compared to the design goals. If the measured
performance compares acceptably with the goals, the design is completed.
If the goals are not met by the design, the design details are iterated to find
a better design, mostly by improving certain aspects of the current design.
In certain cases the performance may not meet the system requirements
but is acceptable to the customer, in which case the user requirement and
system specifications are altered. This process is summarized in Figure 1.1.
There are certain prerequisites that have to be satisfied before any design
work can start. This may sound ludicrous, but these obvious prerequisites
are often not met! The most important are:
Concept study: The concept study is a high-level study into different con-
cepts that may be employed to achieve goals. Extensive use is made
Electro-Optical System Design 5
¢
¢
£
¢
¢DZ
¢
£
this phase. The work during this phase confirms the design approach
at a conceptual level. Furthermore, valuable hardware experience is
gained in the construction of the prototype.
Hardware development, experimental model: The hardware constructed
in the experimental-model phase should be suitable for the final ap-
plication, albeit with low confidence. All functions should be per-
formed, and the size and packaging should be according to specifi-
cation. Formal documentation-change control procedures are intro-
duced to keep track of changes. The output from this phase com-
prises the design data pack, hardware, and software, designed and
built to meet all specifications. In some cases the specifications are
not yet met, and there may be some hardware and software function-
ality not yet finalized.
Hardware development, advanced model: Most of the design issues are
cleared up toward the end of the advanced-model phase. The design
documentation, test methods, and specifications are finalized. Un-
less required for specification conformity, the design will not change
beyond this phase. During this phase the industrialization process is
far advanced, but hardware might still be built in the development
laboratories.
Industrialization: In the industrialization phase, the production processes
are finalized by production specialists to ease future production. The
design is only modified if necessary and then with approval from all
concerned. Although this phase falls late in the development time-
line, the industrialization personnel are involved in the design from
the very beginning. In this phase, all hardware will be built on the
production line.
System acceptance or qualification: System acceptance is the final evalu-
ation testing where the system is evaluated against the requirement
specification. After acceptance and approval by the user, full-scale
production can begin.
Production, Operational Service Support, Maintenance, and Disposal:
These phases are part of the system lifecycle but are not covered un-
der the present discussion.
partake to assess the impact of the design on their own areas. Informal
peer reviews are also very useful in that a quick answer can be obtained
while working on a problem. Even though design reviews are sometimes
regarded with apprehension and fear, they are very useful learning expe-
riences for a well-prepared designer.
Recall that Figure 1.2 shows four parallel activities during product de-
velopment. The first activity, hardware development, was described in
Section 1.2.5. This hardware development activity is supported by equally
important but less-visible activities.
Product evaluation under operational conditions, e.g., during field tri-
als, is an important reality check because it provides real-world feedback of
the system’s performance. Deficiencies and weaknesses can be identified,
and limits of performance can be evaluated. Several operational-condition
tests are typically executed during the development of a complex system.
The nature of the operational tests shifts from initial exploration to evalu-
ation later in the design phases, leading up to the qualification acceptance
testing. Although laboratory tests are valuable, operational tests build
confidence in the system.
Modeling and simulation provide several benefits; for example, it can
provide a development environment for image processing algorithms, pro-
vide the capability to evaluate the system under conditions not possible in
the real world, or evaluate different design options. It stands to reason that
the models and simulation must be comprehensive and validated, relative
to the task at hand. It is essential that the design in the simulation match
the design in the hardware — there must be only one design, just different
instances thereof.
Some system environments require ongoing analysis of elements in
the external environment. One such example, the development of missile
countermeasures, requires an ongoing analysis of the operation of missile
threats. The outputs from these analysis tasks are used to influence some
aspect of the design in order to best respond to the external environment.
To obtain maximum benefit from these four development activities,
there should be a constant flow of information. Each of the activities
should constantly re-evaluate its own position in the context of learning
in the other activities. For example, the modeling and simulation activ-
ity should endeavor to most-accurately reflect the hardware status, and
likewise, the hardware activity can learn from simulation experiments.
8 Chapter 1
¢ ¢
¢
¢
¢
Figure 1.3 The systems engineering V-chart for system decomposition and integration.
1.2.7 Specifications
level, but during the integration process, the higher-level system does not
perform as expected. From the foregoing it is clear that the hierarchical
breakdown provides an ordered approach for the design and integration
of complex systems. However, the success of this approach depends totally
on the discipline and diligence of the people employing these methods.
The designer is frequently faced with choices, and the best selection may
not be clear. In this regard a value system is of great help. A value system
is a hierarchical set of priorities and goals that helps the designer to make
difficult choices, by evaluating design options in terms of predetermined
priorities.
To compile the value system, list all of the important issues concerning
the user need, the hardware to be designed, and the design environment.
The aspects to be addressed should include the importance of development
time scale, product cost, performance, local content, reliability, maintain-
ability, upgradeability, etc. Any real value system is further complicated
by peripheral but important issues such as company policy and mission,
labor situation, profit motives, etc. The project value system should not be
contaminated by ulterior motives such as personal gain, particular inter-
ests or disinterests, and ‘old-boys’ understandings.
When faced with design choices, the designer can allocate points for
each entry in the value system, to each option. The option that best satisfies
the high-priority values is then the desired option.
tive if applied with diligence and careful thought. The procedure is simple:
3. Review and confirm the top-priority assumptions, and ignore the low-
priority assumptions, as there is not enough time to do them all.
5. Keep the complete list — never discard any assumptions, however triv-
ial.
User
need Actions are indicated by solid boxes
Interpret and inputs or results by dashed boxes.
need
Predicted Proposed System
performance system model
Make Compile Detailed
assumptions model analysis
normally start before the hardware design activities, but they continue
while the design process takes place, as shown in Figure 1.2. A model is
used to derive the hardware design details, and hardware evaluation tests
are used to confirm and improve the model, in a never-ending cycle. The
parallel development of models and hardware therefore greatly increases
confidence in the system. This is especially true for systems where human
operators or free atmospheric effects are part of the system.
It should be clear that system design is not a simple one-dimensional
problem. There are many complex issues involved in an iterated design
approach, frequently requiring several experimental models before the fi-
nal design is approved. The design process normally stops when the de-
sign is certifiably demonstrated to meet the requirement. The designer is
normally not allowed to have the final say in the acceptance of the design,
but a review panel or advisors grants the final approval. It often happens
that the design process continues into production if unexpected problems
arises.
14 Chapter 1
Interface Interface
Greater Greater
system system
The Medium, comprising all effects that in any way influence the opti-
cal signal. This may include optical fibers, atmospheric attenuation,
dust, smoke, turbulence effects, etc.
The Sensor, comprising the packaging, optics, optical energy detector, sup-
Electro-Optical System Design 15
Table 1.1 Examples of typical electro-optical systems cast into the generic format.
Ȭ Ȭ
¢
¡ Ȭ
ǯ Ȭ
¢ ǰ ¢ ǯ
ǰ ¢ ¢
ǯ ¢ ǰ ǰ ¢ ¡
ǰ
ǰ ǯǯǰ ¢
ǯ
Ȭ ǯ
ǭ ǯ
ǯ
ǯ ¢ ǯ ¢
ǭ ǯ ¢ǯ ǭ ǯ
ǯ ǰ Ȭ
¢ ¢ǯ ¢ ǰ
¡ǯ ¢
Ȭ ¢ ¢
ǯ ǰ ǰ
ǯ ǯ
ǰ £ ¢
ǯ ¡ ¢ Ȭ
¡ǯ ¡
ǯ ǯ
¢ ǰ
ǯ
¢ǯ
£
ǯ ǯ
Table 1.1 lists typical examples of electro-optical systems. Note how all of
the systems fit into the basic source–medium–sensor model.
1.4 Conclusion
The principles described in this chapter can significantly ease the design of
complex systems with increased confidence in the system performance. In
practice it is very seldom possible to apply all of the principles described
above due to labor or schedule constraints. It is still useful to consider the
complete process and apply certain aspects in a particular situation. An
analysis of this nature is never complete, but the basic approach outlined
here should be applicable to most electro-optical design situations.
Bibliography
[1] Incose, “Systems Engineering Handbook,”
http://www.incose.org/ProductsPubs/products/sehandbook.aspx.
[2] DIRSIG Team, “Digital Imaging and Remote Sensing Image Generation
(DIRSIG),” http://www.dirsig.org/.
2.1 Notation
In this book the difference operator ‘d’ is used to denote ‘a small quantity
of ...’. This ‘small quantity’ of one variable is almost always related to
a ‘small quantity’ of another variable in some physical dependency. For
example, irradiance is defined as E = dΦ/dA, which means that a small
amount of flux dΦ impinges on a small area dA, resulting in an irradiance
of E. ‘Small’ is defined as the extent or domain over which the quantity, or
any of its dependent quantities, does not vary significantly. Because any
finite-sized quantity varies over a finite-sized domain, the d operation is
only valid over an infinitely small domain dA = limΔA→0 ΔA.
The difference operator, written in the form of a differential such as
E = dΦ/dA, is not primarily meant to mean differentiation in the math-
ematical sense. Rather, it is used to indicate something that can be inte-
grated (or summed). In this book, differentiation is seldom done, whereas
integration is done on almost every page.
In practice, it is impossible to consider infinitely many, infinitely small
domains. Following the reductionist approach, any real system can, how-
ever, be assembled as the sum of a set ofthese small domains, by integra-
tion over the physical domain as in A = dA. Hence, the ‘small-quantity’
approach proves very useful to describe and understand the problem,
whereas the real-world solution can be obtained as the sum of a set of
such small quantities. In almost all of the cases in this book, it is implied
that such ‘small-quantity’ domains will be integrated (or summed) over
the (larger) problem domain.
19
20 Chapter 2
2.2 Introduction
Electromagnetic Radiation
SI Units Sources of Wien law Common Frequency Wavelength Size of one SI Units
frequency energy temperature name wavelength wavelength
[K] [Hz] [m]
24
yottahertz 10
exahertz 18
10
x ray 10
6
10
-9 nanometer
sources
soft X rays 1017 protein
5 -8
10 16
10
10
the sun 4 ultra violet -7
10 10 virus
petahertz visible 10
15
light bulb 10
3
10
-6 micrometer
14
10 bacteria
human 2 infrared -5
10 13
10
body 10
1 -4 living cell
10 10
terrahertz 10
12
Microwaves
millimeter
10
0 10
-3 millimeter
radar 10
11 pinhead
radar transmitter -1 -2
10 10
radar 10
microwave 10 tennis ball
-2 -1
oven 10 10
gigahertz UHF TV 10
9
10
-3 10
0 meter
VHF TV 10
8
radio -4 FM radio 1 house
Radiowaves
10 10
transmitters 7
SW radio 10
-5 2
10 10 Eiffel Tower
megahertz 10
6
-6 MW radio 3 kilometer
10 10
5
10
-7 10
4 New York City
10 LW radio
PC monitors 104
-8 5
10 10 British Isles
kilohertz VLW 3
Very low freq
10
10
-9 radio 10
6 megameter
2
10
mains power 10
7 Earth
1
10
hertz 10
0 © CJ Willers 2005
Electric
field
Magnetic
field
Direction
of travel
Artist’s impression 3
Wavelength
(a) (b)
Phase
0 2p 4p 6p 8p 10p
Spherical
Planar wavefront
Wavelength wavefront
(c) (d)
Figure 2.2 Simple light models: (a) electromagnetic wave, (b) photon wave packet, (c) far-
field, plane wave, and (d) near-field spherical wave.
dicular to the wavefront. The wavefront and ray models are mostly used
in aberration calculations during optical design and when laying out an
optical system.
The photon 2 is a massless elementary particle and acts as the force
carrier for the electromagnetic wave. Photon particles have discrete en-
ergy quanta proportional to the frequency of the electromagnetic energy,
Q = hν = hc/λ, where h is Planck’s constant. Attempting to fuse the par-
ticle model and the wave model, consider the photon as a spatially limited
electromagnetic disturbance (a wave packet) propagating through space.
In this book the photon is sometimes graphically illustrated with the sym-
bol shown in Figure 2.2(b). The figure also shows an artist’s impression of
the photon as a wave packet.
Radiometry is the measurement and calculation of electromagnetic flux
transfer 4–10 for systems operating in the spectral region ranging from ul-
traviolet to microwaves. Indeed, these principles can be applied to elec-
tromagnetic radiation of any wavelength. This book only considers ray-
based 9 radiometry for incoherent radiation fields.
Introduction to Radiometry 23
E= dΦ
dA [W/m2 ] M= dΦ
dA [W/m2 ]
10000
Wavenumber [cm-1]
8000
Dλ1 Dλ2
6000
Dν1
4000
Dν2
2000
0
0 2 4 6 8 10 12
Wavelength [mm]
(a)
0
Dν/Dλ [cm-1/mm]
2000
4000
6000
8000
10000
0 2 4 6 8 10 12
Wavelength [mm]
(b)
Figure 2.4 The relationship between (a) wavelength and wavenumber, and (b) dν̃/dλ versus
wavelength.
The flux incident on an object (gas, liquid, or solid) can be reflected, ab-
sorbed, or transmitted so that 19
and, therefore
1 = α + τ + ρ, (2.3)
Consider the two different linear angles ABC or ABD in Figure 2.5. Con-
sider the angle ABC: project the points A and C onto the circle to obtain
the points A and C . Likewise, project the points A and D onto the circle
to obtain the points A and D . Hence, the direction from B to points A, C,
and D, irrespective of distance, are projected onto a circle with radius r.
From the definition of linear angle, the equal value of both these angles is
given by θ = s/r in radians, where s is the arc length between the projected
points A and C or D .
The algorithm for linear angle measurement is therefore: (1) project
the directions of A, C, and D as seen from B onto a circle centered on B,
(2) measure the arc length between the projected points, and (3) divide the
arc length by the circle’s radius.
For a full revolution, the arc length is the circumference of a circle
or 2πr, hence the linear angle of one revolution is 2π rad. A right angle
(90 deg) has an arc length of 2πr/4, leading to a linear angle of π/2 rad.
28 Chapter 2
A’
s
r
θ
B C’ D’
C D
where d2 P cos θ1 is the projected surface area of the surface P in the direc-
tion of the reference point, and R is the distance from d2 P to the reference
point. The integral is independent of the viewing direction (θ0 , α0 ) from
Introduction to Radiometry 29
P P
d2P d2P
θ1 θ1
(θ1,α1) (θ1,α1)
R R
θ0 θ0
(θ0,α0) (θ0,α0)
Figure 2.6 Solid angle definitions: (a) geometric solid angle, and (b) projected solid angle.
the reference point. Hence, a given area at a given distance will always
have the same geometric solid angle irrespective of the direction of the
area.
The projected solid angle Ω of any arbitrary surface P from the refer-
ence area dA0 is given by
P 2
d P cos θ0 cos θ1
Ω= , (2.5)
R2
where d2 P cos θ1 is the projected surface area of the surface P in the direc-
tion of the reference area, and R is the distance from d2 P to the reference
area. The integral depends on the viewing direction (θ0 , α0 ) from the ref-
erence area, by the projected area (dA0 cos θ0 ) of dA0 in the direction of
d2 P. Hence, a given area at a given distance will always have a different
projected solid angle in different directions. The calculation of solid angles
can be seen as a form of spatial normalization.
Consider the left side of Figure 2.7, where a hemisphere with radius r is
constructed around the origin of the coordinate system. The center of the
hemisphere is located at the origin. The geometric solid angle ω, sub-
tended at the origin, of any arbitrary surface P is given by
P
ω= , (2.6)
r2
where P is the projection of the surface P onto a sphere of radius r, as
shown in Figure 2.7. The dimensional unit of the solid angle is [m2 /m2 ],
30 Chapter 2
Figure 2.7 Geometric solid angle ω and projected solid angle Ω of a cone.
indicated with [sr], where the numerator is area, and the denominator is
radius squared.
Note that the geometric solid angle is independent of the direction of
the true area P; only the projected area P is relevant.
For a cone with half-apex angle of Θ (the full-apex angle is therefore
2Θ), the geometric solid angle can be derived as shown below. A small
portion of the projected surface area d2 P can be written as
d2 P = rdθ × hdα
= r2 dθ sin θdα. (2.7)
For Θ = π/2 rad, the cone covers a full hemisphere, and the solid
angle is 2π sr; for Θ = π rad, the full sphere has a solid angle of 4π sr.
Introduction to Radiometry 31
The projected solid angle is calculated relative to a small surface dA0 lo-
cated at the origin (not a single point as for the geometric solid angle). The
small surface dA0 has a vector normal to the surface, as shown on the right
side of Figure 2.7.
The projected area P is weighted with a factor cos θ, where θ is the
angle between the normal vector of the small surface dA0 and the direction
to the projected area P .
The contribution of the projected area P to the solid angle therefore
depends on the direction of the projected area P , relative to the normal
vector of the area dA0 . Restated, the direction of the projected area P
significantly determines its contribution to the solid angle. The term ‘pro-
jected solid angle’ stems from the fact that the projection of the small area
dA0 at the origin weights the contribution of the projected area P . The
reason for doing this is described in Section 2.6.
The projected solid angle is calculated by considering a small area on
the sphere as follows:
d2 P = rdθ × hdα
= r2 dθ sin θdα. (2.9)
Weigh the surface area with a factor cos θ and integrate over 0 ≤ θ ≤ Θ
and 0 ≤ α ≤ 2π rad to obtain
2π Θ
cos θr2 dθ sin θdα
Ω = (2.10)
0 0 r2
Θ
= 2π cos θ sin θdθ
0
= π sin Θ. 2
(2.11)
For Θ = π/2 rad, the cone covers a full hemisphere, but the projected
solid angle is only π sr, one-half of the geometric solid angle. This dif-
ference is due to the fact that the area d2 P is weighted by the cosine of
θ.
For cone half-apex angles Θ < 0.2 rad, the geometric solid angle and
projected solid angle are numerically similar because cos θ is near unity
32 Chapter 2
7
6 2p Geometrical
Apex
solid angle
Solid angle [sr]
5 angle
3 p
2 Projected
solid angle
1
0
0 0.2 0.4 0.6 0.8 1 1.2 1.4 1.6
Half-apex angle [rad]
and the projection effect is negligible. Figure 2.8 shows the geometrical
and projected solid angle magnitude for a cone.
The projected solid angle of a rectangular flat surface, as seen from a ref-
erence area centered above the surface and parallel to the surface, is de-
termined by the integral of the projected area of a small elemental area
Introduction to Radiometry 33
Reference point
H θ
dw
dd
W
D
Figure 2.9 Solid angle of a centered flat surface.
cos θ dd dw, weighted by an additional cos θ across the full size of the sur-
face:
dw dd cos2 θ
Ωs =
W D R2
dw dd cos4 θ
=
W D H2
dw dd H 4
= 2
W D H R
4
dw dd H
= √ . (2.14)
W D H
2
w 2 + d2 + H 2
Note the cos4 θ term in the integral. At large θ, i.e., when two large
surfaces are located relatively close together, the contribution of the ex-
treme areas of the surfaces reduce considerably compared to the central
areas. In optical systems, the cos4 θ effect results in lower image flux to-
ward large field angles — the image flux is not constant across the whole
image field.
S
√
H 2 − r2
r
R
θ
h
t s α P
H = t+s+h
√
H 2 + r2
The solid angle of the cone, defined by the disk at the center of the
sphere with normal vector pointed to P and with half-apex angle α, is
given by
Ωd = π sin2 (arcsin r/ H 2 + r2 ) = π r2 /( H 2 + r2 ). (2.17)
Ωs r2 + H 2
= . (2.18)
Ωd H2
This ratio is always greater than one, hence the solid angle when viewing
a sphere from a distance is always greater than the solid angle of a disk
at the center of the sphere (both referenced from P ). For increasing H the
ratio approaches unity, but at close range the ratio increases.
For the smallest value of H (H = r) the ratio is two. At this lo-
cation, the surface of the sphere, the projected solid angle of the sphere
is π because the sphere appears locally as an infinitely large, almost-
planar surface. At the same location, the disk-projected solid angle is
π sin2 (π/4) = π/2 sr.
θ0
dA0
R01 θ1
dA1
Figure 2.11 Geometrical construction for radiative flux between two elemental areas.
For a lossless medium, the flux flowing between the source and re-
ceiver is given by the product of the (invariant) radiance and the projected
areas of the source and receiver, divided by the square of the distance
between the areas. Note that on the right of the equation, there is only
one radiometric quantity, L; the remaining quantities are all geometric
(nonradiometric) quantities. Radiometry is therefore as much a study of
geometry as it is of optical flux. Equation (2.25) pertains to the flux flow-
ing through the two surfaces; it does not yet include the effects of source
emissivity or receiver absorption (see Section 3.2).
Equation (2.25) can be used to derive all radiometric quantities as fol-
lows:
Irradiance is derived as (note the cos θ1 term)
dΦ L dA0 cos θ0 cos θ1
E= = = L Ω0 cos θ1 . (2.26)
dA1 R201
38 Chapter 2
Intensity is derived as
dΦ R201
I= = L dA0 cos θ0 cos θ1 . (2.27)
dA1
where Φ1 is the flux at dA1 , L0 is the radiance at dA0 , and τ01 is the medium
transmittance between the two elemental areas. In this case radiance is not
invariant because of the medium loss. This simple model is accurate for
most cases where the path radiance contribution is negligible compared to
the flux in the field radiance.
Equation (2.25) calculates the flux flowing between two infinitely small
areas. The flux flowing between two arbitrary shapes can be calculated
by integrating Equation (2.25) over the source surface and the receiving
surface. In the general case, the radiance L cannot be assumed constant
over A0 , introducing the spatial radiance distribution L(dA0 ) as a factor
into the spatial integral. 9 Likewise, the medium transmittance between
any two areas dA0 and dA1 varies with the spatial locations of dA0 and
dA1 — hence τ01 (dA0 , dA1 ) should also be included in the spatial integral.
The integral can be performed over any arbitrary shape, as shown
in Figure 2.12, supporting the solution to complex geometries. Clearly
Introduction to Radiometry 39
A0 A1
θ0
θ1
dA0 R01
τ01 dA1
λ
The
rma
l rad m
iato iu
r
ed
L dl λ M
λ rc
e
u
o
S
Figure 2.13 Describing the electro-optical system as a thermal source and a series of spec-
tral filters.
7. The detector converts the optical flux to an electrical signal by the scalar
value of its responsivity (see Chapter 5).
and projected surface area dA0 cos θ0 , through a receiver with projected
surface area dA1 cos θ1 at a distance R01 , with a transmittance of τ01 be-
tween the two surfaces. The transmittance τ01 now includes all of the
spectral variables in the path between the source and the receiver.
To determine the total flux flowing from elemental area dA0 through
dA1 over a wide spectral width, divide the wide spectral band into a large
number N of narrow widths Δλ at wavelengths λn and add the flux for all
of these narrow bandwidths together as follows:
N
L0λn dA0 cos θ0 dA1 cos θ1 τ01λn Δλ
d Φ= ∑
2
. (2.32)
n =0 R201
Equation (2.33) describes the total flux at all wavelengths in the spec-
tral range λ1 to λ2 passing through the system. This equation is developed
further in Chapter 7.
This equation describes the flux flowing between two small areas. The
source radiates power in all directions, and in order to determine the
power radiated from the source into the hemisphere, one must integrate
the receiver area dA1 cos θ1 over the whole hemisphere. In order to per-
form the integration, construct a hemispherical dome with radius r and
its center at the elemental source area dA0 , as shown in Figure 2.7. The
projected receiver area can be written as
Assume a constant radiance over the small elemental source area and
integrate over the complete hemisphere to obtain
2π π
2 L dA0 cos θ0 r2 dθ0 sin θ0 dα
ΦH = (2.36)
0 0 r2
2π π
2
= LA0 cos θ0 sin θ0 dθ0 dα
0 0
π
2
= LA0 2π cos θ0 sin θ0 dθ0
0
= LA0 π, (2.37)
then
Φ
M= = Lπ. (2.38)
A0
This result indicates that the exitance of a Lambertian radiator is related
to radiance by the projected solid angle [Equation (2.12)] of π sr, not the
geometric solid angle of 2π sr. Why? In Equation (2.36) there is a cos θ0
term describing the projected area of the source. A flat Lambertian source
therefore radiates with a cosine distribution and not isotropically in all direc-
tions. The projected solid angle is effectively calculated by weighting the
projected area of the source with the cos θ0 term; see Section 2.5.3.
The above derivation indicates that one should always use the projected solid
angle instead of the geometric solid angle when dealing with Lambertian sources.
In the event that the solid angle under consideration is less than the full
hemisphere, use the equation presented in Section 2.5.3. Rather than mem-
orizing rules, it is better to perform the calculation from first principles —
it is easier to remember Equation (2.25) than to remember a multitude of
rules.
Consider a Lambertian source as shown in Figure 2.14(a). As the area
dA is rotated, the projected area along the line to the observer decreases
as dA cos θ. For a small source with area dA, when viewed at an arbitrary
angle θ, the intensity varies as I = L dA cos θ. However, in Figure 2.14(b),
Introduction to Radiometry 43
θ θ dA
dA cosθ S
(a) (b)
Figure 2.14 Lambertian sources: (a) finite-size tilted surface, and (b) infinite-size tilted sur-
face behind a finite size aperture.
the area S is infinitely large compared with the window opening dA. In
this case the observer always sees an area dA, independent of the angle
by which the source is rotated. In the case of Figure 2.14(b), the observed
intensity is given by I = L dA.
It is evident from Equation (2.25) that the amount of flux transfer between
two surfaces dA0 and dA1 has a geometrical term and a radiometric term.
Equation (2.30) shows that the geometry can be calculated over areas A0
and A1 . A flux transfer calculation is therefore as much a geometrical cal-
culation as it is a radiance calculation. On condition that (1) the source
spatial radiance is uniform, (2) there are no medium losses, and (3) the
receiving spatial area is uniform, the radiance term can be mathemati-
cally separated from the purely geometrical term. The radiometric term
is a function of the field only (irrespective of space), whereas the geomet-
ric term is spatial geometry only (irrespective of radiance field consider-
ations). The calculation of view factors can be seen as a form of spatial
normalization.
The efficient transfer of heat is of prime importance when designing
furnaces. The heat-transfer community developed a detailed mathematical
concept for spatial geometric integrals, called the configuration factor, view
factor, diffuse shape factor, or angle factor. 23 Tables of configuration factor
values are precalculated for given geometrical configurations.
Assuming diffuse Lambertian surfaces for A0 and A1 , the view factor
is the portion of all of the flux (in the hemisphere, hs) leaving A0 that
passes through A1 and is given by
A L0 cos θ0 dA0 cos θ1 dA1 /R01
2
FA0 → A1 = 0 1
A
2.9.1 A disk
where θ0 is the angle of dΩ1 with respect to the disk normal vector. In
the construction of Figure 2.7, dΩ1 = d2 P /r2 = dθ sin θdα. Integrate over
0 ≤ θ ≤ π/2 sr and 0 ≤ α ≤ 2π sr to obtain
2π π/2
M
= cos θdθ sin θdα
L 0 0
π/2
= 2π sin θ cos θdθ
0
= π sin (π/2) = π,
2
(2.45)
which is just another way to derive Equation (2.38).
2.9.2 A sphere
The normalized spectral response of the eye is called the relative luminous
efficiency. The exact shape of the relative luminous efficiency depends on
the light level. The two extremes of relative luminous efficiency are known
as photopic (high luminance levels) and scotopic (low luminance levels)
luminous efficiencies. Unless otherwise specified, photometric values are
normally specified for the photopic spectral response.
If the luminance exceeds about 3 lm/(m2 ·sr), the eye is light-adapted,
and the cones in the retina are operating. Under light-adapted conditions,
the eye’s spectral response is called photopic. In photopic vision, the eye
has color discrimination and acute foveal vision. The standard photopic
luminous efficiency is shown in Figure 2.15 and Table A.4. The spectral
shape of photopic luminous efficiency is defined in tabular form 24,25 but
can be roughly approximated by
If the luminance is less than 3×10−5 lm/(m2 ·sr), the eye is dark-
adapted. The cones are no longer sensitive, and the rods sense the light.
Under dark-adapted conditions, the eye’s spectral response is called sco-
topic. Under scotopic vision, the eye is not sensitive to color and has no
foveal vision. The standard scotopic luminous efficiency is shown in Fig-
ure 2.15 and Table A.4. The spectral shape of scotopic luminous efficiency
is defined in tabular form 24,25 but can be roughly approximated by
Equations (2.48) and (2.49) are not accurate at the extreme wavelength
limits of the spectral bands. These approximations should be used with
care if the source has significant amounts of flux at the wavelength lim-
its. An example of one such case is the eye viewing a blackbody at low
temperatures.
At luminance levels between 3×10−5 lm/(m2 ·sr) and 3 lm/(m2 ·sr),
the spectral response is somewhere between the photopic and scotopic,
referred to as mesopic vision.
Introduction to Radiometry 47
1
Photopic
Scotopic
0.8
Relative efficiency
0.6
0.4
0.2
0
350 400 450 500 550 600 650 700 750
Wavelength [mm]
where Kλ = Kmax Vλ is the spectral scotopic efficacy, Vλ is the scotopic
efficiency, Kmax = 1700 lm/W is the maximum value of scotopic efficacy
referenced to a 2042-K blackbody standard source, and Leλ is the source’s
radiance.
Literature sometimes shows different values for Kmax and Kmax be-
cause the values changed every time a new standard source was instituted
or a modification was made to the efficiency curves.
Efficacy can be defined as a spectral variable, as in the previous equa-
tions, or over a wide spectral band. The total (wideband) luminous efficacy
48 Chapter 2
where x̄λ , ȳλ , and z̄λ are the color-matching functions of the CIE standard
colorimetric observer, as shown in Figure 2.16. The xyz chromaticity color
coordinates can then be calculated by
X
x = , (2.58)
X+Y+Z
Y
y = , and (2.59)
X+Y+Z
Z
z = = 1 − x − y, (2.60)
X+Y+Z
where x and y define the color coordinates in the xy chromaticity diagram
(Figure A.1). Valid color coordinates are all inside the closed curve. The
U- or dome-shaped part of the perimeter describes the monochromatic
rainbow colors, calculated from the above equations with Lλ = 1. The
color of a Planck radiator (see Section 3.1) can be calculated from the above
Introduction to Radiometry 49
2
1.8
1.6 z (blue)
Tristimilus response
1.4
1.2 y (green)
1 x (red)
0.8
0.6
0.4
0.2
0
0.35 0.4 0.45 0.5 0.55 0.6 0.65 0.7 0.75
Wavelength [mm]
1
Sunlight (5900 K)
Normalized radiance
Incandescent (2850 K)
0.5
Sodium
Fluorescent
0
0.35 0.4 0.45 0.5 0.55 0.6 0.65 0.7 0.75
Wavelength [mm]
1
White
paper
Sample reflectance
0.8
Lettuce
Blue
0.6 Nitrile
0.4 Tomato
0.2
Yellow prune Green leaf
0
0.35 0.4 0.45 0.5 0.55 0.6 0.65 0.7 0.75
Wavelength [mm]
Bibliography
[1] Wikipedia, “Electromagnetic radiation,” http://en.wikipedia.org/
wiki/Electromagnetic_radiation.
[4] Wolfe, W. L. and Zissis, G., The Infrared Handbook, Office of Naval
Research, US Navy, Infrared Information and Analysis Center, Envi-
ronmental Research Institute of Michigan (1978).
[5] Pinson, L. J., Electro-Optics, John Wiley & Sons, New York (1985) [doi:
ISBN 0-471-88142-2].
[7] Boyd, R. W., Radiometry and the Detection of Optical Radiation , John
Wiley & Sons, New York (1983).
[8] Bayley, F. J., Owen, J. M., and Turner, A. B., Heat Transfer, Nelson
Publishers, London (1972).
[20] Palmer, J. M. and Grant, B. G., The Art of Radiometry, SPIE Press,
Bellingham, WA (2009) [doi: 10.1117/3.798237].
[21] Accetta, J. S. and Shumaker, D. L., Eds., The Infrared and Electro-Optical
Systems Handbook (8 Volumes), ERIM and SPIE Press, Bellingham, WA
(1993).
[23] Modest, M. F., Radiative Heat Transfer, Academic Press, San Diego, CA
(2003).
[24] Colour & Vision Research Laboratory, “Colour and Vision Database,”
http://www.cvrl.org/index.htm.
Introduction to Radiometry 53
[26] Schanda, J., Ed., Colorimetry: Understanding the CIE System, Wiley-
Interscience, New York (2007).
Problems
2.2 Given the two equations 4π sin2 2θ and π sin2 θ, explain what they
mean and which is which. [2]
Calculate the geometric solid angle and projected solid angle for
the following half-apex angles: θ ∈ {0, 0.01, 0.1, 1, π/2} rad. Com-
pare these values in a table and explain why they are different.
[2]
2.3 A room has floor dimensions of 5 m by 5 m. The roof is 3.5 m
above the floor. There is a sensor mounted exactly in the center of
the roof. Calculate both the geometric and projected solid angles
of the floor as seen from the location of the sensor. Draw a picture
of the room, showing all of the details. Derive a mathematical de-
scription of the solid angle. Calculate the solid angle numerically
(not analytically). [10]
2.4 Starting from first principles, derive the solid angle through the
optical aperture (opening) of a Cassegrain telescope, as seen from
54 Chapter 2
the detector in the focal plane. Provide all steps of all mathemati-
cal derivations. The telescope is as follows:
dp
ds
Detector
Secondary Primary
mirror mirror
2.5 Describe, in graphical and mathematical terms, how the flux trans-
fer between two arbitrary objects can be calculated. [4]
2.6 A circular disk and sphere of radii r are located at a distance R
from an infinitely large wall, with uniform radiance L. The normal
vector from the disk is normal to the wall. Derive mathematical
formulations that describe the flux flowing from the wall through
the disk and sphere. [6]
dA’1 Disk
dA0 r
θ0
r
dA1
R/cos θ0
dA’1 Sphere
R
2.9.1 Although not shown in the graph, it could well be that a dark-
green leaf has the same color coordinates as an apparently lighter-
green leaf. Consider the concept that artists call the ‘value’ of a
color. To what extent are the color coordinates affected by the
value of the color? [2]
2.9.2 Review Equations (2.57) and (2.60) and consider the effect of ab-
solute irradiance on a sample on the color coordinate of that sam-
ple. Do the color coordinates change with respect to illumination
level? Support your answer with proper mathematical derivation.
To what extent will light leakage into the sample affect the mea-
surement of absolute spectral reflectance? What effect will this
have on the color coordinates of the sample? [4]
The Planck radiation law is derived in detail in several references. 1–3 The
discussion presented here aims to convey key insights, rather than rigorous
mathematical terminology.
Physical matter (atoms and molecules) at nonzero absolute tempera-
ture (T = 0 K) emits and absorbs electromagnetic radiation. Under ther-
modynamic equilibrium, the incident electromagnetic field and the atoms
are in continual energy exchange, mutually sustaining each other’s energy
state by photon emission and absorption. If either the electromagnetic field
intensity or the object’s temperature should change, the energy exchange
will adjust until thermodynamic equilibrium is re-established.
Perfect thermodynamic equilibrium could exist in an enclosed cavity
with the homogenous walls at uniform temperature, such as inside a hol-
low ball [shown in Figure 3.1(a)]. As shown in the figure, there are two
sources of energy: the radiation field [indicated by (1)] and the cavity walls
[indicated by (2)]. At any single frequency ν the electromagnetic radiation
field is sustained by the wall photon emission [indicated by (3)] at a single
energy transition hν at atomic level. Likewise, atoms radiate photons at a
specific frequency corresponding with the energy transition in the atom,
hν. The wall’s kinetic temperature T is sustained by absorption of photons
from the electromagnetic field [indicated by (4)].
Thermodynamic equilibrium means that there is zero energy flow Q̇
and zero mass flow ṁ across the system’s boundary [indicated by (5)].
Hence, the only energy exchange is between the radiation field and the
57
58 Chapter 3
(2)
=1
T
T1
(4) L2
L1
T (1) T1 T2 Q
(3) hν
L1
.
. m=0
. .
. m=0
Q=0 Q=0 Q>0
(5)
(a) (b) (c)
Figure 3.1 Concepts behind Planck’s law and thermal radiation: (a) closed cavity, (b) initial
non-equilibrium, and (c) permanent non-equilibrium.
cavity walls. Einstein showed 4 that the principle of detailed balance re-
quires that the processes of spontaneous emission, stimulated emission,
and photo-absorption [indicated by (3) and (4)] are in equilibrium. These
three processes ensure equilibrium at a single frequency for a given photon
but also collectively for all photons at all frequencies.
The permitted energy at each radiative and absorptive frequency must
be multiples of the photon’s energy hν, produced by the transition between
two energy states in the atoms or molecules. The relative numbers of the
allowed energy transitions are, in turn, related by the Boltzmann proba-
bility distribution p(n) = (1 − exp(− x)) exp(−nx), where x = hν/(kT ),
which is a function of temperature T. The Boltzmann probability distri-
bution provides the link between the wall temperature T and the photon
frequency (spectral) distribution.
The final concept in the thermal radiator discussion involves the reso-
nance modes (standing waves) supported in the cavity, also known as the
density of states. Density of states is also used in the derivation of the
electronic wave function in a crystal (Section 5.5.3) and thermal noise (Sec-
tion 5.3.2). Density of states is beyond the scope of the current discussion.
For more details see complete derivations. 1–3
Planck’s law applies to transition energy levels compliant with the
Boltzmann probability distribution and thus does not apply to lasers, LEDs,
fluorescence, or radioactivity (transition levels not compliant with Boltz-
mann probability distribution). Planck-law radiation is isotropic, spatially
homogeneous, unpolarized, and incoherent.
Sources 59
2πhc2 c1eλ
Meλ ( T ) =
=
, (3.1)
λ5 ehc/(λkT ) −1 λ5 e /(λT )
c 2λ −1
where T is temperature in [K], and λ is wavelength in [m]. Exitance Meλ ( T )
is in units of [W/m3 ] or [W/(m2 ·µm)] (depending on the value of c1eλ ).
This is a spectral exitance in watts per square meter, per wavelength inter-
val. The values of c1eλ and c2λ are given in Table A.2.
The derivative with respect to temperature of spectral radiant exi-
tance, for a given temperature T, as a function of wavelength λ, with
x = c2λ /(λT ), is given by
with units [W/(m2 ·m−1 ·K)] or [W/(m2 ·cm−1 ·K)] (depending on the value
of c1eν̃ ). This is a change in spectral exitance in watts, per square meter,
per wavenumber interval, with temperature.
The spectral photon exitance as a function of wavenumber ν of a black-
body at temperature T is given by
with units [q/(s·m2 ·m−1 ·K)] or [q/(s·m2 ·cm−1 ·K)] (depending on the value
of c1qν̃ ). This is a change in spectral exitance in photons per second, per
square meter, per wavenumber interval, with temperature.
62 Chapter 3
2πhν3 c ν3
Meν ( T ) = hν = c 1eν , (3.9)
c2 e kT − 1 (e 2ν ν/T − 1)
with units [W/(m2 ·Hz)]. This is a spectral exitance in watts per square
meter, per frequency interval.
The derivative with respect to temperature of radiant exitance, for a
given temperature T, as a function of frequency ν, with x = c2ν ν/T, is
given by
2πν2 c1qν ν2
Mqν ( T ) = = , (3.11)
hν
c2 e kT − 1 (ec2ν ν/T − 1)
with units [q/(s·m2 ·Hz)]. This is a spectral exitance in photons per second,
per square meter, per frequency interval.
The derivative with respect to temperature of spectral photon exitance,
for a given temperature T, as a function of frequency ν, with x = c2ν ν/T,
is given by
From Figure 3.2 it is clear that the Planck-law radiation curve has only one
maximum. The equation relating the blackbody temperature and the spec-
tral value at the peak exitance is known as Wien’s displacement law. The
Sources 63
a3 kT
ν̃me = = weν̃ T, (3.14)
100hc
and
a3 kT
νme = = weν T. (3.15)
h
a2 kT
ν̃mq = = wqν̃ T, (3.17)
100hc
and
a2 kT
νmq = = wqν T. (3.18)
h
If Planck’s law is integrated over all wavelengths, the total radiant exitance
from a blackbody is obtained: 3
2k4 π5 4
Me ( T ) = T = σe T 4 , (3.19)
15c2 h3
with exitance Me ( T ) in [W/m2 ], σe the Stefan–Boltzmann constant in units
of [W/(m2 ·K4 )], and temperature T in [K]. Note that the Stefan–Boltzmann
law does not consider energy balance between incident flux and radiated
flux — it assumes the environment is at 0 K, with no incident flux.
64 Chapter 3
Planck’s law can be written in the form of an infinite sum. 5 Consider the
integral of the Planck radiation law as a function of wavelength. Starting
from first principles, where c1 = 2πhc2 , c2 = hc/k, and x = c2 /(λT ),
λ2
2c2 h dλ
L =
λ1 λ 5 ( e x − 1)
λ2
c1 dλ
= . (3.21)
λ1 πλ (ex − 1)
5
where ex − 1 is expanded into the infinite sum through long division. Ap-
plying integral number 2.322.3 in Gradshteyn, 7
x1
c1 T 4 ∞ −mx x3 3x2 6x 6
L = ∑e
πc42 m=1
− − 2 − 3− 4
m m m m
x2
x2
4 ∞
e − mx
c1 T
= ∑
πc2 m=1 m
4 4
( xm ) 3
+ 3 ( xm ) 2
+ 6xm + 6
.
(3.23)
x1
3.2 Emissivity
This section covers the emissivity concept. The work started here is ex-
panded in Section 6.6, which considers the modeling of thermal radiators.
The Planck radiator (or blackbody) is a very convenient basis for modeling
a large class of sources. However, none of these sources behaves exactly
like a blackbody radiator. In order to use Planck’s law with real sources, it
66 Chapter 3
Wavelength Wavenumber
λ in [µm], T in [K] ν̃ in [cm−1 ], T in [K]
x = c2λ /(λT ) = 14387.8/(λT ) x = c2ν̃ ν̃/T = 1.43878 ν̃/T
c1eλ = 3.74177 × 108 c1eν̃ = 3.74177 × 10−8
c1qλ = 1.88365 × 1027 c1qν̃ = 1.88365 × 1015
Planck’s law
c1eλ c1eν̃ ν̃3
Meλ = Meν̃ =
λ ( e x − 1)
5 ( e x − 1)
in [W/(m2 ·µm)] in [W/(m2 ·cm−1 )]
Figure 3.2 Planck’s law and Wien’s displacement law for various temperatures.
6000 K
25
Exitance [q/(s×m2×mm)]
22
6000 K 10
Exitance [q/(s×m2×cm -1 )]
10 3000 K
3000 K 1800 K
23
10
21
1800 K
10 1000 K
650 K
1000 K
21 450 K
10
20 650 K 10 300 K
450 K
300 K 200 K
19 19
10 200 K 10
Wien’s law
Wien’s law
18 17
10 10
2 3 4 5 -1 0 1 2
10 10 10 10 10 10 10 10
Wavenumber [cm-1 ] Wavelength [mm]
Exitance [W/(m2×mm)]
Exitance [W/(m2×cm-1 )]
2 3000 K 3000 K
10 10
5
1800 K 1800 K
1
10 1000 K 3 1000 K
10 650 K
0 650 K
10 450 K 450 K
1
-1 300 K 10 300 K
10 200 K
200 K
-1
10
-2 10
Wien’s law
Wien’s law
Sources
-3
10
2 3 4 5 -1 0 1 2
10 10 10 10 10 10 10 10
Wavenumber [cm-1] Wavelength [mm]
Blackbody photon rate dM/dT Blackbody photon rate dM/dT
Chapter 3
19 24
2
6000 K
1
Blackbody radiant dM/dT 5
Blackbody radiant dM/dT
Exitance derivative [W/(K×m ×cm )]
10 10
-1
0 6000 K
10
2
3
3000 K
3000 K 10
-1 1800 K 1800 K
10 2
10
1000 K
1000 K
-2 650 K 1
10 450 K
10 650 K
300 K 0 450 K
-3 200 K
10
10 300 K
-1
10
200 K
-4
10 -2
2 3 4 5 10
10 10 10 10 -1 0 1 2
10 10 10 10
-1
68
108
6000 K
[W/m2·sr·mm]
surface of
Radiance
6000 K
0.5 surface of 2856 K
the sun 1336 K 800 K jet 500 K 300 K
tungsten tailpipe kitchen
lamp melting warm 250 K
gold oven day (–20 °C)
very cold
night
0.2 0.3 0.4 0.6 0.8 1 2 3 4 5 6 7 8 9 10 20
Wavelength [mm]
Emissivity has many guises, to allow for spectral and directional pa-
rameters, that are necessary to describe emissivity in a more general sense.
Table 3.2 provides a short summary of the various definitions of emissiv-
ity. The source radiance is indicated by the subscript s , and the theoretical
blackbody radiance is indicated by the subscript bb . The source and black-
body temperatures are equal, Ts = Tbb . The two variables θ and ϕ denote
directional zenith and azimuth angles, respectively. A much more detailed
description is given in Palmer and Grant. 11 Directional emissivity and re-
flectance are investigated in more detail in Section 3.4.
0 0
Lbbλ ( θ,ϕ) cos θ dϕ dθ
∞ π/2 2π
L λs ( θ,ϕ) cos θ dϕ dθ dλ
Hemispherical total emissivity ∞
0
π/2
0
2π
0
θ0
dA0 R01 θ1
0
dA1
α1
Figure 3.5 Radiative flux between a source and receiver.
1.0
Blackbody
Emissivity
Grey body
Selective
radiator
0
1
Blackbody
Normalized
Exitance
Grey
body
Selective
radiator
0
0 5 10
Wavelength [mm]
Figure 3.6 Blackbody, grey body, and selective thermal radiators.
0.5
0.5
0
0
0.6 0.8 1 2 3 4 5 6 7 8 9 10
Wavelength [mm]
Bunsen-burner flame CO2 emissivity
1
Emissivity
0
3 3.5 4 4.5 5 5.5 6
Wavelength [mm]
A1 1T1
1 L 1
(1-2) 1 L1
2 L 2
A2 2T2
L2 = 2 Le ( T2 ) + (1 − 2 )1 Le ( T1 ). (3.26)
θ
Reflections
D
d
(a) (b)
Figure 3.9 Multiple reflections in a cavity: (a) simple air-filled cavity, and (b) reflectance of a
light ray entering the graphically unfolded cavity.
1 2
Ds Da Dd
Ra Rd
Figure 3.11 assumes an illumination beam size comparable with the mi-
croscopic scale of a homogenous surface. A similar principle also applies
to aggregate properties of composite surfaces. Consider a satellite cam-
era viewing a crop field. In this case the sun is the illuminating ‘beam,’
and a single satellite-sensor pixel FOV observes an area comprising crop
and soil. The camera pixel FOV footprint covers several rows of the crop.
The observed radiance will vary depending on the sun-field-camera view-
ing geometry, i.e., illumination and/or viewing along or across the rows
of plants. Thus, depending on the application, roughness is not only ex-
pressed in terms of wavelength scale but also in FOV footprint scale.
Sources 77
The geometry describing reflection is shown in Figure 3.12. Let the in-
cident ray be defined by the unit vector I, the surface normal vector by
and the reflected ray vector by the unit vector R.
the unit vector N, It is
shown in Appendix C that the direction of the reflected ray is given by
=
R I − 2( ) N.
I·N In this definition R represents the mirror reflection
vector from the surface. Consider now the direction of the measured flux
The vector S has an angle α with
as reflected along any arbitrary vector S.
(where cos α = S · R)
respect to the mirror reflection vector R and an angle
θs with respect to the surface normal vector N (where cos θs = −S · N).
The incident ray vector I has an angle θi with respect to the surface normal
vector N (where cos θi = − The mirror reflected ray vector R
I · N). has an
angle θr with respect to the surface normal vector N (where cos θr = R · N).
Normal N N
vector R Mirror
reflection
Incident
ray ωs L S
I θr Reflected LI ωi
S ray θi θs
θi α
θs
ϕr
ϕi
ϕs
The Fresnel equation for reflection from a dielectric or metal surface de-
pends on the polarization of the incident light relative to the plane of
the surface. 20,22 For polarized light perpendicular to the surface, the re-
flectance from a single surface is given by
ni cos θi − nt cos θt 2
ρ⊥ = , (3.28)
ni cos θi + nt cos θt
Sources 79
1
Al
Ag
Reflectance
0
0.1 1 10
Wavelength [mm]
where θi is the angle of incidence of the light ray (relative to the surface
normal vector), ni is the refractive index of the medium hosting the inci-
dent ray, θt is the angle between the refracted ray and the surface normal
vector, and nt is the refractive index of the dielectric or metal.
The reflectance of polarized light parallel to the surface is given by
nt cos θi − ni cos θt 2
ρ = . (3.29)
nt cos θi + ni cos θt
The reflectance of unpolarized light is given by
ρ + ρ⊥
ρ= . (3.30)
2
The relationship between incident angle and refracted angle is given by
Snell’s law (see Section 5.5.8),
ni sin θi = nt sin θt , (3.31)
and the transmittance through the surface is given by τ = 1 − ρ because
there is no absorption in the surface itself. Figure 3.14 shows the angu-
lar variation of Fresnel reflectance for a number of dielectric and metallic
materials.
The Fresnel reflection equations, Equations (3.28) and (3.29), provide
the reflection from a single surface, such as from an opaque surface. Trans-
parent dielectric media, such as a plate of glass with two smooth surfaces,
will reflect on both surfaces (see the figure in Problem 3.9). Assuming the
same medium on both sides of the dielectric plate (i.e., air), the second
surface reflects the same as the first surface. The medium’s reflectance can
be calculated by accounting for the successive reflectance by each surface,
as well as the transmittance through the medium.
80 Chapter 3
0.5
Gold: nt = 0.855 + i 1.8955
Figure 3.14 Fresnel reflection as function of incidence angle (single surface in air). 21
The reflection from surfaces with 0 < σ/λ < 1 does not follow the geomet-
ric laws of reflection; a more-complex function is required. The BRDF 11,23
defines how light is reflected at an irregular or rough surface, such as
shown in Figures 3.11(b) and (c). BRDF varies with wavelength — in the
following discussion the monochromatic BRDF at a single wavelength is
considered.
BRDF is defined as the ratio of reflected radiance LS(dωs ) in a small
solid angle dωs along a view vector S to the incident irradiance EI (dωi ) in
a small solid angle dωi along the incidence vector I. Note that an infinites-
imally small solid angle dωi is considered and furthermore that the source
surface with radiance L I uniformly fills dωi ; hence the source surface ori-
entation is irrelevant, and L I dΩi = L I dωi .
As shown in Figure 3.12, the direction of each of the two small solid
angles is defined by the respective azimuth angles ϕi and ϕs and the zenith
angles θi and θs . BRDF is therefore a four-dimensional function. Defined
as L/E, BRDF has units of [1/sr]:
LS(dωs ) LS(dωs )
BRDF = f r (dωi → dωs ) = = . (3.32)
EI (dωi ) L I (dωi ) cos θi dωi
Some BRDFs are isotropic when rotated around the normal N, yielding a
three-dimensional function fr (θi , θs , ϕi − ϕs ), whereas others are anisotropic.
The theoretical requirements for the BRDF function include positivity:
fr (θi , θs , ϕi , ϕs ) ≥ 0; (3.33)
Sources 81
consideration. From Equations (3.32) and (3.37) the reflected radiance from
a mirror is
ρ s ( θ i ) δ ( θ i − θ s ) δ ( ϕ i − ϕ s ± π)
LS(dωs ) = L I (dωi ) cos θi dωi
cos θi
= ρs (θi ) L I (dωi ). (3.38)
ρd = 0.49
ρs = 0.05
n = 30
ρd = 0.3
ρs = 0.2
n = 50
Pure rubber
Pearl paint
Red fabric 2
Fruitwood 241
wood stain
1
1 mm
0.9 4 mm
Reflectivity
0.8 10 mm
0.7 0.5 mm
0.6
0.5
0.5
0.4
Emissivity
0.5 mm
0.3
0.2 10 mm
4 mm
0.1 1 mm
0
0 0.2 0.4 0.6 0.8 1 1.2 1.4 1.6
Incidence angle [rad]
Figure 3.17 Directional reflectance and emissivity for a smooth, gold surface.
noting that τ = 0 except for very thin layers. Figure 3.17 shows the di-
rectional reflectivity and emissivity of a smooth, gold surface, calculated
from refractive index data. 21 It is evident that the surface emissivity is ap-
proximately constant for incidence angles up to 0.8 rad, but the angular
variation at larger incidence angles could increase (at longer wavelengths)
but eventually decrease to zero at zero-grazing angles. Reviewing the Fres-
nel reflectance for dielectrics in Figure 3.14, it is evident that the dielectric
surface emissivity ( = 1 − ρ) stays approximately constant up to incidence
angles of 1 rad, whereafter it decreases to zero for a π/3 rad incidence an-
gle.
Dielectrics and metals both exhibit directional emissivity, as shown
above. It is evident from Figure 3.17 that at longer infrared wavelengths,
the hemispherical emissivity will not differ significantly from the direc-
tional emissivity at normal incidence angle. 41 The ratio of hemispherical
emissivity to directional emissivity r = /(0, 0) for metals is rarely out-
side 1 ≤ r ≤ 1.3 except at large incidence angles (see Figure 3.17). For
nonconductor dielectrics the ratio is generally 0.95 ≤ r ≤ 1. It is clear from
Figure 3.17 that this generalization does not hold in the visual spectral
band. Note that for a Lambertian radiator r = 1.
In Section 3.4 it is shown how (random) surface roughness affects the sur-
face reflectance. Surfaces with directionally structured roughness may also
exhibit directional reflectance and emissivity. An example of such a struc-
86 Chapter 3
Soil 1
Corn 2
ture is a field with rows of corn, shown in Figure 3.18. The soil has emis-
sivity 1 , whereas the corn has emissivity 2 . In the along-row direction
(A), the projected area of the corn is small, and the observer sees mostly
ground. In the cross-row direction (C), at low-elevation angles, only the
corn is visible and not the ground. Directional emissivity is observed in en-
vironments such as crop lands, snow, ground quartz sand, i.e., any surface
with one or more materials in ordered spatial structure.
the surface, and θi is the angle between the surface normal and the sun
vector. The reflected sun radiance from a perfectly Lambertian surface is
then given by Lλ = ρd Eλsun /π, where ρd is the surface diffuse reflectance
function. The reflected radiance is then given by
s Lbbλ ( Ts ) As τs ρd cos θi
Lλ = (3.46)
πR2sun
= ψs Lbbλ ( Ts )τs ρd cos θi , (3.47)
where ψ = Asun /(πR2sun ) = 2.1757 × 10−5 [sr/sr] follows from the geome-
try. The sun geometry factor is an inverse form of the view factor described
in Section 2.8.
Bibliography
[1] Dereniak, E. L. and Boreman, G. D., Infrared Detectors and Systems,
John Wiley & Sons, New York (1996).
[2] Boyd, R. W., Radiometry and the Detection of Optical Radiation , John
Wiley & Sons, New York (1983).
[5] Wolfe, W. L. and Zissis, G., The Infrared Handbook, Office of Naval
Research, US Navy, Infrared Information and Analysis Center, Envi-
ronmental Research Institute of Michigan (1978).
[7] Gradshteyn, I. S. and Ryzhik, I. M., Tables of Integrals, Series and Prod-
ucts, Academic Press, New York (1981).
[8] Mohr, P. J., Taylor, B. N., and Newell, D. B., “CODATA recommended
values of the fundamental physical constants: 2010,” Rev. Mod.
Phys. 84(4), 1527–1605 (2012) [doi: 10.1103/RevModPhys.84.1527].
[11] Palmer, J. M. and Grant, B. G., The Art of Radiometry, SPIE Press,
Bellingham, WA (2009) [doi: 10.1117/3.798237].
[15] Wen, C. and Mudawar, I., “Modeling the effects of surface roughness
on the emissivity of aluminum alloys,” International Journal of Heat and
Mass Transfer 49, 4279–4289 (2006).
[18] Matusik, W., Pfister, H., Brand, M., and McMillan, L., “A Data-Driven
Reflectance Model,” ACM Transactions on Graphics 22(3), 759–769 (July
2003).
[19] Snyder, W. C. and Wan, Z., “BRDF Models to Predict Spectral Re-
flectance and Emissivity in the Thermal Infrared,” IEEE TRANS-
ACTIONS ON GEOSCIENCE AND REMOTE SENSING 36, 214–225
(1998).
[20] Hecht, E., Optics, 4th Ed., Addison Wesley, Boston, MA (2002).
[22] Born, M. and Wolf, E., Principles of Optics, 7th Ed., Pergamon Press,
Oxford, UK (2000).
[23] Nicodemus, F. E., Richmond, J. C., Hsia, J. J., Ginsburg, I. W., and
Limperis, T., “Geometrical considerations and nomenclature for re-
flectance,” NBS monograph 160, National Bureau of Standards (Octo-
ber 1977).
[25] Blinn, J. F., “Models of light reflection for computer synthesized pic-
tures,” 4th Annual Conference on Computer Graphics and Interactive Tech-
niques 192 (1977) [doi: 10.1145/563858.563893].
[27] Lafortune, E., Foo, S., Torrance, K., and Greenberg, D., “Non-linear
approximation of reflectance functions,” SIGGRAPH 97 (1997).
[28] Ashikhmin, M. and Shirley, P., “An Anisotropic Phong BRDF Model,”
Journal of Graphics Tools 5, 25–32 (2000).
[29] Cook, R. and Torrance, K., “A reflectance model for computer graph-
ics,” SIGGRAPH 15, 301–316 (1981).
[32] He, X., Torrance, K., Sillon, F., and Greenberg, D., “A comprehensive
physical model for light reflection,” Computer Graphics 25, 175–186
(1991).
[33] Nayar, S. and Oren, M., “Generalization of the Lambertian Model and
Implications for Machine Vision,” International Journal on Computer Vi-
sion 14, 227–251 (1995).
[34] Brady, A. and Kharabash, S., “Further Studies into Synthetic Im-
age Generation using CameoSim,” Tech. Rep. DSTO-TR-2589, Intel-
ligence, Surveillance and Reconnaissance Division Defence Science
and Technology Organisation (2011).
[35] Dudzik, M. C., Ed., The Infrared and Electro-Optical Systems Handbook:
Electro-Optical System Design, Analysis and Testing , Vol. 4, ERIM and
SPIE Press, Bellingham, WA (1993).
[38] Caillault, K., Fauqueux, S., Bourlier, C., and Simoneau, P., “Infrared
multiscale sea surface modeling,” Proc. SPIE 6360, 636006 (2006) [doi:
10.1117/12.689720].
[39] Maxwell, J., Beard, J., Weiner, S., and Ladd, D., “Bidirectional re-
flectance model validation and utilization,” Tech. Rep. AFAL-TR-73-
303, Research Institute of Michigan (ERIM) (October 1973).
[40] Matusik, W., “MERL BRDF Database,” http://www.merl.com/brdf/.
[41] Incropera, F. P., De Witt, D. P., Bergman, T. L., and Lavine, A. S.,
Fundamentals of Heat and Mass Transfer , 6th Ed., John Wiley & Sons,
New York (2007).
[49] Her, T., Finlay, R. J., Wu, C., Deliwala, S., and Mazur, E., “Microstruc-
turing of silicon with femtosecond laser pulses,” Applied Physics Let-
ters 73, 1673–1675 (1998).
[50] Brown, R. J. C., Brewer, P. J., and Milton, M. J. T., “The physical
and chemical properties of electroless nickel-phosphorus alloys and
low reflectance nickel-phosphorus black surfaces,” Journal of Materials
Chemistry 12, 2749–2754 (2002).
[51] Xi, J., Schubert, M. F., , anbd E. Fred Schubert, J. K. K., Chen, M., Lin,
S.-Y., Liu, W., and Smart, J. A., “Optical thin-film materials with low
refractive index for broadband elimination of Fresnel reflection,” Na-
ture Photonics 1, 176 – 179 (2007) [doi: doi:10.1038/nphoton.2007.26].
Sources 91
[52] Tsakalakos, L., Balch, J., Fronheiser, J., Shih, M.-Y., LeBoeuf, S. F.,
Pietrzykowski, M., Codella, P. J., Korevaar, B. A., Sulima, O., Rand, J.,
Davuluru, A., , and Rapol, U., “Strong broadband optical absorption
in silicon nanowire films,” Journal of Nanophotonics 1 (2007) [doi: DOI:
10.1117/1.2768999].
Problems
3.1 Calculate the color coordinates and show the approximate colors
for blackbodies at the following temperatures: 3200 K, 5000 K,
6500 K, and 9000 K. [4]
3.2 Calculate and plot graphically (more than 10 data points) the ra-
diance from the surface of a Planck thermal radiator at a temper-
ature of 1000 K, over the spectral range of 3–5 µm, as follows: (1)
in the wavelength domain (in units of [W/(m2 ·sr·µm)]), (2) in the
wavenumber domain (in units of [W/(m2 ·sr·cm−1 )]), and (3) con-
vert between the results obtained in (1) and (2) above, using the
conversion defined in Section 2.3.3. [6]
3.3 Calculate the amount of heat energy in joules flowing into a beef
steak on a outdoor barbecue grid from the moment it is put onto
the grid, until it is ready to eat. You may consult any source
(except fellow students) but provide the reference to the source.
Clearly state and motivate all assumptions. Apply the golden
rules to the problem, i.e., dimensional analysis, developing a good
mathematical definition, and drawing detailed diagrams of the
problem statement. Include all code or numerical files used in
the calculation. [6]
3.4 Ironsmiths use the color of a steel sample to estimate the temper-
ature of the sample. Subjective descriptions, such as ‘dull cherry-
red,’ are used to describe the temperature, as in the following ta-
ble: 42
92 Chapter 3
3.5 Refer to Section 2.10.5 and repeat the calculations there with your
own code. The data is available on the pyradi website. 48
3.5.1 Calculate the color coordinates of the six samples when illumi-
nated by the four source spectral radiances.
Calculate and plot your own versions of the graphs in Figures 2.17
and 2.18. Enter the color coordinate values into the table below
and comment on your observations. Apply the Golden Rules
(Chapter 10) in the derivation of the solution. [9]
Sources 93
3.6 Calculate the total flux flowing between two circular disks with
diameter 1 m, separated by 10 m. The first object has a tempera-
ture of 450 K and an emissivity of 0.5, and the second object has
a temperature of 450 K and an emissivity ranging from 0 to 1.0 in
steps of 0.1. Derive a mathematical equation for the net flux flow
and then plot the values. [6]
3.7 Calculate the total flux flowing between two circular disks with di-
ameter 1 m, separated by 10 m. The first object has a temperature
of 450 K and an emissivity of 0.5, and the second object has an
emissivity of 0.5 and a temperature ranging from 300 K to 600 K
in steps of 50 K. Derive a mathematical equation for the net flux
flow and then plot the values. [6]
3.8 An opaque object has a diffuse reflectance of 1.0 for wavelengths
from 0 to 0.5 µm; and 0 for wavelengths from 0.5 µm to infinity.
The object is illuminated by a source with temperatures ranging
from 1000 to 4000 K in steps of 1000 K. The object’s own inter-
nal temperature ranges from 1000 to 4000 K in steps of 1000 K.
Assume Planck law radiation in both cases. Calculate and plot
the object’s color coordinates for all of the temperature combina-
tions. Validate your results against Figure A.1. Comment on your
observations. [10]
3.9 In a thin, transparent dielectric medium, the incident flux is re-
flected between the medium’s surfaces, as shown below. The
medium itself has an internal transmittance τi .
94 Chapter 3
τ ρ
τ ρ τ ρ τ ρ τ ρ
τi τi τi τi τi τi τi τi
τ ρ τ ρ τ ρ τ ρ
3.9.1 Show that the reflection and transmittance magnitudes of the top
and bottom surfaces are the same. [4]
3.9.2 Derive an equation for the reflectance (sum of all components) and
transmittance of the medium. [6]
3.10 Calculate the solar irradiance at the top of the atmosphere. [3]
Calculate the total solar flux absorbed by the earth. Assume the
earth’s diameter to be 12,756.8 km. The earth has an average
albedo (reflectance) of 0.3. [5]
3.11 In Section 3.7 a model is developed for the reflected sunlight from
an object on the earth. Confirm Equation (3.47) by applying the
Golden Rule of dimensional analysis. [4]
3.12 Use Figure 3.4 to estimate the percentage of radiance in (1) the 3–
5-µm spectral band for a jet tailpipe, and (2) the 0.4–0.7-µm visual
spectral band for a tungsten lamp. [2]
Repeat, but calculate more accurately using a spectral integral of
the Planck law between the two wavelengths. [4]
3.13 Derive an equation for reflected sun radiance, similar to Equa-
tion (3.47), but for a diffuse surface modeled with the Phong BRDF
function. [2]
3.14 A hemispherical dome of material with emissivity 1 and temper-
ature T1 encloses a small component with area A2 , emissivity 2 ,
and temperature T2 (Figure 3.8). The objective with this investiga-
tion is to consider the net radiance emanating from area A2 . Both
materials are opaque with Lambertian surface properties.
3.15 A hole with diameter D and depth d is drilled into a metal block,
as shown in Figure 3.9. Assume the drill bit point angle to be
exactly 90 deg, forming a hole as shown in the figure. The surface
emissivity of the block and inside the hole is 0.2.
3.15.1 Calculate the emissivity of a hole of d/D = 5, for the ray shown, at
an angle θ ranging from 0 deg to 90 deg at 5-deg intervals. Ignore
the Fresnel reflection effect. [4]
3.15.2 Calculate the emissivity of a hole of d/D = 5, for the ray shown, at
an angle θ ranging from 0 deg to 90 deg at 5-deg intervals. Include
the Fresnel reflection effect. [4]
3.15.3 Assuming the block’s temperature to be 300 K, calculate the ra-
diance in the hole opening for an angle θ ranging from 0 deg to
90 deg at 5-deg intervals. [2]
The farther reason looks, the greater is the haze in which it loses itself.
Johann Georg Hamann
4.1 Overview
97
98 Chapter 4
The integral
R
R= f ( x)dx (4.6)
0
Lb0 LbR
Lt0 LtR
Lth x Lth
L0 σx Lσ LR
x dx
x=0 Lx L x − γ L x dx x=R
Background Object L x +dx Observer
Lσ L( x +dx )−dx
on a cloudless day. In both cases sunlight is scattered into the sensor’s field
of view, adding flux along the line of sight. The following derivation is
based on the two-flux-Kubelka–Munk theory 15,16 developed for the optical
properties of paint, but the principles apply to any medium.
Path radiance occurs from the medium’s thermal self-exitance or from
flux from another source that the medium scatters into the radiance field.
The total flux is the sum of the source flux and all path radiance contribu-
tions.
The approach set out by Duntley 17 and others 18–20 is simplified by
omitting some scattering sources, but it is extended here to include thermal
self-exitance along the path. All of the variables defined here are strongly
dependent upon wavelength even though it is not indicated as such. Only
the conceptual model development and the results are shown here; for a
detailed mathematical analysis see the sources.
Consider the path geometry shown in Figure 4.1. The ‘line of sight’
is defined as the direction from L R toward the source L0 . The optical field
has a radiance L0 at the source, and after propagating over a distance R
through the medium, it has a radiance L R . At distance x along the path
the field has a radiance L x and after propagating a further distance dx,
the radiance is L x +dx . Source radiance passing through dx will diminish
because of attenuation in the medium, as discussed in Section 4.2.1.
The radiance L x will increase due to the flux scattered into the line of
sight by an amount σx Lσ , where σx is the scattering coefficient in [m−1 ],
and Lσ is the external source radiance. The radiance L x will increase fur-
thermore by the thermal exitance of the medium by an amount x Lth ,
where x is the medium emission coefficient with units [m−1 ], and Lth
Optical Media 101
to the object. The contrast transmittance is the fraction with which the
atmosphere reduces the inherent (close-range) contrast in the scene. Equa-
tion (4.16) has a form similar to Bouguer’s law but with a background
radiance scaling modifier. Contrast transmittance in the atmosphere is
considered in more detail in Section 4.6.9.
For γ = 0.1, the plume has a transmittance of 0.82 through the center,
low optical depth, and hence a low emissivity (0.18). The total volume of
this plume contributes to the radiance field from the plume. The plume
acts as a volume radiator. This also means that the entire plume volume
loses heat by thermal radiation.
At the other extreme, for γ = 2, the plume has a transmittance of
0.018 through the center, high optical depth, and hence a high emissivity
(0.982). It follows that the exitance from this medium emanates from the
outer layers of the medium. A volume of highly attenuating gas therefore
acts as a surface radiator. The plume loses heat only from the surface; the
inner volume retains its heat better than for a plume with low γ.
Conversely, an optically thin medium has high transmittance and low
emissivity. Because of the low emissivity (and good transmittance), the
radiation emanates from the whole volume of the medium — a volume
radiator.
The emissivity of a radiating gas is a spectral variable, as shown in
Figure 3.7. A CO2 gas plume has high emissivity (large γ) at 4.35 µm, but
almost zero emissivity at 3.5 µm. The plume can therefore be a surface
radiator (opaque as brick wall) at 4.35 µm but it is fully transparent at
3.5 µm.
Most media are not homogeneous, f ( x) = c, and Equations (4.8) and (4.9)
must be solved by integration along the optical path. Because the profile
f ( x) is almost never an analytical function, the integral cannot be solved
analytically. Practical solutions to the RTE are all based on discretiza-
tion 4,21 of the continuous function f ( x) into discrete parts f d [ Xdi ]. Each Xd
is a discrete section along the path x, such that the error f ( x) − f d [ Xdi ] is
sufficiently small for all values of x and corresponding values of Xdi . By
decreasing the discrete interval Xd toward zero, the difference between the
continuous function and the discrete approximation approaches zero (at
least sufficiently so for engineering purposes!).
Practical computational considerations limit the number of discrete
intervals and hence determine the coarseness of the approximation. In
some cases the discrete interval is not necessarily a constant value. One
such example is the modeling of atmospheric vertical profiles as 36 discrete
layers (see Section 4.7). A reasonable modeling approach would be to
adapt the sampling interval Xd to the relative magnitude of the profile f .
For path regions with low values of f , the interval Xd could be modified
Optical Media 105
L5
α5 σ5 Xd5
L4
α4 σ4 Xd4
L3
L2 α3 σ3 Xd3
L1 α2 σ2 Xd2
α1 σ1 Xd1
such that the product Xdi f d [ Xdi ] is approximately constant for all values
of i. In this strategy, each discrete path section contributes equally to the
integral along the path.
The discrete model intervals Xdi may be larger than the path length
itself, and the path’s two end points may end up in between Xdi bound-
aries (Figure 4.3). The endpoint section contributions are added pro-rata,
according to the length in each endpoint.
The geometric shape of the discrete intervals depends on the prob-
lem at hand. The earth’s atmosphere on a global scale is modeled as a
set of concentric shells, centered on the earth’s center. On a smaller, local
scale, the atmosphere can be modeled as a set of vertical strata. An aircraft
plume can be modeled as a stacked set of short concentric cylinders. An
arbitrary volumetric radiator can be modeled as a set of voxels, stacked
like a three-dimensional chessboard. The important consideration is that
within each of these individual volumes, the medium is considered uni-
form and homogeneous. The solution of the RTE within these discretized
models require numerical computation.
From Equations (2.26) and (4.4) (and looking ahead to Section 7.2.2) the
irradiance at the sensor from a distant source is given by an equation of
the form as shown in Equation (6.19),
A0 ∞
E = λ Lλ ( Ts )τaλ ( R)Sλ dλ, (4.17)
R2 0
where A0 is the area of the receiver, R is the range from the source to the
receiver, λ is the source spectral emissivity, Lλ ( Ts ) is the Planck law for
106 Chapter 4
1 km 3 km
5
7
9 km
0
3 3.2 3.4 3.6 3.8 4 4.2 4.4 4.6 4.8 5
(b) Atmospheric transmittance, sensor response, and source radiance
1
2300 K
Relative magnitude
Blackbody
Blackbody 330 K
1000 K
Plume
Sensor emissivity
response
0
3 3.2 3.4 3.6 3.8 4 4.2 4.4 4.6 4.8 5
Wavelength [mm]
(c) 3-5-mm effective transmittance: flare, aircraft plume, tailpipe, and fuselage
1
exp(-0.3R)
Effective transmittance
Unity weighted
0.5
MTV flare 2300 K
but only 6.5 km for a 330-K source. Hence, effective transmittance must be
calculated for the correct source temperature.
It is often proposed that the effective transmittance curves versus dis-
tance can be approximated by an exponential curve τa = e−αR . This formu-
lation is Bouguer’s law, as derived in Equation (4.4), at a single wavelength.
When applied to broadband sensors, this approximation yields a poor fit
for grey body radiation observed through a spectrally selective medium.
Figure 4.4 also shows that Bouguer’s law is totally inadequate in approxi-
mating the effective transmittance of a spectrally selective source through
a spectrally selective medium.
100 99.99997% mass is below 100 km Total mass 5.14×1018 kg, with variation
Aurora Borealis
of 1.2 to 1.5×1015 kg due to H2O
85
77 Mesopause Ionosphere D layer
71 Mesosphere
60 Meteors
51 Stratospause
47
40 Very little aerosol Highest balloon
above 30 40 km
Altitude [km]
Single scattering λ
Molecular absorption
Absorption when energy matches Flux can be absorbed out of the optical path.
the energy difference between Absorption is a very strong function of wavelength.
ground and excited states. Small
energy variations result from hν not matching, O
flux not absorbed
additional vibrational energy levels. H H
as shown here O H
(simplified): H H λ
Path radiance
Solar flux
Molecular emission and scattering
add flux to the path. Path radiance
emanates from sources other than Ambient flux Flux scattered into path
the primary source. Path radiance
flux could exceed the primary
source flux. Molecular emission is O H
Molecular emission into path
the inverse of molecular absorption. H
H
O
H
and particles (such as haze, dust, fog, or cloud droplets) suspended in the
air. Figure 4.6 summarizes the medium effects of the atmosphere.
Scattering and absorption by aerosol particles is a prominent factor in
the lower few kilometers in altitude, near the earth’s surface. The molec-
ular constituents in the atmosphere strongly affect attenuation and scat-
tering in the visual spectral range. Atmospheric aerosol particles in the
atmosphere vary greatly in their concentration, size, and composition, and
consequently in their effects on visual and infrared radiation. The aerosols
and molecules are not uniformly distributed along the optical path. Air
density, and hence particle and aerosol density, decreases with altitude and
even varies along paths at constant altitude. Some species of molecules
may vary in concentration, depending on local conditions and air–mass
history.
In terms of the model developed in Section 4.2.1, the atmospheric
attenuation coefficient γ comprises two components γ = α + σ, where α
is the absorption coefficient, and σ is the scattering coefficient, both with
Optical Media 111
1.00
Excited state
Transmittance
Vibrational levels 0.75
Energy
0
1.1
2.7
3.2
1.87
6.27
1.38
16
1
H2O molecular
27 °C, 75% RH
0
1
H2O continuum
27 °C, 75% RH
0
1
CO2
15
394 ppmv
1.4
1.6
2.0
2.7
4.3
9.4
9.5
0
1
CH4
1.66
2.37
3.26
3.31
3.53
3.83
2.2
2.3
7.6
0
1
O3 &
O2
0.63
0.69
1.06
1.27
1.58
4.72
9.01
9.59
14.2
0.76
3.3
0
1
N2O
2.87
2.97
4.06
3.9
4.5
7.9
0
1
All effects combined
0
0.2 0.3 0.4 0.5 0.7 1 2 3 4 5 6 7 8 10 15
Wavelength [mm]
Thermal IR
Microwave
Ultraviolet
Near IR
1 cm Visible al op
tics x = 2pr/l Hail
tric 000
1 mm Ge ome x=2
r is particle radius
Raindrops
Particle radius
100 mm ng Drizzle
sca tteri
10 mm Mie Cloud droplets
.2 g
1 mm x = 0 catterin Dust
Smoke
g h s
0.1 mm lei .002 Haze
Ray x=0 terin
g
10 nm b l e scat Aitken nuclei
ligi
1 nm Neg
Air molecules
0.1 mm 1 mm 10 mm 100 mm 1 mm 1 cm 10 cm
Wavelength
Figure 4.9 Aerosol scattering modes versus wavelength and particle size (used with per-
mission 1 ).
heavier dust particles may drift down and settle out of the atmosphere in
a few months, but photochemically formed sulphuric acid may remain for
up to two years. Only a very small portion of the total aerosol content of
the atmosphere exists above 30 km. 22
Aerosol composition is characterized with a statistical particle density
versus a lognormal size distribution. The statistical modal value of radius
is an indication of the aerosol ‘size;’ note, however, that the size has a
statistical distribution with the modal size occurring most often, but there
is a wide variation in particle sizes present. Indeed, most aerosols are
characterized by several modal distributions, not just a single distribution.
Hence, the concept of aerosol particle size should be understood to mean
the most commonly occurring size in the distribution.
Scattering is a wideband phenomenon, compared to the narrow ab-
sorption lines of molecular absorption. The energy is momentarily ab-
sorbed and immediately re-radiated as if from a point source. The at-
mospheric effects on optical and infrared flux depend on the size of the
particle compared to the wavelength of the light. Simplifying approxima-
tions are made to model the scattering effect. If 2πr/λ 1, where r is
the particle radius, the Rayleigh approximation applies, and if 2πr ≥ λ,
the Mie (sometimes called the Lorentz–Mie) approximation applies. Fig-
ure 4.9 shows the aerosol scattering modes at various wavelengths and
particle sizes. Mie scattering is an approximation of scattering theory, and
Rayleigh is a further special case of Mie scattering. Figure 4.10 shows scat-
Optical Media 115
2pr/λ = 0.063
2pr/λ = 1.88
2pr/λ = 5.03
2pr/λ = 0.63
2pr/λ = 3.77
2pr/λ = 6.28
2pr/λ = 3.77
2pr/λ = 1.26
Figure 4.10 Scattering intensity profiles for various values of 2πr/λ in unpolarized sunlight.
tering intensity profiles for various values of 2πr/λ for a water droplet,
starting with Rayleigh scattering and progressing to Mie scattering. Scat-
tering also depends on the incident light polarization and on the geomet-
rical shape of the particle. The discussion given here pertains only to un-
polarized light and near-spherical and isotropic particles. Comprehensive
coverage is available in Liou’s book. 4
Rayleigh scattering occurs when electromagnetic energy interacts with
aerosol (molecules) of physical size significantly smaller than the wave-
length of the energy field. This process is not a molecular absorption pro-
cess but rather an interaction at energy levels other than molecular absorp-
tion. The resultant scattering attenuation coefficient γ is approximately
proportional to λ−4 , affecting mainly propagation at shorter wavelengths
(ultraviolet, visual, and near-infrared). Rayleigh scattering leads to the
‘blue sky’ observed visually. In the visual and ultraviolet bands, the blue-
sky spectral radiance very roughly mimics a low-emissivity 10,000-K ther-
mal radiator. Rayleigh scattering by atmospheric molecules has little effect
at wavelengths longer than 3 µm. The phase function 4,22 gives the proba-
bility distribution for the scattered light, so that P(θ )dΩ is the fraction of
the scattered radiation that enters a solid angle dΩ about the scattering an-
gle θ. The phase function for Rayleigh scattering (for unpolarized light at
visual wavelengths) can be approximated by P(θ ) = k(1 + cos2 θ ), which
has a component of omnidirectional scatter, with peaks in the forward and
backward directions. Figure 4.8 shows Rayleigh molecular scattering over
a 5-km path length.
116 Chapter 4
300 K
thermal
source
0.5
0
Visible Near IR SWIR MWIR LWIR
0.5 0.75 1 1.5 2 2.5 3 5 8 12 20
Wavelength [mm]
stems mainly from the object’s thermal exitance. Sunlight reflection from
specular surfaces (known as glint) may produce a significant signal in the
LWIR band. 24 See also Sections 8.1 and 8.11 and Figure 8.3 for a discussion
of the effect of sunlight and sunglint on optical signatures.
0.64
8 12 mm 1.5 2.5 mm
Transmittance
3.2 4.8 mm
0.32 Zenith angle
MODTRAN Tropical
6.0 7.0 mm
0
0 10 20 30 40 50 60 70 80 90
Zenith angle [deg]
(a)
100
(1-τ) Lbb(300 K) 8 12 mm
Radiance [W/(m2·sr)]
MODTRAN calculation
10 (1-τ) Lbb(300 K) 6 7 mm
MODTRAN calculation
(1-τ) Lbb(300 K) 3.2 4.8 mm
1
MODTRAN calculation
MODTRAN calculation 1.5 2.5 mm
0.1
0 10 20 30 40 50 60 70 80 90
Zenith angle [deg]
(b)
Figure 4.12 Atmospheric transmittance and path radiance to space, actual and simplified
model: (a) transmittance vs. zenith angle and (b) path radiance vs. zenith angle.
0
0.5 1 3 5 8 12
Wavelength [mm]
(a)
1
Molecular
emissivity
Emissivity
For systems operating in clear air in the LWIR spectral range the scat-
tering coefficient σ and the term Lσ will be zero:
Lth
L R = L0 e−γR + (1 − e−γR ) (4.25)
α
= L0 e−γR + Lth (1 − e−γR ). (4.26)
For systems operating in severe aerosol both terms must be retained, and
the same reservation applies as for MWIR systems.
and water vapor are particularly relevant. Because these molecules absorb
infrared energy, they also emit infrared energy.
The emissivity [Figure 4.13(b)] was determined as the ratio of the
Tropical Modtran™ predicted path radiance (thermal component only),
and a thermal radiator at 300 K (27 ◦ C):
Lpath thermal
= , (4.27)
Le (300 K )
where the path radiance Lpath thermal was calculated by Modtran™, and Le
is Planck’s law for a 300-K source (the temperature of the Tropical atmo-
sphere). In the thermal spectral bands, this is a perfectly legal operation,
as defined in Equation (3.24).
Figure 4.13 shows that, even for the Tropical atmosphere at 27 ◦ C and
75% relative humidity, a 5-km path length results in an atmospheric emis-
sivity of 0.75 in the 8–12-µm band. An 8–12-µm thermal imager is therefore
trying to observe the target, looking through a veiling hot blackbody with
an emissivity of 0.75. The water vapor has much less effect on transmit-
tance in the 3–5-µm spectral band. The emissivity in the 3–5-µm band is
affected by CO2 at 4.3 µm.
Figure 4.14 shows the downward and upward radiance along a slant path.
The Modtran™ Tropical model and Rural aerosol, with 23-km visibility,
was used in this calculation. The path radiance term includes thermal path
radiance and single-scattered sunlight path radiance. The total path length
is 11.3 km, with its two endpoints at sea level and 8 km. Note the atmo-
spheric absorption around 4.3 µm. The Modtran™ Tropical model has
a sea-level temperature of 300 K. When looking down, the warm terrain
is observed in regions with good transmittance, and the cold atmosphere
is observed in spectral regions with poor transmittance. When looking
up, the warm atmosphere is observed in spectral regions of poor trans-
mittance (i.e., high emissivity), and the cold space is observed in spectral
regions with good transmittance. The positive–negative relationship be-
tween the looking-up and looking-down curves illustrate the principle of
atmospheric exitance in the absorption bands. These Modtran™ predic-
tions agree well with published measured data. 1
0
Path radiance and terrain radiance
Tropical atmosphere 27 °C 300-K thermal source
1.0 Terrain background 27 °C
11.3-km slant path Looking up
Radiance [W/(m2·sr·mm)]
Looking down
0.1
Cold atmosphere
0.01
Warm
atmosphere
300 K terrain
0.001
2.5 3 3.5 4 4.5 5 5.5
Wavelength [mm]
25 20 18 15 14 13 12 11 10 9 8 7 6
120
100 (b) 20 km looking down
80
Radiance [mW/(m2·sr·cm–1)]
60 260
K
40 240
K
2 20 K
20 20
180 K 0 K
160 K
0
120
100 (c) Surface looking up
80
60
260
K
40 240
200 K K
180 K 220 K
20
160 K
0
400 500 600 700 800 900 1000 1100 1200 1300 1400 1500 1600 1700
Wavenumber [cm 1]
Figure 4.14 Radiance up and down along a slant path: (a) M ODTRAN™ calculation, (b) mea-
sured (looking down), and (c) measured (looking up). Measured graphs (b), and (c) used
with permission. 1
Optical Media 123
101
Tropical 26.55 °C, 75.6% RH at 0 m ASL
Midlatitude Summer 21.05 °C, 76.2% RH at 0 m ASL
100
Water content [g/m3]
1976 US Standard
15 °C, 46% RH at 0 m ASL
10-1
10-4
0 2 4 6 8 10 12 14 16 18 20
Height [km]
1 200
8-12 mm 95% RH
3-5 mm 60% RH
0.5 CO2 attenuation 3-5 mm 95% RH 100
affects 3-5 mm
Highest
H2O attenuation recorded
affects 8-12 mm
0 0
20 10 0 10 20 30 40 50 60 70
Atmospheric temperature [°C]
Section 4.2.3 derives a general law for contrast reduction. This section in-
vestigates contrast transmittance for paths in the atmosphere. Three path
types are considered here: observations along an upward path, a down-
ward path, and a horizontal path.
In Equation (4.16) the Eb0 term is the sky radiance along the line of sight as
measured from the observer (to space), and EbR is the sky radiance along
the line of sight as measured from the object (to space). Combining Equa-
tions (4.10) and (4.16), the contrast transmittance for upward observations
is obtained as
− γRR∞
CR − γR 1 − e
τc = =e , (4.29)
C0 1 − e−γR0∞
∞
where R R∞ denotes the integral R = R f (r)dr along the line of sight, and
∞
R0∞ is the integral R = 0 f (r)dr [see Equation (4.6)]. For an upward path
the term in the square brackets is almost always less than unity because
there is a shorter path in the numerator and a longer path in the denomi-
nator. If the atmosphere along the path is uniform (which the atmosphere
is not), these integrals degenerate to the path lengths because f (r) is the
same constant value in both cases. The integration limit ∞ is not really
infinity because the atmosphere has finite extent. The symbol ∞ in this
context only indicates that the path runs into space.
Two important observations can be made from Equation (4.29): (1)
τc depends only on γ and R. In an atmosphere isotropic with respect
Optical Media 125
to azimuth angle, R only depends on the elevation angle, i.e., the angle
rising above the horizon. Consequently, for an object moving at a con-
stant elevation angle, the atmosphere will always have a constant contrast
transmittance. (2) The contrast transmittance is independent of the path
radiance phenomenon, scattering, or thermal emission.
1
= , (4.32)
1 − Kν (1 − eRγ )
where Kν is called the sky–ground ratio and to some degree resembles the
ratio of radiance values of the sky (as seen from the ground) and the
ground radiance (as seen from the sky). 26 Typical values of Kν are given
in Table 4.1. Note that these values are only applicable to observations in
the visual spectral band. Duntley 17 and Gordon 18 describe how to deter-
mine Kν . The sky–ground ratio has a very strong influence on the contrast
transmittance τc .
In the MWIR spectral band, Equation (4.30) cannot be simplified be-
cause the scattered and thermal path radiance components are similar in
magnitude.
In the LWIR spectral band, the scattering of the clear sky atmosphere is
insignificant, and Equation (4.30) can be likewise simplified. It is conve-
nient to define a sky–ground ratio for the infrared domain, 26
Lth
Kμ = . (4.33)
γL0
Typical values of the infrared sky–ground ratio are shown in Table 4.2.
If the source object is near the ground, L0 is the earth radiance, and the
126 Chapter 4
Spectral band Kμ Kμ
3–5 µm 0.70 1.42
8–14 µm 0.85 1.17
10–12 µm 0.86 1.17
−10 K +10 K
For horizontal paths the contrast transmittance is the limiting case of both
Equations (4.29) and (4.30), with the same result. For shorter distances
with small R, e−γRR∞ ≈ e−γR0∞ in Equation (4.29). For longer distances
e−γR → 0. In Equation (4.30) the background exitance is replaced by the
sky exitance, and Kν and Kμ become unity. Therefore, on the horizon, the
contrast transmittance is equal to the radiance transmittance:
τc = e− Rγ . (4.35)
100
0.2 km met range convective fog (clouds)
(16 20 mm particle diameter)
Hazy
23 km met range
maritime aerosol 0 90 0.82 0 61 0.37
0.1
50 km met range
tropospheric aerosol
Visual clear sky
1 2 5 10
`met range’ is Koschmieder meteorological range km km km km
0.0001
Ultraviolet Visible NIR SWIR MWIR LWIR
0.2 0.55 1 2 3 5 10 20
Wavelength (mm)
Figure 4.17 Spectral scattering attenuation coefficient for M ODTRAN™ aerosols and indica-
tive transmittance for various path lengths.
eters (most-common particle size for the specific aerosol). Shown on the
right side of the graph is the transmittance over a 1-km path, correspond-
ing to the scattering coefficient on the left side.
In Figure 4.17, from the bottom to the top, the aerosol size varies
from small to large. Clear-sky conditions (blue-sky Rayleigh scattering
and the 50-km meteorological range tropospheric atmosphere) affects the
ultraviolet, visual, and near infrared spectral bands. Light haze conditions
(23-km meteorological range) affect the MWIR and LWIR bands to some
extent. Finally, the large particle fog aerosol affects all of the visual and
infrared bands equally.
Meteorological range is defined in terms of human vision, not infrared
terms. Two aerosols may have the same meteorological range, but have
very different infrared scattering properties. Compare the 23-km Rural and
Maritime MWIR and LWIR aerosol scattering coefficients in Figure 4.17.
Even though the scattering coefficient is the same at 550 nm, it differs by a
factor of five in the MWIR band.
Optical Media 129
4.7.1 Overview
4.7.2 M ODTRAN™
80
70
Height [km]
60
50
40
30
20
10
0
1
150 180 210 240 270 300 10-3 10-2 10-1 100 10 102 103 10-9 10-6 10-3 100
Temperature [K] Pressure [mbar] 3
Water content [g/m ]
Figure 4.18 Standard M ODTRAN™ atmospheric models’ discrete layer altitudes and vertical
profiles.
Bibliography
[1] Petty, G. W., A First Course in Atmospheric Radiation , Sundog, Madison,
WI (2006).
[2] Farmer, W. M., The Atmospheric Filter: Volume I Sources, JCD Publish-
ing, Winter Park, FL (2001).
[3] Farmer, W. M., The Atmospheric Filter: Volume II Effects , JCD Publish-
ing, Winter Park, FL (2001).
[5] Kondratyef, K. Y., Ivlev, L. S., Krapvin, V. F., and Varatsos, C. A.,
Atmopsheric Aerosol Properties, Springer Praxis, Berlin (2006).
[8] Smith, F. G., Ed., The Infrared and Electro-Optical Systems Handbook:
Atmospheric Propagation of Radiation , Vol. 2, ERIM and SPIE Press,
Bellingham, WA (1993).
[9] Andrews, L. C., Field Guide to Atmospheric Optics, SPIE Press, Belling-
ham, WA (2004) [doi: 10.1117/3.549260].
[12] Mayer, B., Emde, C., Buras, R., and Kylling, A., “libRadTran — library
for radiative transfer,” http://www.libradtran.org.
[14] Boyd, R. W., Radiometry and the Detection of Optical Radiation , John
Wiley & Sons, New York (1983).
[24] Palmer, J. M. and Grant, B. G., The Art of Radiometry, SPIE Press,
Bellingham, WA (2009) [doi: 10.1117/3.798237].
[27] Baker, D. J. and Pendleton Jr., W. R., “Optical Radiation from the
Atmosphere,” Proc. SPIE 91, 50–62 (1976) [doi: 10.1117/12.955071].
Optical Media 133
[30] Kneizys, F. X., “Users Guide to LOWTRAN 7,” Tech. Rep. AFGL-TR-
88-0177, Air Force Systems Command, USAF (1988).
[33] Kneizys, F. X., Shettle, E. P., Gallery, W. O., Chetwynd, J. H., Abreu,
L. W., Selby, J. E. A., Clough, S. A., and Fenn, R. W., “Atmospheric
Transmittance/Radiance: Computer Code LOWTRAN 6,” Tech. Rep.
AFGL-TR-83-0187, Air Force Systems Command, USAF (1983).
[34] Roebeling, R. A., Jolivet, D., Macke, A., Berk, L., and Feijt, A., “Inter-
comparison of models for radiative transfer in clouds,” 11th Conference
on Atmospheric Radiation and the 11th Conference on Cloud Physics (2002).
Problems
4.1 Explain what thermal radiation is, why it occurs, how it can be
calculated, and how it is affected by the atmosphere. [5]
4.2 Describe what atmospheric aerosols are and what effect they have
on optical and infrared systems operating in the spectral range 1–
12 µm. Describe the effects of clean air and fog over the full extent
of this spectral range. [5]
4.3 Provide a description of each of the following terms; explain what
they are and how they work: (a) atmospheric transmittance, (b) at-
mospheric path radiance, (c) Rayleigh and Mie scattering, (d) dis-
crete ordinate models, (e) molecular absorption, and (f) aerosol
scattering. [12]
4.4 Start from first principles and derive the transmittance for a path
with length R in a homogeneous medium with attenuation coeffi-
cient γ. [4]
134 Chapter 4
1 − g2
PHG (θ ) = , (4.37)
4π 3/2
1 − 2g cos θ + g2
where θ is the scattering angle, and g is the asymmetry parameter;
where g = +1 provides complete forward scattering, g = 0 rep-
resents isotropic scattering, and g = −1 provides complete back
scattering. Plot this function for various values of g and compare
the results with Figure 4.10. Also plot the simplified Rayleigh scat-
tering function, 33
3(1 + cos2 θ )
PR (θ ) = . (4.38)
16π
Comment on your observations. [3]
The Henyey-Greenstein function is considered to be an inaccurate
model of real aerosol and was replaced with a Legendre polyno-
mial function. 34 Explain why. [2]
Chapter 5
Optical Detectors
Cornelius J. Willers
Ricardo Augusto Tavares Santos, D.Sc.
Instituto Tecnológico de Aeronáutica, S. José dos Campos - Brazil.
Fábio Durante Pereira Alves, D.Sc.
Naval Postgraduate School, Monterey - USA.
135
136 Chapter 5
ǰ ¢ǰ ǰ ǰ ǰ ǯ ¢ ¢ǯ
ǯ ǻǼǯ
Ȭ Ȭǰ Ȭ ǯ ¢ ¡ ǯ
ǰ ǰ ǯ ¢ǯ
ǰ Řǰ Ȭǰ Ȭ¢ ¢ǯ ¢ǯ
Ȭ ¢ǯ ǯ
ǰ Ȧ ǰ ǯ ǯ
¢ Ȧ ¢ ǯ ǯ
Ȧ ǰ Ȧ ǯ ǯ
ǰ Ȧ ¡ǯ
137
138 Chapter 5
Spectral response
Thermal
detector
Photon
detector
λcutoff
Wavelength
Figure 5.1 Spectral response comparison between photon and thermal detectors.
5.3 Noise
(PSD), with units of [W/Hz], [A2 /Hz], or [V2 /Hz]. It quantifies the con-
tent of the noise signal over a small bandwidth (say, 1 Hz) at a particular
frequency. In this case the spectral domain is not optical wavelength but
the temporal variation of a signal expressed in units of [Hz] (cycles per
second). The√noise density√ can also be expressed in volts or current with
units of [A/ Hz] or [V/ Hz].
The PSD of a time signal f (t) is given by
| FT (ω; β)|2
S f (ω ) = lim , (5.6)
T →∞ 2T
where the expected value is taken over the ensemble of signals β (not
time t), and FT (ω; β) is the Fourier transform of the time signal f (t):
T
FT (ω; β) = f (t; β)e−iωt dt. (5.7)
−T
In effect this means that several (as in a very large number) infinitely
long sequences of the time signal f (t) are transformed by the Fourier trans-
formation and multiplied with their complex conjugates. The ensemble
PSD is then obtained by averaging the individual spectra.
White noise has a constant PSD at all frequencies (spectrally flat). In
contrast, band-limited noise may have an arbitrary PSD, depending pri-
marily on the frequency response of the electronic filter used to define
the bandwidth. 1/ f noise has constant noise power per frequency octave
(frequency ratio of 2).
The integral of the PSD over all temporal frequencies yields the total
power in the signal:
∞
P= S f (ω )dω. (5.8)
−∞
Noise can therefore also be expressed in an integrated form with units of
[W], [A2 ], or [V2 ]. Depending on the context and the units used, it should
be clear whether the spectral or integral values are used.
Note that the power discussed here is electrical power in an elec-
tronic circuit, which is not the same as optical flux power. They share the
same unit and fundamental concept of power but have different contextual
meanings.
where vn and in are the rms noise voltage and current power spectral den-
sities in [V2 /Hz] and [A2 /Hz], respectively. k is the Boltzmann constant, T
is the temperature in [K] of the resistive element with value R in [Ω], and
x = h f /(kT ), where f is the electrical frequency. The term
x
(5.11)
e −1
x
describes the frequency spectrum of the noise — note its similarity with
the Planck-law formulation. h f is the energy of a particle at frequency f ,
whereas kT is the kinetic energy of a particle at temperature T. Beacuse
ex = 1 + x + x2 /2! + x3 /3! + . . ., it follows that for small x the term has
unity value. Investigation shows that the Johnson noise spectrum is flat up
to frequencies of the order of 1012 Hz.
The probability density function of unfiltered Johnson noise is Gaus-
sian with variance i2n Δ f or v2n Δ f , where Δ f is the electronic noise band-
width.
where i2 has units of [A2 /Hz]. I is the average current in [A] or [q/s],
and q is the charge on an electron [C]. This equation is easily derived from
the Poisson statistics, which states that the variance in events is equal to
the mean of the number of events Sq (0) = σq2 = aq , where the subscript q
denotes quanta (electrons or photons). If q is the charge of one electron, it
follows that
Si (0) = q2 Sq (0) = 2q2 σq2 = 2q2 a = 2qI, (5.13)
where Si (0) and Sq (0) are the low-frequency PSDs of the current and quan-
tum rates, respectively. The factor 2 is introduced to allowfor positive
ω ω
bandwidths only, i.e., integrating Si as 0 b Si (ω )dω instead of −ωb Si (ω )dω.
b
5.3.5 1/ f noise
in [K2 /Hz], from which the corresponding noise flux Φ can be determined
as 24,25
Φ = 4kGT 2 (5.19)
in [K2 /Hz] for low frequencies, where k is the Boltzmann constant, T is
the object’s temperature, and Δ f is the noise equivalent bandwidth. The
meaning of G and C is discussed in Section 5.4.2.
The detector electronics interface also contributes noise. 14,15,27 This noise
is inherent in the electronic components (shot noise or Johnson noise) or
can be a result of processing (amplitude digitization/sampling noise).
A subtle cause of noise in an imaging system results from the spatial sam-
pling (aliasing) of small objects in an image — this effect is not a noise
source, but it manifests itself in the form of semi-random variations in sig-
nal strength. Section B.4.1 describes this effect in a case-study context. The
performance of imaging systems, including noise effects, are well docu-
mented. 28–30 Noise in imaging systems is best analyzed in the context of
the three-dimensional noise model. 21,28,31
Optical flux generation is a Poisson process, which carries with the aver-
age value an inherent noise variance equal to the mean flux level. The shot
noise fluctuation is smaller at IR wavelengths but becomes significant at
very low flux levels for visual or UV light. Fewer photons are required in
the visual or UV spectral bands than in the IR spectral bands because the
photons have higher energy per photon at shorter wavelengths. The noise
inherent in the signal sets a minimum detectable signal level. The follow-
ing derivation determines this minimum signal level under the assumption
that there are no other signal or noise sources.
An optical flux of Φ p [q/s] results in a current of I = ηqΦ p in a detec-
tor. This current causes shot noise (see Section 5.3.3) with a magnitude
i2n = 2qIΔ f = 2q2 ηΦ p Δ f . (5.20)
Φe = hcΦ p /λ, hence the SNR becomes
ηλΦe
SNR = . (5.21)
2hcΔ f
Optical Detectors 147
The noise equivalent signal power (signal where SNR = 1) then follows as
2hcΔ f (1)2
NEP = . (5.22)
ηλ
Equation (5.22) states that for a given wavelength λ and noise bandwidth
Δ f , the noise equivalent power due to the signal fluctuation is determined by
the detector quantum efficiency η.
It can likewise be shown that, for a wideband photon detector, the
noise equivalent power due to signal fluctuation is
∞
2Δ f ν0 Meν ( Ts )dν
Φ= ∞ , (5.23)
η ν0 Meν (hνTs )dν
where Φ has units of [W], Δ f is the noise bandwidth, η is the detector
quantum efficiency (assumed spectrally constant), ν is frequency in [Hz],
Me is the thermal exitance in [W/m2 ] from a blackbody at a temperature
of Ts , k is Boltzmann’s constant, h is Planck’s constant, and c is the speed
of light. The integration starts at ν0 = c/λ0 , which is the lowest frequency
that the sensor can detect.
The background flux also causes noise in the detector signal. If the sensor
is limited by the noise caused by the background, the sensor is said to be
operating at ‘BLIP’ (background-limited performance). The principle used
to determine the BLIP limit is the same as was used for signal-limited
performance, except that in this case the limit is set by the noise caused by
the background.
For a monochromatic source at frequency ν, the minimum detectable
power (SNR = 1) of an open detector against a thermal radiation back-
ground is 19
∞
hν η (ν)2πν2 exp(hν/kTb )dν
NEP = 2AB , (5.24)
η ( ν) ν0 c2 [exp(hν/kTb ) − 1]2
where Δ f is the bandwidth, A is detector area, η (ν) is the detection quan-
tum efficiency, ν is the frequency, Tb is the background temperature, k is
Boltzmann’s constant, h is Planck’s constant, and c is the speed of light.
an equivalent optical flux power in the detector. This optical noise power
does not have the same units as the electronic noise power (discussed in
Section 5.3) and is called the noise equivalent power (NEP). The NEP is the
optical signal, in [W] or [q/s], required to give the same electrical signal
as the noise signal. From its definition, spectral NEP is given by
id Δ f
NEP λ = , (5.25)
Rλ
√
where id is the detector noise current density in [A/ Hz], Δ f is the noise
equivalent bandwidth in [Hz] , and Rλ is the spectral detector responsivity
in [A/W]. NEP is expressed in units of [W].
NEP is the noise for a particular detector device. Most noise source
contributions scale with detector area, and it is convenient to derive a
more-universal measure of detector performance specific detectivity (D ∗ ),
∗
which is normalized with
√ respect to area √ and frequency. D is normally
given in units of [cm· Hz/W], not [m· Hz/W]. The detector’s D ∗ and
NEP are related by
∗ Δ f Ad Rλ Δ f A d
Dλ = = , (5.26)
NEPλ id Δ f
where Ad is the detector area in [m2 ]. The D ∗ is a more-fundamental
property of the noise processes in detector materials and is independent of
the detector geometry. IR detectors are normally specified in terms of D ∗
at some background and source temperature. The noise of visual and NIR
detectors (e.g., silicon) are generally specified in terms of NEP.
NEP and detectivity can be expressed in spectral terms, NEPλ , or in
wideband terms, averaged over a spectral range. Similar to the spectral
values described above, the wideband values can be defined. Wideband
NEP is defined as
id Δ f
NEPeff = , (5.27)
Reff
√
where id is the detector noise current density in [A/ Hz], Δ f is the noise
equivalent bandwidth of the system, and Reff is the effective detector re-
sponsivity in [A/W], defined by (see Section 7.2.2)
∞
Rλ τa Mλ dλ
Reff = 0 ∞ , (5.28)
0
τa M λ dλ
where Rλ is the detector spectral responsivity, τa is the spectral transmit-
tance of the atmosphere or filters, and Mλ is the reference or calibration
source spectral exitance.
Optical Detectors 149
One or more different noise sources can be present in a sensor system. The
different noise sources are mostly uncorrelated, originating from different
components each with individual, statistically independent processes. It
can be shown that for uncorrelated sources, noise power adds linearly. The
total noise is then given by
N
ieff = ∑ i2n , (5.30)
0
where there are N noise sources in . The√ individual noise sources in this
equation can be either spectral noise [A Hz] or integrated wideband noise
[A].
A little care must be taken with noise expressed in terms of NEP. In
this case optical noise (NEP) corresponds to the signal itself, not the noise
it ‘mimics.’ Hence, when NEP from different noises are combined, the
NEP components must add in the square of NEP:
N
NEPeff = ∑ NEPn2 . (5.31)
0
Given a noise PSD S( f ) as input to a filter with voltage gain Av , the noise
equivalent bandwidth of the filter is defined as
∞
1
Δf = A2v ( f )S( f )d f , (5.33)
max( A2v S) 0
150 Chapter 5
Te, Le
Object Detector Environment
Gsor Gesr
Lo Ls Le
Flux from To Fos Ts Fes Te
the scene Ts, Ls ¥ o Cs s ¥ e
Ωo Gosr Gser
To, Lo Conduction through Pes
the mounting pillars Gesc
and the air gap below
the detector element G ¥
Te
¥
Device substrate
(a) (b)
Figure 5.2 Conceptual model for thermal detector: (a) physical layout and (b) flux flow
model.
because the radiating area is large compared to the thermal mass of the de-
tecting element, the detector quickly cools down once the incident energy
source is removed.
Absorbing plate
suspended above
silicon substrate
Thin film Metal conductor
resistor
hν
Column conductor
on silicon
substrate
Silicon
substrate
Bias current
Row conductor
on silicon
substrate
Bias current
top of the SiO2 , after which the SiO2 is etched away, leaving the membrane
only, supported by the posts. The vanadium oxide film is formed by sput-
tering a layer of mixed oxides on the membrane. The mixed-oxide layer
provides a more-stable detector than would pure oxides. The mixed-oxide
temperature coefficient of resistance α is of the order of −2 to −4 %/◦ C.
Other materials can also be used, and in one case, a forward-biased p-n
diode is used as the temperature sensor.
Microbolometers are available in two-dimensional arrays. The indi-
vidual element sizes can be as small as 17 µm, but common sizes range
from 25–50 µm. Some devices employ a two-story construction to increase
the fill factor. The array structure is built on top of a silicon read-out elec-
tronics interface chip (ROIC).
Microbolometers have achieved relatively high performance with ther-
mal imaging cameras demonstrating noise equivalent temperature differ-
ence of 0.1 K using f /1 optics. 35 Best performance is achieved if the de-
tector is encapsulated in a vacuum (pressure less than 1 mbar) to reduce
thermal conductance through the air, in the small gap between the bottom
of the detector element and the substrate underneath.
The bolometer device has the following noise sources: 35 (1) Johnson
noise (see Section 5.3.2), (2) 1/ f noise (see Section 5.3.5), (3) temperature-
fluctuation noise (see Section 5.3.6), and (4) read-out electronics interface
circuit noise (see Section 5.3.7). These noise sources are uncorrelated and
add in quadrature (see Section 5.3.12).
The pyroelectric effect 25,33,38 is found in many different materials, but fer-
roelectric materials are more commonly used. The ferroelectric detector
senses the changes in the electrical polarization of the material resulting
from temperature changes. The output from the ferroelectric detector is
proportional to the rate of change in input flux — if the scene flux does
not change, the signal disappears. Sensors with pyroelectric detectors rely
on movement in the scene or require a device to ‘chop’ the incident signal,
alternating the scene flux with a reference flux.
The ferroelectric effect is found in materials such as barium stron-
tium titanate, strontium barium niobate, lithium tantalate, and lead titan-
ite. These materials have spontaneous internal polarization, measured as a
voltage on electrodes placed on opposite sides of the bulk of the material;
this forms a capacitor with the sensor material as the dielectric. At a con-
stant temperature, the polarization is equalized by mobile charges on the
158 Chapter 5
Top electrode
Pyroelectric
element
Bottom electrode
hν
Column conductor
on silicon
substrate
Silicon
substrate
Bias current
Row conductor Post with
on silicon contact
substrate
Bias current
Detecting Reference
Metal A junction Metal B junction Metal A
hν
detector has wider dynamic range than other detector techniques. In order
to increase the output signal, more than one junction pair can be combined
on one optical pixel, which is then called a thermopile.
When a metal rod is heated at the one end and cooled at the other end,
the electrons at the hot end have more energy kT and thus greater velocity
than at the cold end. There is a net diffusion of electrons from the hot end
to the cold end, which leaves the positive ions in the hot region. This diffu-
sion continues until the potential field so generated prevents the diffusion
of more electrons. 40 A voltage is generated between the hot and cold ends
of the rod. The ratio of the potential generated across the metal to a unit
temperature difference is called the Seebeck coefficient S = ΔV/ΔT. By
convention the sign of S represents the potential of the cold end relative
to the hot end, but the cold end has a negative potential relative to the hot
end. The Seebeck coefficient in a metal can be negative (Na, K, Al, Mg, Pb,
Pd, Pt) or positive (Mo, Li, Cu, Ag, Au), depending on the diffusion of the
electrons. The Seebeck coefficient in a p-type semiconductor is positive be-
cause in a p-type material the holes are the majority carriers. The Seebeck
coefficient depends on the material temperature S( T ). It can be shown that
the Seebeck coefficient for metals is given by (after some approximation) 40
π2 k2 T
S≈− , (5.50)
3qEF (0)
where k is Boltzmann’s constant, T is the temperature in [K], q is the charge
on an electron, and EF (0) is the Fermi energy at 0 K. The voltage across a
Optical Detectors 161
Measuring the potential across a copper bar with electrical fly leads
made of copper — and connected to the hot and cold end of the bar —
will yield a zero voltage. The copper fly leads form the same potential
difference, opposing the potential in the bar under test resulting in a zero
voltmeter reading. If, however, the fly leads are made of another metal, the
difference in two Seebeck potentials will be measured on the voltmeter:
T1
ΔVAB = (SA − SB )dT. (5.52)
T0
Combining Equations (5.50) and (5.52) yields the metal thermocouple equa-
tion
VAB = aΔT + b(ΔT )2 , (5.53)
where a and b are the thermocouple coefficients, and ΔT is the temperature
with respect to the reference temperature of 273.16 K.
The responsivity of the thermoelectric detector is given by
N (S1 − S2 )
R = , (5.54)
G 1 + (ωτθ )2
where is the surface absorption (emissivity), N is the number of junction
pairs per pixel, S1 and S2 are the Seebeck coefficients of the two dissim-
ilar materials, and G is the effective heat conduction coefficient [Equa-
tion (5.38)].
The thermoelectric detector has three noise sources: (1) Johnson noise
(see Section 5.3.2), (2) temperature-fluctuation noise (see Section 5.3.6), and
(3) read-out electronics interface circuit noise (see Section 5.3.7). However,
the responsivity is relatively low, so the Johnson noise dominates.
1013
Tdetector 0 K
1012
D* [cm·Hz½/W]
Tdetector 77 K
Tdetector 195 K
1011 Tdetector 290 K
1010
109
0 60 120 180 240 300 360 420 480 540 600
Environmental temperature [K]
Figure 5.6 Photon-noise-limited D ∗ for thermal detectors with = 1 and 2π sr field of view
(adapted 25).
where
λ2 2 3
h c ex
Fλ1 ,λ2 = 2 dλ, (5.57)
λ1 λ6 ( e x − 1) 2
where x = hc/(λkTe ), Te is the environment temperature, λ1 is the filter
short-cutoff wavelength, and λ2 is the filter long-cutoff wavelength.
For a detector employed in a focal plane array the photon-noise-limited
Optical Detectors 163
R = R + m 1 a1 + m 2 a2 + m 3 a3 , (5.61)
Orbital
(number of electrons)
2N states Electron band
3s (1N)
Electron energy
6N states 2p (6N)
Forbidden
2N states band
2s (2N)
2N states 1s (2N)
Figure 5.7 Band creation due to the decreasing distance between atoms for sodium
(adapted 45,46 ).
context. 45 The width or spread of the band depends on the overlap of the
individual atoms’ energy orbitals — it affects the outer or higher energy
orbitals (e.g., 3s) first before affecting the inner orbitals (1s). 46,47 Figure 5.7
indicates four orbital energy bands at an atomic spacing of a. Note that for
the 1s state the energy probability distribution is very narrow, whereas for
the 3s band the probability distribution widens considerably.
The bandgap or atomic energy levels are commonly defined in units
of electron volts [eV], the energy gained or lost by the charge of an electron
moving across an electric field of one volt. The charge on one electron is q
C. One volt is 1 J/C, hence one eV is quantity(q) J.
An important parameter in this theory is the Fermi level: electrons
fill the orbital states from the lowest energy level up to a higher energy
level E f , called the Fermi Level. The probability of finding a thermally
excited occupied energy state with an electron in an energy range E to
E + dE (under thermodynamic equilibrium) is given by the Fermi–Dirac
distribution:
1
f ( E) = , (5.62)
1 + e( E− EF )/(kT )
where T is the temperature in [K], EF is the Fermi level in [J], and k is
the Boltzmann constant. Note that kT is the thermal energy of an electron
associated with the temperature T. The Fermi–Dirac distribution satisfies
Pauli’s exclusion principle and hence describes the distribution at all tem-
peratures. 42 At absolute zero temperature (0 K), all of the energy states
are filled from the lowest state, with no vacant or unfilled states, up to the
highest state defined by the Fermi level EF . At nonzero temperatures, the
filled states spread around the Fermi level, according to the Fermi–Dirac
distribution. At higher temperatures, the spread is broader, as shown in
Figure 5.8.
1.0
Spread represents
Occupancy probability
thermal energy kT
0K
0.5
77 K
300 K
0.0
−0.20 −0.16 −0.12 −0.08 −0.04 0.00 0.04 0.08 0.12 0.16 0.20
Energy [eV]
Figure 5.8 Fermi–Dirac energy probability distribution for thermally excited electrons,
around the Fermi (50%) energy level.
h̄2 k2 h̄2
E = = (k2 + k2y + k2z )
2m∗ 2m∗ x
h2
= (n2 + n2y + n2z ) = E0 k F , (5.64)
8m∗ L2 x
where k is the wave vector with its three Cartesian coordinates, m∗ is the
effective electron mass, and h̄ = h/(2π), where h is the Planck constant.
This equation is an approximation 43 of the electron’s energy near the band
edges EC and EV , acceptable within the primary concern of this analysis.
The value E0 = h2 /(8m∗ L2 ) is the lowest energy state in the crystal (longest
168 Chapter 5
wavelength). k2F = (n2x + n2y + n2z ), where ni are the indices of the reciprocal
lattice points (1/distance) inside the sphere with radius k F . This sphere
contains all of the modes associated with indices less than ni — that is,
all of the standing waves with longer wavelengths (lower energy). The
number of modes that can be sustained in a sphere with radius k F is given
by all of the modes starting with the shortest wavelength, then including
all of the modes with longer wavelengths up to wavelength k F .
The number of electrons that can be accommodated in states with
energy E or less is 48
1 4πk3F π E 3/2
N ( E) = 2 = , (5.65)
8 3 3 E0
where the factor of 2 allows for the positive and negative electron spin
for each state, and 4πk3F /3 is the volume of the sphere. The factor of 8
is present because only the positive integers are considered. It can then
be shown 45,48 that the density of electrons with energy less than E per unit
volume is
∗ 3/2
1 2m π 8m∗ 3/2 1/2
D ( E) = E 1/2
= E . (5.66)
2π2 h̄2 2 h2
Having N electrons in the energy band per unit of volume, the condition
determines that EF is given by 49
EF
D ( E)dE = N. (5.67)
0
The band that is normally fully filled with electrons at 0 K is called the
valence band, whereas the upper unfilled band is called the conduction
band.
From Equation (5.66) the density of electrons per unit volume in the
conduction and valence bands becomes 10
π 8m∗ 3/2
DC ( E ) = ( E − EC )1/2 for ( E > EC ) (5.69)
2 h2
π 8m∗ 3/2
DV ( E ) = ( EV − E)1/2 for ( E < EV ). (5.70)
2 h2
Figure 5.9(a) shows the relationship between electron energy as a function
of wavenumber along one dimension. The figure shows a wave vector
Optical Detectors 169
E E
DV(E) density of
states with
Valence band energy above E
-p p k Density of states
Wave vector (electrons per unit volume)
(a) (b)
Figure 5.9 Crystalline material: (a) relationship between electron energy and wave number
ki , and (b) density of states for a free electron in a semiconductor.
for a single atom in the crystal. These wave vectors repeat exactly along
the lattice positions, a requirement for stable electronic wave solutions.
Figure 5.9(b) shows the density of electron states in the conduction and
valence bands of a semiconductor. Both of these functions are parabolic in
shape.
where m0 is the mass of the electron, U (r) is the potential seen by the
electron, ψ(r) is the wave function describing the electronic state, and E
is energy. The first term represents the free electron’s wave function. The
second term represents the dependency on all of the other electrons and
atoms in the material. For crystalline materials, the potential U (r) has
170 Chapter 5
Eg kT Eg > kT
(»6 eV for diamond) (1.12 eV for silicon)
Si Si Si Si Si Si Si Si Si Si Si Si
Si Si Si Si Si Si P+ Si Si Si B Si
Si Si Si Si Si Si Si Si Si Si Si Si
+
Si Si Si Si Si Si Si Si Si Si Si Si
from the crystal. The covalent bonds with the neighboring silicon atoms
require four electrons, whereas the trivalent atom can only provide three.
A valence-band electron becomes ‘caught’ in this fourth covalent bond —
it is not available for excitation to the conduction band. The silicon atom
providing this trapped electron now has one less electron (a hole). The hole
(absence of an electron) in the valence band can freely move through the
material because an electron fills the hole but in the process it creates a new
hole in another atom [see Figure 5.11(c)]. The acceptor-doped material is
called p-type material. In p-type material, the mobile holes are the majority
carriers, and the bound electrons are the minority carriers.
Only a small amount of energy (0.01–0.05 eV) is required to elevate
the ‘extra’ electron from the donor’s energy level ED into the conduction
band [see Figure 5.12(b)]. Similarly, a small amount of energy will excite
an electron from the valence band into the acceptor level E A (the missing
electron in the covalent bond) [see Figure 5.12(c)]. The impurity atoms
therefore introduce ‘allowable’ energy states in the otherwise forbidden
bandgap.
Donor doping introduces electrons in the conduction band but cap-
tures the holes in fixed lattice locations. Acceptor doping introduces holes
in the valence band but captures the electrons in fixed lattice locations. The
free electrons and holes generated by doping do not have their free coun-
terparts available for conduction, as in a metal where both free electrons
and holes are present. Electron-hole pairs formed by thermal excitation
support conduction by free electrons in the conduction band and by free
holes in the valence band.
Figure 5.12 shows electron excitation by photon absorption, but note
that thermal excitation has the same effect in the conduction band. Elec-
tron excitation by thermal means or by photon means are indiscernible
once the electron is excited. Thermally excited electrons normally interfere
with the device’s operation — they must be minimized. For this reason,
Optical Detectors 173
Valence
band
0 0.5 1 0 0.5 1 0 0.5 1
(a) (b) (c)
Figure 5.12 Energy bands and Fermi–Dirac distributions for: (a) intrinsic semiconductor,
(b) n-type extrinsic semiconductor, and (c) p-type extrinsic semiconductor.
When a photon is absorbed, two types of electron transitions can occur: in-
traband and interband transitions, 53 as shown in Figure 5.13 (the detailed
Optical Detectors 175
E E
Conduction band
Interband
transition
Intraband k k
transition f hν
i
hν
i
Valence band
Figure 5.13 Intraband and interband transitions of an electron from an initial state i to a final
state f .
1. The photon energy must exceed the material energy bandgap to excite
the electron across the bandgap Etransition = hν ≥ Eg .
E E
Conduction band
Indirect
transition f
f
Phonon
Photon
Direct
transition
hν k k
hν
i Photon i
Valence band
Figure 5.14 Interband transitions: (a) indirect transition involving a phonon and a photon,
and (b) a direct transition.
5. If the energy gap E(k) between two bands is near-constant over a wide
range in k, photon-induced interband transitions occur more effectively.
It means that there are many initial and final states resonant with this
photon energy.
In photon detectors the transition near the smallest energy gap be-
tween the valence and conduction bands is of main interest. This occurs
where k = 0, which is designated as the Γ point.
The energy structures in indirect semiconductors (e.g., Si and Ge) are later-
ally displaced, hence phonons are also involved in the excitation process.
Indirect semiconductors have absorption coefficients of the form 2,10,43,56
αν =∝ (hν − Eg ± Ep )2 ≈ αλc + α0 (hν − Eg ± Ep )2 for hν > Eg . (5.79)
The free-carrier absorption coefficient is proportional to the electron and
hole carrier concentration because each carrier site acts as an absorption
site. At energy levels below the bandgap, an exponential absorption takes
place, known as the Urbach tail, 7,10,48,52,56,58–60 where
hν − Eg
αν = αλc exp for hν < Eg . (5.80)
kT
The spectral shapes of these absorption processes are shown in Figure 5.16.
The typical absorption coefficient curves shown in Figure 5.15 reflect
specific material samples under specific conditions; different samples and
measurements may differ. The absorption coefficient is a function of tem-
perature, crystal orientation, and impurity concentration (intrinsic or ex-
trinsic material). 61
From Equation (4.4) it follows that the flux at depth d into the detector
will be Φ(d) = Φ(0)(1 − ρ)e−αd . Most of the photons are absorbed within
a distance of 1/α into the absorbing material. For large absorption coef-
ficients, this occurs within a thin layer on the skin of the detector. From
the absorption graphs in Figure 5.15, it follows that photons with longer
wavelength are absorbed deeper into the material, whereas photons with
shorter wavelength are absorbed nearer the surface of the material.
What happens to the photons with wavelengths exceeding λc that are
not absorbed in the material? The material normally has low attenuation
178 Chapter 5
109
Absorption coefficient [m-1]
108 GaAs
GaAs 300 K
107 In0.7Ga0.3As0.64P0.36
6 Ge
10
Si 300 K InP 300 K In0.53Ga0.47As
105 Si InP
104
103
Ge 300 K
102
0.2 0.4 0.6 0.8 1.0 1.2 1.4 1.6 1.8 2.0
Wavelength [mm]
Figure 5.15 Photon absorption coefficient for intrinsic materials (adapted 1,10,62 ).
Absorption coefficient
Ideal semiconductor
absorption
Wavelength
Figure 5.16 Photon absorption coefficient: ideal, Urbach tail, and free-carrier components.
(adapted 56).
Tables A.5 and A.6 contain a summary of physical parameters for selected
intrinsic and compound semiconductors. Material properties are well doc-
umented. 63 It is important to note that all parameters are to be used in
calculations regarding direct transitions near the Γ point, at the minimum
bandgap.
If the energy of the photon exceeds the energy bandgap of the material,
an absorbed photon creates an electron-hole pair in the semiconductor
material. The excited electron-hole pair recombines after some time, the
average value of which is the carrier lifetime. The carrier lifetime is not a
material property but rather a function of the application. Detectors that
employ this phenomenon are called photon detectors.
Photoconductive detectors operate on the principle that photons are ab-
sorbed in the bulk of the detector material, and an electron-hole pair is
formed. The electron-hole pair separates and contributes to electrical cur-
rent conduction, thereby lowering the bulk resistance of the detector. The
photon-induced conductance change in resistance can be detected by an
external electronic circuit.
Photovoltaic detectors operate on the principle that electron-hole pairs
formed by photon absorption in the depletion region of a p-n diode are
accelerated across the depletion region. This electron-hole pair contributes
to current flow by injecting minority carriers, thereby causing current flow
through the depletion region. The photocurrent flows through the deple-
tion region, under the built-in bias that exists in the depletion region. The
photocurrent flow can be sensed by an external electronic circuit.
Ideal detector
1
Quantum efficiency η
Real world detector
External losses:
high surface reflection
Internal losses:
low bulk absorption
0
λcutoff
Wavelength
Figure 5.17 Comparison between ideal and actual photon detector quantum efficiency
(adapted 2).
10
cy
Responsivity [A/W] ffi cien
ume
ant
%Q
u iency
100 antu m effic
5
50% Qu
η=0.7 η=0.5 HgCdTe
η=0.8
Si InSb
0.0
0.0 1.5 3.0 4.5 6.0 7.5 9.0 10.5 12.0 13.5 15.0
Wavelength [mm]
efficiency line [Equation (5.5)]. Note that the photon detector responsivity
increases toward longer wavelengths, as shown in Figure 5.1. This stems
from the fact that the photon energy decreases toward longer wavelengths
and that larger numbers of photons (hence the larger detector current) are
required to provide one watt of optical power.
The quantum efficiency for extrinsic detectors is lower than the quan-
tum efficiency of intrinsic detectors because there are many more possible
excitation locations in intrinsic detector materials than in extrinsic detector
materials. Typical absorption coefficients for extrinsic materials are less
than 103 m−1 , compared to 103 –107 m−1 for intrinsic materials.
Figure 5.19 shows a comparison of various detectors’ spectral D ∗ per-
formance. The caption in the original publication 12 is as follows:
Note that some detectors have quite high D ∗ (high quantum efficiency)
values compared to the theoretical limit.
Thermally excited carriers have energy kT, which may exceed the bandgap
in a detector [Equation (5.62)]. The thermally excited carriers contribute to
detector noise and reduce the performance of the detector. The objective
with detector cooling is to achieve kT Eg . It is evident that materials
with wide bandgaps are less prone to thermal carrier excitation. Long-
wavelength detectors generally operate at temperatures such as 77 K or
even down to 4 K. Modern detector development, and particularly the
use of HgCdTe, enabled LWIR detectors to operate at somewhat higher
temperatures (80 K) than older extrinsic detectors (4 K). Current research
is pursuing detector technologies that would enable even higher operating
temperatures. 7
184 Chapter 5
Figure 5.19 Spectral D ∗ for various detector types. Used with permission from Rogalski. 12
Optical Detectors 185
Window Tcold
Detector
focal plane (3) Liquid cryogen
(cold end of cooler)
Cold finger
(4) Discarded gas
Vacuum cools incoming gas
Tambient
(a) (b)
Figure 5.20 Detector vacuum dewar: (a) with wire feed-through and (b) cooler inserted.
ture is within a few degrees of the cryogen boiling point. Dewars can be
made of glass, metal, or a combination of glass and metal. In some cases,
the dewar is integrated with the cooler, as a single assembly.
A typical open-cycle Joule–Thomson cooler is shown in Figure 5.20(b).
The process (a) starts with a gas at a pressure of 45 MPa at 300 K, and
then (b) the gas is pre-cooled by the discarded gas prior to (c) expanding
through a very small nozzle, such that the pressure drops to around 1–2
bar, and cools the cryogen to liquification (e.g., 77 K in the case of nitrogen),
at which point the liquid cryogen absorbs the heat from the detector and
evaporates, (d) flowing away from the detector. The cryogen is selected on
the basis of its boiling point and cooling capacity. Popular cryogen gases
include helium (4 K), nitrogen (77 K), and argon (87 K), or a mixture of
gases.
The Stirling-cycle thermodynamic process is used most commonly in
closed-cycle coolers. The cryogen is contained in a sealed piping system
and oscillates between the hot and cold ends. The Stirling engine normally
employs an electric motor (rotary or linear) to compress the gas. Stirling
coolers are available in two configurations: the integral cooler and the split
cooler. The integral cooler combines all mechanics and cryogenics into a
single unit. The split Stirling cooler separates the compression unit from
the cooling unit. The split engine has the advantage of being silent with
no vibration coupling from the engine to the detector. This was partic-
ularly important during the earlier generations of coolers. Research has
resulted in new engines that run much quieter, allowing the integration of
the cooling engine with the cooler.
Thermoelectric (TE) cooler modules are solid state heat pumps that
employ the Peltier effect — the phenomenon whereby the passage of an
electrical current through a junction consisting of two dissimilar metals
results in a cooling effect. When the direction of current flow is reversed,
heating will occur. A thermoelectric module consists of an array of p-
and n-type semiconductor elements heavily doped with electrical carriers,
as shown in Figure 5.21. The array of elements is soldered so that it is
electrically connected in series and thermally connected in parallel. This
array is then fixed to two ceramic substrates; a hot and a cold side. Heat
is absorbed at the cold side of the n- and p- type elements. The electrical
charge carriers (holes in the p-type; electrons in the n-type) always travel
from the cold side to the hot side, and heat is always released at the hot
side of thermoelectric element. The temperature differential between the
hot and cold ends is inversely proportional to the pump load. At higher
heat loads the temperature differential reduces.
The vast majority of thermoelectric coolers achieve temperature differ-
Optical Detectors 187
Heat flow in
(cold side)
Ceramic
substrate
+ + + p type
Semiconductor
+ + + n type elements
+ + + p type
Electrical
+ + + n type & heat
conductor
Heat sink
Heat flow out
(hot side)
5.8.1 Introduction
RL
Vb
l
w Rd Vd
d
i
Figure 5.22 Photoconductive detector geometry and bias circuitry.
hence
∂Rd q(μe τe + μh τh )ηΔΦ p
=− . (5.95)
Rd lwdσ
The optimal choice for R L is such that ∂Rv /∂R L = 0, which requires
R L = Rd ; then,
−RL Vd (μe τe + μh τh ) qλη
Rv = (5.100)
R L + Rd σdlw hc
− R L Rd Vd (μe τe + μh τh ) qλη
= (5.101)
R L + Rd l2 hc
= Gc G ph R. (5.102)
The derivation identifies three distinct terms: (a) the bias circuit gain Gc
in units of [Ω] or [V/A], (b) the unitless bulk material photoconductive
gain G ph , and (c) the photocurrent responsivity R in units [A/W]. The
photoconductive gain can be written
Vd (μe τe + μh τh ) ε(μe τe + μh τh )
G ph = = , (5.103)
l2 l
where ε is the electric field across the detector in [V/m]. ε(μe τe + μh τh ) has
units of velocity (carrier velocity), so that l/[ε(μe τe + μh τh )] represents the
transit time the carrier needs to travel along the length of the detector, τt .
The photoconductive gain is then given by
τ
G ph = , (5.104)
τt
which is the shortest carrier lifetime divided by the transit time. G ph there-
fore represents the number of times the carriers can move across the length
of the detector within its average lifetime. Typical values vary from 0.5–1.0
for silicon to 103 –104 for HgCdTe.
Equation (5.100) shows that the responsivity can be increased by re-
ducing the equilibrium conductivity of the detector. The conductivity can
be reduced by reducing the number of free carriers under equilibrium con-
ditions. This can be achieved by cooling the detector and by reducing the
incident background flux on the detector.
where ηΦq is the optically generated carrier rate, η is the quantum ef-
ficiency, τ is carrier lifetime, Φq (t) is the flux incident on the detector in
[q/s], and Δn(t)/τ is the rate of recombination of free (optically generated)
carriers. Taking the Fourier transform on both sides of Equation (5.105),
ΔN ( f )
i2π f ΔN ( f ) = ηΦq ( f ) − , (5.106)
τ
1
ΔN ( f ) i2π f + = ηΦq ( f ), and (5.107)
τ
ηΦq ( f )τ
ΔN ( f ) = , (5.108)
1 + i2π f τ
where f is frequency, ΔN ( f ) is the Fourier transform of the number of
free carriers due to optical excitation, Φq ( f ) is the Fourier transform of
the photon flux incident on the detector, and τ is the carrier lifetime. The
frequency response of the detector is then simply
ΔN ( f ) τ
H( f ) = = . (5.109)
ηΦq ( f ) 1 + i2π f τ
There are four components that contribute to the total noise in the pho-
toconductive detector: Johnson noise, 1/ f noise, g-r noise due to optical
flux, and g-r noise due to thermally excited carriers: 19
Td TL k Iα
it = 4k
2
+ + 1 β + 4q2 ηΦqb G2ph + 4gθ qG2ph . (5.110)
Rd RL f
IR photoconductive detectors are normally cooled as explained in Sec-
tion 5.7. Suppose further that the detector is not used at frequencies
where 1/ f noise dominates, k1 I α / f β i2t , then two performance-limiting
noises are present: generation–recombination-noise-limited operation and
Johnson-noise-limited operation.
192 Chapter 5
where Ad is the area of the detector. Note that the D ∗ depends only
on the background flux Eq ; this condition is also known as background-
limited operation. The dashed line in Figure 5.19 was calculated with
Equation (5.117).
T q2 G2ph ηΦqb
, (5.118)
Reff k
Optical Detectors 193
Diffusion length
Diffusion length
hν Depletion
Diffusion length
Diffusion length
Depletion hν
region region
+ Diffusion + Diffusion
+ + Drift
Diffusion
+ + Drift
Diffusion
Drift Drift
d1 d2 d1 d2
Potential
Potential
Diffusion
carriers
Diffusion
carriers
d d
Carriers contributing Carriers contributing
Radiance
Radiance
to current flow to current flow
d d
(a) (b)
Figure 5.23 Photovoltaic detector construction: (a) p-n diode and (b) pin diode (adapted 68).
with p-type and n-type dopants (less commonly done), the interface be-
tween these two doped regions is called the p-n junction. The n material
has an excess of electrons, leaving a positive ion in the crystal lattice; in
contrast, the p material accepts an electron from the lattice, leaving a neg-
atively charged ion. The free electrons and holes will diffuse across the
junction, leaving a region with positive and negative ions, but with no free
charge carriers — this is called the depletion region. There are no free
carriers in the depletion region, resulting in an internal field across the
depletion region. Real-world detectors are complex devices designed to
locate the depletion layer as close to the detector front surface as possible.
Various techniques are also used to optimize detector size and respon-
sivity, and to reduce noise. Figure 5.24 shows the electronic and energy
state in, and around the p-n junction. Note in particular the concentration
profiles along the depth of the p-n junction. 46
If a photon is absorbed in the depletion region, the resulting electron-
hole pair forms minority carriers in the depletion region. These minority
carriers are accelerated and swept out of the depletion region under the
built-in electric field present across the depletion region. The resultant
minority current results in a measurable current on the device’s terminals.
The operation of a photovoltaic detector and a photoconductive detec-
tor differs in the sense that any carrier pair created in the photoconductive
detector contributes to the lower detector resistance. In a photoconductive
Optical Detectors 195
Majority Majority
carrier (holes) Acceptor ion Donor ion carrier (electrons)
+ Diffusion
+ + + Diffusion
+ +
+ + + +
+
+ + + + + + + +
p type n type nd
na nn
pp
ni ni Concentration
profiles
np pn
Fixed space
+ charge density
0
0 Electric field
intensity
qVd
EC
EF Energy bands:
unbiased
Eg p-n junction
EV
q(Vd+Vr)
EC
EF
Energy-bands:
reverse biased
Eg p-n junction
EV
Figure 5.24 The p-n diode junction, energy diagrams, and energy bands.
196 Chapter 5
The I-V curve describes the operating point of a diode by relating the
voltage across the terminals and the current flowing out of the terminals.
All diodes have this behavior (see also Section 9.3.2.4). The diode current is
related to the voltage across the device by the nonlinear function: 2,3,10,12,16
The photovoltaic detector can be biased in three quadrants and five op-
erating conditions. Strong forward bias is applied to diodes to generate
198 Chapter 5
I
+Vb -
Rf
p n +
-
p n
+
C Open circuit /
F forward biased C
Reverse bias F
V
V
Vos
V
Iph
RL
Sun cell
Poor detector
Figure 5.25 Bias configurations and energy bands for various operating conditions.
light (LEDs and laser diodes). 56,69 This section only considers reverse or
small forward bias as indicated in Figure 5.25. All of the bias conditions
described in the following sections are commonly found in various appli-
cations.
Under reverse bias, the electron-hole pair generated by the photon forms
minority carriers in the depletion region. The hole and electron are swept
out by the electric field in the device. All of the electron-hole pairs are ex-
ternally observed as a current flowing through the device; see Figure 5.26.
In Figure 5.26 minority carriers are indicated by small circles, and major-
ity carriers are indicated by large circles. If the reverse-bias voltage is too
high, the detector may enter the avalanche regime. Some detectors are
designed to operate in the avalanche regime, but special semiconductor
and electronics design techniques are required to construct an avalanche
device that can operate reliably. The avalanche detector is also very sensi-
tive to bias voltage and temperature variations, and special power-supply
Optical Detectors 199
+ + + + +
+ + + + +
+ +
+ +
+ + + + + + +
+
hν
hν q(Vd+Vr)
+
EC
EF
+ hν Eg
+ EV
Open circuit
Lp Reverse bias
Rs i
Solar
cell
Short
circuit
Iph Isat In Il
Cp Cd Rd
v RL Rb
Rs Lp
Figure 5.27 Detector circuit model under all of the bias conditions. I ph is the signal current,
In is the noise current, Il is the temperature-dependent leakage (dark) current, Isat is the
saturation current, Cd is the depletion-layer capacitance, C p and L p are the packaging and
lead capacitance and inductance, Rd is the detector dynamic resistance (partially forward-
biased diode), Rs is the series lead, contact, and bulk resistance, Rb is the detector bias
resistance, and R L is the load resistance.
ure 5.27. Detector data sheets sometimes indicate typical values for these
various circuit elements, but the designer must often estimate values for
some detector parameters. The detector capacitance depends strongly on
the reverse-bias voltage as indicated above. The detector dynamic resis-
tance Rd is equal to the reciprocal of the slope of the I-V curve under
reverse bias. Silicon pin detectors have dynamic resistance values of sev-
eral MΩ. Detectors made from other detector materials, such as InSb, can
have much-lower dynamic resistances. The leakage current is the reverse-
saturation current. Note that if the detector has a low dynamic resistance,
an increase in reverse-bias voltage will also increase the leakage current
through the detector. This is not evident in silicon pin detectors but detec-
tor performance degrades under high reverse voltage for InSb and HgCdTe
detectors.
The photon flux creates minority carriers that are swept out across the
depletion layer. However, once out of the depletion layer, the carriers have
nowhere to go, and an accumulation of charge takes place. The increase
in charge results in an increase of forward voltage across the junction.
As the forward voltage across the junction increases, the forward current
through the device increases, depleting the accumulation of charge. Under
conditions of infinite load impedance, the reverse photocurrent results in
Optical Detectors 201
+
0
+ + + + 0
+ + + +
+ + + +
+ + + +
+ + + +
p type Drift + Diffusion n type
Diffusion Drift
hν q(Vd Vq)
hν
EC
EF
hν
Eg
+ Forward bias
current flow
+
EV
+
exactly the same magnitude forward current, resulting in a zero net output
current. Note that in this case the voltage across the diode is equal to
the depletion-region barrier voltage. The diode therefore sinks all of the
photocurrent under its own voltage; none flows out of the device (see
Figure 5.28).
The diode load is a high impedance, such that the detector external
current approaches zero, I = 0, following from Equation (5.128)
βkT I ph
V = ln +1 (5.133)
q Isat
βkT ηqλΦe
= ln +1 . (5.134)
q hcIsat
Silicon solar cells are operated in a mode to obtain maximum power trans-
fer. The solar cell is allowed to deliver power into a resistor, such that it is
neither short-circuit nor open-circuit. The circuit operation is halfway be-
tween the open-circuit and short-circuit bias conditions already discussed.
A high-value load resistor will cause the internal diode to ‘sink’ the pho-
tocurrent internal to the device. On the other hand, a zero-value load
impedance means that there is no power delivered into the load.
where ηΦq is the optically generated carrier rate, η is the quantum effi-
ciency, Φq (t) is the flux incident on the detector in [q/s], and Δn(t)/τ is
the rate of carriers moving through the depletion layer. Taking the Fourier
transform on both sides of Equation (5.135),
ΔN ( f )
i2π f ΔN ( f ) = ηΦq ( f ) − (5.136)
τ
1
ΔN ( f ) i2π f + = ηΦq ( f ) (5.137)
τ
ηΦq ( f )τ
ΔN ( f ) = , (5.138)
1 + i2π f τ
where f is frequency, ΔN ( f ) is the Fourier transform of the number of
free carriers due to optical excitation, Φq ( f ) is the Fourier transform of
the photon flux incident on the detector, and τ is the time taken for the
carrier to move through the depletion layer. The frequency response of the
detector is then
ΔN ( f ) τ
H( f ) = = . (5.139)
ηΦq ( f ) 1 + i2π f τ
How long does the carrier take to move through the depletion layer?
Consider a silicon diode with the following characteristics. The carrier mo-
bilities for carriers in silicon are 500 (holes) and 1300 (electrons) cm2 /(V·s).
A typical silicon-pin detector depletion width is 50 µm under an applied
20-V reverse voltage. The carrier transit time can be calculated from the
above information to be of the order of 1 ns. This silicon detector should
be able to internally respond to signals well into hundreds of MHz and,
with some care, even into the GHz frequencies.
The operational speed of most photovoltaic detectors are limited by
their operation in real-life capacitive electronic circuits. The depletion re-
gion capacitance, together with the series resistance and load resistance,
provides a very tough design challenge if high operating speeds are re-
quired.
The photovoltaic detector creates shot noise by virtue of its potential bar-
rier across the depletion layer, and Johnson noise in its dynamic resistance.
204 Chapter 5
The first term is the diffusion current, the second term is the thermally
generated current, and the third term is due to the photocurrent. Each
of these terms is an independent current and creates noise statistically
independent from the other currents and must be added vectorially. 2 The
shot noise PSD in [A2 /Hz], created in the detector, is therefore given by
where the integral is calculated over the detector’s sensitive spectral range,
η is the detector quantum efficiency, Ad is the detector area in [m2 ], and Eq
is the background photon irradiance in [q/(m2 ·s)] for a thermally radiating
background at temperature Tb in [K]. The NEP of the detector in [W] is
then
noise
NEPλ = (5.146)
Rλ
hc i2n Δ f
= (5.147)
ηqλ
hc Δ f βkT kTβ
qV βkT 4kT
= 2q e + + ηqΦq + . (5.148)
ηqλ qR0 qR0 R0
Optical Detectors 205
If the noise induced by the background flux exceeds the combined thermal-
generation current noise, diffusion current noise, and Johnson noise,
βkT qV/(kTβ) βkT 2kT
ηEq e + 2 + , (5.153)
q2 A d R 0 q A d R 0 q2 R 0 A d
the D ∗ is given by
∗ ηλ λ η
D = √ = √ λc . (5.154)
λ hc 2 Eq dλ
hc 2 η 0 c Eq dλ 0
Consider a diode with zero-bias voltage eqV/( βkT ) → 1. If the noise induced
by the background is less than the internal detector noise
λc
2kT 2βkT
2
+ 2 η Eq dλ, (5.155)
q R0 A d q R0 A d 0
then the D ∗ is given by
√
∗ ηλ qηλ Ad R0
D = √ 2βkT = . (5.156)
hc 2 q2 R A + 2kT 2hc ( β + 1)kT
0 d q 2 R0 A d
206 Chapter 5
1020
Background temperature 1 K
10 18 2K
λc [cm·Hz /W]
4K
1016
½
10 K
25 K
1014
77 K
D*
1012 195 K
300 K
500 K
1010 0
10 101 102 103
Peak wavelength λc [mm]
Figure 5.29 Theoretical limits to D ∗ for given detector cutoff wavelength and background
temperatures.
Equation (5.158) confirms that the open-circuit bias mode is noisier than
the short-circuit bias mode.
5. Calculate the spectral photon and radiant irradiance on the detector for
a target blackbody source with unity emissivity and temperature Ts =
208 Chapter 5
0.64
Quantum efficiency
0.56
0.48
0.40
106
0.32
0.24
0.16
0.08
105 0.00
1.0 1.5 2.0 2.5 3.0 3.5 4.0 4.5 5.0 5.5 1.0 1.5 2.0 2.5 3.0 3.5 4.0 4.5 5.0 5.5
Wavelength [mm] Wavelength [mm]
2.0
Current [mA]
Test source
101 Photon flux
1.5 induced current
0
10 ( )
1.0 Background
10-1
( ) Dark
0.5 10-2
0 10-3
1.0 1.5 2.0 2.5 3.0 3.5 4.0 4.5 5.0 5.5 −200 −160 −120 −80 −40 0 40 80 120
Wavelength [mm] Voltage [mV]
6.4 1.05
5.6 0.90
NEP [pW]
4.8 0.75
4.0
0.60
3.2
0.45
2.4
1.6 0.30
0.8 0.15
0.0 0.00
1.0 1.5 2.0 2.5 3.0 3.5 4.0 4.5 5.0 5.5 1.0 1.5 2.0 2.5 3.0 3.5 4.0 4.5 5.0 5.5
Wavelength [mm] Wavelength [mm]
Figure 5.30 InSb detector model results: (a) spectral absorption coefficient, (b) quantum
efficiency, (c) spectral responsivity, (d) current–voltage relationship, (e) spectral NEP, and
(f) spectral specific detectivity.
Optical Detectors 209
11. The NEP is determined from Equation (5.25), with the spectral respon-
sivity calculated in item 4 and the noise current calculated in item 10.
The NEP is shown in Figure 5.30(e).
12. The D ∗ is calculated from the area of the detector, the noise bandwidth,
the noise, and the responsivity, by Equation (5.26). The responsivity is
shown in Figure 5.30(f).
susceptible to high background photon flux (but also less sensitive to target
signals) because the detector spectral width is narrower than HgCdTe.
Multicolor detectors are currently in advanced research, and high-
performance detectors should appear on the market in the short-to-medium
term. One approach is to build a HgCdTe shorter-wavelength detector on
top of a longer-wavelength detector. Another approach uses monolithic
structures with different spectral responses as already reported in QWIP
IR photodetectors. 73,74
Several new technologies are being investigated, including quantum
dot detectors, 75 nearly matched lattice detectors, 76 and strained-layer su-
perlattice detectors. 77 Strained-layer superlattice detectors have bandgaps
smaller than any of the constituent materials. Quantum dot detectors con-
fine carriers in three-dimensional space using nanostructures. These tech-
nologies are still in the research phase.
Bibliography
[1] Singh, J., Electronic and Optoelectronic Properties of Semiconductors Struc-
tures, Cambridge University Press, Cambridge, UK (2003).
[3] Rogalski, A., Infrared Detectors, 1st Ed., Gordon and Breach Science
Publishers, Amsterdam, The Netherlands (2000).
[8] Rogalski, A., “Infrared detectors: status and trends,” Progress in Quan-
tum Electronics 27, 59–210 (2003).
[9] Vurgaftmann, I., Meyer, J. R., and Ram-Mohan, L. R., “Band Parame-
ters for III-V Compound Semiconductors and their Alloys,” Journal of
Applied Physics 89(11), 5815–5875 (2001).
Optical Detectors 213
[11] Murdin, B., Adams, A., and Sweeney, S., “Band Structure and High-
pressure Measurements,” Ch. 2 in Mid-infrared Semiconductor Opto-
electronics , Krier, A., Ed., 93–126, Springer, Berlin (2006).
[12] Rogalski, A., Infrared Detectors, 2nd Ed., CRC Press, Boca Raton, FL
(2011).
[13] Palmer, J. M. and Grant, B. G., The Art of Radiometry, SPIE Press,
Bellingham, WA (2009) [doi: 10.1117/3.798237].
[14] van der Ziel, A., Noise in Solid State Devices and Circuits, Wiley, New
York (1986).
[22] Ready, J., “Optical Detectors and Human Vision,” Ch. 6 in Fundamen-
tals of Photonics , Roychoudhuri, C., Ed., SPIE Press, Bellingham, WA
(2008) [doi: 10.1117/3.784938].
[24] Budzier, H. and Gerlach, G., Thermal Infrared Sensors: Theory, Optimi-
sation and Practice, John Wiley & Sons, New York (2011).
214 Chapter 5
[25] Kruse, P. W., Uncooled Thermal Imaging Arrays, Systems, and Applica-
tions, SPIE Press, Bellingham, WA (2001) [doi: 10.1117/3.415351].
[26] Datskos, P. and Lavrik, N., “Uncooled Infrared MEMS Detectors,” Ch.
12 in Smart Sensors and MEMS , Yurish, S. Y. and Gomes, M. T. S. R.,
Eds., 381–430, Kluwer Academic Publishers, Dordrecht, The Nether-
lands (2003).
[28] Biberman, L., Ed., Electro-Optical Imaging: System Performance and Mod-
eling, SPIE Press, Bellingham, WA (2001).
[29] Vollmerhausen, R. H., Reago, D. A., and Driggers, R. G., Analysis and
Evaluation of Sampled Imaging Systems, SPIE Press, Bellingham, WA
(2010) [doi: 10.1117/3.853462].
[32] King, K. F., Bernstein, M. A., and Zhou, X. J., Handbook of MRI Pulse
Sequences, Elsevier Academic Press, London, UK (2004).
[33] Rogatto, W. D., Ed., The Infrared and Electro-Optical Systems Handbook:
Electro-Optical Components, Vol. 3, ERIM and SPIE Press, Bellingham,
WA (1993).
[34] Li, C., Skidmore, G., Howard, C., Han, C., Wood, L., Peysha, D.,
Williams, E., Trujillo, C., Emmett, J., Robas, G., Jardine, D., Wan,
C.-F., and Clarke, E., “Recent development of ultra small pixel un-
cooled focal plane array at DRS,” Proc. SPIE 6542, 65421Y (2007) [doi:
10.1117/12.720267].
[35] Niklaus, F., Vieider, C., and Jakobsen, H., “MEMS-Based Uncooled
Infrared Bolometer Arrays — A Review,” Proc. SPIE 6836, 68360D,
Bellingham, WA (2007) [doi: 10.1117/12.755128].
[36] Song, W.-B. and Talghader, J. J., “Design and characterization of adap-
tive microbolometers,” J. Micromech. Microeng 16(5), 1073–1079 (2006)
[doi: 10.1088/0960-1317/16/5/028].
Optical Detectors 215
[37] Tissot, J., Trouilleau, C., Fieque, B., Crastes, A., and Legras, O., “Un-
cooled microbolometer detector: recent developments at Ulis,” Proc.
SPIE 5957, 59570M (2005) [doi: 10.1117/12.621884].
[39] Vandermeiren, W., Stiens, J., Shkerdin, G., Kotov, V., Tandt, C. D., and
Vounckx, R., “Infrared Thermo-Electric Photodetectors,” Ch. 8 in Laser
Pulse Phenomena and Applications , Duarte, F. J., Ed., 143–164, Intech
Open, New York (2010).
[40] Kasap, S. O., Principles of Electronics Materials and Devices, 3rd Ed.,
McGraw-Hill, New York (2006).
[41] Rogalski, A., “HgCdTe infrared detector material: history, status and
outlook,” Reports on Progress in Physics 68, 2267–2336 (2005).
[42] Ashcroft, N. W. and Mermin, N. D., Solid State Physics, Harcourt Col-
lege Publishing, Fort Worth, TX (1976).
[43] Sze, S. M. and Ng, K. K., Physics of Semiconductor Devices , 3rd Ed.,
Wiley-Interscience, New York (2007).
[49] Tilley, R., Understanding Solids, John Wiley & Sons, New York (2004).
[50] Neamen, D. A., Semiconductor Physics and Devices, 3rd Ed., McGraw-
Hill, Boston (2003).
[51] Nag, B. R., Physics of Quantum Wells Devices , Kluwer Academics Press,
Dordrecht, The Netherlands (2001).
216 Chapter 5
[53] Fox, M., Optical Properties of Solids, Oxford University Press, Oxford,
UK (2001).
[57] Hecht, E., Optics, 4th Ed., Addison Wesley, Boston, MA (2002).
[58] Chur, J. and Sher, A., Physics and Properties of Narrow Gap Semiconduc-
tors, Springer, Berlin (2008).
[59] Huag, H. and Koch, S. W., Quantum theory of the Optical and Elec-
tronic Properties of Semiconductors , 4th Ed., World Scientific, London,
UK (2004).
[63] Rogalski, A., Antoszewski, J., and Faraone, L., “Third-generation in-
frared photodetector arrays,” Journal of Applied Physics 105(9), 1–44
(2009) [doi: 10.1063/1.3099572].
[64] Horn, S. B., “Cryogenic cooling options for forward looking infrared
(FLIR),” Proc SPIE 0245 (1980) [doi: 10.1117/12.959339].
[67] Boyd, R. W., Radiometry and the Detection of Optical Radiation , John
Wiley & Sons, New York (1983).
[69] Ossicini, S., Pavesi, L., and Priolo, F., Light Emitting Silicon for Mi-
crophotonics, Springer, Berlin (2003).
[71] Daniels, A., Infrared Systems, Detectors and FPAs, 2 Ed., SPIE Press
(2010).
[73] Alves, F. D. P., Santos, R. A. T., Amorim, J., Issmael, A. K., and
Karunasiri, G., “Widely Separate Spectral Sensitivity Quantum Well
Infrared Photodetector Using Interband and Intersubband Transi-
tions,” IEEE Sensors Journal 8(6), 842–848 (2008).
[74] Santos, R. A. T., Alves, F. D. P., Taranti, C. G. R., Filho, J. A., and
Karunasiri, G., “Quantum wells infrared photodetectors: design us-
ing the transfer matrix method,” Proc. SPIE 7298, 72980B (2009) [doi:
10.1117/12.820251].
[76] Ting, D. Z., Hill, C. J., Soibel, A., Nguyen, J., Keo, S. A., Lee,
M. C., Mumolo, J. M., Liu, J. K., and Gunapala, S. D., “Antimonide-
based barrier infrared detectors,” Proc. SPIE 7660, 76601R (2010) [doi:
10.1117/12.851383].
[77] Razeghi, M., Hoffman, D., Nguyen, B.-M., Delaunay, P.-Y., Huang,
E. K., Tidrow, M. Z., and Nathan, V., “Recent Advances in LWIR Type-
II InAs/GaSb Superlattice Photodetectors and Focal Plane Arrays at
the Center for Quantum Devices,” Proceedings of the IEEE 97, 1056–
1066 (2009) [doi: 10.1109/JPROC.2009.2017108].
Problems
Material Eg Material Eg
Cds 2.4 PbS 0.42
CdSe 1.8 PbSe 0.23
GaAs 1.35 InSb 0.23
Si 1.12 HgCdTe 0.1
Ge 0.67
eV eV
5.4 This is an extension of the previous problems: draw the spectral
response of the various detectors accurately to scale on the same
graph, ranging from 0–14 µm. There is no need to do extensive
calculations and plots, just plot a few key points on the graph.
Use Matlab® or Python™ to get numerical solutions. [4]
5.5 Do a conceptual design of a two-layer sandwich detector that must
cover the spectral region from 2.5–14 µm. Describe which mate-
rials must be used in each layer. Which layer must be in front?
Draw the spectral response of the combined sandwich, for each of
the two layers, on the same graph. Note that the two detectors are
electrically independent but not optically independent! [6]
5.6 Calculate the response time of a silicon-pin diode with capacitance
of 5 pF, series resistance of 100 Ω, and negligible inductance and
leakage current in each of the four bias modes. [4]
5.7 Calculate the transit time of carriers through the silicon detector
described in Section 5.9.4. [2]
Optical Detectors 219
5.8 Verify Equation (5.143). Make sure that you understand why the
noise currents seem to add together. [2]
5.9 Determine the spectral responsivity of a silicon detector over the
spectral range 0.25–1.2 µm. Use the spectral absorption coefficient
data on the pyradi website. 78 To simplify the calculation you may
approximate the absorption coefficient by two or three straight-
line segments between 0.25–1.2 µm.
Consider the following two depletion layer designs: (1) d1 = 30 µm,
d2 = 60 µm, and (2) d1 = 4 µm, d2 = 34 µm. In both cases the de-
pletion depth is 30 µm, but in the first case the depletion region is
deep inside the detector, whereas the second is close to the surface
of the detector.
Which design is more suited to the detection of IR laser pulses at
0.9 µm? Which design is more suited to the detection of blue or
ultraviolet laser pulses at 0.33 µm? How would you design a sili-
con detector for optimum responsivity at a particular wavelength?
[8]
5.10 Compare the equations for D ∗ between the photoconductive and
photovoltaic cases. [2]
5.11 Use Equations (5.145) and (5.154) (or derive your own) to confirm
any three points in Figure 5.29. [6]
5.12 Use the spectral absorption coefficient data on the pyradi web-
site 78 to determine the coefficients for all of the materials, fitting
Equations (5.78), (5.79), and (5.80). [10]
5.13 Review Section 9.7 and then recalculate Figure 5.29 for a conical
background FOV with an half-apex angle of 15 deg. The cold
shield is at 77 K. [10]
5.14 A popular tea-time discussion topic involves the timing of when
to add the milk to the tea. Given a time lapse T, which cup will be
warmer: the cup with milk added at t = 0, or the cup with milk
added at t = T? Develop a mathematical model for the problem
and find an answer. Ignore heat conduction through the base of
the cup (the tea is served in styrofoam cups). Assume a constant
heat capacity of 4.15 kJ/(K·k). Plot the temperature of the tea
versus time for three cases: (a) milk at t = 0, (b) milk at t = T, and
(c) no milk. [8]
5.15 Derive Equation (5.34) in full, with motivation for the formulation
and derivation. [2]
5.16 Design a bolometer thermal detector element for use in an image
array operating at a frame rate frequency of 20 Hz. The detector
220 Chapter 5
6.1 Overview
The electro-optical system performs two functions. First, the optics focus
the ‘object’ (the thing looked at) into the ‘image’ (a geometrically scaled
representation of the object). 1–6 Second, the sensor collects and converts
221
222 Chapter 6
(Imaginary)
Optical axis
optical energy into a useable signal for subsequent use. In the system
chain, the sensor is the third building block, after the source and medium.
Figure 6.1 shows a conceptual representation of an optical sensor, defining
some terminology for use in this chapter.
The optical flux enters from the left through a window. The window
is normally a transparent optical material; e.g., BK7 glass for visual-band
systems or ZnSe for infrared systems. The window will probably have
a thin, multilayer anti-reflection coating to reduce optical reflections from
the window surfaces (see Section 3.4.3). Infrared window materials may
have a hard coating to protect the soft material of the window against
scratches.
The aperture stop is often designed as a mechanical device located
such that vignetting is reduced or eliminated. Some systems do not have
a dedicated aperture stop, and the diameter of one of the lenses performs
the function.
The optical elements (windows, lenses, and mirrors) can be divided
into two categories: elements with optical power and elements without
optical power. Surfaces without power (plano surfaces) have no focusing
effect, whereas surfaces with power (nonplanar surfaces) affect the diver-
gence or convergence of optical rays. The optical elements are held in
place in the optical barrel by mounting rings, locating the elements accu-
rately in separation and lateral displacement. Some form of strain relief is
required to prevent the optical elements from breaking at temperature ex-
tremes — a typical strain relief is a rubber spacer on a noncritical surface.
Sensors 223
where s is the image distance, s is the object distance, f is the focal length,
h is the image height, and h is the object height. The distances s and s are
positive measured toward the right and negative measured toward the left
(the same as x in the Cartesian coordinate system).
The optical axis is the ‘centerline’ of the optical system, normally
defining zero field angle. Optical systems have ‘planes’ along the optical
axis with special properties. Figure 6.2 shows a system with two principal
planes P1 and P2 , and two focal planes F1 and F2 . These planes are ide-
ally plano (flat) and perpendicular to the optical axis. In real-world optics,
these ‘planes’ are not flat but could be any shape, depending on the de-
sign. In a simple optical system the principal and focal planes tend to be
near-spherical. More-sophisticated systems are designed to achieve near-
flat focal planes because detector surfaces are flat. The planes are often
rotationally symmetric and centered on the optical axis, but this depends
on the design of the system. In this chapter, the term ‘plane’ is used to
denote an optical plane (which is normally not mathematically flat).
A ray passing through the first F1 (front) or second F2 (back) focal
points will be parallel to the optical axis after traversing the lens. A ray
passing through the first or second principal points P1 or P2 will traverse
through the lens with no change in direction. An object at location H1 will
be imaged to the image location at H2 (in the paraxial approximation).
The location of H1 (at distance s) and H2 (at distance s ) are called
conjugates of each other. ‘Infinite conjugates’ is a special but common
case: when the object is located at infinity s → −∞, then s = f , and the
image is formed in the back focal plane of the lens or telescope. Most
real-world optical systems operate at near-infinite conjugates because the
object is normally located a substantial distance from the optical system.
The marginal ray is the ray from the optical axis at the object or image,
through the edge of the pupil. The chief ray is the ray from the edge of
the field stop, through the principal points of the optics. The chief ray
propagates through the lens with no change in direction. The field angle
is defined by the chief ray, as the angle defined the image height h and s
or the object height h and s.
A real optical system has the first and second principal points at sep-
arate locations, as shown in Figure 6.2(a). For first-order studies the op-
tical system is often simplified to the ‘thin-lens’ approximation shown in
Figure 6.2(b). In the thin-lens approximation the front and rear principal
points coincide. The principal points and focal points all lie in planes,
perpendicularly intersecting the optical axes at P, F1 , and F2 .
Sensors 225
Ma
Object Ray parallel to optical axis rgi
na
θ Ray th l ra
rough y
h Field
center
of lens
angle
H1 2F F P Chief r
ay F θ H2 2F h
Thin lens
Approximation
Ray parallel to optical axis Image
(b)
Figure 6.2 Object and image relationship in an image: (a) thick lens and (b) thin-lens ap-
proximation.
2F F P F 2F
h
3F 2F F P F 2F
h
h 2F
3F 2F F P F
A0
Ch A1
M ief
ray
β ar
gin 2 Detector
al M area
ra ar
y2 gi
na
lr
ay
Chie 1
f ray a
1
¥
Figure 6.4 Sensor field of view defined by optical and detector parameters.
could be fixed (e.g., a metal plate cut to specific size and shape) or it could
be variable (e.g., a variable iris for which the diameter of the opening can
be adjusted). Figure 6.5(a) shows the apertures and stops in a real lens.
In the thin-lens approximation, the entrance pupil falls in the plane of the
thin lens, as shown in Figure 6.5(b).
Optical systems often suffer from unintended stray light when observ-
ing high-contrast scenes. The light could be reflected from optical surfaces
or from the inside of the optomechanics. Stray light can be minimized by
locating nonreflective stops or baffles so as to suppress the internal reflec-
tions.
A stop located in an image plane is called a field stop because this stop
limits the field of view of the system. Some systems have more than one
image plane (i.e., a telescope or microscope). By definition, these image
planes are images of each other. The field stop can thus be located at any
image plane and still have the same effect.
The pupil of an optical system is a plane that limits the amount of
light flowing through the system. The lens diameters limit the amount of
light passing through each lens, but because each lens limits the light in
a different ‘plane’ of the system, the light control is not very precise and
can lead to vignetting when some optical elements ‘shade’ other optical
elements. Vignetting is the effect where one or more optical elements limit
the amount of light nonuniformly over the image plane. The usual effect is
228 Chapter 6
Entrance Exit
pupil Pupil pupil
Field
stop
l ray
Margina
H
Object at infinity F2
D θ
s → −∞ F1 P1 P2 f
(a)
Marginal ray
D θ
f F2
Principal plane for
thin-lens approximation
(b)
Figure 6.5 Stops, numerical aperture, and f -number: (a) complex (thick) lens and (b) thin-
lens approximation.
that some (or all) edges of the image are darker than the center of the im-
age. Severe vignetting may even cut off portions of the image completely,
e.g., when the eye is not well aligned with the optical axis of a telescope.
One way to control vignetting is to accurately size the different optical
elements appropriately. A better way to control vignetting is to make all
of the elements slightly bigger than necessary and then to introduce a
mechanical element (aperture stop) to limit the light in a single plane in
a controlled manner. This approach also allows setting the aperture to
different diameters, such as setting the aperture f -stop in a photographic
camera.
The single element that limits the amount of light flow in the optical
system is called the pupil. The pupil is normally a physical device inside
the optics. The optical elements before the pupil image the pupil into
Sensors 229
object space into the ‘entrance pupil.’ Likewise, the optical elements after
the pupil image the pupil into image space into the exit pupil, i.e., seen
from the image, this is the diameter of the optical system. The marginal
ray in an optical system is the ray touching the edge of the pupil. In the
thin-lens approximation, the entrance and exit pupil coincide in the thin
lens and have the diameter of the thin lens, as shown in Figure 6.5.
The f -number (f /# or F#) is a geometric construct comprising the di-
ameter of the exit pupil and the focal length; it has no other meaning. The
f -number is a convenient means to denote the amount of optical flux flow-
ing through an optical system of a given focal length. The f -number of the
optical system is defined as f /# = f /D for a system at infinite conjugates,
where f is the focal length, and D is the diameter of the exit pupil. In
photography the term f -stop is used to denote the size of a lens entrance
pupil, in a progression of doubling areas for each successive f -stop. f -
stop values are f -numbers, usually in a series of the form f /1.4, f /2, f /2.8,
f /4, f /5.6, f /8,... The optical systems encountered in this book are fixed
f -number systems, and the f -number is only used to denote the ratio of
focal length to diameter. Higher f -numbers result in lower irradiance on
the focal plane. Systems with lower f -numbers (higher image-plane irra-
diance) become increasingly difficult to design with high optical perfor-
mance, f /1.2 being a reasonable limit for good imaging quality. Generally
speaking, a low-f -number lens is more complex and more expensive than
a high-f -number lens.
The f -number is only defined for a system at infinite conjugates. For
systems not at infinite conjugates, the ‘numerical aperture’ is used as a
measure of the amount of flux flowing through the system at a given focal
length. The numerical aperture is given by NA = n sin θ, where n is the
index of refraction of the medium in which the lens is located, and θ is
the angle of the marginal ray in image space, as shown in Figure 6.5. For
lenses in air, n = 1. The numerical aperture is not constrained to infinite
conjugates, and hence it can vary depending on the location of the im-
age. The numerical aperture should not be estimated from the size of the
aperture or lens diameter but from the inclination of the marginal ray (i.e.,
from the optical design).
At infinite conjugates the f -number and numerical aperture in image
space are related by
D D
NA = n sin θ = n sin arctan ≈n , (6.4)
2f 2f
and hence
n
F# = . (6.5)
2 NA
230 Chapter 6
Following from the previous sections, it should be clear that there are two
solid angles of concern in an optical sensor. These two solid angles are
indirectly related to each other but have very little to do with each other.
This section explores the field of view (FOV) solid angle and the image-
flux-collecting solid angle, as depicted in Figure 6.6.
The flux-collecting property of an optical system is determined by
the diameter of the pupil (the lens diameter) and is quantified by the nu-
merical aperture or the f -number of the system. These two numbers are
determined only by the marginal ray in the system. The bottom picture
in Figure 6.6 shows the optical system and the marginal ray touching the
pupil. It is evident from the bottom picture that the flux within the pupil
is focused by the lens and collected in the detector on the focal plane.
The irradiance in a point on the focal plane is determined only by the f -
number/numerical aperture/marginal ray. The size of the detector has no
direct effect on the flux-collecting performance of the optical system. The
chief ray maps the field stop into object space, thereby determining the
sensor’s FOV. The size of the pupil has no direct effect on the FOV of the
system.
Consider the top picture in Figure 6.6, the flux collected by the optical
system: not all is collected within the field stop. The field stop blocks the
flux outside of the desired field of view — this flux does not reach the focal
plane. The sensor is only sensitive to flux within the field stop. The chief
ray, determined by the size of the field stop and s , therefore determines
the sensor’s FOV. The size of the pupil has no direct effect on the FOV of
the system.
Sensors
The field stop (detector size) and distance
Object between the pupil and the image plane
Field determine the angle of the chief ray and
Focal
plane Field plane
stop hence field of view of the sensor stop
image
A Chief ray
A’
Chief ray
Detector responds Detector size
only to flux from this area
Figure 6.6 Optical sensor spatial angles.
s s’
dA dA’
Flux radiated by the object Lens collects flux Lens focuses flux from In this solid angle the
outside the pupil solid angle from the object the lens in this solid detector responds to flux
is not collected by the sensor in this solid angle angle on the detector from the sensor housing
231
s s’
232 Chapter 6
For the discussion in this section assume a uniform radiance over an ob-
ject’s area. If the object is smaller than the FOV, the object can be con-
sidered a point target in terms of the sensor FOV: the object’s size has no
effect on its signal in the FOV. If an object is larger than the FOV, it means
that the full sensor FOV is filled by the object (called an extended target or
object in terms of the pixel size) because the object area is bigger than the
sensor FOV. The flux contributed by an extended object is defined by the
size of the footprint of the sensor FOV on the object and not by the object
size itself. See also Section 7.6.
Ideal, paraxial optical systems form perfect images; but practical optical
systems do not because of optical aberrations. 1–5,7 Optical aberrations are
not only the result of poor workmanship or poor quality control but also a
mathematically founded physical process. Elements with optical power
(lenses or curved mirrors) refract or reflect optical rays toward a focal
point; but a single point on the object may not always image to a single
point in the focal plane.
Geometrical aberrations 8 occur as a result of the geometry of the sur-
faces. There are huge families of physical aberrations, including several
variations of spherical aberration, coma, astigmatism, and distortion. Fig-
ures 6.7 and 6.8 depict only the first-order aberrations. Spherical aber-
rations arise when rays distant from the optical axis focus in a different
plane as do paraxial rays. As a result, there is a plane located at the ‘circle
of least confusion’ that provides the smallest optical spot size. Chromatic
aberration occurs as a result of the index of refraction in lenses not be-
ing constant for different colors. This means that blue light does not focus
where green light focuses, and not where red light focuses. Astigmatism is
an aberration when the optical power in the vertical and horizontal planes
differ (e.g., cylindrical deviations from a spherical shape). Comatic aber-
ration occurs when an off-axis spot is focused in different locations in the
Sensors 233
Lens
Paraxial
Circle of least ray focus
confusion
Transverse
spherical aberration
Longitudinal
Marginal
spherical aberration
ray focus
(a)
Lens
(b)
(c)
Figure 6.7 Optical aberrations: (a) first-order spherical aberration, (b) chromatic aberration,
and (c) astigmatism.
234 Chapter 6
Object plane
Lens
Image plane
(a)
Planar
object Lens
Curved
image
(b)
(c)
Figure 6.8 Optical aberrations: (a) first-order coma, (b) field curvature, and (c) distortion.
Sensors 235
focal plane by different parts in the lens. Coma gets its name from the
comet-like shape of the spot. Field curvature is when the image of a plane
in the object space is imaged to a curved surface. A focal surface with
constant distance from the rear principle point (a simple lens) will have a
spherical field curvature. Distortion arises when constant distances in the
object plane are no longer constant in the image plane. The point spread
function of off-axis aberrations (i.e., coma, astigmatism, and distortion) are
not rotationally symmetric.
The main purpose of optical design is to create and employ degrees
of freedom in the optical design so as to minimize the system’s aberra-
tions. These degrees of freedom include the number of optical elements,
curvature of surfaces, thickness between surfaces, and index of refraction.
The optical designer adjusts all of the degrees of freedom in the design
to balance the optical aberrations against each other so as to cancel out,
thereby providing the required optical performance. Optical design is a
very sophisticated topic beyond the scope of this book.
theory it can be shown that the image in the focal plane of an optical sen-
sor will be the convolution of the PSF with the ‘ideal’ image in the absence
of diffraction and aberrations.
The two-dimensional Fourier transform of the optical PSF yields the
optical transfer function 8 (OTF), a complex-value optical frequency re-
sponse. The modulation transfer function (MTF) of the optical system is
the magnitude of the OTF. The MTF of a system with aberrations is not al-
ways rotationally symmetric. MTF is commonly used for the specification,
design, and testing of lenses.
Secondary Primary
mirror (flat) mirror
Field
stop
Secondary Primary
mirror (curved) mirror
The optical power in an optical surface stems from the curvature of the op-
tical surface. In principle, any curved surface can produce optical power,
as is evidenced by studying common, household glass items. Most optical
elements were traditionally manufactured with spherical profiles because
such profiles are easily formed by conventional lapping and polishing tech-
niques. There is also a class of surfaces called aspherical surfaces with
useful optical properties.
The profile of an aspheric lens is neither spherical, cylindrical, nor flat.
During optical design the surface of an aspherical element can be specified
such that it provides optical power, but at the same time minimizes or
corrects optical aberrations caused by other elements. In particular, the
aspherical surface reduces or eliminates spherical aberration. This means
that fewer optical elements are required, and the lens assembly is simpler
and lighter. Aspherical elements are more difficult to manufacture than
spherical optics and are thus used where the total system solution provides
benefits such as weight, complexity, or cost.
238 Chapter 6
Object at Object
long range image
Collimator
fc
Col
l
foca imator Object
l len
gth
Object Collimator Sensor
dc
ds
fc Collimated
(parallel) rays Object
fs
image
Many sensors are focused at infinite conjugates, hence there is need for an
instrument that can create infinite object images within the limited confines
of the laboratory. A collimator is simply an element with optical power
(lens or mirror), focusing an object at infinity. Because the image is located
at infinity, the optical rays are parallel to the optical axis (collimation).
The setup can be depicted as shown in the top picture in Figure 6.11.
In terms of the flux transfer defined in Equation (2.31), the object range
R01 must be replaced by the collimator focal length. The sensor is focused
at infinity and thus can image the collimated object image, in focus, on
the sensor focal plane. The image diameter in the sensor focal plane is
given by ds = dc f s / f c , where f s is the sensor optics focal length, f c is the
collimator focal length, and dc is the object diameter.
The collimator provides an ideal means to perform laboratory work at
infinite conjugate sensors, but there are practical considerations complicat-
ing such work. These considerations include optical axis alignment, lateral
placement in the beam, and beam vignetting. If the collimator and sensor
optical axes are not parallel, the object image will be displaced from the
center of the sensor image. If the sensor is displaced laterally, part of the
flux in the beam will be lost.
One common use of a collimator is to simulate a source at long range
in the laboratory. Figure 6.12 illustrates the vignetting effect that occurs
Sensors 239
3
Dc Source Source
dc image 4 image
A B
1
2
fc Rv
for a large source and small collimator focal lengths. Flux emanating from
the center of the source is collimated to parallel rays (solid lines in Fig-
ure 6.12). Flux emanating from the outer edges of the object is collimated
to parallel rays at an angle to the optical axis (see the two smaller inserts
in Figure 6.12).
From the geometry, it is evident that the only part of the beam pro-
viding flux from the full source area is the conical-shaped zone 1. Zone 4
does not contain any flux from the outer rim of the source, whereas zones 2
and 3 vignette either the bottom or top part of the object area. This sim-
ple geometrical analysis does not take into account the sensor’s FOV and
entrance aperture diameter, both of which place even further restrictions
on the sensor placement within zone 1. If it is important to obtain all of
the flux emanating from the source (i.e., when calibrating a sensor), it is
necessary to locate the whole of the sensor aperture in zone 1. The sensor
located at position A in the beam has no vignetting. The sensor located at
position B in the beam will only receive flux from the central part of the
source.
Using similar triangles, it follows that Rv /Dc = f c /dc , hence Rv =
Dc f c /dc . The depth of nonvignetted zone 1 can be increased by using a
larger collimator (larger Dc and f c ) or a smaller source (smaller dc ). A
smaller source area might not be feasible because the source area is nor-
mally selected to control the total amount of flux from the source.
240 Chapter 6
NEL
NEM
NETD NEE NEP id v
Extended
target object
Ds
α
A0 Pixel field
of view as Optics
defined by A1
detector
β
τa Filter
τs Detector
a Electronics
Object is larger Zt
R01 > f
than pixel field of
view but only flux
b
within the field of
view contributes
to the signal flux f
Figure 6.13 with Figure 6.6, specifically with respect to the role of the optics
in this sensor.
Most sensor systems operate at or near infinite conjugates; for this
model it is assumed that the sensor is focused at infinity s → −∞, and
hence s → F2 . Accurate mathematics require that tan α = a/ f , but for
small angles tan α ≈ α. The respective FOV angles for small angles are then
given by the chief ray in terms of the detector size and optics focal length
by the following simplified equations:
a
α = , (6.6)
f
b
β = , and (6.7)
f
ω = αβ
ab
= . (6.8)
f2
These equations state that a lens with focal length f , in [m], with
242 Chapter 6
a detector with dimensions a and b, in [m], in the focal plane has FOV
angles α and β in [rad]. The sensor pixel FOV has a solid angle FOV ω in
[sr]. For focus at infinite conjugates the optics diameter and clear aperture
area are described in terms of the f -number and detector size by
f
Ds = , and (6.9)
F#
PπDs2
As = (6.10)
4
Pπ f 2
= (6.11)
4F#2
Pπab
= , (6.12)
4ωF#2
where F# is the f -number of the optics, Ds is the optics aperture diameter
in [m], As is the clear aperture area in [m2 ], and the factor P is used to
account for diverse loss effects such as vignetting or central obscuration.
Extending on the optical sensor model developed in Section 6.5, this sec-
tion develops a radiometric model of a simple source–medium–sensor sys-
tem to calculate the signal at various locations in the system. In terms of
Equation (2.33) the source is modeled as a Planck radiator with a spec-
tral emissivity 0λ . The elemental area dA1 is now the entrance pupil of
the optical system, focusing flux onto a detector. It is common practice to
locate an optical filter with transmittance τsλ , near the pupil A1 in order
to selectively control the flux flowing through the detector. An amount
αsλ is absorbed in the receiver area (in most cases this value αsλ is already
accounted for in the detector responsivity). The detector converts the flux
flowing through A1 to an electronic signal. This signal is amplified and
can be used as a measure of the optical flux through the sensor’s entrance
pupil.
regroup the variables, and integrate over all wavelengths, to obtain the
signal voltage as
Zt dA0 cos θ0 A1 ∞
kR
dvS = 0λ L0λ τaλ Sλ dλ, (6.16)
R201 0
244 Chapter 6
where vS is the detector voltage [V] for flux in the band defined by S , and
the elemental source area dA0 has been retained. Equation (6.16) can be
integrated over the source area to obtain the full source signature:
∞
1
v S = k R Zt A 1 0λ L0λ τaλ Sλ dλ dA0 cos θ0 . (6.17)
A0 R201 0
Note that Equation (6.17) can be calculated in either the radiant (watts)
or photon (photons/second) domains. In the radiant domain, use Equa-
tions (3.1) and (5.5). In the photon domain, use Equations (3.3) and (5.4).
It is convenient to define an irradiance, called the apparent irradiance,
or sensor inband irradiance, which is the irradiance from a given source
to a given sensor observed through a given medium. The apparent sensor
inband irradiance is defined as
Φλ
ES = (6.18)
A1
k dA0 cos θ0 λ2
= 0λ L0λ τaλ Sλ dλ. (6.19)
R201 λ1
From Equation (6.19) it can be seen that the source radiance is spec-
trally weighted with the system response and the medium transmittance,
hence the name apparent irradiance: ‘as observed by the sensor.’ Equa-
tions (6.15), (6.16), and (6.19) form the basis of the system model. These
equations calculate the expected signal for any object at any distance. Be-
cause the spectral integrals contain factors of the source, the atmosphere,
and the sensor, it is necessary to calculate the integral for any new set of
sensor/object data or change in atmosphere or distance.
The previous section assumed a small and uniform source with area dA0 .
By integrating Equation (6.16) across the source area A0 , the total detector
signal due to an arbitrary source can be determined as
λ
2 d( A0 cos θ0 )
v S = k R Zt A 1 0λ L0λ τaλ Sλ dλ , (6.20)
source λ1 R201
where the spectral integral is calculated first for each elemental area dA0 ,
and then over the full area A0 . Note that L0λ may be a function of dA0 , i.e.,
the source radiance may vary over the surface of the source.
In the case where the source size is small compared with the distance
between the source and the detector, R01 reduces to the distance between
Sensors 245
the source and the detector. If in addition, the source has uniform radiance
over its total area, and the source solid angle is less than the sensor field of
view, the spatial integral over the source can be separated from the spectral
integral:
λ2
Zt A 1 d( A0 cos θ0 )
vS = k R 0λ L0λ τaλ Sλ dλ . (6.21)
λ1 source R201
The source spatial integral is actually the source solid angle
d( A0 cos θ0 )
Ωs = , (6.22)
source R201
and the detector signal can be written as
λ2
Zt A 1
vS = Ω s k R 0λ L0λ τaλ Sλ dλ. (6.23)
λ1
The mathematical model described thus far only considers one source el-
ement. A complex source may contain a number of different radiators
with highly different characteristics. One portion may be very hot and an-
other portion may be much cooler. For such a source, the total irradiance,
within a single FOV, can be found by summing the contributions from all
N elements as follows:
λ2
Zt A 1 ∑
N
A 0i cos θ 0i
vS = k R 0λi L0λi τaλ Sλ dλ . (6.24)
i=0 R201i λ1
but are appropriately scaled. The different locations are called ‘planes’
following the practice in optical design. As shown in Figure 6.13 there are
four planes of importance: the object plane at the source object, the entrance
pupil or optics plane at the entrance pupil of the optics, the image plane or
detector plane, and the electronics plane.
The mechanism to convert noise is based on the standard radiometry
methodology. The same transformations that are used for optical flux are
used exactly the same way for noise transformations. In particular, note
the following relationships:
LA0 cos θ0 Mω
E= = Lω = , (6.25)
R 2 π
which are extended to the equivalent noise representations
NEL A0 cos θ0 NEM ω
NEE = = NEL ω = . (6.26)
R 2 π
Note that in the above equation, the projected solid angle should be
used because the thermal camera senses Lambertian sources. However, for
the small pixel fields of view generally used, the projected and geometrical
solid angles are numerically equal.
The conventions followed in this book are as follows (the same con-
ventions apply to both noise and signal):
1. Spectral noise equivalent power (NEPλ ) is the optical signal power spec-
tral density [W/µm] required on the detector to yield a signal equal to
the noise in the sensor. NEPλ is always measured in the detector plane
and is expressed as flux in the image plane. See also Section 7.1.3.
NEP λ R201
NEL λ = , (6.28)
A0 A1 τs τa
where A0 is the pixel footprint in the object space, R01 is the distance
between the sensor optics and the object plane, and τa is the atmospheric
transmittance. See also Section 7.1.3.
where NETD λ has units of [K], and dL/dT is Planck’s law temperature
derivative [see Equation (3.4)]. See also Sections 7.1.3 and 9.6.
Wso Wdo
As Wos Optical axis
Ad
Wod
Chief ray
Object
plane Image
plane
Entrance
pupil
Equation (2.31) forms the basis of all radiometric calculations; the flux
flowing from a source to a receiving area can be written as
L dA0 cos θ0 dA1 cos θ1
d2 Φ = ,
R201
= L dΩ0 dA1 cos θ1 ,
= L dA0 cos θ0 dΩ1 , (6.31)
which implies that the total flux is proportional to the (spatially invariant)
radiance times the product of the source area and the receiver solid angle,
or vice versa. It follows that the source could have been on either side; the
total flux transfer depends on the geometrical relationship between the two areas,
irrespective of which is the source.
For the simplified lossless imaging system shown in Figure 6.15, the
argument above can be extended further. It can be shown that the total
flux flowing through the source and detector is given by
d2 Φ = L dΩso dAo cos θo ,
= L dΩos dAs cos θs ,
= L dΩod dAd cos θd , and
= L dΩdo dAo cos θo .
In all of the above equations it is evident that the total flux flowing
through the system is proportional to the product of a solid angle and the
Sensors 249
appropriate projected area. This product of solid angle and area is called
the throughput of the system. System basic throughput, or étendue, is an
indication of the total flux that can pass through the system. It depends
on the FOV and the aperture of the system. It is defined by
T = n2 A Ω, (6.32)
where n is the refractive index at the location where the solid angle Ω is
defined, and A is the area of the aperture at the location where the solid
angle is defined. In terms of the optical system of Figure 6.15, the étendue
is given by n2o Ao Ωs . The units of throughput are [sr·m2 ].
In its most-basic form, the power flowing through the source and re-
ceiver areas is given by the product of the basic radiance and the system
throughput,
L
Φ = n2 AΩ,
n2
LA0 A1
= .
d2
Consider two lenses for use with a staring array detector with a di-
agonal size of 13 mm. The two diagonal field angles required are 5 deg
and 25 deg. Assume that both lenses have apertures of f/1.8. First-order
calculations yield the lens specifications shown in Table 6.1. It is clear that
the wide-angle lens has a smaller diameter, and the narrow angle lens has
a much bigger diameter. This ‘unfortunate’ fact is due to the limitation im-
posed by the throughput of the system (fixed image height and f -number).
When a sensor with given throughput (fixed detector size and f -num-
ber) views an extended target object, the power on the detector stays the
same, irrespective of the FOV. Hence, the sensor gain calibration is identi-
cal for all fields of view (for the given f -number).
250 Chapter 6
Bibliography
[1] Shannon, R. E., The Art and Science of Optical Design , Cambridge Uni-
versity Press, Cambridge, UK (1997).
[2] Fischer, R. E., Tadic-Galeb, B., and Yoder, P. R., Optical System Design ,
McGraw-Hill, New York (2008) [doi: 10.1036/0071472487].
[3] Smith, W. J., Modern Lens Design , 2nd Ed., McGraw-Hill Professional
(2004).
[4] Smith, W., Modern Optical Engineering , 4th Ed., SPIE Press, Bellingham,
WA (2007).
[5] Kingslake, R. and Johnson, R. B., Lens Design Fundamentals, 2nd Ed.,
SPIE Press, Bellingham, WA (2010) [doi: 10.1016/B978-0-12-374301-
5.00003-6].
[6] Walker, B., Optical Engineering Fundamentals, 2nd Ed., SPIE Press,
Bellingham, WA (2009).
[7] Schmidt, J. D., Numerical Simulation of Optical Wave Propagation: With
examples in MATLAB , SPIE Press, Bellingham, WA (2010).
[8] Hecht, E., Optics, 4th Ed., Addison Wesley, Boston, MA (2002).
Problems
6.1.1 Determine the instantaneous solid FOV of the sensor (in radians).
[2]
6.1.2 Determine the geometric solid angle subtended by the lens as seen
from the detector. [2]
6.1.3 Determine the projected solid angle that the barrel (the walls of
the tube holding the lens and detector) of the sensor subtends as
seen from the detector. [2]
6.1.4 Determine the throughput of the sensor. [2]
1.4
1.2
1.0
0.8
0.6
0.4
0.2
0
0 1 2 3 4 5 6
Wavelength [mm]
6.6.1 Plot the detector current versus input power for the four elements.
Observe that there are four different response lines. [4]
6.6.2 A vertical line on the graph represents the optical power from a
uniform source. Predict what the detector current values will be
for an optical power of 0.25 µW. [2]
6.6.3 Comment on how the signals from the four elements can be com-
pensated (by external electronics) such that the sensor output sig-
Sensors 253
nal will be the same for all four detectors, irrespective of the input
power. [2]
6.6.4 Using a block diagram only, do a conceptual design of a circuit
that will perform a nonuniformity correction (NUC) for each of
the four elements in this sensor. [4]
1.2
1
0.8
0.6
0.4
0.2
0
0 1 2 3 4 5 6
Wavelength [mm]
6.8 A square silicon detector is pointed toward the sun (normal vector
is directed to the sun). The detector has an area of 1 cm2 . You may
assume unity atmospheric transmittance.
The sensor silicon detector responsivity can be modeled by Equa-
tion (D.5) with (λc = 1.15 µm, k = 8, a = 3.5 and n = 4.3).
The sensor filter transmittance can be modeled by Equation (D.4)
with (τs = 0, τp = 1, λc = 0.8, Δλ = 0.21 and s = 100).
254 Chapter 6
Spectral response
Responsivity
[A/W]
0
0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 1.1 1.2
Wavelength [mm]
255
256 Chapter 7
In the realist view, the time domain is a dimension with irreversible and
sequential flow or movement from past, through present, into future. Time
is described in a single real number relative to some fiducial epoch (start-
ing event). When modeling time in mathematical terminology, time may
appear to be reversible.
In the Cartesian view, the space domain is a three-dimensional ex-
tent in which objects have freedom of movement and direction. Space is
commonly modeled in a coordinate system with the convention of using
( x, y, z) to denote the three Cartesian dimensions. The origin of this coor-
dinate system is uniquely defined in different problem spaces.
Foreground focus attaches positive importance or priority to an event
or signal, in the context of its application. Foreground is the primary focus
of an investigation.
Background focus attaches a neutral or negative priority to an event
or signal, in the context of its application. Background interferes with the
analysis of foreground. Background is everything that is not foreground.
Signal is a foreground continuous time or discrete function in time or
space domains that conveys information.
Noise is a background continuous time or discrete function in time
or space that interferes with the foreground function or event. Noise has
a connection with unfolding time but can also be expressed in the space
domain. There is a strong connotation that noise has mostly random be-
havior, but randomness is not a mathematical requirement (e.g., electro-
magnetic interference is regarded as a noise but is not random).
Clutter is a special type of noise, often with a connection to the geom-
etry of the real world. Clutter is normally used in the context of image
processing or radar processing, being caused by physical objects in the
Radiometry Techniques 257
The measures listed below are widely used to quantify the performance
of linear (or near-linear) systems. The list is not complete because many
designs have their own unique performance measures. Designers must
identify the relevant measures for their particular designs. Performance
prediction and verification of nonlinear systems are not easily performed
by simple measures such as these; nonlinear system performance is gen-
erally determined by using scenario-based simulation models or the real
hardware item.
Signal-to-noise ratio (SNR): SNR is the ratio of the peak signal value to
the rms value of the noise. Higher SNR values imply better system
performance. SNR is unitless.
This measure depends on characteristics of the source, medium, and
sensor. The noise includes sensor noise or photon noise in the source.
It is normally easy to determine the SNR, both by measurement and
calculation. The SNR is improved by optimizing the bandwidth, in-
creasing source intensity, or decreasing sensor noise.
Signal-to-clutter ratio (SCR): SCR is the ratio of the peak signal value
to the peak clutter value. Higher SCRs imply better system perfor-
mance. SCR is unitless.
This measure depends on the characteristics of the surroundings of
the foreground target object. The clutter could be terrain objects re-
flecting ambient light or infrared radiation from hot terrain objects
such as rocks. The medium and sensor have only limited effect on
the SCR. The SCR is often measured because it is difficult to calcu-
late in analytical form. The SCR is improved by increasing source
intensity or optimizing the sensor FOV.
258 Chapter 7
Noise equivalent power (NEP): NEP is the optical signal power required
to create an electronic signal such that the SNR is one. Lower values
of NEP imply a better performance. NEP has units of [W].
NEP requires the quantification and exact definition of noise and
signal. NEP considers the total wideband noise (not power spectral
density), and hence the bandwidth must be specified. Likewise, the
properties of the source and medium determine the signal and must
be specified.
NEP can be defined for a complete sensor or just a detector on its
own. NEP is mostly calculated because it is difficult to measure di-
rectly. The NEP is improved by decreasing the various noise sources.
See also Section 6.7.
False alarm rate (FAR): For a signal corrupted by noise, the false alarm
rate is the rate by which the threshold is exceeded in the absence of
a signal. FAR is unitless. The false alarm rate is related to Pn by
NPn
FAR = , (7.3)
td
where N is the number of detectors, and td the time on target.
7.2 Normalization
1
1 pole
2 pole
3 pole
Gain
4 pole
0.5 # Poles Df / f-3dB
1 1.57
2 1.11
3 1.05
4 1.025
0
0 0.5 1 1.5 2 2.5 3
Normalized frequency
Figure 7.1 Butterworth filter noise equivalent bandwidth for different filter orders.
where 0λ L0λ τaλ is the source radiance after transmission through the at-
mosphere, and S is the weighting function. In this case, the apparent
source radiance is re-mapped to the sensor’s system response S to repre-
sent the radiance as observable by the sensor.
264 Chapter 7
Section 6.6 describes how to calculate the signal that a given sensor with
response S would receive from a source L, through some medium with
transmittance τ01λ . Suppose that this sensor has a spectral filter τ f with a
narrow (but nonzero) spectral width. This filter is very narrow compared
with its central wavelength, say Δλ = 0.01λc . Such radiometers are used
to determine the spectral radiance of sources or to measure the spectral
transmittance of the atmosphere. The apparent irradiance measured by
such a system can be written [from Equation (6.13)]
∞
Eλ c = k 0λ L0λ τaλ τ f λ Sλ dλ, (7.11)
0
where k accounts for the geometrical factors such as source area, orienta-
tion, and distance. This equation can be written as
λc + Δλ
2
Eλ = k 0λ L0λ τaλ τ f λ Sλ dλ. (7.12)
λc − Δλ
2
By change of variable λ = λc − x,
+ Δλ
2
Eλ c = k 0x L0x τax τ f (λc − x ) Sx dx. (7.13)
− Δλ
2
These equations show very clearly that the irradiance measured with the
filter centered around wavelength λc includes source energy from λc − Δλ 2
to λc + Δλ2 . Apart from the spectral selection, the filter has an additional
effect by smoothing the spectrum being observed because the filter has a
nonzero spectral width.
Equation (7.13) is called a convolution integral because it describes
the convolution between the product (0λc L0λc τaλc Sλc ) and τ f . In linear
systems terminology, the observed spectral source radiance is being con-
volved with the filter spectral transmittance. To investigate the effects of
this convolution consider the two cases: (1) the observed source has lit-
tle variation over the filter passband, and (2) the observed source varies
significantly over the filter passband.
If the product (0λc L0λc τaλc Sλc ) is more or less constant over the filter
passband Δλ, Equation (7.13) can be written to show that the convolution
has little effect other than some insignificant amount of smoothing:
+ Δλ
2
Eλ c = k (0λc L0λc τaλc Sλc ) τ f (λc − x ) dx (7.14)
− Δλ
2
≈ k 0λc L0λc τaλc Sλc τ f Δx. (7.15)
266 Chapter 7
1.0
Transmittance
0.5
0
0.85 0.9 0.95 1 1.05 1.1 1.15
Wavelength [mm]
Figure 7.2 Atmospheric transmittance convolved with 10 cm−1 and 300 cm−1 square win-
dows.
If the product (λ Lλ τaλ Sλ ) varies significantly over the filter passband
Δλ, Equation (7.13) cannot be simplified. In this case, the convolution
attenuates and smears out the finer detail in the spectral information. If
the actual spectral line is very narrow the measured line will approximate
the filter resolution and will be totally erroneous unless the filter spectral
smear effect is compensated by deconvolution.
The effect is best illustrated in Figure 7.2. Modtran™ was used to
calculate the spectral transmittance in the 0.85–1.1-µm spectral range. The
calculation was performed for a 1-km path length at sea level in a Tropical
atmosphere at 27 ◦ C, 75 %RH, 1015 mB, 23-km visibility. The transmittance
data so obtained was convolved with a square filter response. The spectral
filter widths are 10 cm−1 and 300 cm−1 .
As an example of one effect of the finite filter width, consider the case
where a spectral source is observed with a spectral radiometer. The es-
timated source spectrum (a) and measured spectral curve (b) are shown
in Figure 7.3. Suppose Modtran™ was used to calculate the spectral
transmittance (c) of the atmosphere. In order to compensate for the at-
mospheric attenuation, the measured irradiance is divided by the trans-
mittance curve, resulting in the calculated spectrum (d). The sequence on
the left in Figure 7.3 shows what happens if the atmospheric transmittance
was calculated at a much-higher resolution than the spectral measurement
filter width. In the sequence on the right in Figure 7.3, the atmospheric
transmittance (c) was obtained by convolving the high-resolution Mod-
tran™ data with the filter response before dividing the measured source
irradiance with the atmospheric transmittance.
In order to prevent serious errors in narrow-band spectral calculations,
all of the spectral variables must first be convolved to the same spectral
Radiometry Techniques 267
(a)
0 0
Medium resolution measurement Medium resolution measurement
1 1
(b)
0 0
High resolution atmosphere Medium resolution atmosphere
1 1
(c)
0 0
Source / atmosphere Source / atmosphere
1 1
(d)
0 0
1.0 15
γ = 0.08 km-1 0.15 0.20 0.30
13 0.50
Transmittance
Range [km]
11
γ = 0.08 km-1
0.5 9
0 15
0.20
0.50 0.30 7
0 4
11
0 5 10 15 107 108 109 1010 10
Range [km] I/E [m2]
(a) (b)
shown on the right side in Figure 7.4. The vertical lines are for various I/E
values. The intercepts of the vertical I/E lines with the R2 /τa ( R) curves
indicate the detection ranges.
with projected area AB = A B cos θB (limited to the pixel FOV), at range
R B1 , the background has a radiance L B , with atmospheric transmittance to
the observer of τB . The background and the object may not be at the same
distance. The atmospheric path radiance in the direction of the object
is L P . The observer has an instrument with optical aperture A1 , and it
measures irradiance E1 = dΦ/dA1 . The pixel irradiance consists of four
components: the object irradiance, path irradiance in front of the object,
background irradiance, and path irradiance in front of the background:
L T AT τT L B AB τB L Pt AT L Pb AB
dE1 = + + + (7.18)
R2T1 R2B1 R2T1 R2B1
= L T Ω T τT + L B Ω B τB + L Pt Ω T + L Pb Ω B , (7.19)
where Ω T = AT /R2T1 is the object projected solid angle (limited to the pixel
FOV), Ω B = AB /R2B1 is the background projected solid angle (limited to
the pixel FOV), L Pb is the path radiance for the fraction of the pixel filled
by the background, and L Pt is the path radiance for the fraction of the pixel
filled by the target object. In the analysis shown here, L A is the atmospheric
radiance (as applicable to path radiance, see Section 4.2.2). The projected
pixel footprint area A P cos θP = Ω P R2 is defined as the projected area that
the pixel subtends at a given range R. This scenario is further investigated
in Sections 9.2 and D.5.3.
The pixel magnitude as a function of object-to-observer range is shown
in Figure 7.5. Depending on the object size and range, Equation (7.18) can
be recast into several different forms. At ranges corresponding to region
I the object is resolved (object larger than the pixel footprint), whereas at
ranges corresponding to regions II to IV the object is unresolved (object
smaller than pixel footprint).
Region I
The target completely fills the pixel FOV, and no background is vis-
ible. For a resolved object A P cos θP ≤ A T cos θT ⇒ A B = 0, then
A P cos θP /R2 = Ω P and
L T A P cos θP τT
E1 = + L Pt Ω P (7.20)
R2T1
= ( L T τT + L Pt )Ω P
≈ [ L T e−γRT1 + L A (1 − e−γRT1 )]Ω P . (7.21)
For a resolved object the pixel magnitude is given by the object ra-
diance multiplied by the atmospheric transmittance term. Because
the atmospheric transmittance is a function of range, the object mag-
nitude will decrease with e−γR , whereas the path radiance increases
with (1 − e−γR ) [see Equation (4.10)].
270 Chapter 7
10 4 Target Total
10 5
Background
Path
10 6
10 7 2
10 103 104 105
Distance [m]
Figure 7.5 Pixel magnitude as a function of object-to-observer distance, as affected by the
atmosphere.
Region II
The target partially fills the pixel FOV. Then, A P cos θP > A T cos θT ⇒
A B = 0. Now, Ω T = A T cos θT /R2T1 , so that with increasing range,
Ω T decreases, and the background starts filling around the target in
the pixel FOV. Define the solid angle subtended by the background
as Ω B = Ω P − Ω T = Ω P − AT /R2T1 , then
E1 = L T Ω T τT + L B Ω B τB + L Pt Ω T + L Pb Ω B (7.22)
− γR T1 − γR T1
= [ LT e + L A (1 − e )] AT /R2T1
+[ L B e−γRB1 + L A (1 − e−γRB1 )](Ω P − AT /R2T1 ). (7.23)
Region III
The target fills only a very small portion of the pixel FOV, and the
background is the dominant source. For L B Ω B L T Ω T , or Ω T very
small:
E1 = L B Ω P τB + L Pb Ω P
= [ L B e−γRB1 + L A (1 − e−γRB1 )]Ω P . (7.24)
Region IV
At longer range, even the background radiance is severely attenuated
and the only remaining flux is due to path radiance. Hence, for
L B τB L P , the image pixel irradiance is given by
E1 = L P Ω P . (7.25)
At very long ranges the path radiance dominates all other sources,
and the object and clutter are lost in the ‘fog-like’ haze caused by
atmospheric radiance.
If a pixel is only partially filled by the target object, the spectral irra-
diance at the sensor is
ETλ = ( L Tλ τTλ + L Ptλ )Ω T + ( L Bλ τBλ + L Pbλ )(Ω P − Ω T ), (7.27)
where L Tλ is the object radiance, Ω T is the solid angle subtended by the
object, L Ptλ is the path radiance in front of the object, and τTλ is the atmo-
spheric transmittance between the object and the background.
The local radiance variation, or radiometric contrast, is now the differ-
ence between the pixel filled with the object and the pixel filled by back-
ground only:
ΔEλ = ( L Tλ τTλ + L Ptλ )Ω T + ( L Bλ τBλ + L Pbλ )(Ω P − Ω T )
− L Bλ Ω P τBλ − L Pbλ Ω P
= L Tλ τTλ Ω T + L Ptλ Ω T − L Bλ τBλ Ω T − L Pbλ Ω T
= ( L Tλ τTλ − L Bλ τBλ )Ω T + ( L Ptλ − L Pbλ )Ω T . (7.28)
If the target object and background are at the same range, τTλ = τBλ and
L Ptλ = L Pbλ :
ΔEλ = ( L Tλ − L Bλ )τTλ Ω T , (7.29)
hence the signal is proportional to the difference in the radiance values
between the object and the background.
The detection of pulse signals corrupted by noise forms the basis of many
electro-optical systems. The calculation of the probability of detection and
false alarm rate requires information about the peak signal, rms noise, and
threshold setting. The general solutions 13–15 for arbitrary pulse shapes,
noise spectra, and nonlinear detection processes are very complex and are
not considered here. If the signal is a square pulse, filtered by a matched
filter, and the input noise is white with a Gaussian distribution, as is the
case for most natural noises, the detection performance can be readily 13,16
calculated. This is a special case, but, is useful to obtain at least an order
of magnitude indication.
The detection of a square pulse of width t p , immersed in white noise
after passing through a matched filter with bandwidth Δ f = 1/(2t p ), is
shown graphically in Figure 7.6. Detection is the event where the sig-
nal corrupted by noise exceeds the detection threshold. A false alarm is
the event where the noise (with no signal present) exceeds the detection
threshold. The average false alarm rate is given by
1
√ exp−it /(2in ) ,
2 2
FAR = (7.30)
2t p 3
Radiometry Techniques 273
is is
tp Matched filter 2tp
Rectangular Df = 1/(2tp)
signal pulse
Threshold Alarm
Σ detector
is + in > it
in
Matched filter
White noise Df = 1/(2tp) Threshold
Probability of
false alarm
is
it
in
0
tp
Probability of
undetected signal
in in + is in in + is
where it is the threshold value, and in is the rms noise value at the input
to the threshold detector.
When there is a signal present, the probability of detection (signal plus
noise exceeds the threshold) is given by
1 is − it
Pd ≈ 1 + erf √ , (7.31)
2 2in
where erf is the error function:
z
2
e−t dt.
2
erf(z) = √ (7.32)
π 0
1. The false alarm performance must be 1 per 1000 pulses, each pulse
arriving in a 67-µs window. The false alarm rate must therefore be less
than 1/(1000 × 67 × 10−6 ) = 15 pulses per second.
The key design parameters are SNR (determines the probability of de-
tection) and the signal-to-threshold ratio (determines the false alarm rate).
In order to achieve high probability of detection, the designer strives to-
ward a high SNR.
Consider an infrared point target detection system locating hot spots
in a 4 × 105 -pixel image. The system is designed to operate at a SNR
ratio of 12. Using Equation (7.31) it is found that for log(t p FAR) = −22,
Radiometry Techniques 275
Bibliography
[1] Lloyd, J. M., Thermal Imaging Systems, Plenum Press, New York (1975).
[3] Wittenstein, W., “Thermal range model TRM3,” Proc. SPIE 3436, 413–
424 (1998) [doi: 10.1117/12.328038].
Problems
7.1 Explain what the range equation is and why it is important. Derive
a mathematical formulation for the range equation and elaborate
on how it can be solved. [2]
7.2 Explain what happens when a small object is viewed in an image
at different ranges. Draw a diagram that shows the pixel signal for
different ranges from very close to very far. Divide the diagram
into different regions, where different signal sources contribute to
the pixel signal. Explain the dominant source in each region. [5]
Radiometry Techniques 277
7.3 Explain the term ‘effective transmittance’ and show how it is cal-
culated. [2]
7.4 Provide a description (in words or equations, as applicable) of each
of the following terms and explain where it is used: [10]
279
280 Chapter 8
Spectral properties
Temporal
properties τoλ
ρoλ
11 12 1 oλ Sun
9
10 2
3
sλ Lλ ( Ts )
8 4
7 6 5
τsoλ τsoλ
Background
λ Atmosphere
ψ
ψρ bλ Sky
Lskyλ ψρ aλ
ρoλ
bλ Lλ ( Tb ) τaoλ aλ Lλ ( Ta )
τaboλ
Surface
Lpathλ
Atmosphere Lλ ( To ) Environment
τoλ
oλ τaλ
Lpathλ
τaλ
1. Self-emitted radiance (Lself ): The object emits flux according to its di-
rectional emissivity and Planck’s law (Section 3.5). For Lambertian sur-
faces the directional emissivity is simply the diffuse emissivity. The
282 Chapter 8
Symbol Meaning
LS total radiance in the wavelength band S
Lλ ( Ts ) spectral blackbody radiance, sun temperature Ts
Lλ ( Ta ) spectral blackbody radiance, environment temperature Ta
Lλ ( Tb ) spectral blackbody radiance, background temperature Tb
Lλ ( To ) spectral blackbody radiance, uniform object temperature To
Lpathλ spectral atmospheric path radiance: emitted & scattered
Lskyλ spectral sky radiance: emitted & scattered
sλ solar surface’s spectral emissivity
aλ ambient environment’s spectral emissivity
bλ background spectral emissivity
oλ (θv ) object surface directional spectral emissivity
Δ spatial texture variation in emissivity (unity if no texture)
ρoλ object surface diffuse spectral reflection
Δρ spatial texture variation in reflectivity (unity if no texture)
f r ( θs , θv ) object surface bidirectional reflection distribution function
τoλ object spectral transmittance
τaλ object-to-sensor spectral atmospheric transmittance
τaboλ background-to-object spectral atmospheric transmittance
τaoλ ambient-to-object spectral atmospheric transmittance
τsoλ sun-to-object spectral atmospheric transmittance
ψ Asun /(d2sun π) = 2.17 × 10−5
Asun area of the sun
dsun distance to the sun
θa angle between the surface normal and the vertical
θs angle between the surface normal and solar incidence
θv angle between the surface normal and the view direction
Sλ measurement instrument spectral response
Optical Signatures 283
4. Reflected ambient radiance (Lref amb ): It is assumed that the flux from
the environment is incident from all directions; the object is fully en-
closed by the environment. This assumption is approximately valid
when the object is indoors, but less so of the object is outdoors. How-
ever, even outdoors, the object may be immersed in the atmosphere,
which also provides at least some environmental flux (Section 4.6.7 and
Figure 4.12).
6. Reflected solar radiance (Lref sun ): A simple model for reflected sunlight
is presented in Section 3.7. The reflected solar radiance depends on the
orientation of the surface, the BRDF ( fr ), and the transmittance of the
atmosphere from the sun to the surface and then from the surface to the
sensor.
10
[W/(m2·sr·mm)] Reflected sunlight
8 (no atmosphere) Thermal radiance (300 K)
(no atmosphere)
Radiance
6
Reflected 27 °C, 75% RH, 23 km visibility,
4 sunlight Modtran Rural aerosol, 5 km path length
2 Thermal radiance (300 K)
0
2 4 6 8 10 12 14
2
Reflected sunlight Thermal radiance (300 K)
[W/(m2·sr·mm)]
1 Reflected
sunlight
Thermal radiance (300 K)
0
3 3.5 4 4.5 5 5.5 6
Wavelength [mm]
Figure 8.3 Reflected and emitted signature from a target with 80% emissivity.
The Phong Equation (3.39) provides a simple (but not very accurate) model
for BRDF:
ρo ρs (n + 1) cosn α
f r,Phong = + .
π 2π cos θi
Optical Signatures 285
Incident 1.07
Mirror
beam Aircraft white beam
White chalk
Matte blue Matte white
Green leaf
Semi-matte light grey Matte dark earth
Figure 8.4 Phong BRDF for paints and natural surfaces in the NIR spectral band.
Table 8.2 Phong BRDF parameters for paints and natural surfaces in the NIR spectral band.
1. Set a radiance threshold at some level and ignore all pixels in the image
with values below this threshold. The problem with this approach is
that the selected pixels may have widely varying radiance values, from
very cool to very hot. A more accurate approach is to segment the object
image into several regions, each with a limited range of pixel radiance
levels (somewhat similar to a terrain contour map).
2. Set a spatial limit to the size of the radiator and ignore all pixels out-
side the selected size. This approach may reject a significant amount of
radiation. However, if the spatial solid angle is selected carefully, this
approach may yield a good estimate of the useful radiator area.
3. Integrate the flux over a large spatial extent and at all radiance levels
— use as large an area as feasible. Peak normalize (Section 7.2.3) the
measured flux by assuming the highest radiance level for the whole
radiator area, and then calculate an ‘area’ to provide the same flux as
the integrated measured flux.
3 162
316.2
31.6
Figure 8.5 Measured flame radiance in the 3–5-µm spectral band for six different threshold
levels.
0.7
0.6
Flame size [m2]
0.4
0.3 Threshold segmentation and peak normalized
0.2
0.1
0
0 2000 4000 6000 8000 10000 12000
Threshold [W/(m2·sr)]
Inspection of the segmented pictures in Figure 8.5 might lead the an-
alyst to decide that the flame area is approximately 0.2 m2 , corresponding
to a threshold of 2500 W/(m2 ·sr). In this case the area is selected on the
basis of the analyst’s evaluation of the threshold-segmented pictures.
The analysis indicates that the flame has large areas of relative low
radiance, which contributes relatively little to the total signature. Depend-
ing on the application, different flame areas can be chosen. If shape is not
important, the flame could be modeled as a simple uniform radiator with
area of 0.14 m2 . If shape is important, the radiance gradient across the
flame must be accounted for, and the model becomes more complex.
Radiance [W/(m2·sr·mm)] 40
Blackbody source
Instrument setting 150 °C
30 Thermocouple 150.5 °C
200 °C
Scaled to measured
20 155 °C
value at this point
150 °C
145 °C
10
100 °C
Measured
0
2 2.5 3 3.5 4 4.5 5 5.5
Wavelength [mm]
(a)
10
Radiance [W/(m2·sr·mm)]
FOV, spectral response, and the meteorological data during the measure-
ment. This procedure also requires that some estimate of the object spectral
emissivity be known (either by measurement or research investigation).
The purpose of this data analysis calculation is to determine the object
temperature, given the object spectral emissivity, area, and atmospheric
conditions at the time of the measurement. The intention is to calculate the
temperature directly from the measured data with no intermediate steps.
This approach requires a mathematical formulation that can be solved nu-
merically.
Note that the temperature so determined is not necessarily a phys-
ically valid temperature. It is the value mathematically required by the
chosen area and emissivity to provide the measured signal. If the area
and emissivity values are physically valid, then the temperature is also a
physically valid temperature.
Equation (6.16) describes the signal voltage at the output of the ra-
diometer and is simplified to
Zt A 0 A 1 ∞
kR
vS = 0λ L0λ τaλ Sλ dλ. (8.7)
R201 0
The test results are reported in terms of two measurements giving the
same voltage in the instrument. The first voltage is the calibration voltage
when observing a known source. The second voltage is the measurement
of the target source.
Assuming that the calibration is done with a blackbody source with
emissivity close to unity, over a very short distance, the voltage for the
calibration measurement is
∞
Zt Ω p A 1
vc = k R L0λ ( Tc )Sλ dλ, (8.8)
0
where the instrument FOV Ω p is filled by the calibration source, Tc is the
blackbody temperature, and vc is the instrument output voltage during
this calibration.
From Equations (8.1) and (8.7), the voltage for the object measurement
is
thermally emitted Lself
!
Zt A 0 A 1
kR ∞
vS = Δ oλ (θv ) Lλ ( To )τaλ Sλ dλ
R201 0
atmospheric path radiance Lpath
!
∞
Zt Ω p A 1
+k R Lpathλ Sλ dλ
0
294 Chapter 8
Because the radiance during the measurement is the same as the ra-
diance during calibration, now equate the values. Furthermore, if the tar-
get area exceeds the size of the pixel footprint (i.e., an extended source),
As /R201 becomes the instrument FOV Ω p , and then the equation can be
simplified considerably because the geometric factors are the same for all
of the terms. Assuming that all of the other parameters are known, the
target source temperature can now be obtained by solving for To in
calibration radiance
!
∞
0 = − L0λ ( Tc )Sλ dλ
0
thermally emitted Lself atmospheric path radiance Lpath
! !
∞ ∞
+ Δ oλ (θv ) Lλ ( To )τaλ Sλ dλ + Lpathλ Sλ dλ
0 0
transmitted background Ltrn back
!
∞
+ τoλ bλ Lλ ( Tb )τaboλ τaλ Sλ dλ
0
diffuse reflected ambient background Lref amb
!
∞
+ Δρ ρoλ aλ Lλ ( Ta )τaoλ τaλ Sλ dΩdλ
0 env
diffuse reflected sky Lref sky
∞ !
+ Δρ cos θa ρoλ Lskyλ τaλ Sλ dΩdλ
0 sky
reflected sun Lref sun
∞ !
+ Δρ ψ cos θs fr (θs , θv )sλ Lλ ( Ts )τsoλ τaλ Sλ dλ . (8.10)
0
Solving this form of the equation is challenge because of the large number
Optical Signatures 295
Consider the case where the radiometer is observing a flame. The flame
temperature is much higher than the surrounding background and the
atmosphere. It is known that flames have high transmittance and low
emissivity. Suppose further that path radiance can be ignored. Several
terms in Equation (8.10) are therefore discarded. The problem can then be
stated as follows:
calibration radiance thermally emitted Lself
∞ ! !
A0 ∞
Ωp L0λ ( Tc )Sλ dλ = Δ oλ (θv ) Lλ ( To )τaλ Sλ dλ, (8.11)
0 R201 0
where the requirement for an extended target size has been removed. A
smaller target size results in a target solid angle smaller than the instru-
ment solid angle. This may occur during practical measurements where
flames are measured over long range. The solution to Equation (8.11) can
be readily obtained by numerical analysis, but it does require a known
spectral emissivity (see Section 8.4.1).
Two important observations can be made on the methodology given
here: (1) the atmospheric transmittance is compensated for by multiplica-
tion (no divide-by-zero errors)- and (2) the absolute value of the radiometer
system response does not appear in the solution; only the spectral system
response Sλ is required.
The emissivity of shiny metallic surfaces are lower than the emissivity of
rough Lambertian surfaces. The effect of this low emissivity on tempera-
ture measurements should be carefully considered. Suppose the irradiance
from a metal surface is measured over a short distance, as an extended
source, with the intention to determine the temperature of the surface. For
this analysis start with Equation (8.10), assume an extended target, and
ignore the terms for path radiance, transmitted background flux, reflected
sky, and reflected sunlight. Keep the terms for self-exitance and terrain
background. The terrain background in this case is meant to model an
enclosed volume, such as in the laboratory. Note that for an opaque sur-
face, ρoλ = (1 − oλ ). This means that as the self-radiance decreases due to
lower emissivity, the reflected ambient radiance increases due to reflection.
296 Chapter 8
200
True Temperature [°C]
3 5 mm 0.1
temperature
160 0.2
Ambient
120
80 1.0
40
1.0
0
0 10 20 30 40 50 60 70 80 90 100
Apparent temperature [°C]
True Temperature [°C]
200
8 12 mm 0.1
temperature
160 0.2
Ambient
120
80 1.0
40
1.0
0
0 10 20 30 40 50 60 70 80 90 100
Apparent temperature [°C]
Figure 8.8 Apparent versus real temperature for different source emissivity values, ranging
from 1 (straight line) down to 0.1 (most curved line) in steps of −0.1, in the 3–5-µm and
8–12-µm spectral bands.
The objective with this case study is to derive a first-order empirical model
for the radiance in the ‘silver lining’ edge of a back-lit cloud. The problem
is solved by analyzing measured data and compiling a simple model. In
this instance the only available measured information was a ‘temperature’
measurement in imaging-camera images. No other infrared or calibration
information was available. The modus operandi was to first calculate the
radiance from the measured temperature data and second to build a model
around the calculated radiance values.
8.8.1 Measurements
Measurements were made at varying angles from the sun, ranging from
10 deg to 90 deg from the sun. The distance was measured with a laser
rangefinder and varied from one to three kilometers. The measurement
was performed from ground-level at 1500 m above sea level. The ground
level ambient temperature during the measurements was 20 to 25 ◦ C. In
all of the subsequent calculations the Lowtran (precursor to Modtran™)
Mid-latitude Summer model was used. The atmospheric transmittance
and path radiance values are shown in Figures 8.9(b) and 8.9(c).
Measurements were performed with an InSb MWIR camera fitted
with an antisolar filter. This filter suppresses all radiation at wavelengths
shorter than 3 µm. The camera was set to assume a target emissivity of
one. The vendor-supplied software was used to read temperature values
directly from the image, so this analysis starts with a measured tempera-
ture value as input. The camera instrument function was calibrated against
a blackbody simulator at close range under laboratory conditions.
The measurements indicated that a typical side-lit cloud has an appar-
ent temperature of between 10 and 20 ◦ C (Table 8.3). Back-lit clouds had
silver linings and small hot spots with apparent temperatures between
30 ◦ C and 50 ◦ C. This increase in apparent temperature is due to forward
scattering by the silver lining — not a high cloud temperature. The first
step is to calculate the cloud radiance values from the recorded tempera-
tures. The cloud radiance is calculated using the calibration equation
∞
Lcloud = 0λ L0λ ( Tm )Sλ τmλ dλ, (8.13)
0
where 0λ = 1 is the calibration source emissivity, Tm is the cloud temper-
ature as indicated by the camera, Sλ is the normalized camera response
shown in Figure 8.9(a), and τmλ is the atmospheric transmittance during
calibration.
298 Chapter 8
8.8.2 Model
0.8
Cloud emissivity Sensor
0.6
0.4
0.2 Cloud reflectance
0
3 3.5 4 4.5 5 5.5 6
Thermal
2
1
Reflected sunlight
Path + sky
0
3 3.5 4 4.5 5 5.5 6
10-3 (e) Severe back-lit-cloud apparent radiance [W/(m2·sr·cm-1)]
12
Radiance
8 Total
Reflected sun light
Thermal
4
Path + sky
0
3 3.5 4 4.5 5 5.5 6
Wavelength [mm]
Figure 8.9 Cloud model analysis spectral input data and results.
300 Chapter 8
1. Early in the morning, all objects have cooled down by radiation, con-
duction, and convection. Temperature contrast is low.
50
Temperature [°C]
40
Sheet-metal roof
30
20 Grass field
Glass panel
10
0
0 2 4 6 8 10 12 14 16 18 20 22 24
Diurnal cycle [h]
4. In mid-afternoon, all objects are hot, having absorbed heat all morning.
The contrast between the objects is, however, decreasing because most
objects have been heated by now.
5. In early evening, the objects with low heat capacity have lost much
of their energy by radiation, whereas objects with high heat capacity
are still relatively warm. As the temperature decreases, the contrast
increases slightly until all objects have lost most of their excess energy.
A paint with the required visual properties, but which is transparent in the
thermal bands, could be useful in a multi-spectral signature management
302 Chapter 8
Emissivity 1.0
0
2 4 6 8 10 12
Wavelength [mm]
Figure 8.11 Emissivity of a metal plate painted with a thermally transparent paint.
(camouflage) design. The object would have the appropriate visual colors,
but it would reflect the ambient surroundings’ infrared signature from a
shiny metal subsurface.
One such thermally transparent paint was characterized after having
been applied to a clean, polished aluminium plate. The resulting emissiv-
ity value is shown in Figure 8.11. Note the low emissivity in the 3–5-µm
spectral band arising from the fact that the paint is transparent in the 3–
5-µm spectral band, and the polished subsurface is observed. Note also
that the 8–12-µm spectral band emissivity is still too high to be of practical
value.
The dependency of emissivity on paint thickness indicates that the
paint base is not fully transparent. A thick coating (> 150 µm) will ef-
fectively cover any metal surface underneath but with emissivity equal to
other paints.
Radiance [W/(m2·sr·mm)] 10
Scaled solar irradiance
8
4
Thermal
2 self-exitance
0
3 3.2 3.4 3.6 3.8 4 4.2 4.4 4.6 4.8 5
Wavelength [mm]
Figure 8.12 Sunlight reflection of the sea surface (glint), sun at low elevation, looking in the
direction of the sun.
Bibliography
[1] Pyradi team, “Pyradi Radiometry Python Toolkit,” http://code.
google.com/p/pyradi.
[2] Willers, C. J., Willers, M. S., and Lapierre, F. D., “Signature modelling
and radiometric rendering equations in infrared scene simulation sys-
tems,” Proc. SPIE 8187, 81870R (2011) [doi: 10.1117/12.903352].
[3] Wolfe, W. L. and Zissis, G., The Infrared Handbook, Office of Naval
Research, US Navy, Infrared Information and Analysis Center, Envi-
ronmental Research Institute of Michigan (1978).
[4] Accetta, J. S. and Shumaker, D. L., Eds., The Infrared and Electro-Optical
Systems Handbook (8 Volumes), ERIM and SPIE Press, Bellingham, WA
(1993).
[5] Palmer, J. M. and Grant, B. G., The Art of Radiometry, SPIE Press,
Bellingham, WA (2009) [doi: 10.1117/3.798237].
[7] Mahulikar, S. P., Sonawane, H. R., and Rao, G. A., “Infrared signature
studies of aerospace vehicles,” Progress in Aerospace Sciences 43(7-8),
218–245 (October-November 2007).
[8] Ferwerda, J. G., Jones, S. D., and Reston, M., “A free online refer-
ence library for hyperspectral reflectance signatures,” SPIE News-
Room (Dec 2006).
304 Chapter 8
[11] Roblin, A., Baudoux, P. E., and Chervet, P., “UV missile
plume signatures model,” Proc. SPIE 4718, 344–355 (2002) [doi:
10.1117/12.478822].
[12] Neele, F. and Schleijpen, R., “UV missile plume signatures,” Proc.
SPIE 4718 (2002).
[14] Rapanotti, J., Gilbert, B., Richer, G., and Stowe, R., “IR sensor design
insight from missile plume prediction models,” Proc. SPIE 4718 (2002)
[doi: 10.1117/12.478816].
[17] Michalski, L., Eckersdorf, K., Kucharski, J., and McGhee, J., Tempera-
ture Measurement, 2nd Ed., John Wiley and Sons, New York (2001).
[18] Strojnik, M., Paez, G., and Granados, J. C., “Flame thermometry,”
Proc. SPIE 6307, 63070L (2006) [doi: 10.1117/12.674938].
Problems
Nose facing
to front
0.7 m
1.4 m
6m
12 m
Surface B1: tailpipe
Surface A1: side view
Surface A2: bottom view
Surface A3: side view
Surface A4: top view
Transmittance / Emissivity
Atmospheric
transmittance over
1 km path length
0.5
Plume
emissivity
0
3 3.5 4 4.5 5 5.5 6
Wavelength [mm]
8.1.1 Calculate the transmittance from the above equation and confirm
that it agrees with the graph. Calculate and plot the attenuation
coefficient. Then calculate and plot the transmittance at the fol-
lowing ranges: 500 m, 1000 m, 2000 m, and 5000 m. [3]
8.1.2 Draw plots for each component (surface or plume), showing the
irradiance with the atmosphere present for all of the ranges on the
same plot. [3]
Calculate the total (integrated) irradiance received by a sensor
(from all of the surfaces facing the sensor) at all above ranges from
the aircraft, for aspect angles θ ∈ {0, π/2, π} rad, with no atmo-
sphere present. If more than one surface is visible, calculate and
show the contributions separately, and also the sum of all contri-
butions. [3]
8.1.3 Draw plots for each component (surface or plume), showing the
irradiance with the atmosphere not present, for all of the ranges
on the same plot. [3]
Calculate the total (integrated) irradiance received by a sensor
(from all of the surfaces facing the sensor) at all above ranges from
the aircraft, for aspect angles θ ∈ {0, π/2, π} rad, with the atmo-
sphere present. If more than one surface is visible, calculate and
show the contributions separately, and also the sum of all contri-
butions. [3]
1.4
Detector
1.2 responsivity
1
0.8
0.6 Filter
transmittance
0.4
0.2
0
0 1 2 3 4 5 6
Wavelength [mm]
8.2.6 Calculate the current through the detector when it is viewing the
cloud with the optical filter present. [2]
8.2.7 Comment on the use of the 3–6-µm spectral band for observations
during day- and nighttime. What consideration should be made
when performing temperature measurement in this spectral band?
[2]
8.3 A thin circular disk has a temperature of 300 K on the one side
and a temperature of 1000 K on the other side. The emissivity on
both sides is the same with a value of 1. The disk diameter is 1 m.
Assume an atmospheric transmittance of unity. The disk is viewed
against an infinitely large background with a uniform spatial tem-
perature distribution. Consider two scenarios: a background tem-
perature of 270 K and a background temperature of 330 K.
A sensor is located at a range of 1000 m from the disk. The sensor
spectral response is unity in the 3–5-µm spectral band and zero
elsewhere. The sensor FOV is 2 mrad full-apex angle.
The objective with this investigation is to plot the target contrast
intensity from all view directions (similar to the graphs shown in
Figure 8.1) for both background temperatures.
Prior to calculation, try to visualize the spherical intensity, and
draw a spherical contrast intensity diagram in freehand.
Calculate and plot the polar contrast intensity of the disk against
the two backgrounds, in the xy, yz, and zx planes. [6]
Use the tools in pyradi 1 (or any other tool of your choice) to cal-
culate and plot the three-dimensional spherical contrast intensity
of the disk against the two backgrounds. [10]
8.4 Extend the one-dimensional wedge light trap described in Prob-
lem 3.16.2 to two dimensions, with the wedge lines running along
the x axis. The light trap has a temperature of 1000 K. Calculate
and plot the hemispherical radiance distribution of the light trap
(use pyradi 1 or any other tool). [10]
Chapter 9
Electro-Optical System Analysis
Introduction
In this chapter different electro-optical systems are briefly defined and an-
alyzed to demonstrate the application of the radiometric modeling tech-
niques developed in the early chapters of the book. In the work presented
here, the emphasis is on the radiometry and methodology, rather than on
the detail parameters. Some of the case studies may appear somewhat
contrived, but these are still useful hands-on training material. In any real
design, considerably more effort will be expended: the level of detail will
be deeper, and the models much more comprehensive. One example of
a more comprehensive application is the simulation system described in
Appendix B.
The flame sensor must detect the presence or absence of a flame in its FOV.
The sensor is pointed to an area just outside a furnace vent, against a clear-
sky background. The sensor must detect a change in signal indicating the
presence of a flame at the vent. This problem was defined to illustrate the
calculation of spectral integrals. This case study is also a worked example
in Section D.5.1 and on the pyradi toolkit website. 1
The sensor has an aperture area of 7.8 × 10−3 m2 and a FOV of 1 × 10−4
sr. The sensor filter spectral transmittance is shown in Figure 9.1(a). The
spectral transmittance can be calculated with Equation (D.4) and parame-
ters (τs = 0.0001, τp = 0.9, s = 12, Δλ = 0.8 µm, λc = 4.3 µm). The InSb
309
310 Chapter 9
Filter
Atmosphere
0.5
Detector
Emissivity
0
3 3.2 3.4 3.6 3.8 4 4.2 4.4 4.6 4.8 5
-3 2 1
x 10 (b) Path radiance out to space [W/(m ·sr·cm )]
5
4
Radiance
0
3 3.2 3.4 3.6 3.8 4 4.2 4.4 4.6 4.8 5
-6
x 10 (c) Irradiance from flame and path [W/(m2·cm 1)]
1.5
Flame irradiance
Irradiance
1.0
0
3 3.2 3.4 3.6 3.8 4 4.2 4.4 4.6 4.8 5
Wavelength [mm]
detector has a peak responsivity of 2.5 A/W and spectral response shown
in Figure 9.1(a). The preamplifier transimpedance is 1 × 104 V/A.
The flame area is 1 m2 . The flame temperature is 1000 ◦ C. The emissiv-
ity is 0.1 over most of the spectral band due to carbon particles in the flame.
At 4.3 µm there is a strong emissivity rise due to the hot CO2 in the flame;
see Figure 9.1(a). The emissivity can be calculated with Equation (D.4) and
parameters (τs = 0.1, τp = 0.7, s = 6, Δλ = 0.45 µm, λc = 4.33 µm).
The distance between the flame and the sensor is 1000 m. The at-
mosphere is similar to the Modtran™ Tropical climatic model. The path
is oriented such that the sensor stares out to space at a zenith angle of
88.8 deg. The components’ spectral transmittance is shown in Figure 9.1(a)
and the path radiance in Figure 9.1(b).
Electro-Optical System Analysis 311
The spectral peak in the flame emissivity and the dip in atmospheric
transmittance are both centered around the 4.3-µm CO2 band. The strong
spectral variations in both target signature and atmospheric transmittance
necessitates multi-spectral calculation (see Section 2.6.5).
From Equation (6.16), the signal caused by the atmospheric path radi-
ance is given by
∞
Zt Ω A 1
vpath = k R Lpathλ Sλ dλ, (9.1)
0
where the variables are defined as for Equation (6.16). A0 is the flame area
because cos θ0 = 1, and A1 is the sensor aperture area. These equations
were evaluated as described in Section D.5.1, and yielded the results shown
in Table 9.1.
It is clear that the flame signal is several times larger than the path
radiance signal, even though the flame only fills 1% of the sensor FOV.
The severity by which the atmosphere attenuates the CO2 exitance from
the flame is shown in Figure 9.1(c).
Section 7.6 developed the theory for observing a simple target against a
background through a radiant medium. This section describes the model
used to calculate practical values for this scenario. The section closes with a
312 Chapter 9
Hot black target in 0.4 0.75 mm band Cold black target in 0.4 0.75 mm band
10-3 10-4
Tt 300 K Total
Irradiance [W/m2]
t
Irradiance [W/m2]
-4 Tt 1773 K Total 0.9
10 t 0.9 -5 Tb 300
10
Tb 300 b 0 Background
10-5 b 0.05
Background Path
10-6
10-6 Target
Path Target
10-7 10-7
0.1 1 10 100 0.1 1 10 100
Range [km] Range [km]
Hot black target in 3.5 4.5 mm band Hot black target in 8 12 mm band
10-4 10-4
Total
Irradiance [W/m2]
Irradiance [W/m2]
10-5 Tt 500 K Background
t 0.9 Total 10-5
Tb 300 Tt 370 K
10-6 b 0.8 t 0.9
Background 10-6 Path Tb 300
10-7 Path b 0.9
Target
Target
-7
10-8 10
0.1 1 10 100 0.1 1 10 100
Range [km] Range [km]
Figure 9.2 Pixel irradiance as function of object-to-observer distance for varying target tem-
peratures and sensor spectral bands.
the total irradiance in the pixel. The analysis was done for the following
Modtran™ atmosphere: Tropical profiles, 23-km visibility Rural aerosol
(‘MIE Generated’ aerosol phase function), executed in the ‘Radiance with
Scattering’ mode, with multiple-scattering for flux at the observer. The
Isaac’s two-stream multiple-scattering algorithm is used. The observer is
located at sea level, viewing a slant path with 88-deg zenith angle (near
horizontal). The pixel FOV is 1 µsr, and the target area is 0.2 m2 . The target
and background temperature and emissivity properties are indicated on
the graphs in Figure 9.2. The code to calculate these graphs is included in
Section D.5.3.
From Figure 9.2 it is evident that the target determines the pixel irra-
diance for targets at close range, whereas the atmospheric path radiance
dominates the signature for targets at long range. The ‘cold-black’ target
in the visual band presents an interesting case in that at close target range
the pixel irradiance is low, because there is no sunlight reflection from the
target. At intermediate ranges the white reflective background dominates
the signature. At the risk of generalizing based on specifics, at least for
the test cases shown here, the atmospheric path radiance dominates from
10 km onwards. In the visual and LWIR spectral bands the path radiance
magnitude increase exceeds an order of magnitude over the ranges con-
sidered. In the MWIR spectral band, this increase is considerably less than
314 Chapter 9
10
1
Contrast
0.1
0.02
0.01
0.001
0.1 1 10 23.3 100
Range [km]
Figure 9.3 Contrast between target pixel and adjacent background pixel in the visual spec-
tral band.
Sun Lamp
Open-circuit voltage [V] 1.59 1.2
Short-circuit current [A] 0.107 0.06
Load line for fan
Load voltage [V] 0.9 0.54
Load current [A] 0.09 0.054
9.3.1 Observations
Light source
Rsc
0.8 Atmospheric
transmittance
θc
60 mm
Magnitude
0.6
Ac
104 mm 0.4
0.2
Solar cell
responsivity
0 [A/W]
0.2 0.4 0.6 0.8 1 1.2
90 mm Wavelength [mm]
Figure 9.4 Solar panel and lamp geometry, spectral atmospheric transmittance, and spec-
tral solar cell responsivity.
9.3.2 Analysis
104 × 90 × 10−6
Ωpanel from sun = = 421 × 10−27 sr. (9.5)
(149 × 109 )2
The integral is determined over the size of the solar panel: −0.045 m ≤
h ≤ 0.045 m, and −0.052 m ≤ h ≤ 0.052 m. For more details on how to
calculate the integral, see Section D.5.8.
The lamp area is calculated with a little trick: the Stefan–Boltzmann
law (Section 3.1.3) states that the total exitance by a blackbody is given by
M ( T ) = 5.67 × 108 T 4 , but the total power is known to be 60 W (the lamp
rating). For this calculation, assume that the only energy loss is through
radiation with no heat loss due to convection or conduction through the
filaments’ stem wires. From Φ = M ( T ) Alamp , it follows that the total
Electro-Optical System Analysis 317
Following the workflow in Section 6.6 the solar cell current can be written
as follows:
∞
0λ L0λ τaλ Sλ dλdA0 dA1 cos θ0 cos θ1
i ph = ffill R
λ=0 source cell R201
∞
dA1 cosθ0 cos θ1
= ffill R 0λ L0λ τaλ Sλ dλ dA0 (9.8)
λ =0 source cell R201
= ffill Ieff A0 Ω1 ,
where i ph is the photon-induced current, ffill is the fill factor (how much
of the panel is active detector area), R is the detector responsivity scaling
factor with units [A/W], the spatial integrals and cosine terms are impor-
tant in the case of the lamp, A0 is the area of the source (sun or lamp), A1
is the total physical area of the solar panel, and S = Rλ τ f λ is the solar-cell
spectral responsivity.
The fill factor is the fraction of the panel that is able to convert optical
power to electrical power. In this case, the panel consisted of a number of
irregular pieces of silicon that filled only 50% of the area. This implies that
only half of the physical area can respond to the flux and create electricity.
The lamp illumination did not cover the full area of the source because
it was somewhat shielded by the lamp cover. It is estimated that 80% of
the solar cell was illuminated.
Note that the solid angle calculated above is for the whole panel. How-
ever, the panel consists of eight cells, of which some are in series. It is
shown below that the panel was wired as two parallel circuits, each of
four cells in series. It follows that, for the purpose of current generation,
318 Chapter 9
the effective area, and therefore the solid angle, is 0.25 of the totals calcu-
lated above. The exact factor depends on where each cell is located in the
panel, but for this calculation the value of 0.25 is used.
Using Equation (9.8), the calculated values shown in Table 9.3 are
somewhat higher than the measured values. The difference could be at-
tributed to errors in spectral response, fill factor, lamp temperature/area,
or atmospheric transmittance.
The solar cell efficiency is the ratio of flux converted to electricity to
the total incident flux:
∞
0λ L0λ τaλ Sλ dλ
η = λ=∞0 . (9.9)
λ=0 0λ L0λ τaλ dλ
Using Equation (9.9), the efficiency was calculated. In this analysis the
effect of atmospheric transmittance was ignored by setting τaλ = 1. The
value was found to be 1.4% for the lamp and 3.8% for the sun. It is evident
that a very small portion of the incident flux is converted to electrical en-
ergy. The Shockley–Queisser limit for single-layer silicon p-n junction solar
cells sets the theoretical limit at 33.7%. Current commercial and research
solar cells achieve between 22% and 25% efficiency.
9.3.2.3 Configuration
Under ideal conditions, the highest output voltage is less than 0.58 V per
single cell. Because the highest observed output voltage for this panel was
1.59 V, it is inferred that there must be four cells in series. In order to
obtain the best load and generation distribution, the cells are connected
as shown in Figure 9.5. Each cell acts as a current generator on its own.
Small differences in photocurrents lead to small differences in cell voltages,
Electro-Optical System Analysis 319
Iph
Rs
C
R
which are equalized by voltage drop across the series resistors and cells.
It is shown in Section 5.9.2 that the current generated by the solar cell is
related to the cell voltage by the I-V equation of the form [Equation (5.128)]:
Iload = Isat eqV/(kTβ) − 1 − Iph . (9.10)
0.01
Solar-cell current [A]
0.02 Lamp I V
Lamp load line
0.03
0.04
Sun load line Sun I V
0.05
0.06
0 0.05 0.1 0.15 0.2 0.25 0.3 0.35 0.4
Solar-cell voltage [V]
Figure 9.6 Solar cell I-V curve and load line for Isc = 0.03 and Isc = 0.054, lamp illumination,
and sun illumination.
From the measurements and above calculations, the load current for the
sun-illuminated cell is related to the cell voltage as
Iload = 11 × 10−9 e38.6×Vload − 1 − 0.054. (9.17)
Figure 9.6 shows the I-V curves for the two measurements described
in the problem statement (for a single cell). The solar cell is illuminated by
a lamp and the sun, while the same load is applied in both cases.
The analysis in this section indicates how a little information can be
used to derive a model. Once the model is developed, the remaining chal-
lenge is validation that the model is indeed correct (Section 7.9). To some
Electro-Optical System Analysis 321
extent this can be done with the information used to develop the model,
but better confidence requires additional information and analysis.
In this section the range equation for a laser rangefinder is derived. Be-
cause the radiometry techniques developed in this book do not cover co-
herent sources, an explanation is in order. In this rangefinder application
the laser is used as a source with very high radiance. Laser rangefind-
ers operate on the principle that light travels approximately 300 mm dis-
tance in one nanosecond. The time elapsed between the transmission of
the pulse and the reception of the pulse is used to determine the distance.
This specific analysis does not require the laser to be considered as a coher-
ent source, and hence the radiometry techniques in this book can be used
here. These techniques cannot be readily used for problems concerned
with spatial or temporal coherence.
In most laser rangefinders the transmitter and receiver are co-located
and co-axial on the same optical path. If a laser pulse is directed to an
object and reflected back from the object, the elapsed time between the
departure and arrival of the reflected light pulse is an indication of the
distance to the object. The objective with this analysis is to derive an ex-
pression for the SNR for the laser rangefinder. The SNR can then be used
to investigate the effect of several design parameters on system perfor-
mance. For another approach to the range equation for laser rangefinders,
see Kaminsky. 2
Normal vector
θOR R
Sensor
ROR
AO θLO
RLO
L
FL Las
er
The geometrical relationship between the laser transmitter, the object, and
the laser receiver is shown in Figure 9.7. The laser power or flux is de-
noted by Φ L (in watts), the distance from the laser to the object is R LO ,
and the distance from the object to the receiver is ROR . The illuminated
object-surface normal vector makes an angle θ LO with the laser illumina-
tion direction, and an angle θOR with the receiver sightline direction.
Several simplifying assumptions are made in order to simplify the
problem and emphasize the methodology. The receiver and transmitter
fields are co-axial, and the receiver and transmitter are located at the same
position. The receiver and transmitter are coincident, hence the distance
from the object to the laser transmitter is equal to the distance from the
object to the laser receiver, and the same atmospheric transmittance applies
to both optical paths.
The laser beam radiance is calculated from the definition of radiance
in Equation (2.19), requiring the optical power, beam area at the source and
beam solid angle (divergence). In order to use this very simple equation,
the Gaussian shape properties of the laser beam is discarded for two very
simple uniform shapes. The laser beam angular radiance distribution is
assumed to be uniform within the top-hat-shaped beam divergence profile
(e.g., peak normalized divergence). The laser beam power distribution is
assumed to be uniform across the area of the beam (e.g., peak normalized
area). Using Equation (2.19) and these two simplifications, the radiance
Electro-Optical System Analysis 323
can be written as
ΦL
LL = , (9.21)
ΩL AL
where Ω L is the laser beam solid angle, and A L is the laser beam cross-
section area at the laser source. This simplification might not satisfy the
required mathematical rigor, but it does provide order-of-magnitude radi-
ance estimates.
From Equations (2.26) and (9.21), the irradiance on the object is then
L L A L τOL cos θ L cos θ LO
EO =
R2LO
Φ L τOL cos θ LO
= , (9.22)
Ω L R2LO
where it is assumed that cos θ L = 1 because the laser beam radiates per-
pendicularly from the laser mirror. The uncooperative target object can
have any orientation relative to the laser beam, denoted by θ LO .
The laser pulse falling onto the object is reflected by the object. Most
natural surfaces have diffuse reflectance and scatter energy in all directions
(a Lambertian source). The reflected laser spot on the target object has a
radiance of
ρEO ρΦ L τLO cos θ LO
LO = = . (9.23)
π Ω L πR2LO
(ρ cos2 θ LO AO )Φ L τa2
ER = , (9.26)
πΩ L R4
where R is the distance between the object and the rangefinder. Equa-
tion (9.26) is similar to the radar range equation. The product ρ cos2 θ LO AO
can be regarded as the target optical cross-section. In the radar case, the
optical cross section has a fixed magnitude irrespective of distance between
the laser and target object. This is also true for a laser rangefinder illumi-
nating an airborne object where there is no reflective background. If the
object is observed against a terrain background, the terrain background
also contributes to the reflected signal (depending on the geometry).
The rangefinder irradiance SNR is the ratio of signal strength [Equation (9.26)]
to noise [Equation (9.20)], and is given by
ρΦ L τa2 cos2 θ LO AO
ER πΩ L R4
= √ (9.27)
En Δ f Ad
D ∗ A1 τa
ρΦ L τa2 cos2 θOR A0 D ∗ A1 τa
= . (9.28)
πΩ L R4 Δ f Ad
1. Variables and constants that the designer has no control over, such as
the object orientation and reflectivity, D ∗ , and constants.
2. Variables that the designer controls in the design process, such as laser
energy, receiver aperture area, laser pulse width, and detector size.
3. Distance-related factors that the designer has little control over.
The designer can now easily determine that increased laser energy and
receiver aperture improves the SNR linearly, whereas increased detector
area and pulse width decrease the SNR. Contrary to intuition, a longer
pulse width (i.e., a lower electronic bandwidth) decreases the SNR. Why?
If the laser rangefinder is viewing targets against the terrain, the laser
light is reflected from the target object as well as its surrounding terrain.
This implies that the real target area is not of sole importance because the
terrain background also reflects the laser pulse. There are two possibilities
regarding the laser transmitter and receiver beam or FOV sizes.
If the receiver FOV is larger than the transmitter beam width, the
receiver sees the whole laser spot. This implies that the effective laser spot
area is defined by the laser beam width by
AO cos θ LO
ΩL = , (9.30)
R2
and hence Equation (9.26) — the irradiance at the receiver — becomes
ρΦ L τa2 cos θ LO
ER = . (9.31)
πR2
If the receiver FOV is smaller than the transmitter beam width, the
receiver only views a portion of the whole laser spot. This implies that the
effective laser spot area is determined by the receiver FOV as
AO cos θ LO
ΩR = , (9.32)
R2
and hence Equation (9.26) becomes
ρΦ L τa2 cos θ LO
ER = Υ, (9.33)
πR2
where Υ = Ω R /Ω L is the fraction of the laser spot viewed by the laser
receiver FOV, and 0 ≤ Υ ≤ 1. When comparing Equations (9.31) and (9.33),
we see that they are the same, except for the fraction Υ. Equation (9.33)
is the more general case because Equation (9.31) is a special case when
Υ = 1.
326 Chapter 9
D∗ 3 × 1011 √ dλ 0.02 µm
cm· HzW−1
Ad 4.6 × 10−6 m2
The equations in the previous section provide the signal strength obtained
from a laser transmitter at a laser receiver. An estimate of the detection
range can be obtained by solving the range equation
ER ( R )
SNR = , (9.34)
EN
where SNR is the signal-to-noise ratio required to achieve detection. Solv-
ing the detection range problem means finding a value for range R in
Equation (9.29) that would yield the required SNR.
Table 9.5 Background radiance 2 at 1.06 µm, expressed as a detector D ∗ and resultant
operating range.
which in turn determines the noise in the detector. If the sensor is operat-
ing at night, the detector noise is at a minimum. If the sensor is pointed
at a bright, sunlit background, the current and hence the noise in the de-
tector increases relative to the dark night condition. By converting the
background flux noise to D ∗ , the range versus D ∗ graph can be used to
predict degradation in system performance under bright sunlight condi-
tions. Typical background radiance values are shown in Table 9.5. The
background radiance values were used to calculate the current in the de-
tector using a variation of Equation (6.16). Once the current in the detectors
were known, the shot noise at the respective currents were calculated, and
finally, new D ∗ values were calculated using Equation (5.32). These new
D ∗ values now represent the noise performance of the sensor under the
various background conditions. Once the new D ∗ values were known, the
range performance corresponding to the different conditions were deter-
mined from the bottom graph in Figure 9.8. The operating ranges are also
shown in Table 9.5.
If the target has a specularly reflecting surface (see Figure 3.11), the sur-
face BRDF f r is used to calculate the reflection as a function of angle of
incidence and the view direction angle (see Figure 3.12). Then
100
Signal-to-noise ratio 15 km visibility
8 km visibility
10
5 km visibility
3 km visibility
1
0 1 2 3 4 5 6 7 8 9 10
Range [km]
(a)
10
9 Visibility 5 km
Signal to noise ratio 8
8 Total darkness
7 Grass terrain
Range [km]
6
5 Snow terrain
4
3
2
1
108 109 1010 1011 1012 1013
½
D* [cm·Hz /W]
(b)
Figure 9.8 Laser-rangefinder range equation analysis: (a) SNR versus range for several
atmospheric visibility values and (b) expected range as a function of detector D ∗ .
22
Aircraft white paint
21 Semi matte light grey paint
Operating range [km]
1. In Figure 9.9, note the effect of surface properties on range, and in par-
ticular, range as a function of off-normal angle.
2. Except for the white paint anomaly, Figure 9.9 shows that specular sur-
faces support longer detection ranges along the mirror reflection vector
but lower detection ranges elsewhere. Lambertian surfaces, on the other
hand, support a near-constant detection range irrespective of view an-
gle. This is the infrared/optical equivalent of the radar geometric stealth
concept. The white paint anomaly requires that this statement be closely
investigated.
3. The sevenfold drop in diffuse reflectance between white chalk and matte
dark earth paint, 0.77 to 0.11, resulted in a detection-range ratio of
only 1.2. This indicates the relative robustness of the detection pro-
cess against paint variations. The ‘compression’ in detection range is
due to the 1/d2 term as well as the severe atmospheric attenuation at
longer ranges.
This compression effect will be less severe in a moderate atmosphere.
4. The fivefold drop between the specular peak and diffuse reflectance for
the specular paints only results in a detection range improvement of 1.2
times. The argument is the same as above.
330 Chapter 9
The sensor model described in Sections 6.5 and 6.7 is developed further to
predict the performance of a thermal imaging sensor. Much of the work
was already done in these two sections. Despite its simplicity, this model
can be used to predict thermal camera performance in trade-off studies.
Daniels 4 covers the same topic in more detail.
The sensor is regarded as an imaging system that builds up an image
by scanning the FOV with N detectors. The basic sensor configuration is
shown in Figure 6.13. The total number of detector elements are scanned
over the complete image by mechanical or other means. In some cases
the scanning may not be optimal and some time is lost due to the scan-
ning method. This ‘lost time’ is expressed in the scanning efficiency. The
scanning method is not considered here.
unity because the scan velocity is not constant (ηs = v /vmax ) or there
is a portion of the scan that is not available to form the image. The scan
efficiency is effectively the ratio of useful scan period to total frame time.
The scan efficiency for a staring array sensor is one.
The image fill efficiency in the a and b directions ηa and ηb allows for
the situation where the detectors do not cover the total field of regard. In
other words, pixel centerline spacing exceeds the pixel size. This is fairly
commonplace in staring detectors, having fill factors lower than 100% fill.
ηa = ηb = 1 implies exact filling, ηa < 1 and ηb < 1 implies under-filling,
and ηa > 1 and ηb > 1 indicates overfilling.
The electronic bandwidth required to pass the signal is given by
kf k f Ωr FF ηa ηb
f −3 dB = = , (9.39)
τe ωηs N
where k f is the time-bandwidth product (see Section 5.3.14). The noise
equivalent bandwidth can be derived from Equation (9.39) as
kn k f Ωr FF ηa ηb
Δf = , (9.40)
ωηs N
where kn is the ratio of noise equivalent bandwidth to −3 dB bandwidth
(refer to Section 5.3.13).
All of the noise sources in the sensor can be combined into one single
number. It is convenient to express this number in terms of the detector
D ∗ , which can be derived from Equations (5.30) and (5.26).
where NEE S is the inband NEE, D ∗ is derived from Equations (5.30) and
(5.26), Δ f is the noise equivalent bandwidth, Ad is the area of the detector,
As is the area of the sensor’s entrance pupil, and τs is the effective trans-
mittance of the sensor. The optical PSF constant ks is the fraction of energy
from a point source falling onto a single detector element. It therefore rep-
resents the sensor’s capability to gather energy from a point source. It is
assumed that each detector element has uniform responsivity over its area.
332 Chapter 9
π NEE f 2
NETD =
ab dM
dT
ηa ηb kn k f Ωr FF 4( f /#)2
= ∗ P f τ dM
, (9.49)
αβ ηs N Deff s dT
where
∞
dM dMλ ( Tt )
= λ τaλ Sλ dλ (9.50)
dT 0 dT
is the derivative of the source exitance with respect to the source tem-
perature; in other words, the rate at which the source exitance changes
for a given change in source temperature. This derivative is required to
transform a small change in exitance into a corresponding small change in
temperature. It is required that the background temperature be specified
with the NETD value because the NETD value depends on the background
temperature against which it is measured. NETD is generally used to de-
scribe the sensitivity of thermal imaging systems. NETD is only defined
for extended sources, that is, sources that are larger than the sensor pixel
FOV. If the source is smaller than the FOV, any attempt to describe the
NETD is incorrect.
By simplification, the equation for noise equivalent temperature dif-
ference, Equation (9.49), can be reduced to the form of Equation (5.35) in
Lloyd’s classic book: 5
2 Δf
NETD = . (9.51)
αβηD ∗ τs Ds dM
dT
Consider now the application of Equation (9.49) for the performance pre-
diction of a thermal camera. Two spectral bands are investigated: 3–5.5 µm
and 8–12 µm. In addition, the relative value of using a large number of de-
tector elements must be investigated. In particular, we investigate cameras
with N = 1, 256, and 256 × 256 detector elements. For the single-element
detector case it is assumed that the image is formed by a two-dimensional
scanner sweeping the single detector to form the complete image. For
the 256-element detector case it is assumed that the image is formed by a
one-dimensional scanner sweeping a vector of detectors to form the com-
plete image. For the 256 × 256 element detector case it is assumed that the
image is formed by a staring array sensor with no scanning. The design
parameters considered for this analysis are shown in Table 9.6.
334 Chapter 9
For some parameters three values are listed. In these cases they corre-
spond to the detector-element choices of 1, 256, and 256 × 256. Note that
not all of the values assumed above are realistic in practice. It is assumed
that the detector arrays will have perfect responsivity uniformity. This is
not practical or possible in real life. However, from a design comparison
perspective, these values are accepted. The electronic noise bandwidth,
detector dwell time, and NETD values for the three detector choices are
shown in Table 9.7.
1. Practically available technology in the year 2000: staring array for 3–5-
µm imagers and linear scanned arrays for 8–12-µm imagers. All other
parameters are the same. The number of detector elements in the linear
vector is equal to the square root of the number of elements in the
staring array.
Figure 9.10 shows the comparison for three levels of humidity (50%,
75%, and 95%), for atmospheric conditions ranging from −20 ◦ C to +50 ◦ C.
The background temperature is assumed to be the same as the atmospheric
temperature. Four distances are considered: 0 km, 2.5 km, 5 km, and
10 km. The noise equivalent target contrast is calculated for all of these
conditions; see Figure 9.10. The curves show the NETC versus scene/back-
ground temperature. Observe that the 8–12-µm imager performance de-
grades rapidly at higher temperatures and longer ranges. The NETC for a
Figure 9.10 Noise equivalent target contrast required for different 3–5-µm imagers and 8–
Chapter 9
Noise-equivalent target contrast for RH=50% Noise-equivalent target contrast for RH=75% Noise-equivalent target contrast for RH=95%
10 10 10
2000 Technology 2000 Technology 2000 Technology
3–5-mm staring array 3–5-mm staring array 3–5-mm staring array
Target temperature contrast [K]
humidity
recorded
absolute
Highest
10 km 10 km 10 km
5 km 5 km 10 km 5 km
2.5 km 2.5 km 5 km 2.5 km 10 km 5 km
1 0 km 1 0 km 1 0 km
10 km
2.5 km
5 km
2.5 km
2.5 km
Noise-equivalent target contrast for RH=50% Noise-equivalent target contrast for RH=75% Noise-equivalent target contrast for RH=95%
1 1 1
10 km 10 km 10 km
5 km 10 km
5 km 5 km
2.5 km 10 km 5 km
Target temperature contrast [K]
10 km 5 km
2.5 km
humidity
recorded
absolute
Highest
2.5 km
0 km 0 km 0 km
0.01 0.01 0.01
40
336
Flux from
dewar and/or
cold shield
Filtered flux
from optics 1
barrel plus 2
filter flux
Optics barrel
Flux from Thermally-insulating dewar
filter, optics, 3 Detector
and scene Cold finger
Liquid coolant
Cooler
Cold shield
Cold filter
Hot filter
flux on the detector. In order to reduce the unwanted flux on the detector,
a cold shield 8–11 (cold screen or cold cone) is constructed around the detec-
tor, mounted on the cold finger (see Figure 5.21). Because this cold finger
and shield are at the same low temperature as the detector, the thermal
radiation from the cold shield is significantly less than the radiance from
the sensor components at room temperature.
The sensor may also employ an optically selective filter, mounted in
front of the detector. This filter only transmits flux in the spectral band
required by the sensor’s application. Flux outside the transmittance pass-
band is attenuated and never reaches the detector. Note, however, that the
filter also emits flux because the filter is a thermal radiator. Thus, in the
passband, the filter transmits flux from the scene, whereas in the stopbands
the filter radiates with emissivity = α = 1 − τ − ρ, from Equation (2.3.4)
and Kirchhoff’s law (Section 3.2.1). In Figure 9.11 the filter is shown as the
front dewar window, but it can be located anywhere in the optical path.
The filter can be cooled down to reduce its radiated flux. If the filter is
mounted on the cold finger, it is called a ‘cold filter.’
Figure 9.11 depicts the different radiance zones in the sensor. To the
first approximation, the zones are rotationally symmetric even though the
figure depicts these as linear angles. Also, the zones are shown to emanate
from the center of the detector, but any real detector has a finite size with
slightly different zone shapes from every small part on the detector. The
cold shield geometry can become quite complex in a detailed analysis.
Electro-Optical System Analysis 339
In zone 1 the flux on the detector originates on the walls of the de-
war (sensor temperature) and/or cold shield walls (detector temperature).
Clearly, the design objective will be to increase the cold-shield solid angle
so as to decrease the radiation from the hot dewar walls and optics barrel
without reducing the signal flux.
In zone 2 the flux on the detector originates on the inside of the sen-
sor, i.e., the optical barrel and mounting rings. This flux could be ther-
mally emitted or external flux reflected from the barrel. The internal flux
is filtered by the hot/cold filter. If the filter is hot, the flux in the stop-
bands would be of the same magnitude as the internally self-emitted flux
(suppressing the reflected flux). If the filter is cold, the filter flux in the
stopbands can be small.
In zone 3 the flux on the detector is the sum of the scene flux, the
optics flux, and the filter flux. In the event of a sensor with hot optics and
filter, together with a cold scene, the scene flux can be considerably less
than the sensor fluxes.
A key strategy in improving a sensor’s sensitivity is therefore to cool
down the detector environment and filter to reduce the background flux
in zones 1 and 2. A cold filter will also reduce the filter radiated flux
L( T f )(1 − τ − ρ) in zone 3, which can be significant if the filter has a
narrow passband. Optics flux can be minimized by keeping the optics cool,
but more importantly, selecting materials with low emissivity in the sensor
spectral band. For a discussion of the effect of hot optics, see Section 9.8.2.
The radiometry in an imaging system is described in the camera equa-
tion, 12,13 which will be further developed in this section. The derivation
will cover the primary scene radiance as well as (some of) the radiance
sources in the sensor itself. In this analysis, the ideal thin-lens paraxial ap-
proximation is made. The linear angles in the sectional diagrams should
be viewed as rotationally symmetric solid angles, e.g., zone 3 in Figure 9.11
is a conical solid angle. It is also assumed that system throughput is the
same at all field angles, i.e., the on-axis marginal ray cone solid angle has
the same value as the off-axis marginal ray cone solid angle, as in Fig-
ure 9.12.
Figure 9.12 shows the primary flux sources in an imaging sensor. The
contributing source radiance values are the scene focused on the detector
(Lscene ), the optics and window (Loptics ), the filter (Lfilter ), the optics barrel
and inside of the sensor (Lbarrel ), and the detector cold shield (Lcold shield ).
340 Chapter 9
O N M
Pupil
Object plane Focal plane
Lbarrel Lfilter Lcold shield
Off axis light cone
Marginal ray
h α W2b W2a
Chief W1
Lscene ra
(R/cos y Loptics W3 dA¢
dA θ α) θ¢
Ad
Ld
On axis light cone α
h¢
s s¢
where Ad is the detector area, Ao is the optics exit pupil area, α is the
off-axis angle to the object, τoλ is the optics transmittance, τ f λ is the fil-
ter transmittance, Lsceneλ is the scene radiance [Equation (8.1)], Lopticsλ is
the optics radiance, Lfilterλ is the filter radiance, Lbarrelλ is the optics barrel
radiance, and Lcold shieldλ is the detector cold shield radiance. The solid an-
gles Ω1 , Ω2a , Ω2b , and Ω3 are defined in Figure 9.12. From Equation (2.3),
oλ = 1 − ρoλ − τoλ , and f λ = 1 − ρ f λ − τ f λ . Note that Ao /s = Ω3 =
π sin2 θ . Equation (9.52) assumes that the optical system has no vignet-
ting or central obscuration. Analysis of any real system would require that
these factors be taken into account. Not shown in any of the derivations
here is the effect of stray light entering from outside of the ray cone. The
source can be outside the sensor or inside the sensor (e.g., hot rotating
parts). The stray light is normally associated by one or more reflections
from the optical barrel or even optical element surfaces. Stray light is of-
ten suppressed by appropriate baffle design, but some stray effects may
remain. Using Equation (2.12) for Ω3 , the detector flux is then
If the object is not at infinity (finite conjugates), the optics image the
object onto the focal plane with a given magnification m. Combining the
simple lens equations, Equations (6.1), (6.2), and (6.3), with Equation (9.54),
it is found that s = f (1 + |m|). Applying that to the definition of θ it
follows that (for paraxial optics)
D 1
sin θ = = , (9.55)
2(1 + |m|) f 2F# (1 + |m|)
leading to the new formulation for flux on the detector:
πKC K N (α) Ad τoλ τ f λ Lsceneλ cos4 α
Φdet scene λ = , (9.56)
4F#2 (1 + |m|)2
where two new factors are introduced: KC accounts for central obscuration
(if present), and K N (α) accounts for vignetting. 13 Note that the angle under
consideration in the cos4 term is the object field angle, and not the cold
shield angle, which is a function of the numerical aperture (f -number cone)
of the cold shield. Unless the cold shield obscures the optical ray cone, the
cold shield numerical aperture does not come into consideration at all in
the cos4 effect.
The cold shield efficiency is the ratio of scene flux to total flux onto
the detector [Equations (9.52) and (9.56)],
Φdet scene λ dλ
ηcold shield = . (9.57)
Φdetλ dλ
342 Chapter 9
Figure 9.13 Reduced cold shield efficiency and vignetting in practical designs.
In Figure 9.14, consider a small elemental area dAi in the object plane imag-
ing onto a small elemental area dA in the focal (image) plane. The conical
solid angle defined by the marginal rays (solid of revolution) contains all
of the flux flowing from dAi to dA . No flux outside this solid angle con-
tributes to the flux flow. Further, consider a small portion of the solid angle
as shown in the dark shaded area in the top figure of Figure 9.14. All of
the flux flowing from dAi to dA , passing through dAo in the plane Oo ,
Electro-Optical System Analysis 343
Oi Oo N M
Object Out-of-focus Pupil Focal
plane object plane plane
Does not contribute dA’ h’
to detector
flux dAo Contributes
to detector flux
Does not
contribute to
detector flux
h
dAi
h
dAi
θ r dr
Rays between
the two cones
defines radiance-
field solid angle
Ring-shaped
source area
defines
radiance-field
area
has to flow along the shaded solid angle indicated in the figure. Any flux
flowing through the area dAo but outside the dark shaded solid angle will
not contribute to the flux flow between dAi and dA . This is shown more
explicitly in the bottom two pictures in Figure 9.14.
Suppose that an opaque source with uniform radiance is located in
plane Oo . The optics located in N are not concerned with where the flux
emanates, from the plane Oi or Oo . The optics dutifully focus the rays
according to their image forming design, along the same ray paths in both
cases. The focal-plane irradiance of an out-of-focus source dAo of uni-
form radiance provides exactly the same flux on the detector element as
would the in-focus source dAi with the same radiance. This observation
is a re-statement of the principle of radiance conservation: the location
of the source is not important, the spatial properties of the radiance field
determines the flux in the focal plane. The properties (including location)
of the source are important to create the radiance field; the field cannot
exist without the source in its precise location. However, once created, the
radiance field ‘carries’ through space with no further dependence on the
source.
Careful study of Figure 9.14 also indicates that radiance is defined by
a matched set of areas and solid angle directions. The small area dAi in
the object plane, with a full conical solid angle uniformly filled with rays,
provides the same flux on the detector as a large number of small areas in
the plane Oo but with narrow conical sections associated with each small
area.
The practical implication of this observation is that large, uniform
sources do not have to be ‘in focus.’ The requirement for the object location
at a particular plane of focus is only a requirement if the source radiance
is not uniform, and this pattern must be imaged accurately onto the focal
plane, i.e., if a sharp image is required.
¢DZ
ǰ ¢
¢
¢
flame. At the exit of the burner nozzle, the flame is rich in butane/propane
and less rich in oxygen. The gas mixes with the air at the ‘outside’ and top
of the flame, burning away the rich gas concentration at the center of the
flame.
The data analysis workflow is shown in Figure 9.15. The process starts
with the two sets (flame and reference) of measurements. In each case the
instrument calibration data is used to calculate radiance values (a radiance
spectrum and a radiance image). Because (1) the flame only partially fills
the FTIR spectrometer FOV, but (2) the calibration data applies to a fully
filled FOV, the spectral radiance measurement is ‘scaled’ with an unknown
factor.
The temperature and spectral emissivity jointly result in the observed
radiance values. The magnitudes of neither temperature nor spectral emis-
sivity are known. This procedure investigates sets of temperature and
emissivity values (Section 8.4) that would solve the measurement equa-
tion. One set — the one that best matches physical reality — is finally
346 Chapter 9
selected.
Starting from prior knowledge (open literature, past experience, or
physics principles), select a temperature and then calculate spectral emis-
sivity from the measured spectral radiance and Planck’s law:
Lmλ
λm = , (9.58)
Lbbλ ( Tm )
where Lmλ is the measured radiance, and Lbbλ ( Tm ) is the blackbody radi-
ance at the estimated temperature.
Use the radiance image obtained from the imaging instrument and the
spectral emissivity estimate mλ obtained above to calculate a temperature
image (for each pixel in the radiance image) by solving for temperature in
∞
Limage = Lbbλ ( Tm )λm τaλ Sλ dλ, (9.59)
0
where Limage is the wideband radiance value measured by the imaging
camera, λm is calculated above, τaλ is the spectral atmospheric transmit-
tance, and S λ is the imaging sensor spectral response. This process pro-
vides an image where each pixel in the image represents temperature. The
temperature map is critically evaluated to determine if the predicted tem-
peratures are acceptable (judgement call required). If the temperatures are
not acceptable, the FTIR sample temperature estimate is adjusted, and the
process is repeated.
This analysis assumes that the spectral emissivity will not vary sig-
nificantly for different flame temperatures — this assumption is only war-
ranted if the temperature spread is not too wide. This analysis also ignores
the effect of atmospheric transmission, i.e., τaλ = 1 in Equation (9.59), on
the assumption that the path length in the laboratory was short. However,
the atmospheric attenuation in the CO2 absorption band is severe, even
over these short distances.
The model developed here serves to convey the principles involved.
In practice, these principles will be used to develop a model that accounts
for the spatial variation across the area of the flame (texture). In some
cases the model might even account for temporal variations in the texture
and flame radiance.
18000
MWIR 120 ms 42.5 °C
16000
ND 2 filter
14000 500 ms 42 3 °C
20090908
Digital level
12000
10000 500 ms 16.0 °C
8000 120 ms 16.0 °C
6000
4000 30 ms 42.6 °C
2000 Operational
Calibration measurement 30 ms 16.0 °C
0
300 400 500 600 700 800 900 1000
Source temperature [K]
(a)
6000
5000
4000 500 ms 16 0 °C 120 ms 16.0 °C
3000 Hot optical elements,
filter and optical barrel Cold optics,
30 ms 42.6 °C
2000 theoretical curve
30 ms 16.0 °C
1 10 100
Source irradiance [mW/m2]
(b)
Figure 9.16 MWIR imaging radiometer calibration curves for (a) digital level vs. blackbody
temperature, and (b) digital level vs. irradiance.
described in Section 6.6 and Chapter 8. It is also the basis for the analysis
described in this section.
Figure 9.16 shows typical calibration data. The top curve relates source
temperature and instrument voltage (digital level). The bottom curve re-
lates source irradiance on the sensor entrance aperture to instrument volt-
age. For an extended source, the source apparent radiance is related to ir-
radiance by E = Lω, where ω is the pixel FOV. In this case the instrument
was calibrated for three different gain settings (integration times of 30, 120,
and 500 µs) and two different sensor internal temperature conditions (16
and 42.6 ◦ C). Observe the importance of calibration at different internal
temperatures: the curves (for the same instrument and settings) show sig-
nificantly different responses. A good strategy to allow for temperature
changes in the sensor is to measure the calibration curve at several differ-
ent internal temperatures and then interpolate between these, according to
the actual instrument temperature during the measurement.
The effect of hot optics is also shown in Figure 9.16. It is evident
that the hot optics flux sets an asymptotically lower measurable flux limit
(called the ‘floor’). When measuring low-temperature test samples, a small
variation in internal temperature shifts the floor up or down, playing havoc
with calibration. In this particular instrument setting, an ND2 neutral
density filter (0.01 transmittance) was used. Much of the hot optics flux
emanates from this filter. So in all fairness, there is little point in using an
ND2 filter when measuring a low-temperature target. The situation does
arise, however, if there is a requirement to measure both hot and cold test
samples without changing filters.
9.8.3 Measurements
3000
L(3.8 4.8 mm) 552.5 [W/(m2·sr)]
Radiance [W/(m2·sr·mm)]
2500
2000
1500
1000
500
0
2.5 3 3.5 4 4.5 5 5.5 6
Wavelength [mm]
(a)
0.3
Radiance [W/(m2·sr·cm-1)]
0.25
0.2
0.15
Measured reference
0.1
Calculated reference
0.05
0
2000 2200 2400 2600 2800 3000 3200 3400 3600
-1
Wavenumber [cm ]
(b)
Figure 9.17 Spectral radiance for (a) the Bunsen burner yellow flame and (b) the reference
source.
The FTIR instrument does not have the same fine spatial resolution
as the imaging instruments, and the measurement represents some form
of ‘average’ of the flame. The background temperature was much lower
than the flame temperature and was ignored in further analysis. The FTIR
spectrometer’s small FOV was pointed at the bottom of the flame — a
relatively cold part of the flame. Furthermore, the flame radiance was not
uniform in the instrument’s FOV. Because of the uncertainty in percentage
fill and flame nonuniformity, there is some uncertainty in the absolute
magnitude of the measurement. The radiance spectral shape was later
shown to be accurate (see below). If the measured results are accepted as a
scaled version of the full flame radiance, it is still useful because the shape
of the spectral emissivity can be extracted. The top graph in Figure 9.17
shows the FTIR spectral radiance measurement as well as the wideband
350 Chapter 9
500
2000
Temperature [K]
1
1800
Emissivity
1600
1400
1200
1000 0
3 4 5 6
800 Wavelength [mm]
Figure 9.18 Bunsen flame MWIR images in units of digital level, radiance, and temperature.
1
for 650 °C
0.8
Estimated emissivity
for 850 °C
for 1050 °C
0.6
for 1250 °C
0.4 for 1450 °C
0.2
0
2000 2200 2400 2600 2800 3000 3200 3400 3600
Wavenumber [cm 1]
(a)
0.6
0.5
Emissivity
0.4
0.3
0.2
0.1
0
2000 2200 2400 2600 2800 3000 3200 3400 3600
Wavenumber [cm 1]
(b)
Figure 9.19 (a) Bunsen burner estimated emissivity-temperature combinations. (b) Appar-
ent emissivity of the Bunsen burner yellow flame at 825 ◦ C.
The flame area is calculated from the radiance image using the techniques
described in Section 8.4.2. The flame area was determined for nine radi-
ance threshold values, ranging from just above background to 95% of the
Electro-Optical System Analysis 353
10-2
Flame size [m2]
10-5
0 500 1000 1500 2000 2500
2
Threshold [W/(m ·sr)]
(a)
10-2
10-3
10-5
40 60 80 100 120 140
Threshold [W/(m2·sr)]
(b)
Figure 9.20 Bunsen burner yellow flame predicted area as function of threshold for
(a) MWIR, and (b) LWIR.
The Bunsen yellow flame has significant turbulence as the air mixes with
the flame, consuming oxygen and expanding the hot gas plume. The forces
resulting from internal mass flow acting on this plume cause it to become
turbulent, resulting in the ‘dancing’ of the flame. The flame shape varied
considerably in time as the turbulence contorted the flame. Figure 9.21
shows subsequent images in a measurement series; the interval between
354 Chapter 9
Figure 9.21 Bunsen flame sequence with 60-ms intervals between frames.
successive frames was 60 ms. The flame shapes vary considerably between
subsequent frames. The turbulence bandwidth for this flame is higher
than 20 Hz. There is little correlation between the ‘hot spots’ in subse-
quent frames. The hot spots indicate the spatial locations of combustion,
and these locations change very quickly. The flame shapes show huge
differences, but the flame volume (size) appears to be similar between all
of the frames. The flame area is therefore expected to show only a small
variation in time.
Bibliography
[1] Pyradi team, “Pyradi Radiometry Python Toolkit,” http://code.
google.com/p/pyradi.
[4] Daniels, A., Infrared Systems, Detectors and FPAs, 2 Ed., SPIE Press
(2010).
[5] Lloyd, J. M., Thermal Imaging Systems, Plenum Press, New York (1975).
[6] Rogatto, W. D., Ed., The Infrared and Electro-Optical Systems Handbook:
Electro-Optical Components, Vol. 3, ERIM and SPIE Press, Bellingham,
WA (1993).
[7] Rogalski, A., Infrared Detectors, 2nd Ed., CRC Press, Boca Raton, FL
(2011).
[8] Campana, S. B., Ed., The Infrared and Electro-Optical Systems Handbook:
Passive Electro-Optical Systems , Vol. 5, ERIM and SPIE Press, Belling-
ham, WA (1993).
[9] Wolfe, W. L. and Zissis, G., The Infrared Handbook, Office of Naval
Research, US Navy, Infrared Information and Analysis Center, Envi-
ronmental Research Institute of Michigan (1978).
[11] Holst, G., Electro-Optical Imaging System Performance, 5th Ed., JCD
Publishing, Winter Park, FL (2008).
[12] Palmer, J. M. and Grant, B. G., The Art of Radiometry, SPIE Press,
Bellingham, WA (2009) [doi: 10.1117/3.798237].
[13] Slater, P., Remote Sensing: Optics and Optical Systems , Addison-Wesley,
Boston, MA (1980).
[14] Strojnik, M., Paez, G., and Granados, J. C., “Flame thermometry,”
Proc. SPIE 6307, 63070L (2006) [doi: 10.1117/12.674938].
[15] Haber, L. C., An investigation into the origin, measurement and application
of chemiluminescent light emissions from premixed flames, Master’s thesis,
Virginia Polytechnic Institute and State University (2000).
[20] Yildirim, Z., Self-defense of large aircraft, Master’s thesis, Naval Post-
graduate School (2008).
Problems
9.1 The purpose of this study is to verify the proposed FOV and to
optimize the design to achieve maximum SNR. The data for this
problem is given in the DP01.zip data file on the pyradi web-
site. 22
A flame sensor has an aperture area of 0.005 m2 and a proposed
FOV of 10−5 sr. The InSb detector has a peak responsivity of 2.5
A/W and normalized spectral response defined in the data file
(detectorNormalized), shown in Figure 9.22. The sensor filter
Electro-Optical System Analysis 357
9.1.1 Describe the possible means whereby the SNR can be optimized
for this system. [3]
9.1.2 Use the data in the DP01.zip data file, or use Modtran™ to cal-
culate the transmittance and path radiance. Use the path geometry
as defined above and confirm that the calculated values agree with
the graphs shown here. [4]
9.1.3 Compile a mathematical formulation for the signal, the noise, and
the SNR. This formulation will be used to evaluate your optimiz-
ing strategies, so it must be a complete description of the system
with all parameters affecting system performance. [5]
358 Chapter 9
0.2
0
2.5 3 3.5 4 4.5 5 5.5 6 6.5
Wavelength [mm]
(a)
0.02
Radiance [W/(m2·sr·cm )]
Sky radiance
-1
0.015
0.01
0.005
0
2.5 3 3.5 4 4.5 5 5.5 6 6.5
Wavelength [mm]
(b)
Figure 9.22 (a) Spectral data for the detector, filter, atmosphere, and flame emissivity.
(b) Path radiance spectral data.
9.1.6 Calculate the SNR for the following sensor fields of view: 1 × 10−7 ,
1 × 10−6 , 1 × 10−5 , 1 × 10−4 , and 1 × 10−3 sr. [2]
Plot the SNR versus FOV. Plot on log-log graphs and verify that
you get a curve of the form shown below. Explain the shape of the
curve. [2]
Signal to noise ratio: contrast current / noise
100
10
SNR
0.1
10 7 10 6 10 5 10 4 10 3
Field of view [sr]
15
9.2 The data for this problem is given in the DP02.zip data file on
the pyradi website. 22
360 Chapter 9
Band λc n a k
Silicon 1.20 4.30 3.50 8
3–6 µm 6.00 4.30 3.50 30
8–12 µm 12.00 4.30 3.50 60
An optical filter is used to limit the spectral width. The filter re-
sponse is given by Equation (D.4) with the following values:
Electro-Optical System Analysis 361
Band τs τp s Δλ λc
Silicon 0.001 0.90 20 0.20 0.55
3–6 µm 0.001 0.90 12 1.20 4.20
8–12 µm 0.001 0.90 6 2.50 10.00
-4
10
10-6
10-4
-6
10
10-8
10-6
3 3
5 g/m H2O 10 -8 5 g/m H2O 5 g/m3 H2O
-10
10 8 9 10 11 12 8 9 10 11 12 8 9 10 11 12
Irradiance [W/(m2·mm)]
10-6 10-4
-6
10
10-8 10-6
10-8
39 g/m3 H2O 39 g/m3 H2O 39 g/m3 H2O
-10
10
8 9 10 11 12 8 9 10 11 12 8 9 10 11 12
Wavelength [mm] Wavelength [mm] Wavelength [mm]
9.2.5 Calculate the current flowing through each of the detectors when
viewing the different targets through the various atmospheres. Us-
ing the moderate atmosphere as a baseline, calculate the ratios of
currents for the two adverse atmospheric conditions. Compare
your results with the following table (current values are given in
scientific notation and the ratio of currents are given in brackets).
The results will not be exactly the same but should be of the same
magnitude. [10]
Electro-Optical System Analysis 363
Temperature = 300 K
Atmosphere,
Silicon 3–6 µm 8–12 µm
H2 O content
23-km visibility 1.09 × 10−25 1.87 × 10−8 6.47 × 10−7
Rural, 5 g/m3 (1.000) (1.000) (1.000)
2-km visibility 1.09 × 10−25 1.84 × 10−8 6.42 × 10−7
Urban, 5 g/m3 (1.000) (0.982) (0.992)
23-km visibility 1.09 × 10−25 2.20 × 10−8 7.28 × 10−7
Rural, 39 g/m3 (1.000) (1.172) (1.124)
[A] (ratio) [A] (ratio) [A] (ratio)
Temperature = 1000 K
Atmosphere,
Silicon 3–6 µm 8–12 µm
H2 O content
23-km visibility 1.35 × 10−11 4.27 × 10−5 2.30 × 10−5
Rural, 5 g/m3 (1.000) (1.000) (1.000)
2-km visibility 5.97 × 10−13 2.63 × 10−5 1.60 × 10−5
Urban, 5 g/m3 (0.044) (0.617) (0.694)
23-km visibility 1.32 × 10−11 3.15 × 10−5 4.28 × 10−6
Rural, 39 g/m3 (0.972) (0.737) (0.186)
[A] (ratio) [A] (ratio) [A] (ratio)
Temperature = 6000 K
Atmosphere,
Silicon 3–6 µm 8–12 µm
H2 O content
23-km visibility 7.18 × 10−3 1.73 × 10−3 2.79 × 10−4
Rural, 5 g/m3 (1.000) (1.000) (1.000)
2-km visibility 2.33 × 10−4 1.06 × 10−3 1.91 × 10−4
Urban, 5 g/m3 (0.033) (0.614) (0.685)
23-km visibility 7.09 × 10−3 1.29 × 10−3 4.57 × 10−5
Rural, 39 g/m3 (0.988) (0.750) (0.164)
[A] (ratio) [A] (ratio) [A] (ratio)
9.2.6 Analyze the results obtained in the previous two questions. Com-
ment on the relevance of each respective sensor to observe each
target. Review the effect of the different atmospheric conditions
on the observed irradiance values. Make recommendations as to
which spectral bands to use for different sources and different
atmospheric conditions. For example, complete a table such as
shown below. [6]
364 Chapter 9
Figure 10.1 summarizes almost everything one must remember when do-
ing radiometric calculations.
Always view the problem as an application of Figure 2.11 and Equa-
tion (2.31); repeated as Figure 10.1. All problems can be rooted in this sim-
ple model: start here, and extend in wavelength, medium effects and/or
geometry. In particular: (1) consider the source area as some surface in
θ0 θ1
dA0 R01
dA1
365
366 Chapter 10
A0 A0
Ω0 Ω1
A1 A1
A0 Ω0 A0 Ω1
A1 A1
space and the receiver area as some surface in space, and integrate over
both surfaces, (2) consider the spectral properties and integrate over wave-
length or wavenumber, and (3) consider medium effects such as transmit-
tance and path radiance.
Note that in the right of the flux transfer equation there are only two
radiometric quantities L and λ and one medium property τ01 ; the remain-
ing quantities are all geometric (nonradiometric) quantities. Radiometry
is therefore as much a study of geometry as it is of optical flux. Get the
geometry right, and the solution falls in place. If there is not a clear picture
of the geometry, the correct solution is out of reach.
It is easy to get confused by source area and receiver area — when to use
which? The simple rule is to think of it, as if you are standing on one of
the two surfaces and you are viewing the other surface. You cannot view
the one you are standing on, you can only view the other. Your feet and
eyes cannot rest on the same surface. This is shown Figure 10.2.
Derive a mathematical model from Equation (2.31). Start simple and add
components as required by the problem at hand. Using the drawing as
input, add factors for the atmosphere, lenses, optical filters, detectors,
choppers, and amplifiers. Do not add factors for components that are
not specified in the problem statement.
Golden Rules 367
As early as possible, convert problem units to base SI units, 1,2 shown in Ta-
ble 10.1, or directly-derived units. The value of a physical quantity can be
expressed as the product of a numerical value (i.e., 4.3) and a unit (i.e., µm),
both of which are algebraic factors, that can be manipulated by the rules of
algebra. The symbol µm is related to the symbol m, by µm = 1 × 10−6 m,
thus µm/(1 × 10−6 ) = m. One can therefore write λ = 4.3 µm, or divide
both sides to obtain λ/µm = 4.3. Download and study the free IUPAC
Green Book 2 for more information.
It happens frequently that distances and altitudes are given in [km]
(e.g., the sun’s diameter and distance in Section 3.7), temperature is in
Celsius, or that detector D ∗ or sizes contains units of [cm]. Work in base SI
units, rather than problem-domain units. Convert to SI units at the earliest
possible time.
Some of the few exceptions to this rule are that wavelength is normally
−1 ∗
specified in [µm] or [nm],√ wavenumber is specified in [cm ], and D is
specified in units of [cm· Hz/W].
they share the same SI unit: meter [m] (see Table 10.1 for dimensional
symbols). Strict adherence to standards requires the use of the dimension
symbols, such as [L] or [M], but using SI units [m] or [kg] works just as well.
Subscript-mark all length dimensions with the meaning and location of
such dimensions. For example, use [m20 ] or [L20 ] for source area, [m21 ] or [L21 ]
for receiver area, [m2d ] or [L2d ] for detector area, and [mR ] or [LR ] for range.
Use different subscripts for different surfaces, even if they all refer to area.
Note that solid angle has units of [m20 /m2R ] or dimensions of [L20 /L2R ] when
viewing the source (surface 0) from a sensor (surface 1). Once this is done,
do not ‘cancel’ different types of lengths in the dimensional analysis, i.e., a
ms cannot ‘cancel’ a md because they are different types of length. Ensure
that all appropriate units/dimensions are present and cancel correctly.
The dimensional analysis for Equation (2.31) is as follows:
L dA0 cos θ0 dA1 cos θ1 W · m2R & 2 ' & 2 ' 1
d Φ [W ] =
2
m0 m1 → [W ] ,
R2 m20 · m21 m2R
or when considering spectral variables,
Lλ dA0 cos θ0 dA1 cos θ1 dλ
d Φ [W ] =
2
λ R2
W · mR 2 & 2 ' & 2 ' 1 ( µm )
m0 m1 → [W ] .
m20 · m21 · µm m2R 1
dλ 104 ( ) 1
( µm )
−1
=− 2 µm · cm → .
dν̃ ν̃ (cm−1 )2 cm−1
Draw a picture of the system and the spatial geometry of the problem, such
as shown in Figures 10.3 and 10.4. Write down what effect that component
Golden Rules 369
θ0 θ1 F i v
Zt
dA0 dA1
No No No
effect R01 Focal effect
effect length
or loss or loss or loss
Only flux from dA0,
not the whole object
R01 dA1 f
dA0
mR mf md x md
ms × ms
mo × mo
cos θ = 1
cos θ = 1 R
R
cos θ = 1
has on the flow or unit conversion of the flux or signal, and what effect the
medium has on the flux/signal.
The drawing should specify which component provides the flux and
which component receives the flux. It is very easy to confuse the optical
aperture of a lens with the detector area as the receptor of flux.
Ensure that the drawing clearly shows not only the full extent or size
of the source object but also which part is visible to the sensor. The sensor
can only sense the flux radiating within its FOV — the object’s radiation
outside the sensor FOV does not contribute to the sensor signal.
For example, consider a system comprising a spectral source, an at-
mospheric medium, a lens, an optical filter, a detector, and an amplifier.
The diagram could look like Figure 10.3.
In this picture, pay detailed attention to the source and receiving ar-
eas; in particular, make sure which part receives the flux. How is the flux
transferred from one block to the next — are there losses or unit conver-
sions along the way? The picture should also indicate if the cosine factors
in Equation (2.31) can degenerate to unity or if these must be retained.
Figure 10.4 illustrates these concepts.
For each of the objects in the figure, write down the type of com-
ponent, value, the units, and the type of conversion taking place in the
component. Consider the units of the component, i.e., [A/W] means the
Golden Rules 371
dA0
R 01 dA0 cos θ
component receives watts and outputs amperes. An area with units [m2 ]
can convert irradiance in [W/m2 ] to flux [W]. An example is shown in
Table 10.2.
It may also be beneficial to draw the shapes of the source and receiver
true to the real-world object (e.g., the barrel containing the optical elements
and the detector). Such a drawing willimmediately indicate if it is neces-
sary to perform a spatial integral A = A dA over one or both of the areas
in Equation (2.31).
Remember when to remember π. See Section 2.7 for the relationship be-
tween exitance and radiance for a Lambertian radiator. When working
from first principles, there is no need to remember when to use π because
it is taken care of in the mathematics.
The scripting Planck-law functions given in Appendix D provide exi-
tance in [W/m2 ], not radiance in [W/(m2 ·sr)]. When using these functions,
divide the result by π to get radiance.
Simplify spatial integrals where possible. In Figure 10.5 the solid angle of
the optical barrel is integrated over the inside of the box except over the
lens. On the assumption that the optical barrel has uniform radiance, the
barrel geometry can be ‘collapsed’ onto, and integrated over, the portion
of the sphere shown in the figure. The spherical portion is rotationally
symmetrical around the optical axis. In this case the projected solid angle
of the barrel is given by ω = π − π sin2 θb .
372 Chapter 10
1
Detector
Responsivity [A/W]
or Transmittance
Filter
Atmosphere
0.5
0
0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 1.1 1.2
Wavelength [mm]
To confirm visually that the calculation is correct, plot the calculated values
graphically and inspect the graphs very carefully. For example, Figure 10.6
displays filter and detector spectral responses that were calculated and the
atmospheric transmittance loaded from file; all are plotted to confirm that
no error was made.
When you code your problem in a computer language, ensure that you
copy and update the code accurately. It happens too often that code is
copied but the variables are not updated in the new context. The better
solution is to use functions for repeating calculations — the benefit is that
a change made once will apply to all use cases, and varying data can be
clearly defined as function parameters.
Keep track of the value of constants’ exponents and the e-notation in
the computer scientific format: 10e4 is 10 × 104 = 105 and not 104 =
1 × 104 .
Bibliography
[1] ISO, “Quantities and units – Part 1: General,” Standard ISO 80000-
1:2009, International Organization for Standardization (2009).
[2] Cohen, E., Cvitas, T., Frey, J., Holmström, B., Kuchitsu, K., Marquardt,
R., Mills, I., Pavese, F., Quack, M., Stohner, J., Strauss, H., Takami, M.,
and Thor, A., Quantities, Units and Symbols in Physical Chemistry , 3rd
Ed., IUPAC Green Book, IUPAC & RSC Publishing, Cambridge, UK
(2008).
xvii
xviii Nomenclature
375
376 Appendix A
Table A.4 Relative spectral efficiency for photopic and scotopic vision. 4,5
Parameter Material
Ge InSb GaAs InP
Bandgap Eg [eV] 0.742 0.235 1.519 1.4236
Refraction index 4 3.3 4 3.1
Varshni parameter A [eV/K] 0.48 × 10−3 0.5405 × 10−3 0.32 × 10−3 0.363 × 10−3
Varshni parameter B [K] 235 170 204 162
Electron mobility [cm2 /(V·s)] at 300 K ≤3900 ≤7.7 × 104 ≤8500 ≤5400
Hole mobility [cm2 /(V·s)] at 300 K ≤1900 ≤850 ≤400 ≤200
Electron lifetime τe [s] 10−3 10−10 5 × 10−9 10−8
Hole lifetime τh [s] 10−3 10−6 2.5 × 10−7 10−6
Electron effective mass me∗ /m0 0.22 0.0135 0.067 0.0795
Hole effective mass m∗h /m0 0.33 0.43 0.45
Lattice constant a0 [Å] at 300 K 5.6557 6.4794 5.65325 5.8697
Lattice constant temperature coefficient da0 /dT [Å/K] 3.48 × 10−5 3.88 × 10−5 2.79 × 10−5
Electron diffusion constant [cm2 /s] ≤100 ≤2 103 ≤200 ≤130
Hole diffusion constant [cm2 /s] ≤50 ≤22 ≤10 ≤5
379
380
Table A.6 Infrared detector materials (2). 9 10
Parameter Material
Si Ge InSb HgCdTe GaAs
(x=0.22)
Bandgap Eg at 300 K [eV] 1.107 0.67 0.163 0.102 1.35
Bandgap Eg temp. coeff. dE0 /dT [eV/K] −2.3 × 10−4 −3.7 × 10−4 −2.8 × 10−4 +3.0 × 10−4 −5 × 10−4
Electron mobility μe [cm2 /(V·s)] 1900 3800 78000 3 × 105 at 77 K 8800
Hole mobility μh [cm2 /(V·s)] 500 1820 750 1 × 103 at 77 K 400
Electron mobility coeff. dμe /dT [cm2 /(V·s·K)] -2.6 -1.66 -1.6 -1
Hole mobility coeff. dμh /dT [cm2 /(V·s·K)] -2.3 -2.33 -2.1 -2.1
Electron diffusion constant [cm2 /s] 35 100 220
Hole diffusion constant [cm2 /s] 12.5 50 10
Intrinsic carrier concentration [cm−3 ] 1.38 × 1010 2.5 × 1013 2 × 1013 2 × 106
Electron effective mass me∗ /m0 1.1 0.55
Hole effective mass m∗h /m0 0.56 0.37
Refraction index near bandgap 3.42 4 4 3.65
Lattice constant a0 [Å] at 300 K 5.43072 5.65754 6.47877 6.47 5.65315
Absorption coefficient @ Eg αλc [m−1 ] 8 × 104
Absorption coefficient α0 [m−1 eV−1/2 ] 1.9 × 106
Bandgap EΓg at 0 K [eV] from Piprek 10 4.34 0.8893 0.235 1.519
Appendix A
Varshni parameter AΓ[eV/K] from Piprek 10 0.391 × 10−3 0.6842 × 10−3 0.32 × 10−3 0.5405 × 10−3
Γ
Varshni parameter B [K] from Piprek 10 125 398 170 204
Reference Information 381
0.9
0.52 mm (green)
0.8 0.53 mm
0.54 mm
0.51 mm
Monochromatic colors
0.7 0.55 mm (lime)
0.56 mm
0.6 Color coordinates in
unshaded area are valid
0.57 mm
0.50 mm
0.5 0.58 mm(yellow)
y
0.48 mm
(purple)
0.1
0.47 mm Color coordinates in
(blue) shaded area are invalid
0.4 mm (violet)
0
0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9
x
Figure A.1 CIE xy color chart. See Wikipedia 6 for a color rendition.
382 Appendix A
H2O
H2O
H2O
CO2
Rural aerosol
Transmittance
Path 1 km
Visibility 20 km
27 °C
0.5 Visibility 5 km 75% RH
Sea level
Visibility 1 km
0.0
0.1 1.0 10 [mm]
Transmittance with different aerosol types
1.0
Urban 5-km vis, 0.5-mm particles
Transmittance
Path 200 m
Naval aerosol 0.6-km vis, 4-m/s wind 27 °C
0.5 85% RH
Fog 0.5-km vis, 3-mm particles Sea level
Path 1 km
90% RH
T = 5 °C, H2O = 6 g/m3 19 g/m3
0.5 Sea level
T = 24 °C, H2O = 19 g/m3 No aerosol
41 g/m3
T = 38 °C, H2O = 41 g/m3
0.0
0.1 1.0 10 [mm]
Path 10 km
10 000 m
75% RH
0.5 3 000 m 27 °C
Rural aero
1 000 m 23 km vis
0.0
0.1 1.0 10 [mm]
Visible Near IR SWIR MWIR LWIR
0.4 0.7 0.7 1.5 1.5 2.5 3 5 8 12 mm
Bibliography
[1] Mohr, P. J., Taylor, B. N., and Newell, D. B., “CODATA recommended
values of the fundamental physical constants: 2010,” Rev. Mod.
Phys. 84(4), 1527–1605 (2012) [doi: 10.1103/RevModPhys.84.1527].
[4] Colour & Vision Research Laboratory, “Colour and Vision Database,”
http://www.cvrl.org/index.htm.
[7] Levinshtein, M., Rumyantsev, S., and Shur, M., Handbook Series for
Semiconductors Parameters, World Scientific (1996).
[8] Vurgaftmann, I., Meyer, J. R., and Ram-Mohan, L. R., “Band Parame-
ters for III-V Compound Semiconductors and their Alloys,” Journal of
Applied Physics 89(11), 5815–5875 (2001).
Symbols
α, see absorptance
α, see absorption attenuation coefficient
β, see optical thickness
γ, see attenuation coefficient
Δf, see noise equivalent bandwidth
ϵ, see emissivity
η, see quantum efficiency
ηa, ηb, see image fill efficiency
ηs, see scanning efficiency
λ, see wavelength
λc, see cutoff wavelength
ν, see frequency, optical
, see wavenumber
ρ, see reflectance
σ, see scattering attenuation coefficient
σ, see surface roughness
σe, see Stefan–Boltzmann constant
σq, see Stefan–Boltzmann constant
τ, see transmittance
Φ, see flux
ψ, see sun geometry factor
ω, see solid angle, geometric
Ω, see solid angle, projected
Ωr, see field of regard
Cυ, see contrast threshold
D, see pupil diameter
aberrations 232–235
astigmatism 232
chromatic 232
comatic/coma 232
distortion 235
field curvature 235
spherical 232
absolute humidity 123
absorptance
attenuation coefficient 99 110
detector 242
Kirchhoff’s law 69
material property 27
absorption coefficient
direct transition materials 177
extrinsic semiconductor 178 183
free-carrier 177
indirect transition materials 177
intrinsic semiconductor 178 183
refractive index 176
spectral 177
typical curves 178
Urbach tail 177
advanced model, see lifecycle phases
aerosols 112–116
atmospheric transmittance 113
land 112
manmade 112
maritime 112
meteorological range 127
aerosols (Cont.)
Mie scattering 116
Rayleigh scattering 115
scattering attenuation coefficient 128
afocal optics 236 237
aliasing 146 391 396–398
angle
factor, see spatial view factor
linear 27
solid 28–35
aperture stop 222 232
approximation
bandgap 139
BRDF 81
grey body 285
layered atmosphere 104
Planck law 64
responsivity 415
scattering 114
scotopic efficiency spectral shape 46
Seebeck coefficient 160
solid angle 33 437 441
thin lens 221 225–227
time bandwidth 150
transmittance 108
area
clear aperture 242
dimensional analysis 367
elemental 19
estimation of a flame 288–290 352
example calculations 316 441–447
pixel footprint 269
area (Cont.)
projected 28–35
solid angle 28–35 366
spatial integral 407–409
sun 316
areance, see irradiance and exitance
aspheric lens 237
assumption management 11
atmosphere 108–128
absolute humidity 123
aerosols 112
attenuation 108 110
composition 108
contrast transmittance 124–127
definitions 109
effect on image 268–272
effective transmittance 107
looking up/down 121
meteorological range 127
Mie scattering 116
molecular
absorption 111–112
constituents 111
transmittance 113
overview 110
path radiance 118–121 283
LWIR band 120
MWIR band 119
NIR band 118
visual band 118
radiative transfer codes 129
Rayleigh scattering 115
atmosphere (Cont.)
relative humidity 123
scattering 112 127
scattering modes 114
sky radiance 283 398–401
standard profiles 109
transmittance 113 382
water vapor content 121
windows 116
LWIR band 117
MWIR band 117
NIR band 117
visual band 116
attenuation
atmosphere 108
coefficient 98–99
avalanche detector 198
background 256
background-limited operation 147 183 192
205 211
baffle 223
band-limited noise 142
bandgap 138–139
Varshni approximation 139
bandwidth
– 3 dB 262
Butterworth filter 263
noise equivalent 262
best practices 365–373
cavity 57 74
emissivity 74
reflectance 74
chief ray 224 230
cloud model case study
measurement 297
model 298–300
relative signature contributions 300
silver-lining factor 298
®
worked example in Matlab 451
clutter 256
CODATA constants, see constants
cold finger 337
cold shield 338
design 342
efficiency 341 342
collimator 238–239
color
coordinates 48–51
worked example Python™ 430
normalization 48
Planckian locus 49
Ratio 291 398–401
sensitivity to source spectrum 49
space, CIE 1931 48
xy chart, CIE 381
coma 234
complex lens, see thick lens
concept study, see lifecycle phases
conduction band 168
conductors 170
energy bands 171
detection (Cont.)
pulse 272–275
pulse example calculation 436
range 267–268 326
range example calculation 326–327
detectivity 147–149 258
specific 148 183 184
205 258
detector
avalanche 198
conductivity 188
configurations 140
cooling 183–187
gas/liquid cryogen 185
radiative 185
thermo-electric 185
cutoff wavelength 138
detection process 136–140
detectivity, see detectivity
dewar 185
effective responsivity 148
filter 338
history 135–136
intrinsic material 173
material parameters 379–380
noise 140–150 183
normalized spectral responsivity 140 243
peak responsivity 140 243
performance modeling 207–210
photoconductive 179 187–193
detector (Cont.)
photon 138–140
detection process 179–183
quantum efficiency 181–183
photovoltaic 179 193–207
preamplifier gain 243
signal voltage 243
spectral responsivity 182 243
technology impact 210–212
thermal 136–138 151–163
wideband responsivity 261
detector-limited operation 205 206
development
optronic sensor systems 385–386
parallel activities 7
phase, see lifecycle phases
product 4
development model, see also lifecycle
phases
dewar 185
difference
contrast 271–272
noise equivalent temperature (NETD) 247 259 332–333
operator 19
diffuse
reflectance 76 81
example visual spectra 50
Phong BRDF model 82
signature components 279–283
reflectance, Phong BRDF 82
shape factor, see spatial view factor
efficiency (Cont.)
photopic 47
quantum 139
relative luminous 46
human eye 47 378
scanning 330
scotopic 47
spectral shape approximation 46
solar cell 318
Einstein equation 180
electrical frequency 141
electro-optical system
analysis
example 309–364
pyradi toolkit 411
definition 14
examples 15
functions 221
high-level design 15
major components 14
modeling and simulation 16
multispectral 40
simulation application 385–401
electromagnetic
radiation 20–22
particle model 20
wave model 20
spectrum 21
electron-hole pair 179
elemental area 19
emissivity 65–74
absorptivity 69
atmosphere 120 121
blackbody 59
cavity 74
definitions 70
directional 83–86
example curves 85
in nature 85
gas radiator source 103 310
grey body 71
Kirchhoff’s law 69
low 73
measurement 295–296
path radiance 101–103
practical estimation 287–288 344–355
spectral 71
hemispherical 84
temporal variation 390
thermally transparent paint 301 302
energy bands 165–170
bandgap 165
thermal carrier excitation 183
conduction band 168
Fermi level 166 168
Fermi–Dirac distribution 166
interband transitions 174–176
intraband transitions 174 175
orbitals 165
photon-electron interactions 174–176
semiconductor 169–170
valence band 168
flame
area calculation in Matlab® 434
Bunsen burner 344–355
sensor 309–311
®
worked example Matlab 417
worked example Python™ 421
temperature measurement 295
fluctuation noise 146–147
background flux 147
signal flux 146
flux 24
collecting solid angle 230 231
Lambertian source 41
luminous 23
photon 23
radiant 23
system throughput 249
transfer 35–41 70
geometrical construction 36
lossless medium 37–38
lossy medium 38
multi-spectral 39–41
radiative transfer equation 101
worked example 448–451
f -number (f /#) 229–230
clear aperture area 242
optics diameter 242
focal
length 224
plane 223 224
folded optics 236 237
foreground 256
frequency
electrical 141
optical 20
relation to wavelength 20
response
photoconductive detector 190–191
photovoltaic detector 202–203
Fresnel reflectance 77–79
gold surface 85
full-width-half-maximum (FWHM) bandwidth 150
image (Cont.)
modulation transfer function, see
modulation transfer function
(MTF)
object appearance 311–314
object relationship 225
optical aberrations, see aberrations
pixel irradiance 268–271
pixels 268
plane 223 227 246
field stop 227
pupil 227
vignetting 227
point spread function, see point
spread function (PSF)
probability of detection 274
ray tracing 225
rendering 391–398
resolved object 268–271
simulation, see infrared scene simulation
spatial sampling, see aliasing
unresolved object 268–271
image fill efficiency 331
index of refraction 20
atmosphere 97
chromatic abberation 232
complex 78 176
Fresnel reflectance 78
imaginary component 176
metal 78
numerical aperture (NA) 229
real component 176
intensity 24
Lambertian source 42
luminous 23 46
photon 23
radiant 23 38
interface electronics noise 146
intrinsic carrier concentration 174
intrinsic detector, see photon detector
irradiance 24
apparent 244 265
in an image 268–271
see also object appearance in an image
luminous 23
noise equivalent (NEE) 246 258
see also laser rangefinder example
photon 23
pixel 268
radiant 23 37
isolators
energy bands 171
Kirchhoff’s law 69
laboratory
blackbody 59 75
collimator 238–239
Lagrange invariant 249
Lambertian source 41–42
flux, exitance, radiance 41
blackbody 41
definition 41
intensity 42
projected solid angle 42
reflectance 76 81
reflected sun radiance 87
shape 44–45
signature model 279–283
spatial view factor 43
view angle 42
laser rangefinder
detection range 326
example calculation
range equation 326–327
signal-to-noise ratio (SNR) 274
threshold-to-noise ratio (TNR) 274
Lambertian reflective surface 323–325
noise equivalent irradiance (NEE) 321
range equation case study 321–330
signal irradiance 322
specular reflective surface 327–330
medium 14
absorption attenuation coefficient 99
atmosphere 108–128
attenuation coefficient 99
conducting 78
discrete ordinates 104
equivalent path length 99
homogeneous 98
index of refraction 20
inhomogeneous 99 104–105
lossless 37–38
lossy 38
optical 98–104
optical thickness 103
path radiance 99–103
scattering attenuation coefficient 99
transmittance 38 98 108
medium-wave infrared (MWIR) 65
atmospheric aerosol scattering 127–128
atmospheric window 117
contrast transmittance 125
path radiance 119
mesopic vision 46
meteorological range 127
microbolometer 156–157
Mie scattering 116
minimum detectable temperature (MDT) 259
minimum resolvable temperature (MRT) 259
model 12
atmospheric 129
BRDF, see bidirectional reflection
distribution function (BRDF)
model (Cont.)
cloud 297–300
detector 207–210
example 208
discrete ordinates 104
electromagnetic wave 20
imaging sensor 240–245 337–344
light 22
photon particle 22
photovoltaic detectors circuit 200
signature 279–283
solar cell 319–321
solar irradiance 86
source–medium–sensor 14
validation 275
modeling and simulation (M&S) 7 16 385–401
Modtran™
description 129
meteorological range 127
visibility 127
modulation transfer function (MTF) 236 260
multi-spectral 39–41
near-infrared (NIR) 65
atmospheric window 117
path radiance 118
Phong BRDF parameters 285
noise 245–247 256
bolometer 157
considerations in imaging systems 146
noise (Cont.)
equivalent
bandwidth 149–150 262
exitance (NEM) 247 259
irradiance (NEE) 246 258
power (NEP) 147–149 246 258
radiance (NEL) 247 258
reflectance (NER) 259
target contrast (NETC) 335–337
temperature difference (NETD) 247 259 332–333
1/ f 145
fluctuation 146–147
generation–recombination (g-r) 144–145
interface electronics 146
Johnson 142–143
Nyquist, see Johnson noise
photoconductive detectors 191–193
photovoltaic detectors 203–207
physical processes 140
power spectral density 141–142
pyroelectric detector 159
shot 143–144
system 141
temperature-fluctuation 145–146
thermal, see Johnson noise
thermoelectric detector 161
time-bandwidth product 150
normalization 261–263
color coordinates 48
effective value 261–262
peak 262
normalization (Cont.)
spatial 29 261
weighted mapping 263
normalized spectral responsivity 243
n-type material 171
electron concentration 174
numerical aperture (NA) 229 230
object
appearance in an image 311–314
worked example Python™ 424
resolved 268
unresolved 268
open-circuit operation 198 200
optics 223–236
aberrations 232–235
aperture 226
aspheric lens 237
axis 224
chief ray 224 230
collimator 238
conjugates 224
elements 222–224
field
angle 224
stop 226 230
flux collecting 230
f -number 229 230
focal
length 224
plane 223 224
optics (Cont.)
frequency 20
infinite conjugates 224 229
marginal ray 224 229 230
medium 97–104
modulation transfer function (MTF) 236
numerical aperture (NA) 229 230
point spread function (PSF) 235
power 223
principal plane 224
pupil 226–230
ray tracing 225
signature 279–292
model 279–283
rendering 387
spectral filter 240
stray light 227
system 236
afocal 236
Cassegrain 236
Gregorian 236
refractive 236
thick lens 225
thickness 103
thin-lens approximation 224 225
transfer function (OTF) 236 260
vignetting 227 238
Optronics System Simulation (OSSIM) 393
orbitals 165
photon (Cont.)
detector 138–140
noise, see noise
operation 179
quantum efficiency 139
responsivity 139
electon interactions 174
energy 22
wave packet 22
photopic
efficacy 47
efficiency 47
luminance 47
relative spectral efficiency 378
vision 46
photovoltaic detector 179 193
background flux 204
background-limited operation 205
bias configurations 197–202
circuit model 200
open-circuit 200–202
reverse 198–200
short-circuit 202
construction 194
depletion region 194
detector-limited operation
open-circuit mode 206–207
short-circuit mode 205–206
diffusion current 197 204
energy diagrams 195
frequency response 202–203
I-V curve 196–197
Planck (Cont.)
summation approximation 64
radiator 59 65
Planckian locus 49
plume 103
effective transmittance 106
surface radiator 104
volume radiator 104
p-n diode, see photovoltaic detector
p-n junction 194
point spread function (PSF) 235 260
point target 232
Poisson statistics 144
power spectral density (PSD) 141–142
1/ f noise 142 145
band-limited noise 142
combining spectra 149
generation–recombination (g-r) noise 145
Johnson noise 143
shot noise 143
temperature-fluctuation noise 145
white noise 142
principal plane 224
probability of detection 259
probability of false detection 260
prototype, see lifecycle phases
p-type material 172
hole concentration 174
pulse detection 272–275
®
calculation in Matlab 436
calculation in Python™ 436
false alarm rate 272–275
pupil 226–230
diameter 229 230
pyradi toolkit 411
pyroelectric detector 157–159
noise 159
responsivity 159
structure 158
Python™ 409
quanta 20
quantum efficiency 139 181 182
external 181
anti-reflection coatings 182
reflection 181
internal 181
photoconductive detector 187
quantum well detector (QWIP), see
photon detector
radiance (Cont.)
reflected
ambient 283
sky 283
solar 283
self-emitted 281
signature model 279–283
spatial invariance 36
transfer 35–41
transmitted background 283
radiative transfer equation (RTE) 97 101
radiator
gaseous 70 103
grey body 71
Planck 59 65
selective 71
surface 104
thermal 285–292
volume 104
radiometer measurements
atmospheric correction 267
spectral radiance 287
radiometric quantities 375
radiometry 22
definition xxv
nomenclature 23
quantities 24
techniques 255–276
range equation 267
solved in Python™ 435
ray
chief 224
marginal 224
tracing 225
Rayleigh scattering 115
reductionism xxiii
reflectance
bidirectional 76 80–83
cavity 74
diffuse 76 81 82
directional 75–83
in nature 85
Fresnel 77–79
geometry 77
high 73
Lambertian 76 81
material property 27
mirror 81
Snell’s law 77 79 405
specular 76 82
refractive index, see index of refraction
relative humidity (RH) 123
relative luminous efficiency 46
photopic 46
scotopic 46
rendering 387–398
aliasing 391 396–398
rasterization 391
priority fill algorithm 391
side-effects 393
z-buffering 391
super-sampling 392 396–398
requirement allocation 8
response
eye 46
filter 223 240
frequency 150
complex valued optical 236
photoconductive detector 190–191
photovoltaic detector 202–203
impulse 235 260
normalizing 140
spatial frequency 260
spectral weighting 106 244 263–264
system 246
thermal detector 136
unlimited 161
responsivity
bolometer 156
normalized 140 243
peak 140 243
photoconductive detector 189
photon detector 139
pyroelectric detector 159
spectral 140 243
thermal detector 136 152–154
thermoelectric detector 161
reverse-bias operation 198
reverse-bias-saturation current 197 319
review, see design
root-mean-square (rms) 257
scattering
atmosphere 112
aerosols 112
attenuation coefficient 99 110
Mie 116 134
Rayleigh 115 134
scattering modes 114
Schrödinger equation 169
scotopic
efficacy 47
efficiency 47
luminance 47
relative spectral efficiency 378
vision 46
Seebeck coefficient 160
selective radiator, see gaseous radiator
semiconductors
current flow 179
carrier diffusion 179
carrier drift 180
charge mobility 180
diffusion constant 180
diffusion current 180
diffusion current density 180
drift current 180
drift current density 180
energy bands 171
structure 169–170
extrinsic materials 171
concentrations 173
examples 174
Fermi energy level 173
semiconductors (Cont.)
Fermi–Dirac distributions 173
intrinsic materials 171
concentrations 173
examples 173
Fermi energy level 173
intrinsic carrier concentration 174
light absorption 176–178
material parameters 379–380
Schrödinger equation 169
silicon lattice 172
wave equation 176
sensor 14
aperture stop 222
field stop 223
noise model 330–334
optical
elements 222
model 240
throughput 248–250
®
optimization worked example Matlab 459
radiometric model 242–245 337–344
complex source 245
detector signal 242
source area variations 244
signal calculations 242–245
complex source 245
detector 242
source area variations 244
solid angle
field of view 230
flux-collecting 230
sensor (Cont.)
spatial angles 230
spectral
filter 223
response 223 243
stops/baffle 223
terminology 221–223
window 222
worked example 450
sharing xxiii
short-circuit operation 198 202
short-wave infrared (SWIR) 65
shot noise 143–144
interface electronics 146
photovoltaic detectors 203 204
power spectral density 143
signal 256
reference planes 245
electronics plane 246
image plane 246
object plane 246
optics plane 246
voltage 243
signal-to-clutter ratio (SCR) 257
signal-to-noise ratio (SNR) 257
signature
model 279–283
atmospheric path radiance 283
BRDF 284
equation 281
main contributors 280
reflected ambient radiance 283
signature (Cont.)
reflected sky radiance 283
reflected solar radiance 283
self-emitted radiance 281
spatial properties 279
terminology 282
thermal radiator 285–292
transmitted background radiance 283
reflected vs emitted contribution 283
rendering 387
thermal radiation from common
objects 65
silicon detector 139
simulation 385–401
knowledge management 386
validation 386
sky radiance 283 398–401
Snell’s law 79 176 403–405
solar cell analysis 315–321
configuration 318
experimental measurement 315
model 319–321
radiometry 317
solid angles 316
source areas 316
solid angle 28–35
approximation 33
worked example Matlab® 441
Cassegrain telescope example 448
field of view 230
flux collecting 230
geometric 28
spectral (Cont.)
emissivity 71
measurement 288
filter 223 240
filter function 413–415
®
in Matlab 413
in Python™ 414
filtering 39
integral 407
integration, summation 26
mismatch 264
quantities 25
conversion 26
density 25
response
eye 46
filter 240
photon and thermal detectors 138
sensor 223 246
responsivity 243
weighting 106 244 263–264
spectroradiometer 287
specular reflectance 76 82
Stefan–Boltzmann law 63
stopband 240
subsystem 2
sun 86
area 316
geometry factor 87
glint 302
reflected radiance 283 398–401
surface radiator 104
surface roughness 75
scale 76
system 2
acceptance, see lifecycle phases
context 1
engineering 2
noise 245–247
performance measures 255–261
segment, see subsystem
source–medium–sensor model 242–245
V-chart 8
target
extended 268
point 268
technical performance measure (TPM) 10
telescope
Cassegrain 236
Gregorian 236
temperature
apparent 73
cross-over 300
estimation of a flame 290–292
minimum detectable 259
minimum resolvable 259
noncontact measurement 73
radiation 73
temperature-fluctuation noise 145–146
flux 146
power spectral density 145
transmittance (Cont.)
homogeneous medium 98
inhomogeneous medium 99
material property 27
medium 38 98
range 108
two-flux Kubelka–Munk 100
wave model
electronic 166
Bloch functions 170
field strength 176
velocity 176
wave equation 176
light 20
wave packet 22
wavefront 20
wavelength 20
cutoff 138
relation to frequency 20
relation to wavenumber 26
spectral density conversion 26
wavenumber 25
website xxv 411
white noise 142
white point 49 92
Wien’s displacement law 62–63 67