Fundamentals of Instrumentation and Measurement
Fundamentals of Instrumentation and Measurement
Edited by
Dominique Placko
First published in France in 2000 by Hermès Science Publications in two volumes entitled
“Mesure et Instrumentation”
Published in Great Britain and the United States in 2007 by ISTE Ltd
Apart from any fair dealing for the purposes of research or private study, or criticism or
review, as permitted under the Copyright, Designs and Patents Act 1988, this publication may
only be reproduced, stored or transmitted, in any form or by any means, with the prior
permission in writing of the publishers, or in the case of reprographic reproduction in
accordance with the terms and licenses issued by the CLA. Enquiries concerning reproduction
outside these terms should be sent to the publishers at the undermentioned address:
ISTE Ltd ISTE USA
6 Fitzroy Square 4308 Patrice Road
London W1T 5DX Newport Beach, CA 92663
UK USA
www.iste.co.uk
The rights of Dominique Placko to be identified as the author of this work have been asserted
by him in accordance with the Copyright, Designs and Patents Act 1988.
Printed and bound in Great Britain by Antony Rowe Ltd, Chippenham, Wiltshire.
Table of Contents
Introduction. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xvii
Index . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 531
This page intentionally left blank
Introduction
Instrumentation:
Where Knowledge and Reality Meet
instrument of measurement most important to the Greeks was the gnomon, or needle
of a sundial. The gnomon helped the Greek mathematician Euclid, living in the 3rd
century BC, to measure the earth’s radius by simultaneously observing the shadow
cast by the instrument on two points of the same parallel. After this discovery,
developments in mathematics, numerical theory and geometry followed, with
Euclid’s ideas dominating the world of science up until the Renaissance. From the
16th century onwards, Galileo, Newton, and Descartes brought forward new
approaches that were truly objective, which meant that all new scientific theories
had to be verified by observation and experiment. It was in this era that scientific
instruments began to be widely developed and used.
The example we will discuss here will show, without forgetting Euclid’s
contribution as cited above, how instrumentation helped to join knowledge and
reality. In the 18th century, both maritime navigation security and the possibility of
complete world exploration were limited by current imprecision in measuring the
coordinates of a ship traveling anywhere on Earth. The problem of calculating
latitude already had been resolved some time before, thanks to fairly simple
geometric measurements and calculations. Determining longitude presented more
problems. As soon as a relation was established between the idea of time and space,
scientists, especially astronomers, proposed using the movement of the stars as a
cosmic clock: one example was the rotation of Saturn’s satellites, discovered by the
French astronomer Jean-Dominique Cassini in 1669. However, developing this idea
further proved difficult and complicated. Determining longitude by relying on a
measurement of time difference in relation to a given location required a precise
measurement of time that was impossible to attain with the tools then available. To
give an idea of the order of magnitude, let us recall that at the Equator, a nautical
mile is defined as the length of a terrestrial curve intercepting an angle of a minute.
The time zone being equivalent to 15 degrees, the lapse of time of a minute equals
15 minutes of curve or 15 nautical miles. Thus a nautical mile is equal to 4 seconds.
The problem was resolved in 1759 by the English clockmaker John Harrison,
who invented a remarkable time-measuring instrument, a sea clock or chronometer
that was only 5 seconds off after 6 weeks at sea, the equivalent of just 1.25 nautical
miles. This revolutionary clock marked an important step in the search for precision
begun in 1581 with Galileo’s discovery of the properties of regularity in a swaying
pendulum, a principle taken up and developed further in 1657 by the Dutch
physician Christiaan Huygens, inventor of the pendulum clock. John Harrison’s
invention produced a number of other technological innovations such as ball
bearings, which reduced friction that caused imprecision and errors. His
chronometer stimulated progress in a number of other fields, among them
cartography, leading to clearer, more geographically accurate maps. Today the
Global Positioning System (GPS) stills depends on time measurement, but with a
Introduction xix
margin of error of less than several centimeters, thanks to atomic clocks with a
margin of error that never exceeds that of a second every 3 million years!
These kinds of remarkable discoveries became more frequent over time in all
scientific and technological fields, often resulting in new units of measurement
named after their inventors. Instead of the inexact and often anthropomorphic
systems then in use, it became necessary to create a coherent system of measurement
that could be verified by specific instruments and methods from which reproducible
and universal results could be obtained. An example of one older unit of
measurement was the “rope of 13 knots” used by European cathedral builders to
specify angles of 30, 60 and 90 degrees. Other measurements long in use such as the
foot and the inch obviously could not meet the criterion of reproducibility but did
allow for the emergence of standards and the development of somewhat more
regular measurements. The usage of these often varied from region to region,
becoming more widespread over time. The ell, for example, differed not only
according to place but also according to usage. The first tentative step toward a
coherent system was clearly the British Imperial System, adopted in 1824 by Great
Britain and its colonies. The SI, an abbreviation for the International System of
Measurements today in use throughout much of the world, dates from 1960 and
allows scientists to join all measurements in use to a group of specific and carefully
chosen basic measurements, thus giving birth to a new field of science that could not
exist without modern measurement: metrology.
have scientists who can invest time in multi-disciplinary research; the teams
themselves must also serve as conduits between research teams belonging to
complimentary disciplines. This form of interdisciplinary activity, in which research
teams are able to imagine and work out applications of their work beyond their own
fields, is an extremely attractive challenge. But will this necessarily lead to
innovative concepts – and if so, according to which scientific principles?
The reality is that of the vast range of solutions widely available to resolve any
problem of measurement, very few are actually suitable. The emergence of an
innovative and optimum system often appears as the result of an ingenious
combination of a group of methods and technologies drawing on diverse disciplines.
This approach does not necessarily mean a major development has occurred in each
of the involved fields; it does, however, require in-depth knowledge of these fields.
The innovation resulting from this mastery is not less rich, open and dynamic in
terms of scientific, technological and economic terms, resulting as it does from
interdisciplinary exchange.
To give a well-known example of this theme, we look at the car, an object that
has paradoxically retained the same function over decades even as it has never
stopped changing and evolving. We are all aware of how new technologies,
especially in the fields of micro-electronics and industrial computer science, have
changed cars. We notice the continual appearance of new scientific concepts whose
names and acronyms (such as the Antilock Braking System (ABS), the Enhanced
Traction System (ETS) and controller area network (CAN) operating system)
become familiar through widespread publicity and advertising of vehicles. In fact,
the car as a symbol has become more interesting and inspiring than functions such as
airbags or digital motor control which often make use of new, though hidden,
technologies. These technologies usually develop within widely varying constraints
such as safety, reliability, ease with which problems can be diagnosed and repairs
can be made, and cost. Such technologies also are affected by marketing factors like
style and comfort. The car is thus an illustration of an impressive technological
Introduction xxi
expansion that has taken place within the parameters of science and within the
parameters of socio-economics.
This book has been written for technicians, industrial engineers, undergraduate
students in the fields of electronics, electrical engineering, automation, and more
generally those in disciplines related to engineering science who require in-depth
knowledge of how systems of measurement are developed and applied. The chapters
follow a fairly linear progression. However, our text falls into two complementary
but somewhat different halves.
The first half of the book discusses fundamental ideas and issues of measurement
and presents a range of physical phenomena that allow us to obtain measurable sizes
and develop methods of pretreatment of signals. In these early chapters, our
discussion of instrumentation focuses mainly on components. The second half of the
book concentrates instead on the aspect of systems by looking at how data are
processed and used. These two different emphases are linked in Chapter 6, which
presents the carrying out of integrated functions, showing how microtechnologies
have shown great promise in the fields of sensors and instrumentation.
Using the example of the car, the first chapter defines the links between
instrumentation, measurement and metrology, explaining how units and tools of
measurement are developed. Chapter 2 presents the general principles of sensors,
while Chapter 3 gives a detailed description of the general principles of optical,
thermal and mechanical sensors, and how these may be used in developing
measuring tools and sensors. Chapters 4 to 6 discuss a range of methods and
technologies that allow for a complete measuring process, from the conception of an
electronic conditioning of signals, passage through discrete time, data conversion
and quantification, filtering and numerical pretreatment.
Measurement Instrumentation
Whether exploring Mars, measuring the brain’s electrical signals for diagnostic
purposes or setting up robots on an assembly line, measurement is everywhere. In all
human activities, the idea of measurement establishes a relationship between a
natural or artificial phenomenon and a group of symbols, usually numbers, in order
to create the most reliable representation possible. This representation is classified
according to an “orderly” scale of values.
In the short term, this perpetuation guarantees the quality of products and
commercial trade by connecting them to legal standards. Achieved through
instrumentation, measurement is thus the basis of progress in many forms of
knowledge, as well as being essential to production and trade. In the world of
science, it allows us to make discoveries and confirm them. In terms of technology,
instrumentation helps us control, improve and develop production, and in the world
of economics, it makes commercial exchange possible, helping us assign value to
objects and transactions.
Let us look at several examples from history regarding the measurement of time.
The priest-astronomers of ancient Egypt were close observers of natural phenomena,
especially the sky. Simply by observing the natural effects of solstices (including the
floodings and harvests around the Nile coinciding with the rising of the star Sirius)
they were able to invent a 365-day calendar. Their observations also enabled them to
Measurement Instrumentation 3
1 All definitions found in the text in italics come from the International Vocabulary of Basic
and General Terms in Metrology.
Measurement Instrumentation 5
Measurement
Quantified Phenomena
prediction modeling
Quantification Phenomena
by measurement observations
Universal
constants Advances in
knowledge
Metrology Instrumentation
Standards of Standards
measurement
1.5. Instrumentation
Measurement object
under feedback control
B(t)
Return chain
Does a taxonomy of instruments exist [WHI 87]? To the best of our knowledge,
a universal classification of instruments has not yet been proposed.2 The difficulties
of even proposing such a classification are obvious, given that such an attempt
would come up against problems of criteria choice. Would criteria have to be chosen
according to the technologies being used or application fields?3
2 Two excellent books [ASC 87]; [FRA 96] and an article [WHI 87] all dealing with sensors
are exceptions.
3 For the definition of this term, see section 1.7.1.
8 Fundamentals of Instrumentation and Measurement
However, other classification criteria are possible. Tables 1.4 and 1.5 (see also
Appendix 1) give further examples of classification criteria in terms of the nature of
the physical stimulus used and the physical quantity being measured.
Table 1.1. Examples of instrument classification criteria and related application fields
Measurement Instrumentation 9
The concept of the systems approach [ROS 75] is generally used in industrial
design. Looking at the example of the car, it is possible to use this comprehensive
approach to create a subset of instruments in this field. For purposes of brevity, we
can say that the instruments necessary for a vehicle are centered around the driver
and his or her needs [EST 95]. Driving a vehicle through traffic involves
cooperation – and a certain amount of tactical skill. Planning the details of even a
short car trip involves planning an itinerary, departure time and other details, all
requiring strategic skill. Moreover, learning how to drive a car and ensuring its
optimal and safe performance involves operational skills. A useful definition of
instruments in this context would involve a classification by groups of functions,
one flexible enough to accommodate technological changes and cost reduction.
Passenger
Temperature String system
Chassis functions compartment and
function functions
safety functions
From simple sensors and their conditioners to computer data acquisition systems,
instruments must furnish reliable and accurate measurements. We can attempt to
formalize data acquisition chains by using a global model to design an instrumental
system to define different components and measures in use. Modeling an instrument
of measurement depends on quantifiable and qualifiable knowledge of parameters –
but these parameters cannot always be controlled. We can, however, attempt to
Measurement Instrumentation 11
The output phase of a measurement system delivers an “image” value S(t) of the
characteristic being measured. Ideally, this value would be a faithful representation
of the quantity to be determined, with the input and output linked by a characteristic
transfer function of the measurement system or instrument.
These definitions identify the difference between real value M(t) and measured
value S(t). Metrology is a method used to rigorously analyze these differences. The
role of the user is then to critically analyze the results using a thorough knowledge,
by quantifying or qualifying influence quantities so as to estimate any possible
errors they may cause [HOF 83]; [NEU 89].
We can, in certain cases, estimate and deduce errors that occur between
measuring systems and the object to be measured. The measurand can then be
achieved but may not be completely accurate; in such cases we must ensure that
appropriate metrological procedures are followed. In other cases, measurement
cannot be carried out, and being aware of this will help us find another solution to
determining a quantity of interest.
If X(t) is the “true” value of the quantity to be measured when the object of
measurement is not related to the measurement device, then M(t) stands for the
Measurement Instrumentation 13
value of the quantity after measurement. The information conveyed from the object
to be measured to the instrument of measurement represents an image of a
measurand X(t) upon which we superimpose information intrinsic to the energy of
the connection, expressed as X*(t). This energy transfer is a characteristic of
measurement and means that the measured object delivers not only quantity M(t) to
the instrument of measurement but also a second quantity M*(t) (see Figure 1.4).
M
Measurand effort Measurement
X
variable
M* load effect
X* flow
variable
instrument
object to be
or
measured
measurement system
Figure 1.4. Load effect from contact of the object to be measured
with a measurement system
This load effect can be described in terms of energy, a concept fundamental to all
elements of all physical interactions, no matter what the quantity may be. In
engineering sciences, we describe these interactions in terms of pairs of
complementary variables whose product is equal to the power. We will further
discuss the definition and role of these pairs.
P = M(t).M*(t)
W = ∫M(t).M*(t).dt
that of energy.
14 Fundamentals of Instrumentation and Measurement
From the point of view of physics, one of these two variables is extensive: the
flow variable, for example, current, speed, density flow or angular speed. The other
is intensive and is a potential or effort variable: for example, tension, force or
pressure. Sources of flow variables (current or speed) operate in constant flow and
have infinite impedance. Information transfer follows the pair of complementary
variables producing power or the “energy flow” that comes from interaction between
the variables. In all pairs of variables found in classical physics such as electricity,
mechanics, hydraulics and optics, we can define a size as equal to a power or form
of energy:
In both linear and non-linear examples, the course taken by a flow variable or
effort variable expresses the energetic limits of the system and determines an
optimal operating point.
Measurement Instrumentation 15
operating limits
effort source
X
max
operating limits
charge feature
flow source
R
X* X*
max
Both the source feature and the load feature share one or several points of
intersection (see Figures 1.6 and 1.7). These are operating points, determining the
variable values and the load that permits their connection. From an energetic point
of view, two conditions must be met:
– the continuity of shared flow variables;
– the continuity of shared effort variables.
16 Fundamentals of Instrumentation and Measurement
Effort variable
M
These conditions characterize operating points that are stable (see Figure 1.6) if
the source and load features are simple (that is, linear, quadratic, logarithmic and so
on) or unstable (see Figure 1.7) if the features are complex (curved, convex).
Determining the load effect (see Figure 1.4) makes use of the concepts of
impedance and generalized admittance. In non-linear cases, the derivative in relation
to the flow variable can be used but we will not discuss these cases here. We define
the concept of impedance as a relation between intervening quantities in a power
exchange. It is a specific transfer function of the system. The relation of the
derivatives intersecting the associated variables is the determining factor:
effort variable
M
We use this concept in cases when the measurand is an effort variable. Going
beyond equations that fit the model given in Figure 1.4, we here define two forms of
generalized impedance that apply to cases of a power transfer system and an energy
transfer. These are generalized resistance and generalized rigidity4 shown in Table 1.3.
Energy transfer
Power transfer X(t).X*(t)
∫ X(t).X*(t)dt
Generalized
Input Associated Associated Generalized impedance
impedance
variable variable variable Z = Var.Ext/∫Var.Int dt
Z = Var.Ext/Var.Int
Stress
Flow variable Generalized resistance Flow variable Generalized rigidity
variable
X*(t) R = X/X*(t) X*(t) S = X/∫X*(t)dt
X(t)
Electrical
Tension U U/I Electrical charge
current U/∫Qdt
[V] [Ω] Q [C]
I [A]
Force f Speed f/v Displacement d
f/∫d dt
[N] v [m/s] [N/ms-1] [m]
Pressure P Flow volume P/D volume
P/∫D dt
[N/m2] D [m3/s] [Nm/rad/s] V [m3]
Table 1.3. Examples of interactions between effort variables, flow variables and
corresponding generalized impedances in the case of the measuring object (X,X*)
4 The terms resistance and rigidity come from electronics and mechanics terminologies.
18 Fundamentals of Instrumentation and Measurement
M(t) = X(t)-R.X*(t)
This gives us four possible cases according to which we consider that the
quantity to be measured X(t) can be an effort variable or a flow variable. When the
measurement system is connected then X*(t) ≠ 0; if it is not, then X*(t) = 0.
Generally, X(t) depends on X*(t) and the relation between them differs according to
whether they are viewed as the object to be measured or the measurement
instrument. This is most simply and most often expressed as a linear relation.
To estimate the load effect, we write the equation linking the exact value X(t)
and M(t) as:
where R and Rm are the generalized resistances respectively of the measured object
and of the measurement system, with S and Sm being both the generalized rigidity
of the measured object and the measurement system, respectively.
Analysis of the load effect: the pair of linked variables is made up of an effort
variable X(t) that is the tension limit of the battery u(t) and an accompanying flow
variable X*(t) that is the current i(t). The resulting measurement, made by a
voltmeter, would ideally be M(t) = E.
u(t) = E – Rg . i(t)
Rg M(t) = S(t)
E X(t) = u(t) Rv V
Figure 1.8. Contact of an instrument with a measured object and associated variables
E = u(t)(1 + Rg/Rv)
20 Fundamentals of Instrumentation and Measurement
The load effect is thus represented by the term (Rg/Rv). If we want u(t) to be
equal to E, voltmeter resistance must show high resistance to the internal resistance
of the battery; this is in fact the usual result.
The load effect is thus intrinsic to all measurement operations and as such is
inevitable. We might want to think that the load effect is unimportant in view of the
fact that it is responsible for easily tolerated errors. However, this is not always the
case, and practical solutions (such as experimental precautions) and theoretical
solutions (data treatment) are necessary to properly understand and analyze results.
Linearity
Hysteresis
Static Resolution
characteristics
Drift
Instrument
Time constants
Frequencies
As we know the static and dynamic transfer characteristics of all the elements of
a system, we can then combine them to obtain a description of the entire system,
inferring these characteristics from the partial characteristics of various components
described by their transfer functions.
There is no absolute rule for combining static parameters; each case requires
different procedures. Often, contributions of most parameters may be negligible
except those corresponding to a specific element. For dynamic characteristics, we
use transfer functions.
/
T’(p) = T(p) (1 + B T(p))
describe the necessary elements for building the correct instrumental chain for a
given situation, we will discuss ways to approach measurement and the
methodology of measurement.
From these two classes, we deduce three measurement principles, one direct, two
indirect:
– the principle of direct measurement: here, the measured object serves as an
energy vector, carrying information to the measuring system;
– comparison method of measurement: in this indirect system, energy is carried
by an external auxiliary system;
– substitution method of measurement: this system has conditions of the two
above systems.
If no model exists, the need for rigor remains, with the operator experimenting
until finding, identifying and naming a model. Measurement results can lead to
finding a preliminary model that can be improved later by successive approaches.
This section will analyze a measurement situation, again using the example of
the car. Suppose we want to measure the axle load of heavy trucks and similar
vehicles and lighter cars. Using a measurement instrument involves the following
field constraints that may effect measurement acquisition:
– road sensors that must set up and not be interrupted;
– climatic conditions such as ice or heat, among others;
– pressure resistance and other mechanical constraints;
– influence quantities that may be hard to control, such as variations in the main
sensors due to the heat of the road, for example.
This study, carried out on a highway under the control of the Danish National
Road Association (DNRA), has shown conclusively reliable results. With older
systems, the sensitivity of sensors to lateral pressure resulted in a “phantom” axle
count, resulting in an overestimation of the number of vehicles. This example shows
the role of technology and its influence in defining and implementing a
measurement system.
There are two main approaches to error reduction. These are experimental
solutions, such as the use of impedance matching and noise reduction, and statistical
solutions or signal analyses such as error calculation, data analysis, spectral analysis
and others.
Experimental solutions are physical in nature and are closely related to influence
variables and load effects. They are implemented in design before measurement
occurs. Statistical solutions are mathematical in nature. They are part of the analysis
and correction of results and are carried out after measurement has taken place, since
a measurement must be made before it can be corrected. In the section that follows,
we discuss different bases for data analysis and their definitions.
n
∑ ( xi − x )2
σ = i −1
n −1
xi being the result of the ith measurement and x being the arithmetic mean of the n
results considered.
Description of results
To be rigorous, a measurement result must contain the most probable value. The
uncertainty range will include this probable value and the probability associated
with both this uncertainty range and the unit being used.
The above definitions are applicable to all fields such as mechanics, hydraulics
and biology. However, in this chapter, and in the rest of this book, we will discuss
only measurement chains based on electronic technologies. In scientific
measurement and industrial measurement, the observation, interpretation and control
are increasingly carried out by electronic instrumentation. The very important
development of microcomputers has meant that a range of separate devices
dedicated to one or several functions became part of the instrumentation process. A
few of these are the voltmeter, the frequency meter and the oscilloscope. This
Measurement Instrumentation 29
Analog signals
Continuous signals are static or vary slowly. With these signals, the level or
amplification of the signal at any instant constitutes the information. While sensors
(such as the thermocouple) may be used for measurement of these signals, an
analog-to-digital converter (ADC) converts the signal into digital data. The
precision, resolution, reliable pass band, and good synchronization of the ADC
ensure parameters essential to continuous analog signal acquisition.
There are many kinds of temporal signals, such as ECG waves and temperature.
Here, information is carried in the form of waves (amplitude and variation in time).
Temporal signal acquisition means using the largest pass band possible and a precise
time base (to avoid sampling problems), ensuring transfer speed, as well as
beginning and ending the measurement sequence correctly.
Signals
Input/output cards
Acquisition RS 232
and control IEEE 488
VXI
Data analysis
Signal analysis
Analysis Numerical filter
Statistics
Mathematical analysis
Introduction Scrolling graphs
Graphs
Databases
Histograms
Graphic interface
Digital signals
Binary signals and pulse trains: to carry out accurate measurements of numerical
signals, an instrument must generate binary signals (an example is the start-stop
mechanism of machines) and pulse trains (as with sequencing clocks and
synchronization). These measurements are carried out by means of a counting
function.
These classes of signals are not mutually exclusive. A signal can contain more
than one type of information. This is especially true with transient and permanent
states of second order oscillatory systems that transmit signals through microwave
lines. Instruments to measure these signals range from the simple (logic-state
detectors for TOR signals) to the complex (frequency analyzers for spectrum
analyses).
sensors. This means a pressure sensor can act as a surveillance device that gives
instructions to an actuator; it can also be part of a network of several sensors
supervised by a main microcomputer;
– instruments configured around a microcontroller. Another important field of
electronic instrumentation includes measurement and control chains configured
around an intelligent circuit. This pertains to many systems dedicated to a specific
application characterized by reduced size, autonomy and reduced cost. In everyday
living, this type of instrument plays a role in machines such as cars and household
appliances, mostly because of their low cost, portability or small size;
– programmable electronic instruments. These are groups of instruments that
have been configured to carry out customized functions according to the operator’s
needs. They are directed by computer instrumentation software.
Sensor Signal
1 conditioner
M U L I T P L E X E R
Sampler-
Sensor Signal
blocker Systems
2 conditioner
management
NAC and data
acquisition
Sensor Signal
n conditioner
The growth of these systems since the end of the 1980s has occurred because
instrument manufacturers, using SCPI norms, have developed standards to ensure
product quality, and because personal computers have improved and become more
economical.
Computer
Software
Signal
conditioner
The growing use of electronics in the automotive industry has given rise to the
need for operational monotension amplifiers that can function with a supply tension
of +5 V. The example given here has been taken from an application note published
by Texas Instruments. In this note, a piezoelectric pressure sensor interfaces with a
circuit integrated with operational amplifiers to detect rattling in an internal
combustion engine.
Limiting load effects: in order to limit the load effect between the sensor and the
conditioner, a resistor with a high level of power is inserted between the two
devices. The resistor ensures that polarization currents flow from the sensor. The
operational amplifier must have a high driving point impedance to be adapted to the
sensor and very weak polarization currents. These steps require a very precise signal
conditioning.
34 Fundamentals of Instrumentation and Measurement
5V
C4
Rp
V1 1kΩ
+ R5
R3
C A1
- +
220nF C1 C3 A2
R
C2 R1 -
V
R4
1MΩ R2
C2
Piezoelectric sensor Charge effect limitation Signal amplification Pass band filter
charge: q = V.C Gain = 5 Bandpass: 2. 89 kHz
= 21 pC/g C = 300 pF
Band pass: Central frequency = 6.8 kHz
Vi = 24 mV/g R = 1 MΩ
530 Hz – 28 kHz
V = 35 mV/g
C = 600 pF
Functioning principles (see Figure 1.13): when rattling begins, the sensor
produces a series of signals with a frequency that does not occur when the motor is
functioning properly. An operational amplifier (OA) amplifies these signals. The
frequency of the rattling varies according to the size of the cylinders and block
cinder material. A broadband sensor and a very flexible operational amplifier must
be used to be adaptable to all vehicles. Most sensors have band pass of several tens
of kHz, meaning that one type of sensor is suitable for many applications.
Throughout the world, the label “ISO 9000” has become a point of reference for
companies wishing to maintain, guarantee and record the quality of their products
through quality control [DEA 91]; [HOF 83]. These businesses need measurement
instruments and standardized tests based on national norms.
Both the manufacturer of products to be standardized and certified and the client,
who may have access to a laboratory with standardization capabilities, must conform
to certain norms, depending on the quantities involved.
Measurement Instrumentation 35
National institute
Legal document
national laboratory
To give an example, the sensors installed on a line can be calibrated with the
supplier, but some must be calibrated in situ. Measurement is thus a key element in
quality assurance. Figure 1.14 illustrates the positions occupied by the different
levels of instruments of measurement and control in quality assessment processes in
organizations.
1.15. Conclusion
1.16. Appendix
Transduction type
Physical Chemical Biological
Elastomagnetic Electrochemical process Test effect on an organism
Electromagnetic Spectroscopy Spectroscopy
Magnetoelectric Physicochemical transformation Biophysical transformation
Photoelastic Chemical transformation Biochemical transformation
Photoelectric Photochemical transformation
Photomagnetic
Thermoelastic
Thermoelectric
Thermomagnetic
Thermooptic
Table 1.4. Examples of possible instrument classification according to transduction type
Characteristics Definitions
Zero offset Zero offset is true relation of the zero output variable with
the value of the measurand.
Drift Temporal variations in system characteristics.
Dynamic Admissible intervals of variation for input variables (in
decibels).
Hysteresis Maximum difference in output values, when the input
variable is reached from minimum, then maximum
admissible in algebraic value.
Linearity Degree of concordance between the static state diagram and
a straight line used as reference. (A straight line of the
fewest squares calculated on calibration points, the line
joining the farthest points throughout the measurement.)
Relaxation Time lag between the cause and effect of a physical
phenomenon, given in the form of a time constant.
Repeatability Margin of fluctuation in output variable when the same input
variable is applied several times under the same conditions.
Resolution Smallest increase in the input variable leading to a change in
the output variable.
Sensitivity Ratio of change in output variables to the corresponding
change in input variables.
Threshold Threshold resolution is the smallest change of the input
variable relative to zero value.
Response time For a measurable excitation, this is the time required for an
immediate value and a final value to be lower than a
specified value (1%, for example).
1.17. Bibliography
[ASC 87] ASCH G. et al., Les capteurs en instrumentation industrielle, Dunod, 1991.
[ATK 87] ATKINSON J.K., “Communication protocols in instrumentation”, J. Phys Sci.
Instr., 20, p. 484-491, 1987.
[BOI 89] BOISSEAU J.F., Méthodologie de la mesure, Techniques de l’Ingénieur, Mesures et
contrôle, R140, 1989.
[CER 90] CERR M. et al., Instrumentation Industrielle, Techniques et Documentation,
Lavoisier, 1990.
38 Fundamentals of Instrumentation and Measurement
[COM 92] COMMIOT D., “Les miracles de la mesure”, Usine nouvelle, 2373, p. 10-14,
1992.
[DEA 91] DEAN, “Measurement, quality and trade”, Meas. Sci. Technol., 2, 403-404, 1991.
[DRA 83] DRAHEIM H., “Measurement as Science and Technology”, Measurement, 1,
p. 68-74, 1983.
[EST 95] ESTEVE D., COUSTRE A., GARAJEDAGUI M., L’intégration des systèmes
électroniques dans l’automobile du XXI siècle, ouvrage collectif, CEPADUES Editions,
1995.
[FIN 82] FINKELSTEIN L., “Theory and Philosophy of measurement”, in Handbook of
Measurement Science, Wiley Interscience Publications, 1982.
[FRA 96] FRADEN J., Handbook of Modern Sensors, 2nd Edition, Collection Engineering/
electronics, Springer-Verlag, 1996.
[GIA 89] GIACOMO P., Etalons métrologiques fondamentaux, Techniques de l’Ingénieur,
Mesures et contrôle, R50, p. 1-16, 1989.
[HEW 90] HELLWIG H., “The Importance of Measurement in Technology-Based
Competition”, IEEE Trans. on Instr. and Meas., 39/5, p. 685-688, 1990.
[HIM 98] HIMBERT M., “La métrologie: un langage universel pour la science et la
technologie”, in Récents progrès en génie des procédés, vol. 12, Lavoisier Tech & Doc,
Paris, 1998.
[HOF 83] HOFMANN D., “Modeling of errors in measurement”, Measurement, 1, 3,
p. 125-128, 1983.
[JAC 90] JACOMY B., Une Histoire des Techniques, Collection Points, Editions du Seuil,
1990.
[LAF 89] LAFAYE P., Unités de mesure, Techniques de l’Ingénieur, Mesures et contrôle,
R23, 1-13, 1989.
[MAS 90] MASI C.G., “So, You Want to Measure Submicron Dimensions”, Test & Meas.
World, p. 59-68, 1990.
[NAC 90] NACHTIGAL C.L., Instrumentation and Control, Wiley Interscience Publications,
1990.
[NAD 98] NADI M., “Rôle et évolution des systèmes de mesure électroniques”, in Récents
progrès en Génie des procédés, vol. 12, Lavoisier Tech & Doc, Paris, 1998.
[NAD 99] NADI M., “La mesure et l’instrumentation dans la recherche scientifique et dans
l’industrie”, Revue de l’électricité et de l’électronique, no. 3, p. 38-42, March 1999.
[NEU 89] NEUILLY M., Incertitudes de mesure et tolérances, Techniques de l’Ingénieur,
Mesures et contrôle, R280, 1-21, 1989.
[PAR 87] PARATTE P.A., ROBERT P., Traité d’électronique et d’électricité, Systèmes de
Mesure, Dunod, 1987.
Measurement Instrumentation 39
[PRI 89] PRIEL M., Incertitudes de mesure et tolérances, Techniques de l’Ingénieur, Mesures
et contrôle, R285, 1-11, 1989.
[PRI 95] PRIEUR G., NADI M. et al., La mesure et l'instrumentation, collection Mesures
Physiques, Masson, Paris, 1995.
[ROM 89] ROMANI L., Structure des grandeurs physiques, Blanchard Ed., Paris, 1989.
[RON 88] RONAN C., Histoire Mondiale des Sciences, Collection Points, Editions du Seuil,
1988.
[ROS 75] DE ROSNAY J., Le Macroscope, Collection Points, Editions du Seuil, 1975.
[STE 93] STENKON N., “Le standard VXI, un concept d’instrumentation électronique
programmable”, Industronique, 5, p. 54-56, 1993.
[TRA 91] TRAN TIEN LANG, Systèmes de mesures informatisés, Masson, 1991.
[TRA 92] TRAN TIEN LANG, Electronique des systèmes de mesures, collection Mesures
Physiques, Masson, 1992.
[TUR 90] TURGEL R.S., Instrumentation Everywhere, Feb. 1990, Vol. 39, no. 1, p. 1, 1990.
[WHI 87] WHITE R.M., “A sensor classification scheme”, IEEE Trans. Ultrason. Ferroelec.
Freq. Contr. Vol. UFFC-34, no. 2, p. 124-126, March 1987.
[WHI 93] WHITE R.M., “Competitive measures”, IEEE Spectrum, 30, no. 4, p. 29-33, 1993.
This page intentionally left blank
Chapter 2
Sensors are the first components of a measurement chain. They convert the
physical and chemical variables of a process or an installation into electrical signals
that almost always begin as analogical signals. This conversion must also mirror as
closely as possible the involved variables. Only a thorough knowledge of sensor
responses guarantees success; sometimes sensors produce faulty signals due to
interference, the conditions of use, or often because of the processes themselves.
We begin this chapter by discussing some of the basic principles of sensors and
how they work [NOR 99]. These principles are based on calibration, evaluation of
uncertainties, calculation of response time, and conditioning. Our aim is to provide
the reader with a fairly general guide. Some relevant equations and formulae, as well
as many issues relating to instrumentation and signal analysis, will be discussed in
later chapters.
The quantity to be measured being the measurand, which we call m, the sensor
must convert m into an electrical variable called s.
s = S m + s0 [2.1]
F’(m) = constant = S
Of course, we can always define a range of values in which S is constant: that is,
when the sensor is linear.
General Principles of Sensors 43
In this section we will not present methods measuring uncertainty. Here, we only
discuss some general ideas to guide the reader through later chapters where
uncertainty evaluation will be discussed on more detail, particularly in Chapter 10.
Measurement uncertainty is the difference between the true value of the measurand
and the measurement carried out by the sensor. The only known measurands are the
standards with values that have been determined by convention. It is important to
44 Fundamentals of Instrumentation and Measurement
distinguish between systematic errors and random uncertainties: they occur for
different reasons and have very different consequences for measurement.
Systematic errors are always due to faulty sensor knowledge or utilization. This
kind of error is detected by comparing the mean values of the same measurand,
given by two different sensors. The most frequent causes of systematic errors are:
– incorrect or non-existent calibration due to an aging or altered sensor;
– incorrect usage: some examples include failure to reach steady state, a defect
or error of one of the conditioner parts, or a modification of the measurand by the
sensor itself;
– inadequate data processing: examples are error in linearization in the
measurement chain or amplifier saturation in the measurement chain.
We can know the cause of random uncertainties without being able to anticipate
the measurement value. Their evaluation is always statistical. They are due to the
presence of signals or interference; the amplitude of these is random and they are
called, rather vaguely, “noise”. To cite a few examples:
– fluctuating supply sources in the measurement chain or in the conditioner
(such as fluctuation of the electromotive force in a bridge);
– electromagnetic signals produced in an environment and picked up by a sensor
element, conditioner or measurement chain;
– thermal fluctuation, including thermal turbulence of current carriers;
– fluctuation in influence variables, etc.
There are many other causes of random uncertainties, such as reading errors,
defects in sensor mobility and hysteresis. Unlike systematic errors, random errors
can never be completely avoided. We can, however, reduce them by using protection
methods, such as electrical regulations, temperature stabilization, mechanical
isolation and electromagnetic shields. In addition, filtering, synchronous detection
and signal processing can reduce random uncertainties, which must always be
evaluated carefully.
General Principles of Sensors 45
n
∑ si
i =l
s= [2.2]
n
The mean value of the difference si − s is used to statistically analyze the random
measurement of m. The variance v and the standard deviation σ must be introduced:
n
∑ ( si - s )2
i =l
v= [2.3]
n -1
Σ( si − s )2 ( )
n ⎡⎢ s 2 − ( s ) ⎤⎥
⎣
2
⎦
σ= = [2.4]
n −1 n −1
Suppose that the result of 10 temperature measurements are (in °C): 60.1; 60.6;
59.8; 58.7; 60.5; 59.9; 60.0; 61.2; 60.2; 60.2.
46 Fundamentals of Instrumentation and Measurement
First of all, we present the results with one number after the decimal point. This
shows that the uncertainty of these measurements is at best 0.1°C. We also find that
the mean temperature is 60.11°C, and that the standard deviation σ equals 0.63°C.
The mean value of 60.11°C does not mean that it is close to the true value of the
measurand. A systematic error, perhaps in the form of defective measurement
standards, could have produced it, possibly through a deviation of 1°C on all the
values and therefore on the mean value. However, if only the systematic error is
large in relation to the random uncertainty and the sensor is non-linear, the deviation
type does not change after correcting the systematic error. The list of ten values will
not be reproduced if we undertake another measurement, but this second list will
still have something in common with the first. Looking again at the first list, we see
that deviations between 0°C and 0.2°C occur seven times. There is one value that
deviates more than 1°C from the mean value. Therefore, we can classify the
obtained values by their occurrence probabilities. After many measurements have
been carried out on the same measurand, if the uncertainty is truly random, we can
demonstrate that this occurrence probability is a law called the Gaussian
distribution. It expresses, according to the mean value and the deviation type, the
probability density of finding the value s of a measurement. With the Gaussian
distribution, the basic probability of finding the range of s through ds is given as dp
= p(s) ds (where p(s) is called probability density) is given as:
1 (s − s )2
dp = p (s) = e− ds [2.5]
2π σ 2σ 2
+∞ 1 (s − s )2
∫−∞
2π σ
e−
2σ 2
ds = 1 [2.6]
σ.P(σ)
0.4
s-s
-3σ -σ σ 3σ
s +3σ
For s − s = ±3σ , the integral of ∫s −3σ P ( s )ds ≥ 0.99 which means more than
99% of all measurements lead to a value of s expressed as (s – s ) ≤ 3σ.
59.78 ≤ T ≤ 60.44
59.63 ≤ T ≤ 60.26
In this instance, we use statistical results to make our decision. For example,
Chauvenet’s criteria specifies that a non-random phenomenon has modified the
measurand if the probability of obtaining the measurand, calculated with the help of
the Gaussian distribution, is less than 1/2 n, with n being the number of
48 Fundamentals of Instrumentation and Measurement
measurements made up to the point when the anomaly appeared. We can tabulate
the results of this criterion by giving the deviation limit dmax to the mean value;
beyond this the criterion applies (see Table 2.1).
A way to use this criterion is to look at the threshold settings of an alarm system.
Suppose that a presence detection has an analogical output of 0 V without any
environmental interference. We can carry out 500 measurements and find a
deviation type of 265 mV. If one measurement has a value superior to
265 × 3.29 = 872 mV, something has modified the measurand – there has been some
interference. We must then set the threshold of the system to the value of 872 mV,
or set a rule: V < 872 mV = no interference, V > 872 mV = interference.
10 1.96
25 2.33
50 2.57
100 2.81
500 3.29
1,000 3.48
Table 2.1. Chauvenet’s criteria for a Gaussian distribution. Deviation from the mean value
beyond which the engineer must consider if a measurand has been modified
a)
m m
b)
m m
d)
c)
m m
m m
Figure 2.3. Reliability, accuracy, and precision. The dotted lines indicate the
true value (µ). In a) the sensor is neither accurate nor reliable; in b) the sensor
is reliable but not accurate; in c) the sensor is accurate but not reliable;
in d) the sensor is accurate and reliable (from G. Asch [ASC 91])
All industrialized countries have sets of standards. This means they have
laboratories organized for specific purposes, established over time, that link
measurements to basic standards [ASC 91]. Standards and transfer instruments
assure traceability and successive stages from laboratories to industry. In France, for
example, the National Bureau of Metrology is in charge of insuring traceability
according to national standards. Throughout this linking process, successive
operations are carried out within a standardization process that not only joins
measurements to measurands but defines uncertainty levels in sensor measurements.
As far as the legal aspect of some of these operations is concerned, it is important to
remember that processes certified by ISO 9000 standards also require the
traceability of all functioning sensors.
1 ∞ jω t
h (t ) = ∫ S ( ω )e dω [2.7]
2π 0
Lack of knowledge about this response can lead to systematic errors, even when
carrying out stationary measurements.
When using sensors, the idea of band pass can be introduced through a discussion
of distortion phenomena observed during measurement. If the measurand has a
periodic temporal evolution described by Figure 2.4 that can be represented by:
m (t) s(t)
t t
If the values S( 1) are different, or if i is related in some way to θi, a signal s(t)
is obtained with a frequency content that changes in relation to the frequency
content of the measurand. In such cases, we say the signal has undergone a
distortion or that the system is dynamically non-linear.
The band pass is the frequency interval with a value of S( ) that is constant and
in which i differs from θi by a constant additive that can be written as ik, with k
independent of i. Such systems are dynamically non-linear.
Generally, a sensor order is the order of the differential equation that governs its
dynamic sensitivity. The simplest example is that in which an equation linking s to
m in dynamic state is a first order differential equation:
m1 1
s = [2.11]
B ⎡ ⎛ f ⎞2 ⎤
⎢1 + ⎜ f ⎟ ⎥
⎢⎣ ⎝ c ⎠ ⎥⎦
ψ = – arctg ⎛⎜ f ⎞⎟ [2.12]
⎝ fc ⎠
δ(dB)
-3
-20
0.1 1 10 f/fc
ψ(degree)
0°
-5°
-45°
-85°
f/fc
-90°
0.1 1 10
I=kΦ CL RL VL
IC IL
dVL VL
I = IC + I L = CL + = kφ [2.13]
dt RL
and also:
CL dVL 1
φ= + VL [2.14]
k dt kRL
We can also establish that the sensitivity S0 is equal to kRL, which means the
product (gain × band pass) is constant. Again, this result is quite general. We see
that usually large band pass and strong gains do not exist together in most sensor
measurement chains.
When the measurand is a mechanical variable [HAN 99], we often find second-
order sensors. With these sensors, the equation joining the measure s to the
measurand m is of this type:
d 2s ds
A +B +C = m [2.15]
2 dt
dt
1 C
f0 = [2.16]
2π A
and
B
ξ= [2.17]
2 CA
fc/Hz
107
105
103
R/Ω
s 1
=
m1 2
⎛ ⎛ f ⎞2 ⎞ ⎛ f ⎞
2
[2.18]
C ⎜ 1 − ⎜ ⎟ ⎟ + 4ξ 2 ⎜ ⎟
⎜ ⎝ fo ⎠ ⎟ ⎝ fo ⎠
⎝ ⎠
and
⎡ ⎤
⎢ ⎥
⎢ ⎥
−2ξ
Ψ = arctg ⎢ ⎥ [2.19]
⎢ ⎛ ⎛ ⎞ ⎞ ⎥⎥
2
⎢ fo ⎜1 − ⎜ f ⎟ ⎟
⎢ f ⎜ ⎝ fo ⎠ ⎟⎠ ⎥⎦
⎣ ⎝
Dynamic sensitivity can show a resonance with weak damping and damping
beyond the frequency fo equals 40 dB per decade. The phase lags are doubled
compared to those of first-order sensors.
ξ<√2
⏐s/m1⏐
critical damping
1 ξ=√2
10-1 ξ>√2
10- 2
slope -2
f/f0
phase ξ (d°)
0°
ξ>√2
ξ<√2
-90°
ξ=√2
-180°
f/f0
10-2 10-1 1 10 102
The response time of sensors can also be deduced from the differential equations
presented above. For a first-order sensor, if m = 0 for t < 0 and m = m0 for t > 0, we
get the solution:
A
s = so (1 − e −t / τ ) with τ =
B [2.20]
and
m0
s0 =
B [2.21]
1
Here is sometimes called static sensitivity.
B
General Principles of Sensors 57
s0 = m0/B
0.9 s0
τ 2τ 3τ t
– if ξ<1/ 2 ,
1−ξ 2 t 1−ξ 2t
s = k1e −ξωot e jωo + k2 e−ξωot e− jωot [2.23]
– if ξ=1/ 2 ,
s = k1 (1 + ωo t )e −ω o t [2.24]
58 Fundamentals of Instrumentation and Measurement
ds
If we take as initial conditions s = 0 and = 0 with t = 0 , we get the following
complete solutions: dt
mo ⎡ exp(−ξω t ) ⎤
ξ < 1/ 2 ⇒ s (t ) = ⎢1 − o
sin ⎡( 1-ξ 2 )ωo t + ψ ⎤ ⎥
C ⎢ ⎣ ⎦⎥
⎣ 1−ξ 2 ⎦
mo
ξ = 1/ 2 ⇒ s (t ) = [1 − (1 + ωot ) exp − ωot ]
C
⎡ +ξ + ξ2 − 1
exp ⎡ −ξ + ξ 2 − 1 ⎤ ωo t
mo
ξ > 1/ 2 ⇒ s (t ) = ⎢−
C ⎢ 2 ξ −1
2 ⎣⎢ ⎦⎥ [2.25]
⎣
+ξ − ξ2 − 1 ⎤
exp ⎡ −ξ − ξ2 − 1 ⎤ ωo t ⎥
2 ξ2 − 1 ⎣⎢ ⎦⎥ ⎥
⎦
s (t)
2m 0 /C
ξ= 0
m 0 /C ξ= 0.2
ξ= 1/ √2
ξ= 10
0
2π ω0t
Second-order sensors are usually built with a damping = 0.6 (see Figure 2.12).
With this the steady state m0/C to almost 10% is reached for a period equal to
General Principles of Sensors 59
2.4/ 0. This value must be taken into account to be certain that the sensor conforms
to the calibrations set by the manufacturer.
Passive sensors convert the measurand into an impedance variable. They must
always be used with a circuit that has a source current or tension and, generally,
several additional impedances. The circuit is called the conditioner. There are two
main groups of impedance conditioners [BAU 61]. In the first group, the sensor’s
impedance variation is converted by a variation of potential difference. In the second
group, the impedance variation is used to modify an oscillator frequency. In this
case, the sensor reading is actually a frequency measurement. Here we will only
discuss the first group.
Zc
Vm = e [2.26]
Zc + Z K
This result can be found in the case of a simple polarization by a source current
as well.
60 Fundamentals of Instrumentation and Measurement
Zk
Zc
I vm
e
Ζc vm
On the other hand, bridge conditioners help eliminate the noise very efficiently
(see Figure 2.14).
C
Z1 Z3
Vm
e A B
Zc
Z4
D
It is easy to show that the tension Vm appearing with the measurand is increased
by ∆Vm when e changes to e + ∆e following:
or:
Z1∆ZC
∆Vm # ∆e [2.29]
( Z K + Z1 )2
General Principles of Sensors 61
For the same polarization instability, the noise generated by the measurement in
the case of a bridge and a potentiometer are in the relation:
that is, in the variation order relative to the impedance. For measurements in which
impedance varies from the order of % to around Zco, the bridge is 100 times less
sensitive to random variations of e than with potentiometric conditioners or the
direct polarization by current source.
Vm = f ( RK Rc ) [2.31]
For example, in the case of a potentiometric conditioner (or when one of the
assembly resistances is sensitive to the influence variable g) and when the
62 Fundamentals of Instrumentation and Measurement
sensitivities to g of the two resistances Rc and Rk are the same, the condition dVm = 0
is equivalent to:
∂Vm ∂V
=− m [2.34]
∂Rk ∂RC
∂Vm Rk
= [2.35]
∂Rc ( R + R )2
c k
and:
∂Vm − Rc
= [2.36]
∂Rk ( R + R )2
c k
influence of g
Rk
e
influence of g Rc
Vm
With resistive bridges (see Figure 2.14), we see that sensitivity to influence
variables is also minimal when:
R1 = RCO = R3 = R4 [2.37]
Nernst’s bridge is used for sensors with an impedance that can be represented by
an impedance Zc:
Rc
Z c= [2.38]
1 + jRc C ω
For the value m0 of the measurand we adjust the following impedances of the
bridge:
Re = Rc = Rc0 [2.39]
and
Ce = Cc = Cc 0 [2.40]
e ∆Z
Vm ≈ . c [2.41]
4 Zc0
Rc R
Cc
Vm
e
C c0
R
Rc0
Re
R
Ce
Vm e
Lc
Rc
R
Zc = R c + jLc [2.42]
R2
Rc 0 = [2.43]
Re
and:
so we then get:
R.∆Z c
Vm ≈ e. [2.45]
( R + Z c 0 )2
Direct readings of active sensors are rarely satisfactory, whether or not these
sensors are equivalent to tension, currents or charge sources. This is because this
kind of reading presupposes a correction that is not always easy to evaluate.
General Principles of Sensors 65
Zi
Vm = ec [2.46]
Zi + Z c
Zc
Zi Vm
ec
The sensor can also appear in a form equivalent to a current source (ic) in parallel
with an impedance Zc. The electrical signal Vm is then given as in Figure 2.19.
im
ic Zc
Vm
Zc
vm = Zi im with im = ic [2.47]
Zi + Z c
Z i << Z c [2.48]
Lastly, with sensors that are also charge sources, it is clear that a simple
measurement by the difference potential to the resistance limits affects the signal,
since the measurement discharges the sensor.
Here we will not discuss in any detail the many schema used for signal
processing of active sensors that resolve the problems presented in this section (see
the following chapters for a more detailed presentation of these). Instead, we explain
the three basic assemblies that correspond to the three types of equivalencies found
in active sensors. These are basic to using operational amplifiers [FRA 93].
First of all, let us suppose that the sensor is equivalent to a tension source ec in
series with an impedance Zc. With the assembly shown in Figure 2.20, we can easily
see that, if we are close to the operational ideal for the amplifier, we get:
⎛ R ⎞
Vs = ⎜ 1 + 2 ⎟ Ve = GVe [2.49]
⎝ R1 ⎠
R2
-
+
Zc
R1 v RL vs
e
ec
We can see that the sensor does not produce any current (i+ = i- = 0 in an ideal
amplifier) which means that it is connected to infinite impedance. The non-influence
condition of the sensor’s internal impedance Zc is fulfilled when:
– Vs, as output, is independent of the current transmitted in the charge RL. The
tension Vs transmitted by the amplifier acts as a tension source of a zero internal
impedance;
– the choice of R1 and R2 help regulate the desired gain G.
µo
Gωc = [2.50]
τo
where µ0 is the open circuit gain and k0 is the response time of the operational
amplifier.
Suppose now that the sensor is equivalent to a current source placed in parallel
with a resistance Rc. We can then use the assembly seen in Figure 2.21.
ic
E - S
ε
+
ic Rc Ve
RL Vs
M
Figure 2.21. Assembly type for a sensor equivalent to a current source
Since the input of an ideal amplifier would not transmit current, and since input
differential tension is zero i ≈ 0, the potential difference between E and M is zero
68 Fundamentals of Instrumentation and Measurement
and no current circulates in the resistance Rc of the sensor. The output tension VS is
expressed as:
VS = −Ric [2.51]
As with amplifier tension, we will give a few basic descriptions of how basic
assembly works:
– the value chosen for the feedback resistance R does not influence a sensor
equivalent to a current source;
– input resistance is zero because source limits are maintained at the same
potential to the input of an ideal amplifier;
– output provides a source of tension whose resistance is zero (VS is independent
of the charge resistance placed at output).
RR CR
CS
dq ve -
RS +
dt
vs
Because no current can come through amplifier inputs, all charge variations
within the sensor’s limits are found within the limits of CR. Here we get:
Q
VS = − [2.52]
CR
General Principles of Sensors 69
In reality we often have to take into account the resistance to leakage of capacity
CR (RR in parallel with CR) especially always with low frequencies. It is easy to
show that VS becomes:
Q jωRR C R
VS = − . [2.53]
CR 1 + jωRR CR
This is the expression of a high pass filter that shows a converter of this type
does function properly at low frequencies and cannot transmit current.
2.7. Bibliography
There are now so many sensors available [NOR 89] that it would be impossible
to discuss the principles of all of them in a single chapter. We have therefore limited
ourselves to three classes of measurands: optical, thermal and mechanical. However,
even with this restriction, we still must limit our scope, and will only present the
most frequently used laws for these types of physical sensors.
One important class of sensors detects electromagnetic beams. Within this group,
we will restrict our discussion to those optic sensors that are sensitive only to beams
with wavelengths of 10 nm – 1 mm, that is, frequencies of between 1016 and 1011 Hz.
In the specific case of light sensors, for reasons relating to the sensitivity of the
human eye, it is necessary to introduce specific concepts when discussing visibility
(0.4 µm to 0.8 µm). After a brief recapitulation of the variables that act as
measurands for optical sensors, we will define the reference light source used in
making calibrations. We will then discuss the principles of sensors that are used in
constructing semiconductors.
S
Figure 3.1. Flux coming from a volume V across a surface S
In the specific case of a plane wave, it is known that H = E/µv where v is the
speed of the wave in the middle of the index n (v = c/n with c the light speed in
traveling in a vacuum). The flux is reduced to:
From this we can deduce that the energetic flux is proportional to the square of
the amplitude of the electric field. This relation is very commonly used by all
sensors sensitive to energy and therefore to the square of the field, that is to say that
any phase information is lost. The phase can only be retrieved by interference
phenomena such as holography or speckle.
Physical Principles of Optical, Thermal and Mechanical Sensors 73
In optical sensors with at least one capacity of visual sensitivity, measurands can
be either energetic or luminous variables. Energetic variables are measurements that
are completely independent of the sensitivity of the human eye. With luminous
variables, on the other hand, the effect of the eye is taken into account as a variable
according to the radiation wavelengths. It is important to remember that when
choosing between these two measurands, the practical application is very important.
For example, when studying photomulitplicator sensors, within the framework of
spectroscopy, it becomes clear that energetic variables are the only measurands we
need to take into account. However, if we want to measure the flux coming into a
camera producing images for the human eye, luminous variables are indispensable.
As well, measuring the light in a room or road can only be expressed in units of
luminous flux.
We have seen how energetic flux is directly related to the square of the
electromagnetic field. Luminous flux is defined with respect to retina sensations.
This kind of flux is a measurand, that is, a measurable variable, because we can
define the equality of these two fluxes (the same sensation produced by two adjacent
zones of the same shield) and the sum of several fluxes that superimpose their action
on the eye. These measurements can also be made with a photoelectric cell with a
spectral sensitivity set as close as possible to that of the human eye. No matter which
methods are used, the measurements must be carried out with monochromatic
waves, which means that the measurements must be taken in the brief interval d
around the wavelength of the measurement. These are spectral variables.
“Normal” human eyesight is not the same for everyone across the spectrum of
visual experience or for a single visual sensation. Even for one person, this sensation
varies according to psychological and physical factors. Therefore, we define
luminous efficiency of the eye by citing a large-scale statistical study carried out on
people with “normal” eyesight. The results of this study have led to a definition of
the average eye. The sensations of this standard eye, which we call the luminous
flux F , are at each wavelength proportional to the received spectral energetic flux
f . The proportionality factor k of course depends on . If this average eye receives
the spectral energetic fluxes f and f ’ (to and ’) so that the luminous spectral
fluxes Fλ and Fλ’ are equal, we express them as:
Fλ K φ
= λ λ =1 [3.3]
Fλ ' K λ 'φλ '
74 Fundamentals of Instrumentation and Measurement
Kλ
Vλ = ≤1 [3.4]
Km
F =Km V f [3.5]
The numeric value attributed to Km leads to defining the relation between the
unities of energetic flux and luminous flux. The unity of energetic flux, called the
lumen (lm), has been set since 1979 as Km = 680 lm.W-1.
Three other variables are important for optical sensors. These are intensity,
luminance and illumination. Of these, intensity is certainly the most well known
because it is the most frequently used. Its origin resides in the fact that most
luminous sources transmit fluxes that depend not only on surface points but also on
the emission angle in a normal relation to the surface. For this reason, it is necessary
to evaluate the elemental flux transmitted by an element dS of the surface around the
point 0 in a small solid angle d around a given direction x. This leads to the
definition of transmitted intensity in this direction (see Figure 3.3). The unity of
energetic intensity df/d is clearly the W.sr-1 and that of the luminous intensity is
dF/d is the candela or lm.sr-1.
Vλ
λ Vλ
1 400 410-4
450 0.038
500 0.323
555 1
600 0.63
650 0.107
700 410-3
750
10-4
0
400 500 700 nm
600 µm
0.4 0.5 0.6 0.7
Now, if we are within the range of a sensor transmitting waves in the direction x
(see Figure 3.3), we see that the visible transmission surface is no longer dS but d ,
which is the projection of dS on the plane usual to 0x. As long as dS stays small, the
luminous flux perceived in this direction is proportional to d . We then introduce
the luminance L as:
dI dI
L= = [3.6]
dΣ dS cos θ
dΣ
→
n
θ
0 x
dS dΩ
Let us look at the flux d2f transmitted by the surface dS in the direction of a
sensor, delimiting the solid angle d by the surface dS´ of the sensor.
dS cos θ
where is the solid angle d ’ by which the sensor observes the source. We
( 00' )2
then get:
The quantity dS cos d = dS’ cos ’ d ’ is called the geometric area (see Figure
3.4). This quantity is conserved in stigmatic (that is, having to do with images)
optical systems.
For most sensors, and especially optical sensors, the important measurand is the
flux received by the unity surface called E:
d 2φ
E= [3.10]
dS'
Now that we have some knowledge of measurands, the next step is learning to
calibrate optical sensors. To do this, we must be able to produce energetic and
luminous fluxes whose properties are both completely known and reproducible. This
can only be achieved through thermal radiation of black bodies. How these black
bodies react depends on the temperature of the body and the universal constants.
Physical Principles of Optical, Thermal and Mechanical Sensors 77
⎡ ⎤
Lλ = ε λ ⎢ ⎥ = ε L0 ( T )
C1
( )
λ λ [3.11]
⎢ 5 c2 / λT ⎥
λ e −1 ⎥
⎣⎢ ⎦
hc
C2 = = 1.438 10−2 m.K = 1.438 104 µ m.K [3.12]
k
2,898
M = [3.13]
T
The spectral luminance of the black body, shown in Figure 3.5 for three adjacent
temperatures of the surrounding ambient, shows that the total luminance (integrated
with the variant of 0 to +∞) is a rapidly growing function of the temperature. We get:
∞ σT 4 σ
L° =
∫0 L°λ d =
π
with
π
= 1.8 10-8W/m2 K 4 [3.14]
78 Fundamentals of Instrumentation and Measurement
Lλ
ª 200°C
≈ 100°C
ª 0°C
λ (µm)
6.1 7.8 10.6
We note here that by explaining how energetic exchanges between matter and
light occur in the form of photons, the discovery of the black body law marked the
beginning of quantum physics.
g +k + =1 [3.15]
To analyze the black body, the first idea is to look at opaque materials. An
opaque object does not transmit energy at any wavelength and for all incidences
(k = 0), which means it is a black body (i = 1) if it does not reflect energy for all
wavelengths and incidences ( = 0). To see how this works, we will look at opaque
materials.
Dielectrics are not helpful to realize black bodies because their behavior largely
depends on the wavelength. For example, white paper has very weak remote infrared
reflectance, which makes it close to being a black body beyond 6 µm (an
transmissivity of the order of 0.92). Unfortunately, it becomes a much stronger
reflector in the visible, with almost no emissions at all.
Physical Principles of Optical, Thermal and Mechanical Sensors 79
In the case of metals, k = 0 at any , but we know that the reflection factor,
independent of incidence except when it becomes low-angled, is expressed by:
4n
ρλ = [3.16]
(1 + n )2
where the index is high for a metal. The transmissivity is written as:
4n
ε λ = 1− ρλ = [3.17]
(1 + n )2
Figure 3.6 shows that the use of opaque metallic bodies or dielectric bodies does
not give the transmissivity of black bodies.
0 0°
Given these findings, it becomes necessary to look at transparent bodies and treat
them as opaque artefacts. The transmission coefficient k is written as:
k = exp( − x) [3.18]
where is the extinction coefficient and x is the distance light traveled in the
material. If we choose a dielectric with the weak reflection coefficient, we get:
i = 1 − k = 1 − exp( − x) [3.19]
80 Fundamentals of Instrumentation and Measurement
φemitted = dS ε σT 4 [3.20]
where T is the supposed uniform temperature of the cavity. This surface dS also
receives, from another element of the surface dS of the cavity, a flux i jT4 dS, with
one part shown as:
φreflected = (1 − ε ) ε dS σT 4 [3.21]
When we consider only dS, the radiative flux that comes from this term is:
φ = ε (1 − ε ) dS σT 4 + ε dS σT 4 = ε ( 2 − ε ) σT 4 dS [3.22]
Thus, we have a large transmissivity i (2-i) that, for i = 0.9, already improves
transmittance by 10%. Taking into account successive n reflections, we see that
transmissivity becomes 1-(1-i)n+1. Beyond three reflections, noticeable
transmissivity is nearly that of the black body, having become independent of the
wall transmissivity; that is, of its nature.
In addition, a black body means the internal wall temperature will be uniform.
This is realized with walls of excellent thermal conductivity (for instance, with
copper), and in perfect thermal isolation from the exterior. The exact form of the
internal cavity depends on the temperature at which the black body functions. For
example, when a circulating liquid assures temperature uniformity, the cavity
appears in the form of tubes that appear much longer than they do wide. Interior
copper walls can be covered with a strongly diffusing dielectric (with layered
painting, for example), to assure a transmissivity almost isotropic to the cavity’s
interior.
Physical Principles of Optical, Thermal and Mechanical Sensors 81
Radiation detectors are sensitive to the sum total of received radiative fluxes, that
is, to the difference between entering fluxes and exiting fluxes. This leads to a
measured variable which differs from the desired variable. Let us look at the
example given in Figure 3.7, in which a light sensor is used to measure the
transmitted flux by the object facing it. We call i the object’s transmissivity, is the
detector’s transmissivity and φbo the flux transmitted by the object to be measured.
When the detector receives φbo, part of φbr, the reflected flux is expressed as:
Furthermore, the detector itself transmits φso and a part φsr that comes back to it
after reflection on the object. The state of received and measured flux f is given as:
( )
Φ = ⎡εσ T 4 − ε (1 − ε S ) σ T 4 − ε S σ TS4 + ε S (1 − ε ) σ TS 4 ⎤ S = σεε S T 4 − TS4 S [3.25]
⎣ ⎦
We see that the measured flux f depends on the detector’s temperature Ts and
on its transmissivity is. To find the flux φbo, that is, the measurand, the
transmissivity is must be very close to 1 and that Ts << T. With light sensors, this
condition is often not met. Not taking this fact into account can sometimes lead to
significant systematic errors.
Object Sensor
Φ
Φbo
output
T Φ br TS
ΦSr
Φ So
ε εS
λ2 ⎛ d φ (λ ) ⎞
St =
∫λ
1
S (λ ) ⎜
⎝ dλ ⎠
⎟ dλ
[3.26]
λ2 ⎛ d φ (λ ) ⎞
∫ ⎜
λ1 ⎝ d λ ⎠
⎟ dλ
signal
measurand
t t
Sensor
m(t)
b(t)
t
We say that b(t) is of external origin when it has an origin coming from the near
environment (see Figure 3.8). It is then captured, either directly (an example is
radiation interference in the detector band) or indirectly by the entire electric circuit,
as with antennae. It is always possible to reduce noise of an external origin by
placing the sensor in interference-free areas, by putting them in Faraday cages or
shields, and by limiting the sensor’s interference sensitivity. When these external
noises are eliminated, we see that considerable noise levels still remain. These kinds
of noise, called ultimate or fundamental, have their source in the corpuscular nature
of electrical currents [BAU 61]. Every free charge is stimulated by random
movements, resulting in an output current variation around a mean value. The
following list gives two sources of these kinds of fundamental noises:
– thermal or Johnson’s noise is the product of the collisions of the carriers with
the lattice structure. The greater the number of these collisions, the more the mean
quadratic current increases. It appears in the passive component and grows with
their resistance. Related to thermal agitation, thermal noise increases with higher
temperatures. The equivalent schema of this kind of noise is given in Figure 3.9
where f is the band pass of the sensor and of the sensor’s electronic system where
KT ∆f
we get e1 = 2 RKT ∆f or I1 = 2 ;
R
– shot noise or Schottky noise characterizes the discrete nature of the current and
obeys Poisson’s law, which, unlike the case with Johnson’s noise, is present when
there are few carriers – that is, in charge-free zones. Shot noise appears in junctions
or in a vacuum. The mean quadratic current of this kind of noise is expressed by:
I S2 = 2q I ∆f [3.27]
84 Fundamentals of Instrumentation and Measurement
eI or II R
When the current delivered by the sensor is the sum of light and obscurity
currents, (I = I0 + IL), the smallest Schottky quadratic current is:
I S2 = 2q I ο ∆f [3.28]
These first noise sources that do not usually depend on frequency are called
“white” noises. The other sources of noise are dependent on frequency. They
decrease very quickly with lowered frequencies and are often described as 1/f noise
(which is a very rough approximation) or “pink” noise. Pink noises are products of
material defects and random recombinations of carriers on the irregularities of the
crystalline lattice structures.
For all random uncertainties, the mean quadratic noise currents have to be added
to give the total quadratic noise current I b2 and we call iB the spectral noise current
(i.e. in a band pass of 1 Hz):
I b2
iB = ( A/ Hz ) [3.29]
∆f
i
NEP = B (in W Hz -1/2 ) [3.30]
Sλ
Physical Principles of Optical, Thermal and Mechanical Sensors 85
where S is the spectral sensitivity of the sensor (in A.W-1). Some manufacturers
sometimes use detectivity D (the reverse of NEP), but because iB is generally
proportional to the root of the sensitive surface A of the sensor, the specific
detectivity D* is more often used and is expressed as:
A
D* = (in W -1cm Hz1/2 ) = D A [3.31]
NEP
For example, for a photodiode we find S = 0.6 µA/µW, A = 1 mm2, Io = 150 pA,
D* = 81012 cm Hz1/2 W-1. From this we can deduce the mean quadratic value of the
total spectral noise current:
2
⎛ A Sλ ⎞
ib2 = ⎜ ⎟ = 56 10-30 A 2 Hz −1 [3.32]
⎜ D* ⎟
⎝ ⎠
We can also obtain the mean quadratic value of the spectral current of Schottky’s
noise as I S2 = 2q I o = 4,810–30 A2Hz–1/2. From this formula we see that Schottky’s
noise is predominant in this photodiode. Apart from this kind of calculation, the user
can verify that the assembly is not adding too many complementary noise sources to
the already existing fundamental noise; in other words, he can check that the
assembly is not compromising the performance of the light sensor.
We know that the simple elements of the periodic table, peripheral electrons
remain stable when the peripheral electronic strata called s (two electrons) and p
(four electrons) are saturated. This means that the 2 + 6 electronic peripheral states
86 Fundamentals of Instrumentation and Measurement
of the element must be occupied. This result is only obtained with rare gases (He,
Ne, Ar, Kr, Xe and Ra). All other elements tend to group together, that is, they join
with the peripheral electrons so that the strata s and p are complete, like those of the
rare gases (eight electrons s and p). For example, with carbon (C, atomic number
Z = 6), there are only four peripheral electrons s and p. Four electrons are lacking
that are necessary to attain a chemical stability identical to neon (Z = 10). Carbon
thus tends to combine with the atoms which are capable of “lending” it four
electrons. This can occur through covalent band.
Atom H, which has only one electron, therefore tends to gain a second and forms
H2 to attain the stability of He. The molecule CH4 acquires the requisite structure in
bringing the electron of each atom H to the central atom C. C itself takes each of its
four peripheral electrons from H atoms placed around itself (see Figure 3.10).
As a result of the formation of this kind of covalent molecule, the energy levels
of the electrons reorganize themselves: we speak of orbital states or molecular
states.
In other cases, the grouping reaching chemical stability will release many atoms
that regroup in the form of a solid. This happens with silicon when a monocrystal is
created in a solution. Silicon (Z = 14) contains 14 electrons, of which four are
valence electrons s and p (this situation is analogous to that of carbon). Silicon tends
to group with atoms that give it the four electrons s and p it lacks to achieve valence
saturation (argon stability). In preparing a monocrystal, the atoms of Si group
together in a diamond configuration (the crystal of C of the face-centered cubic type)
as shown in Figure 3.11. We notice that this structure has the same tetraedic
arrangement of CH4, but in the monocrystal, this basic figure is reproduced to
infinity (in practice, just to the surface where, in fact, chemical stability is no longer
a given).
H
H H
The Schrodinger equation that regulates the states of the isolated atom Si is itself
modified and the eigenstates of this equation become more numerous than for the
isolated atom (Z times the number of atoms in the crystal). The energies regroup in
bands in the solid. In a band, the energy levels are very close to each other [ZIM 72].
Figure 3.12 shows that in the case of silicon, two bands appear that are separated by
a zone where there are no authorized levels.
88 Fundamentals of Instrumentation and Measurement
Silicon
crystal Atoms separated
from silicon
Electrons
Conduction
band Empty
1s2 2s2 2p6 3s2 3p2
1.09 e V Empty 3p
GAP Full
Electron energy
3s
Valence
Electron
energy
band
6 8 R(10 -1)
r 0 = 2.34% 10 -10 m Distance between
the aggregates of silicon
Figure 3.12. Modification of electron energy states occurring if the interatomic distance is
diminished. A r0 =2.34 indicates a formed solid. The energy levels are grouped in two
bands. The valence band corresponds to electrons bounded to a silicon lattice. The
conduction band corresponds to free electrons (from J.-J. Bonnet [BON 84])
Conduction
band
Eg
Fermi level
µ
f(ε)
0
Valence
band
The valence band is the band of lowest energy (that is, greatest absolute value).
The conduction band has the weakest absolute values and the band gap is the zone
without allowable states [PHI 73]. The probability of the presence of an electron at a
given energy ε is given by the Fermi function f(ε) (see Figure 3.13). This law is
temperature T dependent. At T = OK, f(ε) = 1 when ε ≤ Fermi level and |ε| = o when
ε > Fermi level. The Fermi level in pure silicon is exactly in the middle of the
bandgap when the temperature reaches absolute zero (T = OK). The valence band is
thus saturated and the conduction band is empty (see Figure 3.13).
However, other elements of the periodic table, called transition elements, in the
isolated state have many peripheral electrons that occupy states other than s and p.
These electrons cannot combine in interatomic bands when solids are created. The
crystals that come from these elements have an energy state structure in which the
bands are not separated by a gap (see Figure 3.14). The crystal has an enormous
number of free electrons and nothing stops their displacement when a potential
difference is applied to the crystal. In this case, we are describing a metal.
Conduction
band
Ev
EF
EC
Valence
band
In all solids, when the temperature increases electrons can occupy higher
energetic states (less negative) than at absolute zero. Unoccupied places then appear
in lower energy levels. The Fermi-Dirac statistics of electron’s energetic states
shows their probability P(E) occupies an energy state comprised of E and E + dE.
We get:
dn( E ) 1
P( E ) = = [3.33]
dN ( E ) ⎛ E − EF ⎞
1 + exp ⎜ ⎟
⎝ KT ⎠
90 Fundamentals of Instrumentation and Measurement
when dn is the number of electrons whose energy is comprised of E and E + dE, and
dN is the number of possible energetic states (states specific to the Schrodinger
equation) between E and E + dE. The Fermi level EF is therefore the energy for
which the probability of an electron existing at this (possibly imaginary) level is 1/2
(see Figure 3.15). Of course, at absolute zero, the Fermi level gives the highest
energy that an electron can attain (if it is authorized). In other words, all the energy
states lower than Ef are occupied and the higher states are empty.
P(E)
T = 0K
T
wf
In the case of metal, the Fermi level is an eigenstate and for this reason is the last
state occupied at absolute zero. With semiconductors and pure (or intrinsic)
insulators, the Fermi level is not a proper state but is found exactly in the middle of
the restricted band. The Fermi energy is not reached at absolute zero and the
electrons saturate the valence band. We see in Figure 3.15 that the probability of
finding electrons in the conduction band becomes greater as the temperature
increases. These electrons no longer take part in crystalline bands and become free
carriers, as in metal. This increase of the number of electrons in the conduction band
also frees an equal number of spaces in the valence band. These spaces act as
positive mobile charge carriers.
The practical uses of semiconductors derive from the fact that the position of the
Fermi level inside the gap can be moved by introducing carefully selected impurities
into the crystal [SAP 92]. When these impurities have five peripheral electrons
(donors), the Fermi level moves towards the top of the gap. The probability of
finding electrons in the conduction band increases and very few holes remain in the
valence band. This kind of semiconductor is called type N to remind us that most
current carriers are electrons. On the other hand, when impurities are elements with
only three peripheral electrons (acceptors), the Fermi level moves towards the
bottom of the gap. The probability of finding electrons in the conduction band
becomes very low, and there are many empty spaces in the valence band. In this
case, the carriers are mostly holes (semiconductor type P).
Physical Principles of Optical, Thermal and Mechanical Sensors 91
J = ρv = − ne v [3.34]
In addition, the relation between the speed v of the electron mass me and the
field E is given by the Coulomb law and the fundamental relation of the dynamic,
and is expressed by:
dv
F = −e.E = me [3.35]
dt
dv −e
= E [3.36]
dt me
The resolution of this equation leads to a speed v that linearly increases with
time:
–e
v= E t (+ constant = 0 if v = 0 at t = 0) [3.37]
me
Actually, this law is completely unrealistic because the electron (e–) undergoes many
collisions with the crystalline lattice. If we call k the mean time between two collisions,
the average velocity is limited to the value it achieves at the end of this time k:
−e
v= Eτ [3.38]
me
92 Fundamentals of Instrumentation and Measurement
e
The quantity τ is called mobility µe of e–:
me
eτ
µe = [3.39]
me
qτ
µ= [3.40]
m
where the quantity q is the absolute value of the mobile charge. The speed v is thus
an average velocity between two successive collisions and this velocity is equal to:
⎛ + : charges > 0⎞
v= ±µ E ⎜ ⎟ [3.41]
⎝ - : charges < 0 ⎠
J = + ne µ E [3.42]
J =σE [3.43]
which is the local Ohm law with j the given electric conductivity, expressed in the
case of conduction by e– by:
j = neµ [3.44]
J = J n + J p = e (nµn + p µ p ) E [3.45]
Physical Principles of Optical, Thermal and Mechanical Sensors 93
where n and p are the densities of e- and of e+ and where µn and µp are given by:
e.τ .n .
µn =
mn
[3 .46]
e.k.p.
µp =
mp
σ = e ⎡⎣ p µp +n µn ⎤⎦ [3.47]
10 6
N a, N d ≤ 1012 cm-3
10 5
4 × 10 13 T -2.20
.
.
T -2.42
.
10 4
.
1.3 × 10 17
2 × 10 17
10 3
µe
µt
10 2
1 10 10 2 10 3
Temperature (K)
Figure 3.16. Mobility variation (cm2/V s) of electrons µe and of the holes µp in silicon
according to the temperature for materials with different dopings. The mobility of e– is
represented in continuous traits, with those of e+ in discontinuous traits. The indicated
gradients correspond to the best linear adjustment of the experimental curves
wave
Ip
Surface A
L
Two phenomena occur in the semiconductor. These are: (i) carrier generation
through the photoelectric effect and (ii) carrier recombination on the crystalline
lattice.
Physical Principles of Optical, Thermal and Mechanical Sensors 95
(i) As we saw before, light must be represented in the form of photons that are
energy grains hv. If hv > Eg (the gap energy) G electrons are created per second
throughout the volume of the sensor or g electrons per second and by unit of
volume:
G 1 (1 − R) 1 (1 − R)
g= = f= f [3.48]
AL AL hν AL hc
where A.L. is the sensor volume, is the photon conversion into electrons, R is the
coefficient of optical reflection of the receiving surface, f is the flux incident of
light on the sensor, and is the wavelength of the light flux.
(ii) The electrons freed by the photoelectric effect leave in the crystal many
charged atoms that can trap them after their displacement. The variation by unit time
of the number of free electrons occurring because of this recombination is
proportional to the number of free electrons produced, as well as to the number of
charged atoms. We get:
∂n
= −r n2 [3.49]
∂t
where r is called the recombination rate. At equilibrium, there are as many electrons
created as recombined:
∂n g
= 0 = - r n 2 + g from which we get n= [3.50]
∂t r
σ = e µn n [3.51]
since the majority carriers are electrons. In the following equation, we see that by
replacing n with its expression as f (see [3.49] and [3.51]), we get:
1 η (1 − R )
σ = eµ n λΦ [3.52]
AL hc
96 Fundamentals of Instrumentation and Measurement
When the sensor is polarized by the tension V, the current going across the
sensor is equal to:
V V σAV
.
I= = = [3.53]
R L L
A
In fact, the experimental dependence of resistance with the flux f is not exactly
the same as with the model shown above. This dependence, however, is not linear
but is like the following:
The number of carriers recombined by unit time can also be expressed with the
help of the lifetime of the carriers kn (which are the e- in the examples given above):
∂n n
=− [3.55]
∂t τn
At equilibrium we get:
n
g= [3.56]
τn
V A G A V µn k n
I= = V σ = Vq µn k n = q G = FqG [3.57]
R L AL L L2
Physical Principles of Optical, Thermal and Mechanical Sensors 97
τ µ
F= n n V [3.58]
L2
F can attain several tens of thousands following the applied tension V and the
geometric form of the sensor. In addition, this equation shows that it is necessary to
use semiconductors with excellent mobility and lifetime.
R 0 Rc p
R= # R cp = a l- if R 0 >> R cp [3.59]
R0 R cp
The relation between the light current Ip and the flux f is not linear:
V V
Ip = = l [3.60]
R a
∂I p V −1
S= = φ [3.61]
∂φ a
Furthermore, the spectral sensitivity can be deduced from the above formulae:
k n µnV (1 − R )
S (λ ) = q [3.62]
2 hc
L
This is applicable up to maxof the gap. If there is an abrupt gap, the spectral
hc
sensitivity ends after ≥ max = , shown in Figure 3.18.
E
98 Fundamentals of Instrumentation and Measurement
S(λ) λmax
Response time is directly related to the lifetime of the carriers kn. Like the
lifetime, response time is connected to parameters such as the temperature and
doping of the semiconductor. If we know the nature and operating mode of the
semiconductor, we can obtain values that range from 0.1s to 10-7s. Response time is
noticeably reduced when the luminous flux is high; this is because the lifetime
decreases with the number of free carriers.
The ultimate noise of these sensors is of the Johnson noise type and its minimum
value depends on the value of the darkness current I0. This obscurity current comes
from the creation of carriers (for example, of electrons) by thermal agitation. In
calling n0 the number by unit of volume of the thermal carriers, the electric
conductivity in obscurity is j0:
j0=eµ n0 [3.63]
1 L
R0 = [3.64]
eµn n 0 A
KT
iB = 2 [3.65]
R0
1010 W-1 cmHz1/2, clearly inferior to what can be obtained with photodiodes, as we
will see in the following sections.
WC P >WC N [3.66]
that is:
VP < VN [3.68]
qVA
I ≈ I majority = I 0 exp [3.69]
kT
where q is the electron charge 1.6 10-19 C and k is the Boltzmann constant + 1.36
10-23 JK-1.
Physical Principles of Optical, Thermal and Mechanical Sensors 101
P N Energy
+ + + – + – – –
+ + + – + – – –
+ + + – + – – –
P
N
– +
We get:
qVA
I = I0 exp − I0' [3.72]
kT
I’0 = I0 [3.73]
⎡ qV ⎤
I = I0 ⎢exp A − 1⎥ [3.74]
⎣ kT ⎦
Figure 3.22 shows us that a high negative current appears when reverse bias
becomes strongly negative. This is called the breakdown voltage and is produced by
a collision of minority carriers, which take on a high kinetic energy in crossing the
barrier, ionizing the fixed centers of the crystalline lattice. This effect benefits some
photodiodes by amplifying the photo current and reducing response time of light
sensors.
We can see that the photoelectrical current acts as a current of minority carriers
and therefore is negative.
incident flux
hν
P
0+ VA
zone of the field E e-
N
– in order for the photons to penetrate the field zone in large numbers, the
incident flux must not be weakly absorbed by region P. Assume x is the thickness of
region P and g is its optical extinction coefficient at the frequency of the photons
of the monochromatic flux. The part 0 that comes to the junction after reflection on
the face in front and transmission by the doped zone P (see Figure 3.23) is equal to:
- αx
(1 − Ropt )φ 0 e [3.75]
where Ropt is the reflection coefficient of the photodiode surface. This transmitted
flux will be accordingly greater than α ⋅ x and will be weaker if the entire flux
coming into the junction is absorbed mainly in the field region;
– a field region must be designed that is sufficiently thick to allow the total
absorption of the light. This becomes possible, for example, by creating the type of
structure often called junction type PIN (P-intrinsic N).
qη (1-R opt )λ
Ir = φ0 e−α x [3.76]
hc
⎡ ⎛ qV ⎞ ⎤
I = I 0 ⎢exp ⎜ A ⎟ − 1⎥ − I r [3.77]
⎣ ⎝ kT ⎠ ⎦
or:
⎡ ⎛ qV ⎞ ⎤ qη (1- R opt )λ
I = I 0 ⎢ exp ⎜ A ⎟ − 1⎥ − φ0 e−α x [3.78]
⎣ ⎝ kT ⎠ ⎦ hc
Equation [3.78] shows that if the dark current I0 is weak with respect to Ir, I is
proportional to 0 when VA is negative.
I VL
C R
Φ Φ
E
V V
I >0
I >0
RL RL
B A B A
With the photoconductor mode, there is always a darkness current that produces
an internal noise of the Schottky type, which means that this mode is not especially
favorable for detecting very weak fluxes. On the contrary, with the photovoltaic
mode, the straight line of the charge goes through the origin I = V = 0, and the
darkness current no longer limits the very weak flux measurements. In the
photoconductor mode, photodiodes are linear, but in the photovoltaic mode,
photodiodes behave in a logarithmic fashion, except under very weak charges when
a photovoltaic generator is used with a battery charge as for solar cells (see Figure
3.26).
photoconductor mode
V
Φ= 0
photovoltaic mode
Φ=1
Φ= 2
Φ=3
Since the equivalent final schema can be reduced to the one shown in Figure
3.24, photodiodes are first order sensors (see Chapter 2). The cut-off frequency
given by fc = 1/2ヾRC depends on the charge resistance R (see Figure 3.27).
Physical Principles of Optical, Thermal and Mechanical Sensors 107
fc (Hz)
107
106
105
104
103
102
102 103 104 105 106 107 Rm (Ω)
Depending upon the structure of the different doped regions of a photodiode, the
performances are different. Table 3.2 reviews some of these possibilities. In the
schemata of Table 3.2, we see that the front face of the sensors, excepting the
Schottky structures, is covered with a fine protective layer of transparent SiO2. The
first semiconductive layer is generally of type P. This doping is often preferred for
its simplicity of preparation and transparency. The first schema (called planar type)
corresponds to the junction already described. The interfaces are flat and establish a
potential barrier between P and N. For those applications that concern the visible
light and are in use, this technique, used in silicon, is sufficient. When response time
needs improvement, the sensor’s thickness must be augmented, which reduces its
capacity. To effect this, a structured layer can be created by successive dopings
(weak planar capacity). A progressive passage (type P-N-N) can also be created to
realize the P-N junction. The spectral sensitivity is then modified by ensuring that
certain wavelengths are not absorbed into the depletion zone. This allows better
sensitivities in the infrared or ultraviolet. As mentioned above, the best way to
reduce diode capacity and improve the separation of photogenerated couples is by
creating an enlarged depletion zone. This is done by inserting an intrinsic stratum
between P and N. This produces rapid photodiodes (type P-I-N). Flux detection in
the ultraviolet field is difficult with semiconductive junctions because their fluxes
are rapidly absorbed in the P region when the wavelength decreases. To avoid this
problem, metal semiconductive Schottky junctions are realized. These junctions are
similar to P-N junctions, especially relating to their potential barriers. Although this
kind of junction has a narrower depletion zone, the metallic stratum can be made
very transparent to ultraviolet rays. Often, the width of the charge zone of space can
be augmented by creating a dopage gradient in the semiconductive area. In addition,
an internal amplification can be produced by a correct and controlled breakdown
stage.
108 Fundamentals of Instrumentation and Measurement
P
Weak darkness current, rapid
Planar, low capacity N response, strong sensitivity in Silicon
UV and IR rays
N
P N
Weak darkness current, strong
PNN sensitivity in UV rays, Silicon
N
insensitivity to IR rays
N P I
P N
Very high sensitivity in UV GaAsP.
Schottky
N rays GaP
The breakdown stage also influences the noise current so that photodiode
detectivity does not increase proportionally to the internal gain. In fact, the resulting
strong bias from the breakdown stage augments the width of the charge zone,
considerably reducing response time. Avalanche photodiode structures always
thicken gradually (zone ヾ), so that the electrical field is moderated, even under
strong tension, sometimes up to 1,000 V. Photodiode structures can also be created
heterogeneously with different kinds of semiconductors. In these cases, we speak of
heterojunctions as opposed to homojunctions created by dopings of a single
semiconductor. Since we have been for the most part discussing visible effects, we
have been describing photodiodes essentially as silicon homojunctions.
Physical Principles of Optical, Thermal and Mechanical Sensors 109
Passive sensors are mostly used in mechanics. In Chapter 2, we saw that these
sensors can be resistive, capacitive or inductive. Inductive sensors are often used for
displacement measurements. On the other hand, resistive sensors are often used for
deformation measurements and are sometimes called, somewhat incorrectly,
constraint gauges.
The piezoelectric effect [CAD 64] is the most widely used basic principle in
active mechanical sensors. It is used in its simplest form with force and deformation
sensors. In the following sections, we will explain the principles of piezoelectricity
and some methods used to analyze the signals it generates.
It is important to remember that very little difference exists between force and
constraint measurements. In order to measure a force, a kind of dynamometer is
generally used. This is a tool that helps us establish an equilibrium between the force
we want to measure and the constraint produced by the deformations undergone by a
solid that makes up a part of the sensor under the action of this force. Working in the
elastic domain of deformations, we see that constraints and deformations are
proportional. A simple calibration allows the same sensor to carry out both a
constraint measurement (or force by surface unit) and a deformation measurement.
Resistive gauges are simply resistive circuits that can be attached to a structure to
determine its local deformations. These kinds of resistances represent an important
percentage of deformation sensor sales. Their ability to function in many conditions
and their low prices explain their widespread usage. In addition to these long-known
assets, more recently gauges have been developed that aid in producing very small,
high-resolution sensors. As well, these resistive gauges are associated with proof
bodies and conditioners that improve sensitivities and signal to noise ratio.
Following these processes, the measurable relative elongations go from 10-7 to 10-1.
The relative deformation error, seldom below to 10-3, is more often of the order of
510-3 to 10-2. These gauges are sometimes in the form of wire, sometimes thin layers
of some material, or sometimes they are created by doping in the semiconductors.
For a wire of section S and the width 1 made of a material of resistivity ρ, the
∆R ρl
relative variation of the resistance R given by R = is written:
R S
110 Fundamentals of Instrumentation and Measurement
∆R ∆ρ ∆l ∆ S
= + − [3.79]
R ρ l S
∆R ∆l ∆l
= ⎡⎣(1 + 2ν ) + C (1 − 2ν ) ⎤⎦ =k [3.80]
R l l
where k is the gauge factor and C is Bridgman’s constant. The resistance values are
usually from several hundreds to several thousands of ohms. With metals, when
= 0.3 and C is of the order of 1, we get k of the order of 2. With semiconductors, C
can reach 200 and the gauge factor is high, of the order of C. This means that
measuring very weak deformation must be done with semiconductive gauges, but in
this case it is important to remember that R is very dependent on the temperature,
which, in practice, limits the use of these gauges to temperatures below 200˚C.
The main shortcoming of these gauges is that they must be attached to the
structure. This limits their use to medium temperatures (up to 500-600˚C for
metallic wire gauges). At higher temperatures, resistive gauges are increasingly
being replaced by optical methods using coherent light beams, among them speckle
interferometry.
Piezoelectricity derives its name from the Greek word “piezo”, meaning “to
press”. The piezoelectric effect is the conversion of pressure into electricity. To be
more precise, the term describes the appearance, due to the action of microscopic
deformations, of charges on the surface of a solid. These charges are produced by
local displacements of centers linked to the crystalline mesh.
In fact, the piezoelectric effect only exists in crystals, ceramics and polymers that
are anisotropes; that is, that have no symmetrical center in the elementary mesh.
Physical Principles of Optical, Thermal and Mechanical Sensors 111
Although the Curie brothers discovered piezoelectricity in 1880, it was only put to
practical use during World War I, first by Paul Langevin, who developed sonar, then
in 1918 by Walter Cady, who built the first quartz oscillator. Today, the popularity
of high quality quartz oscillators makes piezoelectricity an integral part of
electronics. The production of products using piezoelectricity employs hundreds of
thousands of people throughout the world.
When the solid has a center of symmetry, the deformation is very weak and is
proportional to the square of the applied electric field: this is electrostriction.
When the solid does not have a center of symmetry, the charge displacement,
which is clearly more important, is proportional to the applied filed. This is the
inverse piezoelectric effect. Often, when there is no center of symmetry, the material
already has a permanent piezoelectric bias. This dielectric bias varies not only
according to the applied field but also according to temperature: this is the
piezoelectric effect.
The following four points summarize these effects: in metals there is no effect; in
dielectrics with a center of symmetry, the weak effect is called electrostriction; in
dielectrics without centers of symmetry, their effect becomes stronger and is called
piezoelectricity; and in anisotropic dielectrics with permanent bias, there is a
piezoelectric disturbance of the piezoelectricity because the temperature is in this
case an intruding influence variable.
-
+
the total charge of the ensemble remains zero. Thus, the electric induction vector D
is also zero (Gaussian law), and we get:
Finally, the potential V that appears at the limits of the piezoelectric crystal is
deduced from:
P
E = −grad V = − [3.82]
ε0
Let us now suppose that we have short-circuited the two electrodes by means of
a metallic wire (see Figure 3.30b). The potential and the field between the electrodes
become zero:
which means that the flux of P across the system shown in Figure 3.30b (the
crystal, the electrodes and the metallic wire) is equal to the total sum of the
contained charges. In the volume of the crystal = 0 and the specific charges that
produce the flux of P are the electrode charges M1 and M2 that we call the images of
internal bias, that is, constraints.
+++++++++++++++ M1
--------------------------
a) b)
Figure 3.30. Quartz with constraints in open circuit (a) and in short circuit (b)
114 Fundamentals of Instrumentation and Measurement
z z
y
y
x
x a) b)
In elasticity [GER 62] we note jij, the constraint components in a solid, the index
i giving the direction of the component, and the index j showing the normal of a
facet to which this component is applied (see Figure 3.32).
σzz
σyz
σxz
σzy
σzx σyy
σxy
σyx
σxx
It has been shown [ROY 99] that constraint tensors (jij) are symmetrical
(jij = jji) and we note that jii = ji and jij = jk. In general, we describe the
Physical Principles of Optical, Thermal and Mechanical Sensors 115
piezoelectric effect through the linear relations which bind the constraints to charges
by surface unity (qi) appearing on normal of facets in the direction i when the crystal
is short-circuited:
For this Curie cut of crystal, we metallize the faces perpendicular to Ox (see
F
Figure 3.33). If we apply a force of module F in the direction Ox so that σ1 =
L`
the charge by unity of surface qi that appears on the electrodes equals:
F Q1
q1 = d11σ1 = d11 = [3.86]
`L L`
Optical (3)
axis z
L
e
l
y (2)
0 Mechanical
Electrical (1) axis
axis x Contacts
If this same pressing force is applied following the Oy axis, the total charge Q1
L
increases in the relation :
e
F L
q'1 = d12 σ2 = − d11 σ 2 = − d11 = Q1 [3.87]
e e
Studying the piezoelectric matrix also makes clear that no charge can appear on
the faces perpendicular to Oz. In addition, the sensor can be used for measuring
shearing constraints of large surfaces (L, l), but not for hydrostatic measurements
(pressure P). Indeed, such a constraint leads to:
⎛ 0, 0, 0, 0, d15, 0 ⎞
⎜ ⎟
⎜ 0, 0, 0, d15 0, 0 ⎟ [3.89]
⎜ d , d , d , 0, 0, 0 ⎟
⎝ 31 31 33 ⎠
⎛ 0, 0, 0, 0, d15 , 0 ⎞
⎜ ⎟
⎜ 0, 0, 0, d 24 0, 0 ⎟ [3.90]
⎜ d , d , d , 0, 0, 0 ⎟
⎝ 31 32 33 ⎠
Physical Principles of Optical, Thermal and Mechanical Sensors 117
The values of dij are slightly above those of quartz and are sensitive to and
dependent on preparation procedures.
a b
Figure 3.34. Assemblies for measuring the three components of a force (a) or of the
component following z and of the couple around z (b) (from G. Asch [ASC 91])
118 Fundamentals of Instrumentation and Measurement
dQ
Vm
dt Rc Cc Rd Cd
Req
Ceq
If we measure the tension to the limits of the sensor or even the output of a
tension amplificator (which is equivalent), we get:
Q R r Cr jω
Vm = [3.91]
Cr 1+R r Cr jω
This is a first order low pass function with a cut-off frequency of c=1/2ヾReq Ceq
and the permanent value Q/Ceq depends on the impedance of the connecting cables
and the input impedance of the tension amplificator. This situation is less promising
than using a charge-tension convertor (see Figure 3.36) [BAU 61].
Rr
Cr
-
dQ ve +
dt C eq
R eq Vm
If we take the operational amplificator as an ideal with Vi zero, we carry out the
short circuiting of electrodes and the transfer function depends only on the counter-
reaction impedance:
Q R r Cr jω Q jωτ
Vm = = . [3.92]
Cr 1+R r Cr jω Cr 1 + jωτ
Physical Principles of Optical, Thermal and Mechanical Sensors 119
The response time k = RrCr of this kind of assembly no longer depends on the
sensor or on the converter connection. For the values of Cr of the order of several
hundreds of pF, and of Rr of the order of several 109 , we reach time constants of
the order of several hundred ms and measurement tension of several mV/pC. For
values of djj of the order of about 100 pC/N, we see that it is easy to reach sensitivity
of the order of V/N.
Our bodies can qualitatively evaluate the concept of hot and cold objects, but in
this sense, though this concept is assimilated to the concept of touch in the normal
five senses, it is both non-linear and residual: it depends on prior experience. Like
the other human senses, hot and cold cannot be measured – we do not even know
what it is we are trying to measure. The basic thermal sensor is, as we said above, a
temperature sensor. The first question to be answered is how to define this odd
measurand, which is not well understood in the physiological sense.
There are two ways a body can transfer its kinetic energy to another body. One is
by contact that communicates the agitation of the first system to the second system;
the second is electromagnetic radiation exchanged between the particles of the two
120 Fundamentals of Instrumentation and Measurement
systems. The particles of the first body play the role of an electromagnetic source
and those of the second body play the role of receiver. Temperature sensors are thus
systems that can transform the kinetic energy of agitation communicated by contact
or radiation into another form of energy, usually electrical.
We will only describe transfer by contact, since transfer by radiation really only
occurs with optical sensors. Temperature sensors were called thermometers for the
first time in 1624.
Looking for this universality, Carnot stated that energy takes two forms: thermal
agitation (heat) and organized energy (existence of privileged directions of speed)
which is called work. This means there is a relation between the concept of
temperature and that of heat conversion in work by means of a motor. Carnot
introduced the idea of the ideal or reversible motor able to constantly operate the
conversion of heat in work or the reverse (reversibility). He showed that the
production of such a motor, functioning between two heat sources of temperatures
1 and 2, is independent of the technology used to construct it. This production
depends only on 1 and 2:
F(θ1 )
η = 1− [3.93]
F(θ 2 )
T1
η = 1− [3.94]
T2
Two other temperature scales are in use today. Much of the English-speaking
world uses the Fahrenheit scale for measuring thermodynamic temperature. This
scale differs from the Celsius system used through most of the world. R values are
different in the two scales. The R value of the Fahrenheit scale is related to a
thermodynamic temperature scale called the Rankin scale.
– Fahrenheit temperature, used mostly in the USA and the UK, takes from the
Rankin scale the value 32˚F as the freezing point of water. The relation between ˚F
and ˚C is expressed as:
Following the example of the perfect gas thermometer, thermal dilation is the
physical variable most useful for making a thermometer. Thermal dilation is at the
basis of many thermometers used today. In all the bodies, the increase of mean
kinetic energy, that is, of temperature, is expressed by a modification of the mean
distances separating the elementary particles (atoms or molecules). In solids, this
modification is often different depending on the direction, and taking into account
this anistropy, we define the linear dilatation g1 in the following way:
1∂
αe = [3.97]
∂T
1 ∂V
α v = 3α = [3.98]
V ∂T
Many thermometers, among them perfect gas thermometers, make use of fluid
dilatation for temperature measurements, but this does not mean the measurand is
easily converted to an electric variable. Using the phenomenon of dilatation helps us
understand that it is crucial that thermometers carry out an efficient transfer between
the system we want to measure and the thermometer. We will discuss this further in
the following section.
3.3.4.1. Conduction
The origin of heat (of thermal agitation) cannot really be known. We can say that
the increased temperature of a system does not retain the memory of whatever
produced the increase. This is due to the fact that it is impossible to contain heat
within a system. The agitation inevitably spreads to the outside environment by
contact or radiation. This process can be slowed to some extent by isolation
procedures, but the “leakage” is, in fact, inevitable. These phenomena are explained
Physical Principles of Optical, Thermal and Mechanical Sensors 123
by heat transfers; and we must understand these transfers in order to establish the
equilibrium temperature of a system, which is fundamental to procedures of taking
and recording temperatures.
Figure 3.37 schematizes the group of thermal transfers that can occur between a
temperature source T0 and the environment surrounding the temperature Ta. The
source is maintained at a constant temperature T1, partly to compensate for losses it
undergoes, and partly to give it the highest possible calorific capacity C. Three
successive plates of different materials have been attached to this source, and Figure
3.37 shows the temperature distribution from the source to the limit of the material
stacking making contact with the exterior air.
The transfer that occurs in the plates P1, P2, and P3 is transfer by conduction. It
lowers the temperature of T0 to Tb. The last temperature decrease (from Tb to Ta) is
produced by two other possible types of transfers: convection, which is a specific
form of conduction, and radiation. It is important to note that Figure 3.37 does not
show the order of importance of different thermal transfers. Transfer by conduction
is not systematically more efficient than other types of transfer.
Each of the gradient plates, which are made of different materials, show transfer
by contact, that is, conduction. With the notations of Figure 3.37, we get:
dS
grad T
ϕ # −λ grad T
1 dQ ⎛ T −T ⎞ ⎛ T2 − T3 ⎞ ⎛ T3 − T6 ⎞
= λ1 ⎜ 1 2 ⎟ = λ2 ⎜ ⎟ = λ3 ⎜ ⎟ [3.99]
S dt ⎝ L1 ⎠ ⎝ L2 ⎠ ⎝ L3 ⎠
This equation shows that the quantity of heat leaving the source by time unit and
the surface crosses without modifying the plates P1, P2, and P3. In other words, the
∆T
heat flux has been conserved. The factor i, which defines the gradient of the
∆x
temperature distribution in each material, is the thermal conductivity (Wm-1K-1).
More generally, thermal conductivity is defined through Fourier’s law. This
vectorially expresses the heat flux transmitted by conduction:
f f f
ϕconductive = − grad ( T) ≅ - grad T [3.100]
This expression only holds absolutely for the first equality because thermal
conductivity is weakly dependent on temperature (almost always increasing).
A a B
A B
T1
T2A
T
T2B
T3
Lg
f 1 dQ T2 A − T2 B
ϕ = = [3.101]
S dt Ra
Physical Principles of Optical, Thermal and Mechanical Sensors 125
Taking into account the weak conductivity of air, the contact zone between the
two surfaces is responsible for the value of the thermal resistance. This explains why
the value Ra cannot really be calculated precisely. This value is deduced from the
T −T
relation 2 A f 2 B , which is measurable.
ϕ
3.3.4.2. Convection
Convection takes as a starting point the fact of a mobile fluid taking part in a
thermal transfer. In this transfer, the transmitted thermal flux increases considerably
in relation to the flux produced by the conduction between a solid and a fluid. The
fluid particles, their kinetic energy having been increased by contact (conduction)
with the solid wall, displace and are then replaced by other molecules of weaker
kinetic energy, capable of harnessing the heat of the wall. The movement of the fluid
components permanently renews the fluid molecules in contact with the solid.
Solid
Kinetic energy transfer
T∞
Convection appears of its own accord when different temperature zones coexist
in a fluid. Actually, its volumic mass = PM/RT decreases with the temperature, so
that the hot fluid tends to rise and the cold fluid tends to drop with the effect of
Archimedes’ principle (or law of buoyancy). This type of convection is called
natural convection. Of course, we can also force the movement of fluid by using a
turbine. We then speak of a forced convection. The speed of the particle group
becomes much higher and the flux exchanged by forced convection becomes higher
by several orders of magnitude to that of natural convection. Calculating convection
is difficult and often must be carried out by means of numerical calculation. In this
text we will limit ourselves to phenomenological expressions of the transmitted flux
by convection of a solid to a fluid. This occurs by means of a proportionality
coefficient hc that exists between the exchanged flux and the temperature
126 Fundamentals of Instrumentation and Measurement
differences, between the surface S of the solid Ts and the fluid far from the surface
T∞ (in practice to several tens of thousands of µm):
1 dQ
= h C (Ts − T∞ ) [3.102]
S dt
In the air, hc is clearly equal to 5 W/m2K for natural convection and can reach
some hundreds of W/m2K for forced convection.
3.3.4.3. Radiation
We have seen in our discussion of optical sensors (see section 3.1) that the
surface dS of a body carried to a temperature T produces a radiation. The energetic
flux of the radiation transmitted in a solid angle d around a direction n (see Figure
3.41) is expressed by:
∞
dφ = ∫0 (ε λ L0λ ( T ) dΩ dS cos θ ) dλ [3.103]
If the transmissivity i does not depend on (i = i is the gray body), the flux
transmitted by radiation can be calculated in the half space above the surface dS by
using the expression of the black body luminance L0λ (T ) (see Figure 3.41):
d φ = ijT4dS [3.104]
dS
C 2 = hc / k
We can see that the flux exchanged by radiation grows very quickly with the
temperature. In section 3.1 we also saw that with two facing objects of
transmissivity i and ia that are at two temperatures T and Ta, the flux lost by the
surface dS at the temperature T is:
4
dφ = ε εa σ (T − Ta ) dS
4
[3.105]
hR = 4 εε a σ T
3
[3.107]
For a body at ambient temperature, we find hr of the order of five. So, in air
without movement, at a temperature close to ambient, the sum of the convective and
radiative transfers can be expressed as:
This differential equation is of the first order. The response time k of the
thermometer is expressed by:
C
τ= [3.109]
1
hs +
R
with S being the surface of the thermometer in contact with the ambient fluid.
For the usual values of C, S, and R, we get k of the order of the second. This
order of magnitude shows that temperature sensors usually are very slow. To
improve this response time, the most frequently used solution is reducing the caloric
capacity of the thermometer by reducing to the minimum its geometric dimensions.
However, it is important to keep in mind that response time is not intrinsic to a
thermometer. It depends on how the thermometer is used.
With metal, electric conductivity decreases with the temperature because the
number of current carriers in a metal does not, in practice, depend on the
temperature. The only dependence on temperature comes from mobility, which
decreases with the number of collisions per second and thus with the temperature.
This decrease is approximately linear.
i>0
Metal B
Heat T T Heat
source 0 source
Metal A
i>0
We can express the creation of this electric current by saying that maintaining
constant temperatures from two heat sources leads to the appearance of an
electromotive force in the loop metal A + metal B. This A-B pair is called a
thermocouple. We make EAB (T, T0) the algebraic value of this electromotive force
in considering it to be positive because it makes the current circulate from A towards
B in the junction M to T0.
EAB (T, T0) = – EBA (T, T0) = – EBA (T0, T) = EAB (T0, T) [3.111]
Metal A
T0 M N T
Metal B Metal B
P Q
Voltmeter
maintained to a
uniform temperature T ’
If T = T0, we observe that the deviation of the voltmeter is canceled and we can
also verify the sign (Vp – Vq) by reversing T and T0 or the metals A and B described
by equation [3.111].
The laws of thermodynamics (the first principle and the Onsager relation for
irreversible processes) allow us to establish the fundamental laws governing
thermocouples [MAC 62]. Thermodynamics does not give any explanation for the
basic physics of the phenomenon (which can be found by studying, through solid
state physics, the electron distribution in the different energies in the two metals
[KIT 83]). However, the description physics gives us of the phenomenon is
sufficient for studying temperature sensors. Thermodynamics shows us specifically
that the Seebeck effect is the result of the Peltier and Thomson effects.
Physical Principles of Optical, Thermal and Mechanical Sensors 131
i
A B
Ensemble at uniform T
Suppose that two different metals A and B are welded and traversed by a current
i (see Figure 3.44). When the current i traverses the welding in the direction AsB,
with the ensemble maintained at a temperature T, a certain power is freed in addition
to the Joule effect. This power dQ/dt is called the Peltier effect. It is proportional to
the current intensity i which traverses the welding and its sign depends on the
direction of i, helping us to differentiate it from the Joule effect, which is always
positive:
dQ
= π A − B (T ) i [3.112]
dt
π A− B (T ) = − π B− A(T ) [3.113]
During a time interval dt, a certain quantity of heat is transmitted by the Joule
effect. But we also observe an emission or a heat absorption dQ of a different
physical nature. This quantity dQ is positive (heat emission towards the outside)
when i circulates from P towards Q and is negative (the metal absorbs the heat)
132 Fundamentals of Instrumentation and Measurement
iiiif f d 2Q
d 2 Q = σ A (T ) grad(T). i.dx.dt or again = σ A (T ).dT .i [3.114]
dt
d E AB (T , T0 ) d π AB
= + (σ A (T ) − σ B (T ) ) [3.115]
dT dT
d E AB (T , T0 ) π AB (T )
= [3.116]
dT T
d π AB (T ) d E AB (T , T0 ) d 2 E AB (T , T0 )
= + [3 .117]
dT dT dT 2
d 2 E AB (T , T0 ) σ A −σ B
= [3.118]
2 T
dT
These equations clearly show that EAB is not a linear function of T. With the help
of the relations just presented, we establish the Seebeck effect created in the circuit
shown in Figure 3.43. For this, we maintain M to T0 and progressively raise N from
T0 to T1 so that the integration of dEAB(T, T0) from T0 to T1 leads to:
This fundamental relation shows that we can deduce the electromotive forces of
the two couples A-B and C-B from the couple A-C:
This relation is used when T'0 is the ambient temperature and we want to deduce
the EMF value in relation to T0 = 0˚C, the latter being the reference temperature
found in standardized tables. The electromotive force is often created by
compensation housing that allows for thermocouple use without a reference source.
It is important to remember that all isothermal metals C introduced into the loop
A-B do not modify the EMF of the thermocouple A-B. This allows us to construct
couples by heterogeneous welding. This fact also helps explain measurement with
the help of a voltmeter, as we have already described in section 3.3.5.2. Figure 3.43
can be schematized by Figure 3.45, in which we see that the branch PQ develops a
zero EMF if the metal C is at a uniform temperature T', so the only EMF is that of
the thermocouple A-B. We can introduce as many metals as we wish between P and
Q provided the temperature of the ensemble remains uniform.
Even though many alloys can be used in making thermocouples, less than ten or
so are currently used for this purpose. The couple platinum-platinum plated with
10% rhodium with sensitivity of the order of 10 µVK-1 offers the lowest uncertainty
(< 0.1˚C) because of the attainable purity of its components and their chemical
stability. It can be used from 300 to 1,800 K but more often is used above > 500 K
in which its stability and manageability often make it preferable to other
thermometers, excepting platinum resistance thermometers, which are considered
the reference.
Thermocouples made of metal alloys have very weak sensitivities. The most
sensitive of these is the type E thermometer (chromel-constantan); this has a value
of around 80 mVK-1. We can replace the metal base couples with semiconductive
junctions (for example tellurium, bismuth or germanium with different dopings).
Although the transmitted electromotive forces are in this case clearly higher, the
manageability of these semiconductive thermocouples is still today limited, in any
case, for industrial usages. It is important to keep in mind that the calibration curve
of a thermocouple, whatever its composition, is never linear – but only within a
narrow temperature range. This means that sensitivity is strongly temperature-
dependent.
The welding and connecting wires are usually placed within a protective metallic
sheath. The welding is sometimes in electrical contact with this sheath. The wires
are always isolated by a silica powder or a compacted alumina. When the structure
to be tested is metallic and can be grounded with the instrumentation, the assemblies
Physical Principles of Optical, Thermal and Mechanical Sensors 135
with which the welding is connected to the sheath are preferred for suppressing
usual disturbances.
E AB (T , T0 ) = E AB (T , Ta ) + E AB (Ta , T0 ) [3.122]
this is the value we would have by using a second couple at the reference
temperature T0. The EMF EAB (Ta, T0) (compensation casing) can be produced by
the disequilibrium tension of a Wheatstone bridge containing a thermistor sensitive
to Ta. Regulating different impedances and tensions of the bridge gives us
EAB (Ta, T0) with a weak uncertainty if the ambient temperature does not vary more
than 50˚C during measurement. When the compensation casing cannot be directly
connected to the thermocouple, intermediary cables, called compensation cables,
must be used. This avoids systematic errors that can occur by creating parasite
EMFs at A-B junction connections if the connection was made without proper
precautions. Obviously, these compensation cables are dependent which A-B
couples have been used. Their sheath is usually the standard color of the
thermocouples currently in use.
3.4. Bibliography
4.1. Introduction
variable digital
sensor filter amplifier processing
measurable & data analysis
analog processing
Noise sources that come from inside components have many origins and fall into
five categories. We discuss these below.
I
U
R RR
It is possible to use another, equivalent schema for a noise source using a current
or Norton generator. In this case, it is enough to divide the expression given above
by R2 to get a clear expression of noise current.
⎛T ⎞
I 2 = 4k .⎜ ⎟.∆f [4.2]
⎝R⎠
Φ U+ ( f ) = 4.k B .T .R [4.3]
(V2/Hz)
4kTR
0 f
i(t)
iG(t)
ID
0 t
The sum j corresponds to the crossings that occur per second. The probability
density of the time interval T1, separated by two successive crossings, is expressed
by (this is a Poisson process):
where represents the average number of carriers crossing the barrier by unit of
time. We then see that the intercorrelation function Ri(k) of i(t) can given as:
Ri (τ ) = (q.λ )2 + q 2 .λ.δ (τ ) = I D
2
+ q.I D .δ (τ ) [4.6]
Analog Processing Associated with Sensors 141
Here ID represents the average value of the current i(t). The spectral density of
the power of the current i(t) is calculated with the help of the Wiener-Khintchine
theorem applied to equation [4.6]. With a direct current we get:
+∞
Φi ( f ) = ∫ Ri (τ ). exp(− j 2πfτ )dτ = I D + qI D
2
[4.7]
−∞
The second term of this expression corresponds to the spectral density of the shot
noise ( iG). This means that shot noise, like thermal noise, is white noise. Because
of this, shot noise cannot be differentiated from thermal noise as soon as they both
become present in the same electrical circuit. On the other hand, shot noise only
exists when a current ID crosses the potential barrier.
In conclusion, an effective shot noise current (IG) depends on the direct current
(ID) and the range of working frequency (Bb) according to the relation:
I αD
I F2 = K 1 . .Bb [4.9]
f
2
(A /Hz)
f
0
β
I P2 ID
= K2. . [4.10]
Bb 2
⎛ f ⎞
1 + ⎜⎜ ⎟⎟
⎝ fc ⎠
(A2/Hz)
-40dB/dec
f
0 fC log scale
Un
In
Quadrupole
without noise
10-13 10-23
(for 741) (for 741)
1/f gradient 1/f gradient
10-14 10-24
fC f fC f
0 (130 Hz for 741) log scale 0 (1,600 Hz for 741) log scale
These noise specters have both a white noise component and a noise component
in 1/f.
True
Un quadrupole
R=0
In Quadrupole
or without noise Umes Specter
analyzer
AV(f)
R = RE
– When the input is short-circuited (R = 0), the spectral density of the noise
voltage equivalent to the input is simply expressed as:
u i2 = u n2 [4.11]
Analog Processing Associated with Sensors 145
2
2
u mes = Av ( f ) .u n2 [4.12]
u i2 = u n2 + u E2 + R E2 .i n2 + 2C.R E .u n .i n [4.13]
where the term uE corresponds to the thermal noise of the source resistance, we get:
u E2 = 4k .T .R E [4.14]
where C represents the correlation coefficient (a number between -1 and +1) of the
noise sources un and 1n.
By taking a very high value of RE, the third term of the equation is preponderant
even if the measured spectral density is as follows:
2
2
u mes = Av ( f ) .R E2 .i n2 [4.15]
This allows us to establish the source of the noise current In. An intermediary
resistance value for RE helps us find the correlation coefficient C.
Let Ps be the power of the useful signal at the output of an electronic chain and
Bs is the power of the corresponding noise. We then define the signal-to-noise ratio
(S/B)s at the output of the chain with the relation:
⎛S⎞ PS
⎜ ⎟ = 10. log [4.16]
⎝ ⎠S
B B S
146 Fundamentals of Instrumentation and Measurement
In the same way, the signal-to-noise ratio at the input of the chain can be written as:
⎛S⎞ P
⎜ ⎟ = 10. log E [4.17]
⎝ B ⎠E BE
These two relations help us establish the quality (with regard to noise) of an
electronic chain by defining the noise factor as:
FdB =
(S / B )E ⎛ P .B
= 10. log⎜⎜ E S
⎞ ⎛ B
⎟ = 10. log⎜ S
⎞
⎟⎟ [4.18]
(S / B )S ⎟ ⎜ G.B
⎝ PS .BE ⎠ ⎝ E ⎠
BS = G.BE [4.19]
In this case, the noise factor is equal to the unity (0 dB). Another definition of
the noise factor is:
u i2 . Av2 ( f ) u n2 + u E2 + R E2 .i n2 + 2C.R E .u n .i n
F= = [4.21]
u E2 . Av2 ( f ) u E2
u n2 + R E2 .i n2 + 2C.R E .u n .i n
F =1+ [4.22]
u E2
When the noise sources un and in are weakly correlated, the speed of F according
to the source resistance is shown in Figure 4.10.
Analog Processing Associated with Sensors 147
un2/4kT 4kT/in2 RE
The above expression shows that the noise factor of a quadrupole is always
higher than the unit. Consequently, signal analysis always involves some
degradation of the signal-to-noise ratio (see equation [4.18]). With reception
systems, or with weak signal amplification coming from sensors, the additive noise
at the first stage of reception or at preamplification plays an essential role in
conditioning the signals to be analyzed.
4.3. Amplifiers
US saturated
Usat+<Ualim+
AV gradient
input output
+
ε
IS
ε0' ε = (U +-U-)
_ 0 ε0
U+
US = Av.ε
U-
Usat->Ualim -
reference potential saturated
This is why, from the perspective of linear functioning, the operational amplifier
is always used with a retroaction feature. This kind of amplifier can also be used in a
non-linear functioning mode, with or without a positive reaction. In this instance, the
feature is used for a broader input voltage dynamic. For these applications
(comparator, trigger, astable, etc.), the output can only take two values: Usat+ or
Usat–.
In order to study all possible feedback types, the operational amplifier can be
assimilated to a quadrupole of simplified design. This is shown in Figure 4.12.
I1A I2A
Zs
U1A U2A
Ze
A v.U 1A
Ze represents the input impedance of the operational amplifier, with Zs its output
impedance, and AV the amplification of the differential voltage.
amplifier
input output
signal
signal
ε
+_ Av
BR
retroaction l oop
I1 I1A I2A I2
Amplifier
U 1A U 2A
Av
U1 U2
I 1R I 2R
counter
U1R feedback block U2R
BR
connections
Following the serial or parallel linkage of input dipoles of the two quadrupoles,
we can distinguish four possible configurations:
– serial input dipoles and parallel output dipoles;
– serial input dipoles and serial output dipoles;
– parallel input dipoles and parallel output dipoles;
– parallel input dipoles and serial output dipoles.
I1 I 1A
Zs I 2A I2
U1A
Ze
U 2A I1 Zs/(1+T) I2
Av .U 1A
U1 U2 U 1 U2
I1R I2R Ze.(1+T)
U 1R BR.U 2 U 2R A v.U 1/(1+T)
(T=Av.Kv )
I1=0
+ I2
U1A
_
R2
I1=0
RRchch U2
U1
U1R RR11
I1 I1A I2A I2
Ze
U1A 1/Zs U 2A
g.U 1A I1 Zs(1+T) I2
Ze.(1+T)
U1 U2
I 1R I2R U1 U2
U 1R B R.I2 U2R g.U 1 /(1+T)
(T=g.ZR)
I1=0
+ I2
U1A
_
RRch
ch U2
I1=0
U1
U1R RR11
As with the output current I2 = U1R/R1 ≈ U1/R1, this has servo-control of the input
voltage U1: this is how voltage current convertors or transconductance amplifiers are
made. The transfer function of the return block is consistent with an impedance
BR = ZR = U1R/I2, (equal to R1 if we disregard I1). This assembly is similar to the one
shown before it, except that the charge resistance fluctuates because it replaces R2.
The input and output impedances are very high; they are both multiplied by (1 + T).
The term T corresponds to the transfer function in open loop of the assembly. It is
written T = g.ZR. Here, g represents the assembly transconductance. If g is high
enough, ideally infinite (or if I1 can be disregarded), the input impedances are
infinite. This means the voltage-current convertor is perfect and the
transconductance equals 1/ZR = 1/R1.
I1 I1A I2A I2
Ze Zs
I1 Zs/(1+T) I2
z.I 1A
U1 U2 U 1
I1R I2R Ze/(1+T) U2
I1 .z/(1+T)
B R .U 2
(T=z.YR)
I1
RR22 I1R
I1A
_
U1=0
+ RRch
I1 ch
U2
I1 I1A I2
Ze Zs
I1 Zs(1+T) I2
Ai .I1A
U1 U2 Ze/(1+T)
I1R U1 U2
B R .U2 I1 .Ai/ (1+T)
(T=Ai.Ki)
I1 I1A _
I2
U1A=0
+
RRch
ch U2
RR22
I1R
U1R RR11
With this assembly, the output current I2 has servo-control of the input current I1.
This current crossing the charge resistance is then shared among the resistances R1
and R2, which are both linked in parallel because U1 = 0. Under these conditions, we
can easily show that I2 = I1R(R1 + R2)/R1 ≈ I1(R1 + R2)/R1. The return chain is thus a
coefficient of current transfer without dimension BR = Ki. The loop transfer function
is T = Ai.Ki with Ai being the amplification in current of the direct chain. Since this
is a retroactive effect in current, the input impedance is divided by (1 + T), and the
output impedance is multiplied by this term. When Ai is very high or if we suppose
that U1 = 0, the input impedance is zero and the output impedance is infinite, the
current generator is then ideal and the current amplification equals
1/Ki = (R1 + R2)/R1.
These assemblies are important for conditioning signals before the measurement
chain is introduced, especially with signals coming from sensors. According to the
nature of the available signal, we can make use of one or the other of these counter-
feedbacks.
Analog Processing Associated with Sensors 153
4.3.1.2.1. Bandwidths
When the input signals are of high speed and amplitude, the linear functioning of
the amplifier is limited by its maximum variation speed (also called the slew rate).
This slew rate corresponds to the maximum gradient of the output signal, so that
when the input signal exceeds this value, there is a reduction in amplitude and major
distortion. In such cases, we speak of a “high signal” bandwidth. With these circuits,
the term varies from 0.5 V/µs to 20 V/µs, but specific amplifiers can reach several
hundred V/µs and 1,000 or 2,000 V/µs for hybrid components.
When the input signals are of low amplitude (“small signals”), the slew rate no
longer appears. However, sometimes disturbances can create a bandwidth through
the Miller effect. At the first order, we can see that the transfer function Av in the
open loop of an operational amplifier (the relation between the output voltage Vs and
the differential voltage of input i in sinusoidal regime) is given as:
A0
Av = [4.23]
f
1+ j
fo
The asymptomatic forms of the module (which is that of a first order bandwidth
filter) of this function, in discontinuous features, are:
A0 . f 0
– Av = for f >> f0 [4.25]
f
154 Fundamentals of Instrumentation and Measurement
20 log(AV) AV real
AV asymptotic
2.105
A0 A'V asymptotic
-20 dB/dec
A'0
10 Hz f(Hz)
0 f0 f'0 fT log scale
A0
f
Av 1+ j f A0 A' 0
Av' = = 0
= ≈ [4.26]
A A / A' 0 f A f
1+ v 1+ 0 1+ j + 0 1+ j
f f 'o
A' 0 1+ j f 0 A' 0
f0
A0
f '0 = . f0 [4.27]
A'0
This bandwidth gain corresponds to the transition frequency fT, for which the
voltage amplification in open loop becomes equal to the unity (gain of 0 dB). This
expression shows that the multiplication term between gain and cut-off frequency
called “bandwidth” is conserved. We see that the bandwidth of an operational
amplifier assembly is accordingly reduced when the static amplification of the
closed loop is high.
Analog Processing Associated with Sensors 155
2 β .U T 2
Red = 2h11 = = [4.29]
IC 40 I B
Ip-
Zed
V-
Ip+ Ip+=I0 Ip-=I0
V+ VS
Zmc 2I0
The input resistance in common mode (as seen on inverting and non-inverting
inputs in relation to the ground), in the same conditions, is expressed by:
R mc = 2 β .R E [4.30]
where RE indicates the polarization resistance that links the emitters to the power
source. Rmc is in practice much higher than the differential input resistance
(Red < <Rmc).
156 Fundamentals of Instrumentation and Measurement
The polarization current (Ip) is then defined as the average value of the currents
Ip+ and Ip¯. Their difference is called the gap current (Id):
I +p + I −p
Ip = and Id = I +p − I −p [4.31]
2
It is important to limit these currents. In practice, they bring about an output gap
voltage whose value is dependent on the assembly being used. This is why
amplifiers with FETs at the input stage are by far the best performing and most
popular amplifiers currently used for instrumentation purposes.
-
1 Because the terms Ip+ and Ip correspond to base (or gate) currents of the differential stage,
their input or output direction depends on the transistors being used (PNP, NPN, N-channel or
P-channel).
Analog Processing Associated with Sensors 157
Ud
ε' ε
_
U+
Us
U-
Because of the evolution of the Ube and the Ugs with the temperature of
transistors of the differential stage, the thermal drifts of the gap voltage are of the
order of several µV/˚C. These drifts can be minimized by using the following
techniques:
– we can try to create two differential stages to compensate for the drifts
(LM121, for example);
– we can stabilize the substratum temperature, close to the differential stage, by
means of a transistor that heats it (to a certain extent).
There are also self-switching circuits (one example is the ICL 7605 of Intersil)
that use two differential amplifiers functioning alternately, switching to the rhythm
of a clock. These are similar to the chopper-stabilized amplifiers like the circuit ICL
7650, also made by Intersil. The gap tensions we get are of the order of
2 µV and the temperature drift of the first circuit is of the order of 0.1 µV/˚C.
However, the use of these components is limited to frequencies below that of their
internal clocks.
In any case, most circuits have an interior adjustment that provides for an offset
compensation.
158 Fundamentals of Instrumentation and Measurement
U + +U −
U mc = [4.32]
2
⎛U + +U − ⎞
Us = Av .U d + Amc .U mc = Av . U ( +
−U −
) ⎜
+ Amc .⎜
⎜ 2
⎟
⎟
⎟
[4.33]
⎝ ⎠
It is clear that the higher the relation Av/Amc, the smaller the gap voltage is at the
amplifier’s output. We call this common mode rejection ratio the Common Mode
Rejection Ratio (CMRR). It is expressed in decibels in manufacturers’ specifications
and can vary from 80 dB to 140 dB depending on the circuits.
We see that this ratio decreases with the utilization frequency and, in practice, is
only a factor in assemblies with fairly high to high common mode voltage (for
example, in differentiating assemblies and instrumentational amplifiers). This is
because with assemblies having one input directly linked to the ground (as with
reversed assemblies), we have: U+ ≈ U¯ = 0V.
4.3.1.2.6. Noise
We model the noise sources appearing at the input of a differential amplifier
assembly by three generators of basic noise (Un, In1 and In2), and by two other sources
(UTHR1 and UTHR2). These are relative to the equivalent thermal noise resistances
analyzed through inverting and non-inverting inputs, as shown in Figure 4.21.
I n1
R1 U THR1 Un
+
R2 U THR2
_
Us
In2
The generators of noises specific to the amplifier (Un, In, and In2) are defined
from the respective spectral densities (un, in1, and in2) for a frequency utilization band
Bb = fmax - fmin:
f max
Un = ∫f min
u n df , [4.34]
f max
I n1,n 2 = ∫f min
i n1,n 2 df [4.35]
If we let In = In1 = In2, the voltage of total noise at the input of a differential
amplifier is written as:
U s = A'.U nT [4.38]
Manufacturers indicate the spectral density values (un, in) of these components
whose order varies from several nV.Hz-1/2 to about 100 nV.Hz-1/2 for un, and of
0.01 pA.Hz-1/2 to 1 pA.Hz-1/2 for in.
This demonstrates the importance of the first stage of amplification in the global
noise level of the amplifier device.
The name of these amplifiers comes from the fact that they are designed for the
amplification of very low measurement signals (of the order of µV or of mV)
coming from sensors, transducers (constraint gauges, thermocouples) and
measurement bridges such as Wheatstone’s bridge, among others. They must
perform well and must have the following features:
– a significant static amplification (> 106);
– a very low offset error (≈ 1 µV);
– a lower drift versus temperature and time (< 1 µV/˚C and < 1 µV/month);
– efficient values of noise, low in voltage and current, (of the order of 1 nV/Hz1/2
and of 1 pA/Hz1/2);
– a high common-mode rejection ratio (> 100 dB);
– a high input impedance in well-functioning high impedances;
– polarization currents below 1 pA;
– a bandwidth and a high slew-rate according to the frequency and nature of the
signal.
The first two assemblies offer a very high input impedance and regulation of the
gain with one variable element (P). However, the common-mode error is, relatively
speaking, higher for a structure with two amplifiers. The commuted capacity
assembly can function up to frequencies of close to MHz and the common-mode
rejection ratio reaches 120 dB. In addition, manufacturers have developed other
schemata and specialized components.
162 Fundamentals of Instrumentation and Measurement
Signal
Device
or Measurement
sensor device
Mass tension
There are three types of isolation amplifiers that we define according to the
physical principle being used. “Galvanic” isolation is achieved by any of the three
following ways:
– it can be achieved by electromagnetic coupling with the help of a transformer.
In this case, a high frequency carrier is modulated in frequency or in impulse width
by the signal that must be isolated. In particular, this principle is used in the isolation
amplifiers made by Analog Devices. This company has demonstrated the advantages
of using only one supply voltage that is shared between the “emission” and
“reception” points of the transmission;
– it can be obtained by optical coupling (with a DEL emitter and a photodiode
receiver). This technique does not require a high-frequency carrier and is well used
by Burr-Brown. This company has succeeded in reducing linearity defects by using
a retroaction mechanism with a second photodiode in the emission point;
– it can be obtained by capacitive coupling of a high frequency carrier
modulated in frequency by the signal to be transmitted (the ISO 122 model made by
Burr-Brown is an example).
The isolation tensions are of the order of 4 kV for isolation amplifiers with
magnetic and capacitive coupling, and of several hundreds of volts for those using
optical coupling. The bandwidths are lower by about 100 kHz.
Analog Processing Associated with Sensors 163
To construct this type of amplifier (see Figure 4.24), we use the feature of a P-N
junction with an equation (Ebres-Moll equation) in the following form:
⎛ ⎛ qU ⎞ ⎞
i = io ⎜⎜ exp⎜ ⎟ − 1⎟⎟ [4.42]
⎝ ⎝ kT ⎠ ⎠
⎛ qU ⎞
i ≈ io .exp ⎜ ⎟ [4.43]
⎝ kT ⎠
R R
_ _
Ue Input + Ue +
Output Us Input Output Us
4.3.5. Multipliers
Ue1
Logarithmic
lnUe1 lnUe1 + lnUe2
amplifier
Ue1*Ue2
Exponential
Adder
amplifier
Output
Ue2
Logarithmic
lnUe2
amplifier
4.4 Bibliography
[BIC 92] BIQUARD M., BIQUARD F., Signaux, systèmes linéaires et bruit en électronique,
Ellipses, 1992.
[CHA 90] CHATELAIN J.D., DESSOULAVY R., Electronique, tome 1 – Traité d’électricité
d’électronique et d’électrotechnique, Dunod, 1990.
[COR 99] CORVISIER P., “Filtres à ondes de surface”, Mensuel Electronique, no. 94, July
1999.
[GRA 93] GRAY PAUL R., MEYER ROBERT G., Analysis and Design of Analog
Integrated Circuits, Wiley International Edition, 1993.
[HAS 89] HASLER M., NEIRYNCK J., Filtres electriques – Traité d’electricité
d’electronique et d’électrotechnique, Dunod, 1989.
[HOR 96] HOROWITZ P., HILL W., Traité de l’électronique analogique et numérique,
Techniques Analogiques Elektor, 1996.
[LET 87] LETOCHA J., Circuits intégrés linéaires, McGraw-Hill, 1987.
[PAR 86] PARATTE P.A., ROBERT P., Systèmes de mesure – Traité d’électricité
d’électronique et d’électrotechnique, Dunod, 1986.
[PRI 97] PRICE T.E., Analog Electronics: An Integrated PSpice Approach, Prentice Hall,
1997.
[TIE 78] TIETZE U., SCHENK CH., Advanced Electronic Circuits, Springer-Verlag, 1978.
[TRA 96a] TRAN TIEN LANG, Electronique analogique des circuits intégrés, Groundon,
1996.
[TRA 96b] TRAN TIEN LANG, Circuits fondamentaux de l’électronique analogique, 3rd ed.,
TEC & DOC, 1996.
This page intentionally left blank
Chapter 5
Analog Filters
5.1. Introduction
Analog filters are also used in the output stages of instrumentation systems when
it is necessary to reconstruct an analog signal from the conversion of a digital signal.
Filters are complex and expensive mechanisms, with features that are often
crucial to the overall performance of a system.
The role of a filter is to separate the useful frequencies in a signal (those that
carry information) from unwanted frequencies, such as noise or other signals.
2 H( j )
H ( p) = 0
with Q = 0
2 p H (0)
p2 + 0 + 02
Q
Electronic resonators respond badly to the first two constraints, especially if the
filter must be produced in the form of integrated circuits. That is why high quality
filters are basically made with mechanical resonators with surface or volume waves.
In this filter type, mechanical resonance is transformed into electrical resonance by
piezoelectricity [DIE 74]. However, this technique does not work except with high
frequencies (more than a few MHz). For lower frequencies that mostly exist in
instrumentation systems, we get around this problem by using the following three
techniques:
– we can use inductors and capacitors. If the L-C resonators individually have
mediocre stability, the filters made with the help of this technique have, overall, an
excellent stability (with a few restrictive conditions). This paradoxical property is
directly turned to good account when technological constraints making adequate
inductors are not prohibitory. Otherwise, we can make electronic copies of filters
from L-C models;
– we can use active filters that have only resistors, capacitors and amplifiers.
These kinds of filters are made by putting basic filters into cascade or by copying L-
Analog Filters 169
These three basic techniques are presented in this chapter, after a brief summary
of general calculation methods. These techniques are still evolving, connected as
they are to rapid, constant and unforeseeable progress in electronic component
technologies.
V e ( jω )
A = 20 log = 20 log H −1 ( jω )
V s ( jω )
ϕ = arg( H ( jω ))
In order for a filter to transmit a signal without deformation, the phase difference
must vary linearly according to the frequency. If this occurs, the derivative of the
phase difference in relation to the frequency has a constant value. This variable is
the group delay of group k defined by the relation:
∂ϕ
τ =−
∂ω
In practice, the values of A and k become written within the attenuation gauge
and the group delay gauge, as shown for the attenuation in Figure 5.1 and the group
delay in Figure 5.2.
170 Fundamentals of Instrumentation and Measurement
A(dB)
Amin
Amax
0 ωp ωa
τ (S)
τmax
τmin
0 ωp- ωp+
ω (Rd/s)
Figure 5.2. Gauge in group delay of a band pass filter
A(dB)
Amin
Amax
There are four types of gauges that measure attenuation, with corresponding
parameters:
– low pass, with parameters ωp, ωa, Amax and Amin (see Figure 5.1);
– high pass, with parameters ωp, ωa, Amax and Amin;
– band pass, with parameters ωp+, ωp- ωa+, ωa-, Amax and Amin (see Figure 5.3);
– band reject, with parameters: ωp+, ωp-, ωa+, ωa-, Amax and Amin.
In instrumentation systems, band pass filters are most widely used (for anti-
aliasing, improvement of signal-to-noise ratio, and for reconstruction). However,
band pass and band reject gauges are also used when a major amplitude disturbance
must be eliminated before conversion into digital signals.
In all cases, the usual method of calculating is to determine, from the attenuation
gauge, the transfer function of a band pass filter which is called the prototype. From
this we deduce, by conversion, the transfer function of the filter being planned. The
synthesis is then carried out from this function. We see that the transfer function of
the prototype is always normalized; that is, it is calculated by taking the cut-off
pulsation p as unity.
172 Fundamentals of Instrumentation and Measurement
There are four types of important and well identified attenuation functions of
prototype filters. These give rise to transfer functions that can be physically
produced and verified:
– Butterworth filters:
2
H −1 ( jω ) = 1 + ε 2ω 2n
2
H −1( jω ) = 1 + ε 2Tn2 (ω )
Tn (ω ) = cos (n a cos (ω ))
In these relations, n is the order of the filter, that is, the degree of its transfer
function. i is a parameter depending on attenuation tolerance in the pass band Amax:
ε 2 = 10 A max/ 10 − 1
For these two types of filters, the inverse of the transfer function is a
polynominal of variable . They are thus called polynomial filters. The attenuation
curve of Butterworth filters is said to be maximally flat because all the attenuation
derivations are zero at pulsation 0. However, the pass band response of Tchebycheff
filters fluctuates n + 1 times between the values 0 and Amax (in dB). These are
called equiripple filters.
In order to meet the requirements of a given filter, the necessary order n is much
higher for a Butterworth filter than for a Tchebycheff filter. Polynomial filters are
easy to make because their transfer function is simple.
( )
n/2
2 2
∏ ω − ω0i
2 i= 1
H −1 ( jω) = 1 + ε# 2 if n is even
( )
n/2
2 2
∏ ω − ω∞ i
i= 1
Analog Filters 173
( ω2 − ω02i )
n −1/ 2
ω2 ∏
2 i =1
H −1 ( jω) = 1 + ε# 2 if n is odd
∏ ( ∞i )
n −1/ 2
2 2
ω − ω
i =1
ωa
with: ω0i ω∞i =
ωp
where 0i and ω∞i are respectively normalized pulsations for which the attenuation is
zero and infinite and ε is a constant. The numerical values of these parameters can
be calculated analytically, as Cauer has shown. However, nowadays it is simpler and
more efficient to calculate these values by direct digital optimization; the existence
and unicity of the solution have been shown, leading to a rapid and certain
convergence of the algorithm for the calculation.
The attenuation curve of Cauer filters fluctuates n + 1 times between the extreme
values allowed by the gauge. This is true both for pass bands and attenuated pass
bands. Of all filters, these meet the requirements of a given gauge with a transfer
function of minimal order n. Unfortunately, their group propagation time is
extremely irregular, which means they cannot be used when the temporal form of a
signal must be preserved.
2 1
H −1 ( jω) =1 +
⎛1⎞
ε 2Tn2 ⎜ ⎟
⎝ ω⎠
Tn ( ω) = cos ( n a cos ( ω) )
ε is a scale constant.
These filters are among the best available in terms of stiffness of the attenuation
curve and pass band group delay regularity. These properties make this kind of
filters very useful in instrumentation systems.
174 Fundamentals of Instrumentation and Measurement
Attenuation functions define the square of the module of the transfer function. A
standard, but sometimes difficult to realize, mathematical calculation helps us
deduce the transfer function. It is carried out as follows, by noting the transfer
function H(p) = P(p)/E(p):
E ( p ) E (− p ) F ( p) F (− p) 2
H −1 ( p ) H −1 (− p) = =1+ = H −1 ( jω )
P( p ) P(− p ) P( p) P(− p )
E ( p) E (− p ) = P( p ) P(− p ) + F ( p) F (− p )
Resolving the Feldtkeller equation requires a computer (there are many software
programs available that can do this, such as Matlab and Mathcad), except for
Butterworth filters, which allow for a simple analytic solution:
n/ 2⎛ ( 2i − 1) π 2 ⎞
E ( p ) = H −1 ( p ) = ε ∏ ⎜ p 2 + 2δ p cos + δ ⎟ if n is even
i =1 ⎝ 2n ⎠
( ) ⎛
n − 1 / 2
iπ ⎞
E ( p ) = H −1 ( p ) = ε ( p + δ ) ∏ ⎜ p 2 + 2δ p cos + δ 2 ⎟ if n is odd
i =1 ⎝ n ⎠
with δ = ε ( −1/ n )
ω0
– low pass high pass transformation: p →
p
In these formulae, 0 is the central pulsation of the band pass or band reject,
expressed in radian/second, with B the width of the band pass or band reject, in the
same unity. These transformations are complicated and difficult to carry out,
necessitating the use of a computer. Many commercial software programs contain
these transformations (Matlab, Mathcad and Matrixx are examples, as are most
signal analysis programs).
Two synthesis methods of an analog filter, done from transfer functions, are
widely used.
Cascade synthesis
This is based on the decomposition that can always occur of H(p) in biquadratic
terms (and of a first degree term when order n of the filter is odd):
α m p m + α m −1 p m −1 +...+ α 0
n/2
n/2 ai' p 2 + bi' p+1
H ( p) =
β n p n + β n −1 p n −1 +...+ β 0
=G ∏ a p 2 + bi p +1
∏
=G Bi ( p)
=
i 1 i i 1 =
if n is even, and:
n −1 n −1
2 2 2
a0' p +1 a' p + bi' p +1 a ' p +1
H ( p) = G
a0 p +1 ∏ aii p 2 +bi p +1 = G a00 p +1 ∏ Bi ( p)
i =1 i =1
if n is odd.
To construct this kind of filter, we need to cascade basic biquadratic circuits with
transfer function Bi(p) (and an additional first order circuit when n is odd). Here we
176 Fundamentals of Instrumentation and Measurement
must ensure that no circuit interacts with another (see Figure 5.4). This condition is
met if the output impedance of each basic circuit is zero, or very low in comparison
to the input impedance of the next circuit.
B1 B2 Bi
V1 V2
V2
H ( p) = = B1 ( p)A × Bi ( p)
V1
Each biquadratic element depends on only four parameters. These are ai, bi, a’i
and b’i, and can be created by standard circuits that are easy to adjust individually
and available in the form of hybrid or integrated modules from many manufacturers.
The best way to adapt these elements to specific needs is to construct them, with the
help of resistors, capacitors and operational amplifiers, using methods which we will
describe later. These methods are very easy to implement, both in terms of
calculations and from the point of view of adjustments and maintenance. They are
universally applicable and work no matter which transfer function needs to be
synthesized. Unfortunately, this structure does not have a good sensitivity in relation
to component value variations. In particular, it does not allow us to construct narrow
band pass filters (< 5%) as soon as the frequency exceeds several tens of kHz.
Comprehensive synthesis
This method involves synthesizing the filter with one network in order to
minimize sensitivities. Calculations, adjustments, and optimizations are much more
complex. Since L-C filters have an optimal sensitivity (see section 5.4), most of
these methods directly or indirectly stimulate L-C structures.
Analog Filters 177
Passive filters using inductors L and capacitors C were first developed in 1923.
These are called Zobel filters [ZOB 23]. They are made of a quadrupole that only
contains this type of element, inserted between two resistors R1 and R2, respectively
the generator resistor and the charge (Figure 5.5).
R1
L-C V2
E1 R2
Quadripole
However, these filters require the use of inductors, components that are costly,
take up much space, and cannot be adjusted for industrial purposes. This is why, in
mass production, L-C filters have been increasingly replaced by active filters, at
least for in low and medium frequency applications. However, the technologies
involved in manufacturing inductors have improved in recent years. Miniaturization,
quality and costs have continued to improve. Laboratories have created higher
quality inductors that can be integrated and electronically regulated.
Moreover, the very low sensitivity of L-C filters cannot be matched. Therefore,
we use these filters as models to produce “electronic copies” with the same
sensitivity features, but without inductors. These techniques will be discussed in
sections 5.4.2 and 5.4.3. In practice, the only structures that are still used are ladder
filters, as shown in Figures 5.6 and 5.7.
178 Fundamentals of Instrumentation and Measurement
The basic schema of an L-C filter is shown in Figure 5.5. The voltage generator
E1 of internal impedance R1 can supply charge R2 with a maximum power
2
P1m = E1 / 4R1 .
2 2
P1m E1 R2 R2 E1 2
= = = H −1 ( j )
P2 4 R1 2 4 R1 V 2
V2
E1 R2
with: = H -1 ( p )
V2 4 R1
A( ) = 20 log H −1 ( j )
Let Xi be the value of an element of the filter (inductor or capacitor), and ∂A/∂Xi
the partial derivation of A in relation to the value of this element. If Xi diverges from
its nominal value by a quantity Xi (because of temperature variation or a time drift,
for example), the corresponding variation of attenuation A moves near the first order
by means of:
∂A
A( X i + X i ) = A( X i ) + X i
∂X i
Xi
∂A
(
= A Xi + Xi ≥ 0 )
∂X i
Analog Filters 179
Since variation Xi is a rather vague sign, the relation can only be verified with
an equal sign. From this we get the following theorem: the partial derivation of the
L-C filter attenuation inserted between resistors in relation to the value of each of
the elements is annulled to the attenuation zeros.
However, active filters do not have the advantages of this property. That is why,
at this time, constructing a high performance filter using an L-C prototype must be
done by electronic simulation. We must remember the restrictive conditions of
Orchard’s argument when carrying out this procedure:
– it is not valid except for the pass band (and not for the stopband where
admissible tolerances for attenuation are notably higher);
– it only applies if attenuation zeros correspond to the transmission of maximum
available power. Filters taking in a constant non-zero pass band attenuation are
therefore excluded;
– it applies only to attenuation and not to propagation delay.
In practice, most L-C filters have a ladder structure. This topology avoids the use
of transformers and gives better results. From a given gauge, we must establish the
schema and calculate the value of the elements of this schema.
The transmission zeros of Cauer and low pass inverse Tchebycheff filters are
produced by trapped circuits in series in parallel branches. Only odd order filters can
be produced by L-C technology (the explanation for this is given later). We thus
have one schema in T (and its equivalent in ) as shown in Figure 5.7.
The schemata shown help produce standard low pass filters, such as the
Butterworth filter, the direct and inverse Tchebycheff filter, and the elliptic filter.
These give excellent results in most practical applications.
The schema being established from the transfer function, we need only calculate
the value of the elements. This calculation can be carried out analytically following
Darlington’s decomposition method, or it can be done directly by modern methods
of digital optimization, if the necessary software is available.
b) By knowing polynomials E(p), P(p), and F(p), we can deduce these L-C
quadrupole impedances: z11, z22, y11, and y22. The results of this calculation are given
in Table 5.1. Indices p and I indicate the even and odd parts of polynomial E(p) and
F(p). This means Ep indicates the even part of polynomial E(p). These results have
been established from Darlington’s original calculation [HAS 81].
c) From any one of these four impedances, we calculate the successive values of
the branch impedances by a procedure of iterative extraction. For polynomial filters,
this procedure is very simple; it is only a development of continued fractions of the
initial impedance. For example, for a structure in T shown in Figure 5.6, it is:
1
z11 = L1 p +
1
C2 p +
1
L3 p +
C4 p + ........
Structure in Ei + Fi Ei −Fi E p + Fp E p + Fp
R1 R2 R1 R2
Π Ep − Fp E p −Fp Ei − Fi Ei + Fi
Structure in Ei − Fi Ei + Fi E p −Fp E p − Fp
R1 R2 R1 R2
T E p + Fp E p + Fp Ei + Fi Ei − Fi
This decomposition can occur element after element by making p tend towards
infinity:
L1 p = z11 p → ∞
1
C2 p =
z11 − L1 p p → ∞
etc.
Analog Filters 183
For elliptic and inverse Tchebycheff filters, the procedure is a bit more
complicated. This is because the frequency of L-C branches must be positioned to
their values, that is, to frequency pulsations ω∞i, which annul P(p) and correspond to
a transmission zero. For example, for the structure in T shown in Figure 5.7, we get:
1
z11 = L1 p +
1
Y2 +
1
L3 p +
Y4 + ........
Ci p Ci p
with: Yi = =
2
1 + Li Ci p p2
1+ 2
∞i
L1 p = z11 p 2 → − 1/ L C
2 2
1
Y2 = − Yr
z11 − L1 p
etc.
We see that in all cases, we must use a calculator to carry out these operations
precisely in order to avoid cumulative errors at each stage. We illustrate this
mechanism with two examples.
For filters with attenuation equal to Amax at zero frequency, and infinite at
infinite frequency (such as direct even order Tchebycheff filters), the resistor of the
generator is taken as unity and the terminal resistor is calculated in the following
way. By writing that the attenuation at = 0 is obtained by bridge divider R1, R2,
the L-C filter acting as a short circuit between the two resistors is expressed as:
Pm E 2 V22 (R1 + R2 )2
= 1 =
P2 4 R1 R2 4 R1R2
184 Fundamentals of Instrumentation and Measurement
⎡ (R + R )2 ⎤
A max = 10 log ⎢ 1 2 ⎥
⎢⎣ 4 R1 R2 ⎥⎦
R2 = (2α − 1) ± (1− 2α )2 −1
In this relation, the sign + is used for structures in T and the sign – for those in .
P( p) = 1
Since F(p) is even, F(p) =Fp and Fi = 0. This gives us z 11 = y22 and z22 = y11.
Ei = 15.115 p5 + 19.57 p3 + 5 p
By carrying over the values in Table 5.1, we get parameters z and y of the filter
in normalized form (R1 = 1 for the calculation of z11 and y11 and R2 = 1 for the
calculation of z22 and y22):
In order to obtain the real values of the components, we must determine the
unitary values of the capacitor and the inductor, after having determined an
impedance level by choosing resistor R1, which we take as unity. For example, if we
choose R1 = Runitary = 50 , we deduce R2 = 133 , and from relations RuCu u = 1,
and Lu u = Ru:
1 Ru
Cu = = 3.18nF and Lu = = 7.958µ H
2π fu Ru 2π fu
Figure 5.8. Schema of a low pass filter calculated in Example 1. The resistors are in Ohms
Example 2
We can construct a low pass elliptic filter of order 5, Amax = 0.2 dB and
Amin = 40 dB, with a cut-off frequency of 50 kHz. The structure is in T. The
calculation follows the method presented in section 5.3.2 and gives characteristic
polynomials E(p), F(p), and P(p) in normalized form.
P ( p ) = p 4 + 6.12 p 2 + 8.14
In carrying over the values in Table 5.1, we get parameters z of the filter (if we
take R1 = 1 and R2 = 1):
or L1 = 0.8965
1 0.8179 p
Y2 = − Yr =
z11 − L1 p 1 + 0.5115 p 2
1 R
Cu = = 3.183 nF and Lu = u = 3.183 mH
2π fu Ru 2π fu
Eventually, the schema for creation shown in Figure 5.9, where we see that real
component values have been obtained by multiplying the normalized values by the
unitary values.
For elliptic and inverse Tchebycheff filters, the variable is also a vector of n
values of capacitors and inductors; that is, a value by branch. For branches with a
tuned circuit, the value of the second element is calculated based on knowing
transmission zeros ω∞i of the transfer function, which means of the zeros of P(p):
L j C jω ∞2 i = 1
This procedure is shown by the direct numerical calculation of the two same
filters that were previously calculated by the analytic method, using Matlab software
and its “optimization toolbox”:
Step 1: we have the initial values of the components. The unity value is always a
good choice for L and C elements. The generator resistor is taken as equal to unity.
The terminal resistor is calculated as shown above.
Step 2: we choose m pulsation values, for which we calculate the filter
attenuation. Between 100 and 1,000 values included between normalized pulsations
0 and 5 constitute a correct order of magnitude.
Step 3: for these m values, we calculate the filter attenuation with the help of a
function written for this purpose.
Step 4: for these m frequency values, we calculate the ideal attenuation from the
transfer function.
Step 5: we calculate the distance between these two series of values with a
quadratic criterion.
Step 6: we minimize this distance by using a universal optimization function
made by Matlab. Here, we use the “constr” function which allows us to introduce a
positivity constraint on the element values. This helps us to avoid arriving at a
solution that cannot be put into practical use.
The calculation takes only several dozen seconds working with a Pentium III and
much less time at a workstation. We usually get a correct convergence. We see that
if the calculation is not carried out very precisely, we can get a value set quite
different from the values obtained by analytic calculation, even if the filter has a
satisfactory response. This fact proves the weak sensitivity of L-C filters, since
equivalent results can be obtained with quite a different value set.
For an elliptic filter, the procedure is almost exactly the same, with one
difference. It is necessary in this case to calculate the values of the second resonant
branch elements from the initial values.
Analog Filters 189
Example 1
The same Tchebycheff filter of order 6 calculated by this method gives, after
1,800 iterations, the following normalized values (without unities):
We notice that the values are more or less different from those calculated by
Darlington’s exact method, even though the precision required by the algorithm of
the calculation was very high However, the difference between the response curves
is undetectable.
Example 2
The elliptic filter of order 5 calculated by the direct numerical method gives the
following normalized values (without unities) after 800 iterations:
Even if the values are quite different from the exact values, the response curves
differ by less than a thousandth of a dB.
When the filter is not a band pass, we first calculate the transfer function of the
corresponding low pass prototype. Then, we carry out the synthesis of this
prototype. The last step is deducing the final schema by applying the frequential
transformations shown in section 5.3.3 to the impedance values of the low pass
schema. In this way we get the schema of the transformed filter, with element values
expressed in real unities. This transformation is simple to effect.
Example
Suppose we want to produce an elliptic band pass filter of order 10 with
Amax = 0.2 dB and Amin = 40 dB. Its central frequency is of 100 kHz and its band
width B = 20%, that is, 20 kHz.
The low pass prototype of this filter is exactly the filter calculated in Example 2
of the above section (see Figure 5.9). Transformations of low pass impedances s
band pass are seen in Figure 5.10. Each impedance is converted into a capacitor in
190 Fundamentals of Instrumentation and Measurement
parallel with an impedance. In this schema, the elements appear with normalized
values.
0.3196 0.95
3.13 1.05
1
5.7
4.09
0.2445 0.175
Figure 5.10. Schema of a band pass filter obtained by conversion of the low pass prototype
shown in Figure 5.9. The elements are in normalized values (without unities)
The schemata obtained by the method just described are not always compatible
with the technological constraints presented by capacitors and, most importantly,
inductors. We can then obtain schemata derived from network transformations.
Here, there are many possibilities, such as the Norton transformation and the
Star/Triangle conversion. Implementing them is a delicate task and needs to be done
by an engineer. These techniques are described in [HAS 81]. We present here one
very common example of conversion, universally used in creating band pass filters.
This consists of replacing the four elements of parallel branches with two resonant
circuits in series. We then have a schema similar to that shown in Figure 5.11, but
less sensitive to disturbances and with values that are less dispersed. In this schema,
the values of elements are in real unities, by taking a value of 1 k as the generator
resistor.
Analog Filters 191
Optimization
The calculations presented above are carried out by supposing that the capacitors
and resistors have no losses. In reality, even if this hypothesis is generally valid for
capacitors, it is not as valid for inductors whose coefficients of quality Q are limited.
The response curves diverge so much from ideal curves that these losses become
significant.
Active filters are made of capacitors, resistors and active elements (almost
always operational amplifiers). Less bulky, easier to produce, and thus less costly,
active filters are useful when frequencies are not too high (typically up to several
MHz). Nevertheless, active components introduce noise, limiting the maximum
voltage that can be filtered, and requiring a power supply.
Active filters are usually produced by putting basic second order cells in cascade,
as shown above. Simple and very widely used, this method does have the limitation
of producing filters that are very sensitive to imprecisions or variations in
component values. To avoid this problem, we also make active filter copies of L-C
filters, following several very good methods. The following sections will present
these different approaches.
192 Fundamentals of Instrumentation and Measurement
Their general schema is given in Figure 5.12. Quadrupoles (RC)1 and (RC)2 are
respectively inserted into a positive and negative reaction loop. In order for this to
be cost-effective, quadrupoles (RC)1 and (RC)2 must not have more than a minimal
number of capacities, at most two or sometimes three. From there, we have three
circuit families, each one having approximately the same qualities.
V2 − R2 C p
=
V1 R1 R2 C 2 p 2 + 2 R1 Cp +1
1 R2 1
Q= and ωo =
2 R1 C R1 R2
Figure 5.13. Biquadratic band pass cell in negative reaction (Rauch cell)
C1 C2
Q=
2C2 + (1 − K ) C1
1
o = with K ≈ 1
R C1C2
The major advantages of these simple structures are their low number of
components and reduced energy use. But there are many disadvantages. Some are:
– difficult adjustments and settings (not independent for Q and 0);
2
– high dispersion of component values (proportional to Q );
– high active sensitivities (proportional to Q2);
– structures specific to a response type.
Manufacturers have presented several schemata for these cells. All are based on
the theory of state variables. According to this theory, it is always possible to
decompose a transfer function of the order n in an ensemble of n functions of the
first order that have been simulated by integrators and combined with adders-
subtractors. As an example, let us look at the high pass function of the second order
with a transfer function as follows:
V2 a' p 2
=
V1 ap 2 + bp +1
a' b V2 V2
V2 = V1 − −
a a p ap2
This equation can be carried out by the analog circuit shown in Figure 5.16,
which has two integrators and an adder.
Figure 5.16. Principle of producing a biquadratic state variable high pass transfer function
196 Fundamentals of Instrumentation and Measurement
Several schemata derived from this principle have been developed. We will
present two of them, both of which are widely used and important.
V3 2 Q −1 R 2C 2 p 2 1
= with RC =
Ve Q 2 2 2 RCp
R C p + + 1 ω0
Q
These cells are sold by several manufacturers under the name “Universal
Filters”. A fourth amplifier is also sold along with these cells, allowing the user to
adjust transmission zeros using a technique we will discuss below.
Tow-Thomas cell
Because of a small modification generalized in the Tow-Thomas cell (Figure
5.18), amplifiers have their positive input linked to the mass. By supposing three
Analog Filters 197
amplifiers, whose gain in open loop is A >> 1, to be identical, the transfer function
of this kind of cell is written:
V2 1 1
= with ωo ≈
V1 RC RC
R2 C 2 p2 + p +1
Q
We simultaneously get a band pass output at point V3. This circuit has excellent
qualities, low passive and active sensitivities, and gives the possibility of producing
circuits with very high quality coefficients Q.
This is why component manufacturers sell KHN cells with an amplifier, which
makes summation possible. These cells are very easy to use, in addition to being
adaptable and affordable.
198 Fundamentals of Instrumentation and Measurement
V2 a ' p 2 + b' p + c
=
V1 ap 2 + bp +1
a a a a
R= a R1 = R2 = R3 = Q=
a a ' − b' a 'b c' b
Analog Filters 199
5.5.4. Elevated order active filters (elevated by putting biquadratic cells in cascade)
The principle of these cells has been shown in section 5.3.4 (Figure 5.4). In order
to minimize noise and maximize the dynamic, three precautions must be taken:
– we must ascertain that the order used to put the cells in cascade are such that
the cells with high overvoltage coefficients are placed near output. This preserves
the filter dynamic;
– for filters with transmission zeros, we should ensure that the poles and zeros
of biquadratic functions are optimally matched. Practically speaking, we must often
be satisfied with linking the closest pair of zeros to a pair of poles;
– we must ensure that the gain repartitions of the different biquadratic cells
preserves the ensemble dynamic. A good way to do this is to equalize all cell
responses.
It is important to note that this type of synthesis is limited to certain filters. These
are:
– low pass and high pass filters with a degree not exceeding eight to ten;
– band pass filters with a band width not much below 20%.
200 Fundamentals of Instrumentation and Measurement
If these limitations are not considered, the synthesis may produce excess
sensitivity to component value variations.
Orchard has shown that L-C filters inserted between two resistors have the best
possible sensitivity. This means that an active filter with low sensitivity is produced
by “copying” an L-C filter, which is used as a model. Among the solutions proposed
for this, three methods yield a workable way of doing this, and are easy to
implement. All three techniques produce a filter with a sensitivity considerably
lower than their homologues using the cascade technique.
FDNR-based simulation
Another method for simulating L-C filters is based on an impedance conversion
called the Bruton transformation. This transformation suppresses the inductors while
activating the Frequency Dependent Negative Resistor (FDNR). This element is an
electronic ensemble that converts the capacitor impedance into a negative resistor
whose value is inversely proportional to the square of the frequency. The best FDNR
schema is obtained with the help of the GIC shown in Figure 5.21, in which Z1 and
Z are capacitors of value C and the other impedances are resistors of value R. The
−1
input impedance of this mechanism is then Z e = . Here, we are describing
RC 2ω 2
a negative resistor dependent on the frequency.
Analog Filters 201
V Z Z
Figure 5.21. Generalized impedance converter (GIC): Ze = 1 = 1 3 Z
I1 Z2 Z4
The overall response is not affected by this transformation. This method works
particularly well with low pass filters (see Figure 5.23), if we begin with a schema in
T in which resonant circuits are serial circuits, in parallel branches. As with
202 Fundamentals of Instrumentation and Measurement
Figure 5.23. Low pass L-C filter (a) and a filter converted by
Bruton’s method (b) producing two FDNRs
Producing a filter requires very precise and stable components. This condition is
incompatible with the technological constraints of integrated circuits. An ingenious
and very useful device invented by Friend in 1972 [FRI 72] avoided this problem
and produced excellent filters that were completely integrated and required no
adjustments. Unfortunately, at the time of writing this type of filter can only be
produced at frequencies lower than a few MHz.
Analog Filters 203
The basic assembly, shown in Figure 5.42, consists of replacing the resistors of
an active filter with an assembly that only contains capacitors and interruptors that
alternately open and close to the rhythm of a clock of period T = 1/fh. Each T period
decomposes in two non-overlapping phases 1 and 2. If this mechanism is
connected to two voltage sources V1 and V2, we can write that a charge Q is
transferred from the output source at each period: Q = C1(V2-V1).
During time t >> T, the transferred charge and the mean current are:
t
Q= C1(V 2 - V1)
T
Q C
I mean = = 1 (V 2 - V1) = f h C1(V 2 - V1)
t T
- 1 T
R equivalent = V 2 V1 = =
I mean .
C1 f h C1
This resistor of value T/C1 can be easily integrated, since MOS transistors can be
excellent analog interruptors. This mechanism becomes extremely useful when it is
used as the resistor of an analog integrator (see Figure 5.25). The corresponding
transfer function is:
V2 −1 C
= =− 1
V1 ReqC2 P TC2
C2
Φ1 Φ2
V1 C1
V2
The time constant of this integrator only depends on the relation of capacities C1
and C2, and not on their individual value. If these elements are produced on the same
substratum, this relation depends solely on the relation of their physical surfaces,
which can be established by construction, with excellent precision (of the order of
0.1%).
Analog Filters 205
Since it is possible to make active filters that use only integrators (the KHN cell
is one example), we can create very precise integrated filters without carrying out
adjustments. Another advantage is that the time constant depends on the clock
period T, so we can modify cut-off frequencies by digitally programming the clock.
In practice, this is a very helpful property.
As such, the basic assembly cannot be used in an integrated filter. Actually, the
values of C1 and C2 must be very low (lower by several pF) in order to conserve the
silicon surface. This means that it is of crucial importance that inevitable stray
capacitances, which have values that are difficult to control, should not influence the
charge transfer. We can achieve this by using an integrator with four interruptors
(see Figure 5.26). This assembly is insensitive to the principles of the stray
capacitances of MOS interruptors. It can be used with two phasings of separate
clocks. The following steps describe two different phasings:
– interruptors 1 and 4, as well as 2 and 3, are simultaneously activated. The
transfer function is that calculated previously. We will call this a “type 1” integrator;
– interruptors 1 and 3, as well as 2 and 4, are simultaneously activated. The sign
of the transfer function is reversed, because capacitor C1 is re-set before ceding its
charge to C2:
V2 1 C
= = 1
V1 ReqC2 P TC2
C2
C1
Φ1 Φ1
V1 Φ2
Φ2
V2
This dual possibility facilitates filter synthesis by avoiding the use of inverters.
We will see that, strictly speaking, these two circuits do not share the same transfer
function in z, which is an added advantage.
C2
C1
Φ1 Φ2
V1
Φ2 Φ1
V2
Attempting to overcome this problem, we see that, aside from input nodes,
currents are zero and voltages are completely stationary, except for moments of
interruptor switching. If we restrict our observation of circuits to instants that
immediately follow these transitions, we can write equations from divergences that
create the charge conservation between two consecutive clock periods. From this we
can deduce a transfer function in z, allowing for a rigorous analysis of the circuits.
For this method to be effective, we must block the input voltage on each T period in
order to make it stationary as well.
Analog Filters 207
V2 C 1
C2 V2 (1 − z −1 ) = − C1V1 or: =− 1
V1 C2 1 − z −1
V2 C1 z −1
C2 V2 (1 − z −1 ) = C1 V1 z −1 or: =
V1 C2 1 − z −1
Synthesis of biquadratic cells is done only from state variable cells. Only these
cells use nothing but adder-subtractor integrators. As for the last two applications,
they use all the operational simulations of L-C filters, in order to take advantage of
the low sensitivity of these filters. This method will be described in the next section.
An L-C filter is a ladder filter with impedances of branches in parallel. These are
noted as Zi, with the admittances of serial branches Yj (Figure 5.27). We can express
recurrently the node voltages and branch currents as follows:
I1 = (V0 − V2 ) Y1
V2 = ( I1 − I 3 ) Z 2
I 3 = (V3 − V4 ) Y3
V4 = ( I 3 − I 5 ) Z 4
... = ...........
I1 I3 I5 I7
V2 V4 V6
Y1 Y3 Y5 Y7
V0 Z2 Z4 Z6 Vn
-1 -1
V2 I3
I1
E1 Y1 Z2 Y3 Z4
-1
To illustrate this method, we have to use a third order filter, knowing that the
process is iterative (Figure 5.29). The iterative equations are modified in order to
introduce terminal resistors and to avoid voltages (we multiply the intensities by a
scaling resistor of arbitrary value R). We then introduce the appropriate signs for the
integrations to end up with a simple schema. We get the following equations:
− 1 Ve − V 2
− V2 = ( − I2)
C1 p Re
+ R 2 (V s − V 2 )
− RI 2 =
L2 p R
− 1 Vs
Vs = ( − I2 )
C3 p Rs
I2
Re
V2
L2
Ve C1 C3 Rs Vs
1/Re
1/Re −1
Ve -V2
C1 p
1/R 1/R
2
RI
R2
L2 p
1/R 1/R
−1
Vs
C3 p
1/Rs
Figure 5.30. Operational graph of the L-C filter shown in Figure 5.29
This set of equations is symbolized by the graph shown in Figure 5.30, where
there are only integrators and adder-subtractors. Integrators 1 and 3 are type 1 and
the central integrator is type 2. The ensemble can be created entirely with switched
capacitor integrators. We see that many interruptors have two uses. After
suppressing the unnecessary elements, we get the final schema, shown in Figure
5.31.
In this assembly, capacitor values C2 and C4 are the same as those in the first
schema. The other capacitor values are established after choosing a clock frequency
of period T and a resistor of arbitrary value R:
T T T L2
C= Ce = Cs = C3 =
R Re Rs R2
Φ2 Ce Ce
Ve
Φ1
C2
Φ2 Φ1
Φ1 Φ2
C C
C3
Φ1 Φ2
Φ1 Φ2
Φ
C C
Φ1 Φ2 C4
Φ2 Φ1
Vs
Cs
Figure 5.31. Switched capacitor filter deduced from graph in Figure 5.30
To show this, Figure 5.32 presents a schema of this type of cell. The interruptors
shown there are in another form which is also widely used. The reference phase is
212 Fundamentals of Instrumentation and Measurement
that which corresponds to the position of the interruptors on the schema. In the next
phase, all interruptors switch. The transfer function of this cell is:
V2 DI + z −1 ( AG − DJ − DI ) + z −2 DJ
=
V1 BD + z −1 ( AC + AE − 2 BD) + z − 2 ( BD − AE )
These switched capacitor filters perform very well from the point of view of
flexibility and complete integration possibilities. However, we should remember that
their applications are limited in frequency to several hundred kHz. In addition, the
filters have been sampled and are subject to aliasing. This means they must have a
continuous filter that eliminates this unwanted phenomenon.
5.7. Bibliography
[ALL 78] ALLSTOT D.J., BRODERSEN R.W., GRAY P.R. “MOS swiched capacitor ladder
filters”, IEEE Journal on Solid-State Circuits,12/1978, p. 806-814.
[BIL 80] BILDSTEIN P., Filtres actifs, Editions de la Radio, 3rd edition, 1980.
[CAU 41] CAUER W., Theorie der Linearen Wechselstromschaltungen, Akademische
Vergas, Gesellschaft Becker, 1941.
[DIE 74] DIEULESAINT E., ROYER D., Ondes élastiques dans les solides, Monographies
d’électronique, Masson, 1974.
[FRI 72] FRIED D.L., “Analog sample data filters”, IEEE Journal on Solid-State Circuits SC
7, August 1972.
Analog Filters 213
[HAS 81] HASLER M., NEIRYNCK J., Filtres électriques, vol. XIX, Editions Georgi, 1981.
[HUE 76] HUELMAN L.P., Active L-C filters, theory and applications, Dowden, 1976.
[SAA 58] SAAL R., ULBRICH E., “On the design of filters by synthesis”, IRE Transactions
on Circuit Theory, vol. CT 5, December 1958.
[SED 79] SEDRA A.S., BRACKETT P.O., Filter theory and design, active and passive,
Pitman, 1979.
[ZOB 23] ZOBEL O.J., “Theory and design of uniform and composite wave filters”, Bell
System Technical Journal, January 1923.
This page intentionally left blank
Chapter 6
6.1. Introduction
When carrying out numerical analysis in real time, it is not enough to sample the
signal to be analyzed at a rhythm that complies with Shannon’s criterion. It is also
necessary to calculate at a speed compatible with the flow of the incoming samples.
If, for a sampling frequency fe, we make M basic operations1 between each
sampling, the necessary calculation power is M · fe, expressed in MOPS (millions of
operations per second). This calculation power is a datum of the processor being
used and is a compromise between the sampling frequency and the sophistication of
the chosen analysis. But the lower the sampling frequency, the higher the order of
the anti-folding filter;2 thus it will be difficult to integrate into the numerical system.
The popular technique currently used to solve this issue (over-sampling and
decimation) is discussed at length in this chapter. Part of this discussion will deal
with - analog-to-digital converters using this technique. This will lead us into a
presentation of implanting digital filters in cabled or programmed models (these can
be comb or half-band filters).
In this chapter, we will only discuss “real” signals, that is, those with a physical
existence. It is also quite often a question of limited spectrum signals. These are
signals with a spectrum assumed to be zero outside a band [fmin · · · fmax]. From a
mathematical point of view, limited spectrum “real” signals do not exist. Actually, a
+∞
∫ X(f )
2
df [6.1]
0
f max
∫ X(f )
2
df [6.2]
f min
In real-time, the Nyquist frequency is the minimum frequency that can be used
for signal sampling. The first criterion is Shannon’s criterion, which shows that, for
a limited spectrum signal [0…fmax], we can take3:
The second criterion is more constraining than the first. It is the anti-folding
filter4, and it is indispensable, even if a priori the signal to be analyzed is present at
input. Here, we must always take noise into account. If we consider that the signal at
input is spoiled by a uniform density noise in a band [0···K · fe/2], with K being even,
a sampling at fe will fold down in the band [0···fe/2] the K-1 bands for
[ ( ) ]
nf e A n + 12 f e for − K ≤ n < K . This means the noise power in the band [0···fe/2] is
2 2
therefore K times more significant after sampling than before, degrading the noise-
signal ratio accordingly.
3 We will see in section 6.2.4 that there is another, less constraining version in the case of
band signals limited around a carrying frequency.
4 Or, in symmetrical fashion, the smoothing filter as a analog-to-digital converter, if the
analyzing device has an analog output that functions at the same frequency as the sampling.
Real-time Data Acquisition and Processing Systems 217
It becomes clear that we are not trying to eliminate folding in the band [0···fe/2].
Consequently, there has to be some folding in the band [fmax ··· fe/2]. We assume from
the start that the numerical analysis we will carry out will sufficiently attenuate this
band.
We see that the width of the transition zone of this filter is fe - 2fmax. For a given
attenuation, the more narrow the transition zone, the higher the order of the filter;
and since this is an analog filter, the higher the phase distortion in the band [0···fmax]
will be. By way of an example, we sample an order 9 elliptical filter at 256 kHz, and
do an 18 bit quantification on a useable band signal heard from 0 to 100 kHz. We
understand how difficult it will be to produce such a filter and the stability problems
that will occur over time.
This is why another approach is being used increasingly. In order to limit the
order of the analog filter, we will sample at a frequency significantly higher than
Nyquist’s frequency. This will give us a good flow of samples, which we will reduce
by digital filtering operations,6 called decimations (analyzed in section 6.3.2). We
will see in section 6.2.3 that over-sampling yields a better resolution than an analog-
to-digital converter. This technique is fundamental to - converters.
5 “Sufficiently” here depends, in this case, on the quantum of the analog-to-digital converter
being used. The folded data will be of a lower level than the quantum.
6 These filters, even if they are often of a higher order than their analog equivalent, will be
stable over time and more often will be linear-phase filters.
218 Fundamentals of Instrumentation and Measurement
Te error(t)
However, for a high-amplitude signal, this error presents the speed of a noise
(Figure 6.2). This means we consider the error to be an independent random input
signal called quantification noise. This is characterized by it power spectral density.
This approach is only valid if the quantum is small in relation to the maximum value
of the input signal. Since this value defines the full range of the converter, the
concept of independent quantification of the input signal only has meaning for
converters with a fairly high number of bits.
not depend on the signal frequency, and the spectral density can be considered as
uniform. In order to give a value to this density, and to respect Shannon’s criterion
as much as possible, we limit the input signal frequency to the interval 0 to fe/2. We
then over-extend the uniformity domain of the spectral signal throughout [0···fe/2].
With this hypothesis, the spectral density of the quantification noise power equals:
σ q2 q2 1
Sq = = ⋅ [6.4]
fe 6 fe
2
From here, we establish a reference signal ratio to quantification noise where the
reference signal is an amplitude sinusoid equal to the full range Vref of the converter,
whose frequency is in the band [0···fe/2]:
2
Vref
2
⎛ S ⎞
⎟ = 2 = 6⎛⎜ ref ⎞⎟
V
⎜ [6.5]
⎜ Bq ⎟ q2 ⎜ q ⎟
⎝ ⎠ ref ⎝ ⎠
12
2Vref 2Vref ⎛ S ⎞
q= ≈ ⇒⎜ ⎟ ≈ 3 ⋅ 22 M [6.6]
M
−1 M ⎜ Bq ⎟ 2
2 2 ⎝ ⎠ ref
⎛ S ⎞
10 log ⎜ ⎟ = 1.76 + 6.02 ⋅ M [6.7]
⎜ Bq ⎟
⎝ ⎠ref
6.2.3. Over-sampling
over-sampling is to enlarge the width of the transition zone of the anti-folding filter
that goes through 2 f a (α − 1) . This helps lower the order of this filter and improves the
phase linearity in the useful band. Again, there will be noise folding in the band
[ f a A f a (α − 1)] , but we already know that numerical filtering will eliminate this
problem.
q2 1 q2 1
⋅ → ⋅ [6.8]
6 2 fa 6 α ⋅2 fa
The power noise quantification in the only useful band is reduced by a factor a:
fa
q2 1 σ q
2
q2 1
∫ ⋅
6 α ⋅ 2 fa
⋅ df = ⋅ =
12 α α
[6.9]
0
For a full-scale sinusoid in the band [0···fa], the signal to quantification noise
ratio becomes:
⎛ S ⎞
10 log ⎜ ⎟ = 1.76 + 6.02 ⋅ M + 10 log α [6.10]
⎜ Bq ⎟
⎝ ⎠ref
We can write:
by taking:
1
Me = M + log 2 α
2 [6.12]
0.01
0.008
0.006
0.004
0.002
-0.002
-0.004
-0.006
-0.008
-0.01
0 0.002 0.004 0.006 0.008 0.01
noise
Te
fa fe/ 2
fa
CAN
Over-sampling of an entire factor N consists of, for the first time, inserting N – 1
zeros between each sample. This clearly does not change the spectrum of the
sampled signal (since we have added only zeros), but it does displace the sampling
Real-time Data Acquisition and Processing Systems 223
frequency of the first image of the spectrum in the basic band to that of the Nth
power. A low pass numerical filtering, carried out at the new sampling frequency,
helps eliminate all the intermediary images (Figure 6.6).
sampled signal
fe
fe
L −1
y ( n) = ∑ a i ⋅ x(n − i ) [6.13]
i =0
( L −1) N
y ( n) = ∑ a N ⋅i ⋅ x(n − N ⋅ i) [6.14]
i =0
224 Fundamentals of Instrumentation and Measurement
( L −1) N ⎫
y (n + 1) = ∑ a N ⋅i +1 ⋅ x(n − N ⋅ i) ⎪
⎪
i =0
⎪
( L −1) N ⎪
y (n + 2) = ∑ a N ⋅i + 2 ⋅ x(n − N ⋅ i ) ⎪
⎬ [6.15]
i =0
⎪
B ⎪
( L −1) N ⎪
y (n + N − 1) = ∑ a N ⋅i + N −1 ⋅ x(n − N ⋅ i)⎪⎪
i =0 ⎭
Then the filter initially of length L, working at a sampling frequency Nfe, requires
a power calculation L · N · fe, which is divided into N filters of length (L-1)/N. These
are calculated alternately. The necessary power calculation is therefore only Lfe.
6.2.4. Under-sampling
This does not mean we cannot take a sampling frequency lower than fmax or fmin.
In general, we speak of under-sampling as soon as the sampling frequency is below
2 fmax, this being the minimum frequency for standard sampling.
In Figure 6.7, we give the spectral effect of an ideal sampling of a basic band
signal at limited spectrum.
If we number the “Nyquist zones” that have a width fe/2, we find that:
– the images of positive frequency are in the odd zones;
– the images of negative frequency are in the even zones.
Real-time Data Acquisition and Processing Systems 225
1 2 3 4 5 2N 2N+1
Figure 6.7. Spectra before and after an ideal basic band sampling
If now, for a given sampling frequency, we start from a “high frequency, narrow
band”7 signal (HF-NB) that is located in an odd Nyquist zone, after under-sampling
we find absolutely the same spectrum as in the previous case (see Figure 6.8). In
zone 1 we find same spectral information as in the signal of origin at an interval
(f s f – Nfe). In other words, if the signal frequency is not useful information, the
under-sampling preserves the information. The major under-sampling application
domain is that of transmitting information by amplitude modulation and/or phase, or
carrier frequency. This falls within the field of communication, but some sensors
function according to the same modulation principle as a measurement carrier8.
1 2 3 4 5 2N 2N+1
If the HF-NB is in an even zone (see Figure 6.9), we find, after under-sampling,
the same spectrum, but with a frequency turn-up (f s Nfe – f).
7 This means the band is still lower by half than the sampling frequency.
8 For example, sensors using Foucault currents in measurement applications of thickness and
distance in front of a moving object. The frequency used is linked to the electromagnetic and
geometric properties of the object, and this influences the module impedance and the sensor’s
phase. But the development rapidity of this impedance, and thus the band width of the signal,
depends mainly on the speed of movement.
226 Fundamentals of Instrumentation and Measurement
1 2 3 4 5 2N 2N+1
-fe fe 2fe N * fe
N ⋅ f e < f min ⎫
⎪
⎛ 1⎞ ⎬ ⇒ f e > 2( f max − f min ) [6.17]
⎜ N + ⎟ ⋅ f e > f max ⎪
⎝ 2⎠ ⎭
⎛ 1⎞ ⎫
⎜ N − ⎟ ⋅ f e < f min ⎪
⎝ 2⎠ ⎬ ⇒ f e > 2( f max − f min ) [6.18]
N ⋅ f e > f max ⎪
⎭
We see that under-sampling can lead to using a sampling frequency much lower
than those in the useful field domain. We can thus use an analog-to-digital converter
with a fairly long conversion time. However, the sampling must be “ideal”, which
means it must be very short as concerns the signal frequency domain. In addition,
the sampling clock must have a very weak jitter, always in keeping with the signal
frequency domain. This means that under-sampling requires using sampling
structures that perform, as well as those of “standard” sampling. The importance is
in the fact that the flow of numerical data to be analyzed by the processor may be
low and is not connected to the carrying frequency. To demonstrate this (see Figures
6.10 to 6.15), we present the structure and the different analysis steps of a GSM
receptor functioning by under-sampling of the first intermediary frequency
(69.875 MHz) for a band width of 200 kHz (corresponding to the juxtaposition of
8 channels).
Real-time Data Acquisition and Processing Systems 227
69.875
Figure 6.12. After sampling at 6.5 MHz, zones 1 and 2 and quantification noise
-6.5 0 6.5
-6.5 0 6.5
-0.406 0 0.406
The anti-folding filter is still necessary, even if only to limit the noise band to
[fmin ··· fmax]. The anti-folding filter then becomes a pass band (see Figure 6.16).
We get from this a relation between the sampling frequency, the central
frequency of the signal to be sampled, and the number of the Nyquist zone in which
we find the signal:
even zone ⇒ Z = 2 N ⎫ 4 fc
⎬ ⇒ fe = [6.21]
odd zone ⇒ Z = 2 N + 1 ⎭ 2Z − 1
– gradient converters, both the high resolution and low speed varieties, have
been supplanted by - converters.
In this section we have only touched on “new” points, introducing the idea of
number of effective bits and, most importantly, the - converter, which uses the
more modern techniques of over-sampling and digital decimation.
RSBqmin + D − 1.76
M = [6.23]
6.02
From this point, the converter can be characterized in terms of dynamic or static
non-linearities, missing codes and so on. These features can be obtained by applying
a sinusoid to the converter’s input. This sinusoid must be of very high spectral
purity. We then analyze the incoming samples:
– by histogram, that is, mainly in order to detect the missing code;
–by FFT for non-linearities.
The spectral analysis of an ideal converter gives a single line to the input
frequency13 emerging from a lower limit of quantification noise. With non-linearity,
lines will appear at harmonic frequencies. We then call SINAD the ratio, in power,
of the fundamental line to the highest disturbance lines or of the ground noise. If
there is no distortion, the SINAD is equal to the reference signal-to-noise ratio
defined in section 6.2.2, depending only on the number of bits. If not, we define the
number of effective bits by:
SINAD − 1.76
ENOB = [6.24]
6.02
The non-linearities depend on the input frequency and increase with it. At low
frequencies, the number of effective bits is equal to the number of bits of the
converter; but this decreases at higher frequencies. This means an 8 bit converter can
have an ENOB of 6.7 bits for an input signal of 100 MHz. A general rule is that we
get curves of the type shown in Figure 6.17.
Figure 6.17. Development of the ENOB depending on the input frequency, for two converters
At a higher input frequency, the best ENOB does not necessarily correspond to
the highest number of low frequency bits.
6.3.2. - converters
13 There must be a precise relationship between the sampling frequency and the sinusoid
frequency.
232 Fundamentals of Instrumentation and Measurement
audiodigital technology was this: it used a base Nyquist frequency (44 kHz for a
pass band of 20 kHz) in order to reduce the number of samples to be memorized and
analyzed. This was done with a quantification of at least 16 bits for 90 dB of desired
dynamic and a signal-to-noise ratio of the order of 8 dB, leading to very high order
anti-folding filters. This type of filter is not only tricky to produce; in an essentially
analog technology it cannot be integrated on the same chip as the rest of the digital
part.14 This resulted in higher costs. To overcome this problem, manufacturers have
tried to reduce as much as possible the analog part by transferring most of the
analysis towards digital processes. Over-sampling has provided this solution. Today,
64 the most widely used over-sampling factor, bringing the anti-folding filter to an
order of 2 to 4. However, there are not (or were not) 16 bit converters at 2.8 MHz
(64 times 44 kHz). But we can make use of the fact that over-sampling helps
improve a converter’s resolution (see section 6.2.3). To improve resolution even
more, we proceed to - modulation that helps reject quantification noise at high
frequencies, that is, outside the useful band. We can then significantly lower the
number of converter bits, and even use a 1 bit converter (a simple comparator) that
has the advantage of being perfectly linear. An inherent part of the converter, digital
analysis for filtering and decimation helps return the initial high flow and low
resolution to the Nyquist flow with a resolution of 16, 18, …, 24 or even 28 bits!
That is the range of techniques discussed in this section.
14 The production and purchasing levels reached by the field of audio technology has lead to
the creation of specific integrated circuits.
Real-time Data Acquisition and Processing Systems 233
The transfer functions in z of the integrator and of the time-lag are respectively
1 and z-1. In these conditions, we get:
1 − z −1
(
S * ( z ) = S ( z ) + Bq ( z ) 1 − z −1 ) [6.25]
are independent), but rejects it at higher frequencies. We say that the noise has been
“put into form”. Its power spectral density is obtained by calculating:
2
q2 1 2
⎛π f ⎞ q 1
2
Nq ( f ) = 1 − z −1 = 4 sin ⎜ ⎟ [6.26]
6 fe z = e j 2 πf / f e ⎝ fe ⎠ 6 f
e
fe
2α 2 2
q 2 ⎡ 1 1 ⎛ π ⎞⎤
4 sin⎛⎜ π f ⎞⎟
q 1
∫ ⎝ fe ⎠ 6 f
e
⋅ df = ⎢ − sin⎜ ⎟⎥
6 ⎣α π ⎝ α ⎠⎦
[6.27]
0
q2 1 π 2
⋅ ⋅ [6.28]
6 6 α3
We can express the signal-noise ratio with the reference signal according to M
and g.
⎛ S ⎞
⎜ ⎟ = 3 ⋅ 3 ⋅ 22 M ⋅ α 3 [6.29]
⎜ Bq ⎟
⎠ref 2 π
2
⎝
⎛ S ⎞ ⎛ 3α 2 ⎞
10 log ⎜ ⎟ ≈ 1.76 + 6.02M + 10 log α + 10 log ⎜ 2 ⎟ [6.30]
⎜ Bq ⎟ ⎜ ⎟
⎝ ⎠ref ⎝ π ⎠
So as soon as α > π 3 ≈ 1.8 , the signal-to noise ratio increases much faster with g
than in the case of simple over-sampling, without appearing as noise. We gain 9 dB,
that is, 1.5 bits of resolution each time that g doubles.
(
S * ( z ) = S ( z ) + Bq ( z ) 1 − z −1 )
N
[6.31]
Real-time Data Acquisition and Processing Systems 235
The power spectral density of the noise put into form is then:
2N
⎛ f ⎞ q2 1 [6.32]
N q ( f ) = 22 N sin⎜⎜ π ⎟⎟ ⋅ ⋅
⎝ fe ⎠ 6 fe
Integrating this function is not simple, and we must be content with a first order
development. This is reasonable because we integrate from 0 to fe/2α, a field for
which f << fe, since we over-sample from a significant factor.
2N 2N
⎛ f ⎞ f
sin ⎜⎜ π ⎟
⎟ ≈π [6.33]
⎝ fe ⎠ fe
We then get:
⎛ S ⎞
⎜ ⎟ = 3 ⋅ 2 N + 1 ⋅ 22 M ⋅ α 2 N +1 [6.34]
⎜ Bq ⎟
⎠ref 2 π
2N
⎝
⎛ S ⎞ ⎡ α2N ⎤ [6.35]
10 log ⎜ ⎟ = 1.76 + 6.02 M + 10 log ( α ) + 10 log ⎢( 2 N + 1) ⎥
⎜ Bq ⎟ π2 N ⎦⎥
⎝ ⎠ ref ⎣⎢
16 That is, the gap between the number of effective bits and the number of converter bits used
in the modulator.
236 Fundamentals of Instrumentation and Measurement
We thus have a significant flow of quantified data with a low number of bits,
representative of a basic signal frequency and a significant noise level; rejected
beyond, however, the useful band of the signal (Figure 6.23).
Spectral density in dB
Digital filtering suppresses this high frequency noise in order to conserve only
the useful band [0A fe 2α ] . Since at the end of this digital filtering, there is nothing
left in the band [ fe 2α A fe 2] , we can only take a sample on α without losing
Real-time Data Acquisition and Processing Systems 237
information. This brings the sampling frequency to the Nyquist frequency. This
ensemble of operations, filtering and lowering the sampling frequency, is called
decimation. This process will be discussed in detail.
2 ⎛ 1 ⎞ f
2M + 1 = log⎜ ⎟⋅ e [6.36]
3 ⎜ 10δ pδ a ⎟ ∆f
⎝ ⎠
where:
– ha and hp are undulations in attenuated bands and pass bands;
– f is the width of the transition band of the filter.
Looking at the example in Figure 6.23, we see that we want to bring the noise
level to that in the pass band (–100 dB is the “resolution” of a 16 bit converter) for a
transition width of 22 – 20 = 2 kHz. This means we need an attenuation of 110 dB.
We can choose an undulation in the pass band lower than ½ quantum. Let this be
152 µV (16 bits, full range 10 V). The sampling frequency is 64*44 kHz = 2.816
MHz. All this leads to a filter order of around 13,000. With a processor capable of
analyzing 200 million operations per second, at 2.816 MHz, we cannot produce a
filter of an order above 70. This means we have to proceed in several steps, by
beginning with a filter that can be made by using a cabled structure: the comb filter.
The comb filter is an FIR filter, with all coefficients equal. It is therefore a linear
phase filter, and there is a recursive way of expressing it. For a filter of L length, we
get:
k = L −1 k = L −1
1 − z −L
h ( z ) = h0 ⋅ ∑ z −k = h0 ⋅ 1 − z −1 ⇒ y(n) = h0 ⋅ ∑ x(n − k ) [6.37]
k =0 k =0
The response in frequency of this filter is given by the following equation and is
shown in Figure 6.24.
⎛ L ⎞ ⎛ fL ⎞
L
− jω T sin⎜ ω T ⎟ sin⎜⎜ π ⎟
2
⎝ 2 ⎠ ⇒ h( f ) = h ⋅ ⎝ f e ⎟⎠
h( jω) = h0
e
[6.38]
sin (ωT )
0
L
− jω T ⎛ f ⎞
e 2 sin⎜⎜ π ⎟
⎟
⎝ fe ⎠
238 Fundamentals of Instrumentation and Measurement
Its continuous gain is therefore h0 L. For a unitary gain we take h0 = 1/L. This
filter carries out the arithmetic mean of the last L input samples. If L = 21, the
division by L in turn performs l intervals to the line of the accumulated total of the
samples. Under these conditions, the filter can be made with the help of simple
accumulators and registers, that is, by means of a hard-wired logic that can function
at a very high sampling frequency.
10
-10
-20
-30
-40
-50
-60
1/16 2/16 3/16 4/16 5/16 6/16 7/16 8/16
1 − z −L
L
1 − z −1
1
1 − z −L L
1 − z −1
1 −1
L 1− z
1 − z −1
Figure 6.25. Three equivalent ways of producing comb filtering and one decimation
Real-time Data Acquisition and Processing Systems 239
To enlarge the zones of high attenuation around all the fe/L areas, we increase the
filter order by cascading the accumulators17 and differentiators before and after
decimation (Figure 6.26). The frequency response for an order 4 is compared to that
of an order 1 in Figure 6.27.
N integrator
N differentiator
-50
-100
-150
1/16 2/16 3/16 4/16 5/16 6/16 7/16 8/16
Figure 6.27. Response in frequency (in dB) of comb filters of 16 lengths and
of order 1 and 4, between 0 and fe/2
17 This cascading does not require accumulators with a high number of bits.
240 Fundamentals of Instrumentation and Measurement
This kind of filter is used as the first decimator filter at the output of the -
modulator. It is most frequently of an order 4 and a length 16; this means it has a
decimation of factor 16. This filter does not perform well enough to be able to go as
low as Nyquist frequency. This phase is completed by FIR programmed filters. For
these filters, we must minimize the necessary power calculation.
Looking again at Figure 6.23, we see that the noise remaining after a comb filter
of 16 lengths and of order 4, but before decimation of a factor 16, is given in Figure
6.28. We see that, except for the first two lobes, all the others are below –80 dB. The
decimation will move back all these lobes for a total level of –64 dB. To proceed
further in decimation (there remains a factor 4), we must bring this level to the
required –100 dB. By calculating the order of the necessary filter, we get the
following parameters:
– 36 dB and 152 µV undulation;
– sampling frequency of 176 kHz and transition width of 2 kHz.
18 By assuming that the processor can make a product and an accumulation in one operation.
Real-time Data Acquisition and Processing Systems 241
The decimation, done in two steps with half-band filters, helps reduce the
necessary power calculation.
A half-band filter is really a linear phase low pass filter with a unitary gain in the
pass band, so that H(f) presents an odd symmetry around the point fe , 0.5 : ( 4 )
H ⎛⎜ 4e − f ⎞⎟ = 1 − H ⎛⎜ 4e + f ⎞⎟
f f
[6.39]
⎝ ⎠ ⎝ ⎠
These kinds of filters are of odd order, with even coefficients, apart from the
central coefficient, which is zero. For a given order of a filter, the required power
calculation is two times lower than for an ordinary FIR filter. In addition, a = p.
This kind of filter is especially useful to the decimation of a factor 2, as we see in
Figure 6.29. Since the transition width is relatively large, this leads to a slightly
higher order. Still for the same example:
– for a first order half-band filter:
- ha = hp = 50⋅10-6 (1/3 of the previous undulation because we will cascade
three filters);
- sampling at 176 kHz and transition width (88 – 20) – 20 = 48 kHz;
– for a second order half-band filter:
- ha = hp = 50⋅10-6;
- sampling at 88 kHz and transition width (44 – 20) – 20 = 4 kHz.
This gives filters of order 19 for 111 for the second. So the total power
calculations of 10*176 103 + 56*88 103 ≈ 6.7 MOPS.
Zone folded by
useful band decimation by 2
1
1/2
fa
Fe/4 Fe/2 Fe
Before decimation by 2
1/2
fa
Fe/2 Fe
After decimation by 2
We have to complete the procedure with a standard FIR filter that compensates
for the attenuation in the pass band resulting from the comb filter and carries out a
last filtering in the transition zone. Of an order of around 100, at 44 kHz, it adds
about ten MOPS, which is within the range of all specialized processors.
Today, we find integrated converters20 that have a modulator, digital filters and
decimation capabilities. With the resulting high number of bits, the interface is
serial, often connected to a processor unit.
Specialized processors used in signal analysis are basically conceived for linear
analyses of the convolution type or the equivalent. These can be expressed by:
y= ∑ a i ⋅ xi [6.40]
i
A calculation unity expressed in floating point is more complex, and thus more
costly21; above all, it uses more energy. There are also processors that use both types
of notation, but the ones that use fixed-point notation tend to be preferred when
energy consumption is a concern.
19 For example, HSP50016, HSP43220, HSP43168, etc., from Harris [HAR 95].
20 Often with two in the same circuit, since the audio is stereo.
21 But which especially depends on the production volume.
Real-time Data Acquisition and Processing Systems 243
− b N −1 ⋅ 2 N −1 + b N − 2 ⋅ 2 N − 2 + … + b1 ⋅ 21 + b0 ⋅ 2 0 [6.41]
The most weighted bit is representative of the sign, but there is a weighting since:
− 2 N + 2 N −1 = 2 N −1 ⋅ (− 2 + 1) = −2 N −1 [6.42]
− b N −1 ⋅ 2 N + b N −1 ⋅ 2 N −1 + … + b1 ⋅ 21 + b0 ⋅ 2 0 [6.43]
(− b N −1 ⋅ 2
N −1
)
+ bN − 2 ⋅ 2 N − 2 + … + b1 ⋅ 21 + b0 ⋅ 20 ⋅ 2 − k [6.44]
For a coding in fractional numbers, the dynamic23 is always (2N-1 – 1). The
quantification error (1/2 LSB or 2-k-1) is a constant, and relatively more important for
low values.
b N −1b N − 2 b0
[6.45]
e E −1e E − 2 e0 m M −1 m M − 2 m0
For a reason we will explain below, we exclude the possible values by exposing
the value – 2E-1. We then get:
− 2 E −1 + 1 ≤ exp ≤ 2 E −1 − 1 [6.46]
for a fixed point notation on the same number of bits. For 16 bits having 4 exponents
and 12 of mantissa, the dynamic is 3.35⋅107 in floating notation as opposed to
3.28⋅104 in fractional.
The notation of the mantissa is in QM-2, but one of the most weighted of the 2 bits
is necessarily non-zero:
( )
– the highest positive value is 2 2 −1 ⋅ 2 ( M −1) − 1 ⋅ 2 −(M −2 ) ,
E −1
E −1
– the lowest positive non-zero value is 2 −2 +1 ⋅ 1 .
( )
So the dynamic is 22 − 2 ⋅ 2 − 2−(M −2) . This is around 2(M - 2) times lower than in
E
the previous example. For 16 bits, 4 of exponents and 12 of mantissa, the dynamic is
of 3.28⋅104, practically the same as for fractional notation. In terms of the dynamic,
we have lost the advantage of the floating point, but the quantification error is no
longer a constant. It equals 2exp ⋅2- M + 1. This means that for the relative error, we
have, no matter what the value is:
2− M +1 δx 2− M +1
= 2− M < ≤ = 2− M +1 [6.48]
2 x 1
– For negative numbers, the mantissas between -1 and -2 coded as QM-2 are
written as 10.xxxxxxx.
The exponent on 8 bits is in offset binary of 126. The binary mantissa is signed
on 23 bits, with the first “1” implicit. The notated value is then:
In certain cases:
– 0 if exp = 0 and mantissa = 0;
– (-1)s ∞ if exp = 255 and mantissa = 0;
– NaN (Not A Number) if exp = 255 and mantissa ≠ 0;
– (-1)s·2-126 .(0.mantissa) if exp = 0 and mantissa ≠ 0 (denormalization).
Using a processor in fixed point requires special attention in the area of data
framing, in order to obtain the best precision by avoiding overflow problems. In
floating point, the problem of framing is not of crucial importance. The only
remaining issue is the effect on the final result of the variable quantification.
24 In TMSC3x processors, made by Texas Instruments, the bit of the mantissa sign is kept and
the following bit is omitted.
246 Fundamentals of Instrumentation and Measurement
The examples in this section are drawn from processors of TMS320C54x (fixed
point) and TMS320C3x (floating point) made by Texas Instruments, the uncontested
leader in DSP processors. What we present here is general information: the
structures described here can be found in all these processors.
the xi are successive samples x(n-i); they are numbers noted in fixed point, coming
from a M bits converter, thus in QM-1.
y( z ) a0 + a1 z −1 + A + a N z − N Waiting time of
h( z ) = = sampling
x( z ) 1 + b1 z −1 + A + bN z − N
Acquiring x(n)
Looking at Figure 6.30, we see that for efficient analysis, we must have:
– a multiplication/accumulation structure (see section 6.4.2.1);
– a time lag or data aging structure (see section 6.4.2.2).
Another structure that is not explicitly shown in Figure 6.30, but is nevertheless
necessary, for both DSP processors working in fixed point and floating point is data
reframing. It will be described in section 6.4.2.3.
In the first case, the operands can be introduced simultaneously if distinct buses
exist, and the multiplication/accumulation can be carried out in a single cycle. In the
second case, a temporary register T is necessary, and there must be a supplementary
cycle.
We have already discussed multiple buses. These have two types of structures:
– the modified Harvard type of structure with a memory space “program” and a
memory space for “data”, all accessible by several bus addresses and data (Figure
6.32). The “program” and “data” spaces can be internal and/or external. This kind of
configuration allows, means of a pipeline, an instruction “fetch”, two data readings,
and a data writing, all in a single cycle (see Figure 6.37). The one condition that
must be met for this to happen be that the two “program” and “data” spaces are
respectively internal and external;
– a single memory space “simplified”, but with multiplication buses to allow for
simultaneous access to the different memory units (Figures 6.34 and 6.35).
27 It is under the name MAC that the aptitude of a general-interest processor or micro-
controller is designed to carry out signal analysis.
248 Fundamentals of Instrumentation and Measurement
Temporary register T
Multiplier
Accumulator
CNTL PC ARs
M
D
INTERNAL U M EXTERNAL
X C U
MEMORY E X MEMORY
S
E
D AT C P DA
s/u s/u D = Data Bus
MPY C = Coefficient Bus
A P = Program Bus
FRCT B
A = A accumulator
0
ADD B = B accumulator
T = Temporary register
s/u = signed/unsigned
acc A acc B FRCT = Fractional mode bit
Figure 6.33. Operand choices for a MAC operation (doc. Texas Instruments, C54x)
Time shifts carried out through memory stimulate high bus activity. However, if
the convolution is calculated in “inverse” order, we get:
N
∑ ai ⋅ x(n − i) = aN ⋅ x(n − N ) + aN −1 ⋅ x(n − N + 1) + A + a0 ⋅ x(n) [6.51]
i =0
Once the sample x(n-i) is loaded in the multiplier, it can be rewritten at the next
address x(n-i-1), which already has been used and been shifted. This operation is
called MACD (Multiply, Accumulate, and Delay: see Figure 6.37). The modified
Harvard structure seen in section 6.4.2.1 helps us produce it in a single cycle if the
pipeline is used correctly.
Real-time Data Acquisition and Processing Systems 251
x(n-2) x(n-1)
x(n-N) x(n-N+1)
Garbage
Pointers are the only solution to circular addressing (that is, of ARn address
registers). But these pointers must allow us to carry out modular operations
throughout the length of the table. This requires a dedicated arithmetical unit, such
as the Address Register Arithmetic Unit (ARAU). This allows for efficient address
tables (see Figure 6.39).
252 Fundamentals of Instrumentation and Measurement
Beginning
Beginning x(n) x(n)
x(n-1) x(n-1)
Figure 6.39. Extract from addressing mode by register of C3x (doc. Texas Instruments)
y ( n) = ∑ a i ⋅ x(n − i ) [6.52]
i
Real-time Data Acquisition and Processing Systems 253
Here, y(n) and the x(n-i) are samples directly represented as whole numbers,
since they come from or will go into AN or NA converters. The ai are coefficients
that can be determined by one of several techniques, and are therefore real numbers
that we will represent as Qk:
(
ai → Ai = round ai ⋅ 2k ) [6.53]
n
combinatory
H
-k
28 A multiplication by 2 is carried out by k shifts, to the right or left according to the sign of
k.
254 Fundamentals of Instrumentation and Measurement
x1 = m1 ⋅ 2exp1 ,1 ≤ m1 ≤ 2
x2 = m2 ⋅ 2exp21 ,1 ≤ m2 ≤ 2
exp 2 = exp1 + k , k > 0
( )
x = x1 + x2 = m2 + 2− k m1 ⋅ 2exp 2 = m ⋅ 2exp
if m ≥ 2, x = m ⋅ 2exp 2 +1
2
This structure is linked to an intensive pipeline. In most cases, this allows for
functioning according to the following rule: one instruction = one cycle (Figures
6.44 and 6.45).
256 Fundamentals of Instrumentation and Measurement
TIME
P1 F1 D1 A1 R1 X1
P2 F2 D2 A2 R2 X2
P3 F3 D3 A3 R3 X3
P4 F4 D4 A4 R4 X4
P5 F5 D5 A5 R5 X5
P6 F6 D6 A6 R6 X6
Usually these are serial, and interface directly with the analog interfaces of the same
family. These interfaces have converters, anti-folding filters and switched smoothing
capacities. The counters help determine the filters’ sampling frequencies and break
frequencies (see Figure 6.46).
The interfaces also ensure data transfer to and from the DSP. This makes the
exchanges completely transparent for the programmer. Almost without interruption,
an input register can be read and an output register can be filled in when a transfer
has taken place. The main program resumes at initializations and while waiting for
interruptions (see Figure 6.47) and analysis is done during the interruptions (Figure
6.48).
the result is in R2
Only six words are necessary, whatever the filter order. 11 + (N – 1) cycles must
be anticipated.
FIR filters in linear phase have a central coefficient symmetry, so that the
number of products can be divided by two:
L −1 ( L −1) 2
y ( n) = ∑ ai ⋅ x(n − i) = ∑ ai ⋅ [x(n − i) + x(n − L + i + 1)] [6.55]
i =0 i =0
we gain nothing from the symmetry feature, since it requires two instructions and
thus two cycles. The C54x makes use of the FIRS instruction:
– total of the accumulation A in the accumulation B;
– multiplication of the accumulation A by a coefficient of a memory-based
program;
– addition of the two samples of a memory-based program in the accumulation
A.
y( z) a 0 + a1 z −1 + A + a N z − N
h( z ) = = [6.56]
x( z ) 1 + b1 z −1 + A + b N z − N
N N
y ( n) = ∑ a i ⋅ x(n − i) − ∑ bi ⋅ y ( n − i ) [6.57]
i =0 i =1
a0i + a1i z −1 + a 2i z −2
h( z ) = ∏ 1 + b1i z −1 + b2i z −2
[6.58]
i
The most straightforward way of producing the effect described above on second
order cells in shown in Figure 6.51, but since the analyses are linear, we can
exchange them (see Figure 6.52), leading to the effect shown in Figure 6.53, which
has the advantage of reducing the number of samples to be memorized.
Real-time Data Acquisition and Processing Systems 261
found outside format. In this case, we use a scale factor before each cell (Figure
6.54), such that:
(
– for the first cell: SF1 = max h1 ( z ) z =e jwTe )
⎛ ⎞
– for the second cell: SF2 = max⎜ 1 ⋅ h1 ( z ) ⋅ h2 ( z ) ⎟
⎜ SF1 ⎟
⎝ z =e jwTe ⎠
The memory structure for a second order cell is given in Figure 6.55. Setting up
the circular address (here, modulo 3) with masking creates the alignment of an
address table to the addresses of the power of 2 (here 4).
; result in R0
For the cascading operation, we can use the structure shown in Figure 6.57. The
empty compartment is necessary for the alignment of the multiple addresses of 4
tables. After calculating a cell, the pointer on the coefficients jumps by 4 in order to
go from one table to the next.
30 The assembler of the C3x makes the idea of parallel instrumentation explicit, here a
“MAC” written as mpyf (operands) ャaddf (operands). But this corresponds to one instruction,
coded on one word and carried out in one cycle, if the pipeline functions well.
264 Fundamentals of Instrumentation and Measurement
6.5. Conclusion
In this chapter, we have tried to show the efficiency and relative simplicity of
digital signal analysis as carried out by modern converters and specialized
processors. Their use for instrumentation purposes will continue to grow because
manufacturers are increasingly offering “standard” microcontrollers with added
signal analysis features. We thus have mechanisms with input/output that sometimes
have defects on DSPs. As well, the increased capacities of FPGAs allow us to set up
filters with order that are too high and coefficients coded on 8 or 12 bits. In this way,
they can function at fairly high sampling frequencies, but we should not forget
integrated digital filters that work with coefficients and data with significant word
capacities.
Real-time Data Acquisition and Processing Systems 265
6.6. Bibliography
ANALOG DEVICE, Practical Analog Design techniques, Compte rendu du séminaire 1995.
ANALOG DEVICE, Sigma-delta ADCs and DACs, Note d’application AN-283.
AZIZ P.M. et al., “An overview of sigma-delta converters”, IEEE signal processing
magazine, January 1996.
BAUDOING G., Les processeurs de traitement de signal, famille 320C5x, Dunod.
DE FATTA D.J. et al., Digital Signal Processing: A System Design Approach, Wiley.
HARRIS Semiconductor, A brief introduction to Sigma Delta Conversion, Note d’application
AN9504, 1995.
HERSCH R.D., Informatique industrielle, Presses polytechniques et universitaires romandes.
IFEACHOR E.C. et al., Digital Signal Processing: A Practical Approach, Addison Wesley.
KUC R., Introduction to Signal Processing, McGraw-Hill.
KUMAR M.S., Digital Signal Processing: A Computer Based Approach, McGraw-Hill.
MARVEW C. et al., A Simple Approach to Digital Signal Processing, Texas Instruments.
MOTOROLA, Principles of Sigma-Delta Modulation for Analog to Digital conversion, Note
d’application APR8/D, 1990.
PARKS T. et al., Digital Filter Design, Wiley.
SENN P. et al., “Convertisseurs analogique numérique CMOS à haute résolution pour les
circuits VLSI audio”, L’écho des Recherches, no. 153, 1993.
This page intentionally left blank
Chapter 7
7.1. Introduction
In this sense, these shared acquisition and analysis structures are superior to
equivalent centralized systems in at least three ways:
– in terms of reliability, because of redundancy and reduction of connective
wiring;
– response time;
– flexibility.
These shared structures are based on data capturing, carried out as closely as
possible to their source, accompanied by a local conditioning and pre-analysis. The
goal of these operations is to make a quick decision, most often of the reflexive
kind. The exchanges with the central unit, made through a serial transmission bus,
are then reserved for slowly changing data analysis (door closing indicator, trunk,
hood, tire pressure, different alarms, to name a few features). Some of these features
require fairly complex calculations.
Sensitive
cells Auto
Communication
Servo-control
test
Information
Signal
analysis
Auto-
Actuators calib .
These properties help us use the experience gained from previous versions and
apply it to the newer technologies. They lead to higher profits for the manufacturer
and better quality for the consumer through lower production and advertizing costs.
The electronic aspects will be discussed, along with some of the relevant
architectures for the conditioning of signals coming from three different sensitive
elements (capacitive and pressure piezoelectric cells, and capacitive acceleration
cells). These are important because of their diversity and complementarity.
All the examples we have chosen come from the automotive field.
7.2. Microtechnologies
Car motors have been affected by microtechnological advances since the 1970s,
when monitoring pressure sensors appeared, and since the 1980s, with the
development of accelerometers that released airbags. Today, apart from these two
applications, microsystems are no longer employed in this general way. However,
within the next decade their use may increase greatly, both in the replacement of
older technologies and in new applications.
These and other microsystems have developed because of the growing use of
electronics, which has the higher calculation power necessary for the car of the
future. In order to use the increasingly miniaturized electronic systems throughout a
car, we also must be able to send these systems adequate information through
sensitive elements or very small, even miniature, sensors.
However, a harsh environment can present a problem with using sensors in a car.
High temperatures, shocks, vibrations, humidity, conditions that cause corrosion,
electromagnetic interferences and radio frequencies can cause problems. This type
of environment makes more demands on design and manufacturing. As well,
production volume must be high (usually one million units or more) in order to
absorb research and manufacturing costs and follow the market demand for new
vehicles. The lifespan of vehicles must be at least 10 years/250,000 km and their
prices must remain low. Generally, cars need the hardy qualities of military vehicles,
at mass-market prices. These qualities (high production volume, low prices,
reliability, and durability) are inherent to microsensors and microsystems.
The mass manufacture of many units at the same time, all with the same Si
layers, functioning like integrated units, leads to very high production volume,
overall low price and high reproducibility. The reliability of these microsystems is
due to:
– expertise in manufacturing processes;
– using materials (especially silicon) with well-known (mechanical, thermal,
electronic) properties;
– relatively simple assemblies with few units and few or no mobile parts.
mechanism of analog and/or digital analysis. These capacities can even extend to
monolithic integrated circuits merging DSP and Si microsensors.
In summary, microsystems are used in cars to reduce sensitive cell size, to lower
their production costs, to improve their performances, and to integrate them
simultaneously with their electronics and/or other microsensors. This is to enlarge
the application of the cells, leading to the creation of an “intelligent car”.
Figure 7.2. Design of the principle (not to scale) of a piezoresistive microsensor in micro-
machined silicon, integrated monolithically with a MOS transistor
The Contribution of Microtechnologies 273
Si:P
1. Type P Si substrate
Si:N
5. Grid oxidation
6. PolySilicon layer
Si:P
Figure 7.3 shows all the necessary steps needed to produce this integrated
pressure microsensor with its electronics. As the succession of steps show, the
principle of producing integrated circuits and microsystems is to repeat, as many
times as is necessary, the following two basic steps:
– remove or modify the thickness of the structure or material layer;
– use photolithography to establish the geometry of shapes.
a)
b)
Our discussion will be limited to the general principles and essential points of
these technologies. For a more detailed discussion, the reader can consult more
specialized texts, including:
– for microelectronics [SZE 81], [GHA 94];
– for microstructures [SZE 94], [RIS 94], [GAR 94], [ELW 98] and [FUK 98].
7.2.1.1. Si substrate
Integrated circuits and silicon microsensors are mass-produced in
monocrystalline substrate in the form of a disc (see Figure 7.3, step 1 and Figure
7.4). The Si layers are made from a monocrystalline bar obtained by Chzochalski’s
method. The bar is sawed, then polished to obtain a finished mirror.
7.2.1.2. Si epitaxy
The Si monocrystalline substrate can be continued by the ordered or epitaxial
layering of a thin Si layer (see Figure 7.3, step 2). Epitaxy helps control the active
layer in which the devices are made. In addition, epitaxy enables the stacking of Si
276 Fundamentals of Instrumentation and Measurement
The Si atoms deposited on the surface must then organize themselves following
the substrate atoms in order to form a perfect crystalline film. In general, this step is
difficult to perfect technologically, and therefore is rather costly. In this case, we can
use mechanisms to attain better performances. For these reasons, Si epitaxy is used
mainly in bipolar and BiCMOS technologies, and optionally in some CMOS
industries.
Wet oxidation occurring with water steam (H2O obtained by H2 with O2) is faster
and helps us obtain thicker oxides. It can provide lateral isolation between the
different mechanisms of a circuit (see Figure 7.3, step 3).
Dry oxidation with O2 is slower and results in layers of oxide that are thinner but
of better quality. These oxides are used as grid dielectrics in MOS transistors, the
building blocks of very high density integrated circuits such as VLSI-ULSI (Figure
7.3, step 5).
7.2.1.4. Photolithography
In order to manufacture integrated circuits and microsystems on a significant
scale, the patterns that make up mechanisms are transferred to the wafer by
photolithographic techniques whose main steps are shown in Figure 7.5.
For each of these steps, a glass mask with opaque patterns in chrome has been
designed from a computer-generated layout.
This PolySi layer is deposited by a technique called Low Pressure CVD by silane
pyrolysis. It is also possible to deposit PolySi with types N or P doping by
introducing certain gases: phosphine (Ph3) and diboraine (B2H6) respectively.
7.2.1.6. Etching
To transfer defined photoresin patterns on the Si wafer, photoresine acts as a
protective mask for the parts of the wafer we want to retain (thick oxide in step 4
and PolySi in step 7). The rest of the wafer that is not protected is removed by
etching.
Two types of etching are currently used. These are listed below:
– Moist etching, an operation in which a liquid agent marks the different layers
selectively, in relation to both photoresine and lower existing layers (see Figure 7.3,
steps 4, 11, and 15, SiO2 marking by HF; step 13, Al marking).
– Dry etching, an operation in which a plasma is used to mark the layer by
physical effect (chemical reaction and conversion of the layer) or by a combination
of two effects of reactive pulverization called Reactive Ion Etching (RIE) (see
Figure 7.3, step 7 for PolySi etching; step 13 shows Al etching).
278 Fundamentals of Instrumentation and Measurement
UV exposing
c. Aligning the glass mask with glass mask
chrome pattern
the chrome patterns
and
UV exposing
d. Development of exposed
photoresist. The patterns of the
mask are transferred
into the photoresine
Figure 7.5. Details of photolithography steps. Definitions of active zones for a circuit’s MOS
transistors and for a pressure microsensor (see also Figure 7.3, step 4)
Dry etching gives better control of the dimensions of etched patters, as well as
high reproducibility. Because of this, it is used in VLSI and ULSI technologies.
The Contribution of Microtechnologies 279
7.2.1.7. Doping
With conductors, we can modify and control their electric conductivity with the
addition of miniscule amounts of impurities. This is called doping. When a
semiconductor is doped type N, the current transfer is made by electrons, while with
a semiconductor doped type P, the charge carriers are the positive holes. By
juxtaposing zones of different dopings, we can make semiconductive electronic
mechanisms such as PN junction diodes, bipolar transistors and MOS, thyristors,
and optoelectrics such as light transmitting diodes, lazers, photodetectors and many
others.
Once introduced, the doping profile can be modified by using the following
steps. At high temperatures (T > 800°C), impurities diffuse significantly in the Si.
Depending on the situation, this diffusion can be desirable to obtain an adequate
profile for the diffusion annealing and/or oxidant (see Figure 7.3, step 10). Or this
diffusion may be an inevitable interference that we try to limit as much as possible
following a strict thermal regime.
280 Fundamentals of Instrumentation and Measurement
To ensure the connections between the different parts of the circuit (or
microsystems), and the environment, the metallic lines are defined by the wafer
(Figure 7.3, steps 12 and 13). Several levels of metallic interconnections are
sometimes necessary for complex circuits, in which case an electric layer is inserted
between each level of metal for isolation.
This kind of etching has Si erosion speeds that are highly dependent on the
directions of the crystalline planes. This means the planes <100> and <110> are
etched much more rapidly than the planes <111>, which stay almost intact. The
chemical agents that stimulate this anisotropic etching are inorganic alkalines such
as KOH or organic solutions like EDP (ethylene diamine pyrocatechol), to name a
few of the best-known. Table 7.1 gives these two anisotropic etching agents the
etching speed of SiO2. Figure 7.6 shows the details of the different steps of the
process and the resulting microstructures.
Table 7.1. Main features of liquid anisotropic etchings for Si and SiO2 (from [GAR 94])
282 Fundamentals of Instrumentation and Measurement
SiO 2
+
Etch-stopping
Couche d' arrêt Player
a. Si (100) wafer before P+
etching
Si (100)
c. Intermediary time
structure of the etching.
The planes (111)
determine the Si volume
54.74°
to be etched
Planes {111}
d. Structure at end
of etching
Membrane Hole
Membrane
with central mass
Figure 7.7. Examples of structure made possible by liquid anisotropic etching of Si (100)
To form the Si membranes, bridges and beams, we must not etch the entire
thickness of the wafer. Instead the etching must be controlled, then stopped to leave
the desired fine layer of Si.
The simplest way of controlling the anisotropic etching is to stop it after a certain
time period, but because the structures of Si wafers are inherently inhomogeneous,
requiring different etching speeds from one point to another, this technique is not
sufficiently reproducible and reliable. In practice, four etch-stopping techniques are
used. We will discuss them here.
Etch-stopping on an Si layer that has been highly doped with boron or Si:P+ (see
Figure 7.6 for how the membrane is made). The liquid anisotropic etching speed
drops quickly for the Si doped with boron, with concentrations above 3 1019
atoms/cm3. The problem of layers that have been stopped with Si:P+ is that the high
boron concentration induces significant voltage constraints in the Si, modifying the
mechanical properties of the membrane. It also stops the creation of semiconductive
mechanisms in the layer, even if these are simple piezoresistances.
Etch-stopping at SiO2 or Si3N4 protected by the opposite facing layer (see Figure
7.6d before HF piercing). The membranes or multi-layered beams of SiO2/Si3N4,
SiO2PolySi/ SiO2 and so on can be created in opposite layer after the volume of the
subjacent Si wafer has been etched.
Accelerometers and above all other inert sensors (like the microgyrometer shown
in Figure 7.9) can be created on the surface of the Si wafer with their electronic
command and analysis integrated monolithically. In general, the principle of
detection and/or excitation is capacitive, the different Si or PolySi levels being
conductors. These microstructures are called micro electro-mechanical systems
(MEMS).
The Contribution of Microtechnologies 285
SiO 2
a. Si layer before
micromaching with SiO2 layer
(Type 2 µm)
Si
f. HF etching of sacrificial
SiO2; Bridge or membrane;
Bridge or Beam for
beam for comb structure
membrane comb structure
Figure 7.9. Microgyrometer made by surface micromachining (source: IEF [VER 99])
286 Fundamentals of Instrumentation and Measurement
In particular, deep anisotropic etching devices have appeared (Deep Reactive Ion
Etching (DRIE)). With these, a dense plasma and specific operating conditions
(having to do with the nature of gases, wafer temperature, etching sequencing) are
conducive to high-speed etching Si carried out vertically and very deeply (Figure
7.10, example a). The form of these etched structures no longer depends on patterns
defined by photolithography; and this leaves complete freedom in designing the
microdevice (Figure 7.10, example b). In using the masking layers and adequate
stopping layers, the Si wafer can be locally etched, either partially or completely, as
is shown in the examples of Figure 7.10.
b. Etched microstructures in Si on a
thickness of 40 µm
a. Test structure for deep anisotropic etching of c. Supported and pierced membrane of
holes 15 µm in diameter at a depth of 150 µm 15 µm thickness created by etching
across the entire Si wafer
by forming an eutectic or
400°C a layer of gold or other metal
metal alloy
average
direct Si-Si another Si wafer
≥ 300°C
From these five technologies, the glass-Si anodic bonding was developed
specifically within the framework of microsystems.
The glass-Si anodic bonding helps us assemble an Si wafer with a sodium glass
wafer that can be micromachined, or it can have metallic patterns for the electrodes.
After having connected the two wafers, the ensemble is placed at a moderate
temperature (350-500°C), at which the sodium ions become mobile in the glass. By
288 Fundamentals of Instrumentation and Measurement
applying a voltage of between 400 and 700 V according to the temperature, the
sodium ions leave the interface zone. This produces a significant electrical field and
generates high electrostatic pressures that are sufficient to establish very close
contact between the two wafers, even when their flatness is not perfect. The bonding
takes several minutes, usually with the formation of a very fine layer of SiO2 as a
connecting layer. Today, the reliability and reproducibility of anodic bonding means
it can be used in industrial settings, especially in the assembling of Si pressure
sensors. However, the use of high voltages is not always compatible with the
presence of integrated circuits on the same wafer.
In addition, other substrates can be used. Some of these are listed below:
– silicon-carbon (SiC) can be used in creating microsystems in harsh
environments (high pressure, high temperature, corrosion);
– gallium-arsenide (GaAs) and, less often, indium phosphide (InP). Both are III-
V mixed semiconductors used for HF and electronic devices and circuits.
– mechanical materials like metals (Cr and W, among others) and shape memory
alloys (Ti/Ni and others);
– magnetic materials (NiFe, CoFe and others);
– thermoelectric materials (SiGe, Bi2Te3);
– inorganic chemical materials (SnO2, metallic oxides and others) and organic
chemical materials (polymide, polypyrroles and phtalocyanines).
Moreover, the robust architectures, as regards the material variability, must have
the following characteristics.
– they must not be significantly affected by the length of their connections to the
sensitive cells and by the majority of the interference capacities. This suggests that
low impedance inputs of the virtual mass type;
– they must be sensitive to the single performances of a minimum number of
components, especially to the sensitive element. This means that a feedback loop
must be used systematically in order to reduce significantly the influence of the
component variability that is part of the direct chain. Then we proceed to
290 Fundamentals of Instrumentation and Measurement
This last concept is only valid for transducers with a relatively slow response
speed (limited to around 10 kHz). As for the how the process is carried out, this
value rarely is an obstacle. In a vehicle, however, this limitation prohibits the use of
a - modulator in applications requiring a dynamic measurement of the steering
wheel angle. Conversions of the “flash” or “weighted” types that occur outside the
loop are, in this case, preferred.
Autotest, either total or partial (we will look at the accelerometer for an example
of this difference), must be at all times superimposed on the measurement, which
itself must remain fully operational.
Lastly, taking into account these important objectives: material flexibility; low-
cost integration of analog and digital parts on the same chip with CMOS or
BICMOS technology; low sensitivity to supply voltages; and good low noise
frequency functioning. Without a doubt, switching capacity techniques are the best
means of carrying out these architectures.
7.3.2. Conditioning electronics for capacitive cells that are sensitive to absolute
pressure
Pressure Pressure
Counter-electrodes
Cmes(P ) =
Cmeso
P
1−
P max
where Pmax is the pressure at which the thin membrane touches the substrate; P is
in principle always below Pmax.
⎡ Cref ⎤ P
The voltage: Vs = Vref ⎢1 − ⎥ = Vref
⎣ Cmes ( P ) ⎦ P max
However, this model is not very realistic, since it does not take into account the
embedment effect that immobilizes the thin membrane, keeping it to the limits of the
enclosure. This gives us the new expression of Cmes(P) (see Figure 7.12b):
Cmes(P ) = Coffset +
Co
with: Coffset + Co = Cref [7.1]
P
1−
P max
P P
We will keep the formula shown in [7.1] as the definitive expression of the
capacity dependent on the pressure. To obtain a measurement voltage linearly linked
to the pressure, it must be capable of calculating:
⎡ Cref − Coffset ⎤ P
Vmes = Vref ⎢1 − ⎥ = Vref [7.2]
⎣ Cmes(P ) − Coffset ⎦ P max
switching capacity techniques allow us to easily carry out the above formula.
We note here that this technique is based on analysis of analog signals sampled
periodically (Figure 7.13). It requires discrete time circuits made of capacities,
switches, and amplifiers sequenced at the sampling frequency Fe by a clock f (in
general of ½ cyclic ratio) of period T.
Here, the amplifier is ideal (infinite gain in open loop and infinite pass band); the
charges are always calculated on the plaques that can be part of electrostatically
isolated systems. To be able to apply the Z conversion, Cmes(P) will be assumed to
be invariant in time [BAI 94-1].
In the architecture shown above, the two capacities {Cref – Coffset} and Coffset
are very easily produced. This is done by dividing the counter-electrode from the
capacitive structure that is insensitive to pressure into two electrically independent
parts (see Figure 7.11). There are two subsystems: S1, which has the three capacities
Coffset, {Cref – Coffset}, Cmes(P); and S2, which is made of C2.
7.3.2.2.1. Switching capacities integrator: the first phase (see Figure 7.15)
The switch that short-circuits C2 only has an initialization function. In normal
functioning, it stays permanently open.
Interval ] (n – 1/2)T, nT [:
⎛ Cmes(P ) ⎞ ⎛ Coffset ⎞
VS ((n + 1)T )⎜⎜1 + ⎟⎟ = VS (nT )⎜⎜1 + ⎟
⎝ C2 ⎠ ⎝ C 2 ⎟⎠
[7.3]
⎛ Cref − Coffset ⎞
+ ⎜⎜ ⎟⎟Vref
⎝ C2 ⎠
VS (Z ) Cref − Coffset
= H (Z ) =
Vref ( Z ) C 2 + Cmes ( P) + (C 2 + Coffset )Z −1
If the value of C2 conditions the response time of the device, the sensitivity of the
sensor, on the other hand, is completely independent of the choice, since the relation
that links the voltage VS to the pressure is expressed in stabilized regime:
⎛ Cref − Coffset ⎞
VS ((n + 1)T ) = VS (nT ) n⇒∞ = Vref ⎜⎜ ⎟⎟ = Vref ⋅ (1 − P P max) [7.5]
⎝ Cmes − Coffset ⎠
Here, we see the advantage of the feedback loop that eliminates the variability
influence of the components inside the loop. It also enables the use of a zero method
The Contribution of Microtechnologies 297
that generates a signal in the form of a product of a stable value of the chosen
electrical variable, here voltage, using the relation between the measured variable
value and the reference value. By taking VS from Vref, we get Vmes (equation [7.2]).
This subtraction operation, not shown in the schema, is easily carried out with a
switching capacities amplifier. The voltage Vmes can then be digitized with an A/N
converter used at the end of the chain.
Let us look again at the example shown above, substituting the discrete analog
feedback {-VS(nT)(Cmes – Coffset)} with a quantified feedback: {0, –Vref(Cmes –
Coffset)}, driven by a conditional clock (see Figure 7.17). As we will see, we keep
the advantages of the zero method by proceeding to digitization by using some of
the properties of first order - modulators with one bit [BAI 96].
easily understand the relevant principle. In effect, we compare (see Figure 7.17) the
capacities Cref – Coffset and Cmes – Coffset to the tub vref at a liquid level. The
integrator then becomes a tub, and the comparator (a one-bit quantifier) is a
measurer of the logical output level (SL = 0/1). In this direct mode, at each clock
cycle, (f) the tub is refilled with the help of Cref - Coffset, which is always smaller
than Cmes – Coffset. When the liquid level in the tub rises above a certain level, it is
partly emptied with the help of Cmes – Coffset by means of commanded
commutators, especially by the conditional cycle SL(f). During N clock cycles, the
number of times it makes use of Cmes – Coffset:
i = N −1
∑ SL(i)
i =0
i = N −1
N ⋅ V r e f ( C r e f − C o ffs e t ) − ∑ {S L ( i ) ⋅ V r e f ( C m e s − C o ffs e t )} = 0
i=0
i = N −1
The mean 1
N ∑ SL(i) represents this quantity with a fractional number between
i =0
{0 and 1}, which can be linked to the numerical pressure measurement:
i = N −1 i = N −1
Cref − Coffset
∑ ∑ SL(i)) =1 − Cmes − Coffset = P max
1 1 P
numberd = (1 − SL(i )) = [7.6]
N N
i =0 i =0
There can be errors when this schema is being carried out. These can occur when
the result of the transferred charges in the integrator by charge injection (of liquid,
for example) due to the clocks, by the interference capacities that are part of the
circuit is not zero “numberd”. To eliminate this systematic error, we can proceed to a
new sequence in reverse mode. In this case, the tub is emptied at each clock cycle
(f) by Cref – Coffset and is refilled, under the effect of the conditional clock SL (f)
with the help of the capacity Cmes – Coffset. By alternating the two modes, we see:
P P
numberd = + err.........numberi = − err [7.7]
P max P max
The error is naturally eliminated by the addition of the two direct and reverse
modes. It disappears easily through a simple decimation filtering.
The Contribution of Microtechnologies 299
Quite often, we must be content with a lower quality correction, with the goal of
simplifying the architecture. The cause of errors can be assimilated to a capacity that is
added to or subtracted from Cref - Coffset. This means it is enough to add or subtract a
physical capacity of the same value to find the quantity “number” we need to find.
(QS )(n+1 / 2)T = −C2 ⋅ VS ((n + 1 : 2)T ) − SL((n + 1/ 2)T ) ⋅ Cmes((n + 1/ 2)T ) ⋅ Vref
SL((n + 1/ 2)T ) = SL(nT )
In this expression, Cmes is a variable that constantly varies over time according
to the measured pressure but slowly with the recurrence frequency of the clock f
(Fe), or “over sampling” frequency chosen to avoid problems of spectrum folding.
That is why a certain degree of leeway is possible in selecting the associated
sampled variable; so we write: Cmes ((n+1/2)T) = Cmes(nT).
and:
C 2 ⋅ V S (( n + 1)T ) = C 2 ⋅ V S ( nT )
[7.8]
+ ( C ref − C offset ) V ref − SL ( nT ) ( C m es ( nT ) − C offset ) V ref
C r e f − C o ffs e t P (nT )
w ith : = 1−
C m e s ( n T ) − C o ffs e t P m ax
To create number (nT), we can set up, at each clock cycle, the accumulation of
consecutive N values of SL:
1 ⎧i = N −1 ⎫ 1 ⎧i = N −1 ⎫
number (nT ) = ⎨ ∑ (1 − SL(n − i )T ) ⎬ = ⎨ ∑ SL ((n − i )T ) ⎬ [7.9]
N ⎩ i =0 ⎭ N ⎩ i =0 ⎭
300 Fundamentals of Instrumentation and Measurement
⎛ ⎛ C 2 ⋅ V S (( n + 1)T ) ⎞⎞
⎜ ⎜ − ⎟⎟
1 ⎜ i = N −1 ⎛ P (( n − i )T ) ⎞ 1 ⎜ ( C mes − Coffset ) nT ⎟⎟ [7.10]
number ( nT ) = ∑ ⎜ ⎟+
N ⎜ i = 0 ⎝ P max ⎠ Vref ⎜ C 2 ⋅ V S (( n − N + 1)T )) ⎟⎟
⎜ ⎜⎜ ⎟⎟ ⎟
⎜ ⎟
⎝ ⎝ ( C mes − Coffset ) ( n − N )T ⎠⎠
so that Cmes((n - i)T) is very close to Cmes(n – i - 1)T), which is even coherent with
the principle of “oversampling”.
Formula 7.10 shows, in a certain way, the increase in resolution produced by the
averaging operation. However, since it contains two unknowns, this equation, like
equation [7.8], does not have any predictive value. To obtain this value, we must
eliminate an unknown by expressing VS according to SL. They are connected by the
nonlinear relation:
VS (n + 1) > 0 ⇒ SL(n + 1) = 1
VS (n + 1) < 0 ⇒ SL(n + 1) = 0
For a stabilized value of the observed measurement capacity, the average voltage
of integrator output <VS>, is established at:
This approximate expression (the rigorous equation is not linear) comes from the
simple observation of the output voltage of the integrator for diverse values of
(Cmes – Coffset) (see Figure 7.18). The relation between VS(nT) and SL(nT) is then
established easily. By combining equations [7.6] and [7.9], equation [7.11] becomes:
C 2 < VS > 1
≈< SL > −
Vref < Cmes − Coffset > 2
The Contribution of Microtechnologies 301
C2 ⋅ VS ((n + 1)T )
= (SL((n + 1)T ) − δS ((n + 1)T ) ) −
1
Vref (Cmes((n + 3 / 2)T ) − Coffset ) 2
[7.12]
C2 ⋅ VS ((n + 1)T ) ⎡ ⎤⎡ 1⎤
⎥ ⎢(SL((n + 1)T ) − δS ((n + 1)T ) ) − ⎥
P max
≈⎢
(Cref − Coffset )Vref ⎣ P max− P((n + 3 / 2)T ) ⎦ ⎣ 2⎦
Then, after taking into account the results and rearrangement, equation [7.8]
becomes:
⎛ Cref − Coffset ⎞
SL (( n + 1)T ) = ⎜⎜ ⎟⎟ + δ S (( n + 1)T ) − δ S ( nT )
⎝ Cmes ( nT ) − Coffset ⎠ [7.13]
⎛ ⎞
⎟ − (δ S (( n + 1)T ) − δ S ( nT )
P ( nT )
S L (( n + 1)T ) = ⎜
⎝ P max ⎠
⎡ P (Z ) ⎤
SL (Z ) = Z −1 ⎢ ⎥
⎣ P m ax ⎦
⎡ P (Z ) ⎤ i2πf [7.14]
SL (Z ) = H SL (Z ) ⎢ ⎥ ⇒ H SL (e ) = e −i2πf
⎣ P m ax ⎦ Fe Fe
⎛ C r e f − C o ffs e t ⎞ ⎛ P ⎞
w ith : ⎜ ⎟ (Z ) = ⎜1 − ⎟ (Z )
⎝ C m e s − C o ffs e t ⎠ ⎝ P m ax ⎠
brSL ( Z ) = (1 − Z −1 )δ S ( Z ) [7.15]
∆2 Fe Fe
dspbrS = with: − ≤ f ≤ [7.16]
12 Fe 2 2
with ∆, the quantification step (which is equal to 1, since we evaluate on the basis of
the capacitive ratios and, by extension, the pressure ratios included between {0,1})
and that the quantifier has one bit. A noise power corresponds to this density:
∆2
PbrS =
12
Relations [7.13] and [7.15] show that the - modulator carries out the discrete
differentiation of the quantification noise of the comparator. This operation
transforms the “white noise” of the one-bit quantifier to a colored noise concentrated
in high frequencies (Figure 7.19). It is characterized by a spectral density of power:
∆2 ⎛ πf ⎞ Fe Fe
dspbrSL = 4 sin 2 ⎜ ⎟ with: − ≤ f ≤ [7.17]
12 Fe ⎝ ⎠
Fe 2 2
The Contribution of Microtechnologies 303
One bit N
√ dspr SL(kT) SL ((k-N+ 1)T) registers
SL-SL Fe stack
Σ -∆
capacitor
measurement
modula tor UAL Ν√ dspr
number
Fe
N-1
Σ SL((k-i) T)) = N.numbe r(kT)
i=0
Stack of N
Fe M bits
registe rs:
M
=N
2
UAL
N.numbe r(k-N+ 1)T)
Fe
p Fc = p N 2 <num ber> Ν 2√ dspr
N <number>
3.5
Noise spectral density of the
output of N modulator
3
2.5
2
1.5
1
0.5
0
0 0.1 0.2 0.3 0.4 0.5
N 2.dspbr
< numbe r > (f/F e) Freque ncy in f/Fe for N= 10
One bit
Η SL SL(kT) SL ((k-N+1)T) N registers
Fe stack
Σ-∆
capacitor
measurement
modulator UAL ΝΗ number
Fe
N-1
Σ SL((k-i)T)) = N.numbe r(kT)
i=0
Stack of N
Fe M bits
registers
M
=N
2
UAL
N.numbe r(k-N+ 1)T)
Fe
pFc = p N 2 <number> N 2H <number>
N
0.8
H SL ( )
ei2πf
Fe
transfert function
0.6 i2πf
H number
(e Fe)
0.4
i2πf
0.2 H < number >
(e Fe)
0
0 0.1 0.2 0.3 0.4 0.5
Frequency in f/F e for N=10
The first filter is a simple averaging filter that helps establish the number. Its
schema of principle is shown in Figure 7.20. It has a shift register that contains an
adder and its accumulator on M bits: 2M = N at each instant kT N, consecutive values
of SL (SL(kT) to SL(k-N+1)T). These are sequenced at the oversampling frequency
Fe. This filter is regulated by the recurrent equation shown below, deduced from
equations [7.9], [7.10] and [7.13]:
1 i = N −1 ⎛ P (( n − i )T ) ⎞
num ber (( n + 1)T ) = ∑ ⎜ ⎟ .....
N i = 0 ⎝ P m ax ⎠
{δ S (( n + 1)T ) − δ S (( n − ( N − 1))}
... −
N
to which we can link the transfer function for the corrected useful signal of the
quantification error:
Z −1⎛1− Z −N ⎞⎛ P ⎞
num ber ( Z ) = ⎜ ⎟⎜
⎟ ⎝ P m ax ⎟⎠
(Z )
⎜ −1
N ⎝ 1− Z ⎠
⎛ P ⎞
= H num ber ( Z ) ⎜ ⎟ (Z )
⎝ P m ax ⎠
(1 − Z − N ) 1 ⎛1− Z −N ⎞
brnumber ( Z ) = δS ( Z ) = ⎜ ⎟ brSL ( Z )
N N ⎜⎝ 1 − Z − 1 ⎟
⎠
For the useful signal, this filter presents a frequency response more or less in
cardinal sinus:
i 2 πf ⎛ sin c ( πfN ) ⎞
( )
sin(x)
H number (e Fe ) ≅ e −i πf ( N + 1) ⎜ Fe ⎟ with: sin c( x) = [7.18]
Fe ⎜ ⎟
⎜ sin c ( πf ) ⎟ x
⎝ Fe ⎠
306 Fundamentals of Instrumentation and Measurement
The module of the above equals the unity at zero frequency, which has
Fe
transmission zeros at all the ± and presents a pass band: {–Fe/2N to Fe/2N},
N
shown in Figure 7.20.
The second filter is an averaging filter of the same type as described above, but it
has as input number and for output <number>:
1 ⎛ i = N −1 ⎞
< n u m b er ( n T ) > = ⎜ ∑ n u m b er (( n − i ) T ) ⎟ T h erefo re:
N ⎝ i=0 ⎠
⎡ 1 ⎪⎧ i = N − 1 j = N − 1 ⎛ P (( n − i − j ) T ) ⎞ ⎪⎫ ⎤ [7.20]
1
⎢ 2 ⎨ ∑ ∑ ⎜ ⎟⎬ − 2 × ⎥
⎢ N ⎪
⎩ i=0 J =0 ⎝ P m ax ⎠ ⎪
⎭ N ⎥
< n u m b er (( n + 1) T ) > = ⎢ ⎥
⎢ ⎪⎧ j = N − 1 j = N −1
⎪⎫ ⎥
⎢⎨ ∑ δ S (( n + 1 − j ) T ) − ∑ δ S (( n − ( N − 1) − j ) T ) ⎬⎥
⎣ ⎪⎩ j = 0 j=0 ⎪⎭ ⎦
For the useful signal, this filter presents a frequency response approximately in
squared cardinal sinus:
2
i 2 πf ⎛ sin c ( π fN )⎞
H num ber ( e Fe (
) ≅ e − i 2 π fN )
⎜
Fe ⎜
⎜ sin c ( π f
Fe ⎟
⎟
) ⎟
[7.21]
⎝ Fe ⎠
which clearly attenuates the useful signal (see Figure 7.20) outside the pass band
≈ {–Fe/2N ⇔ + Fe/2N}.
4
2 ⎛ sin c ( N πf )⎞
4 sin 2 ( πf )⎜ Fe ⎟ Fe Fe
dspbr< number > ≅ − ≤f≤ [7.22]
Fe ⎜ ⎟
12 Fe ⎜ sin c ( πf ) ⎟ 2 2
⎝ Fe ⎠
At this filtering level, the signal-to-noise ratio gains another factor √N in relation
to the previous step. This is because its non-zero spectral density (see Figure 7.19) in
more or less concentrated from: ~ {–Fe/2N to Fe/2N}; that is, a band N times more
narrow than before the previous filtering.
π
P br< num ber > ≅ P brS [7.23]
3
N 2
Doped silicon
Silicon “substrate”
constraint gauge
The temperature range that can support this type of device can be increased by
placing the piezoresistive elements in oxide shells. This eliminates any escape
currents. These sensitive cells are used in electronic speed boxes or they are used to
measure oil pressure (20-50 bars, more than 200 degrees).
However, the sequencing differs from that used in the capacitive pressure sensor,
since here we measure a differential pressure that can be positive, negative, or zero.
From this we see that at each clock cycle (f), the quantity of charge Cref ⋅ Vref,
must be brought to or sampled by an integrator according to the state of the
comparator. According to the principle schema, the state “SL = 1” of the comparator
is a sampling and counting order, while the state “SL = 0” constitutes a carrying and
deducting order. As well, during N clock cycles of the clock f, the balance of
charge transfers is established as:
The Contribution of Microtechnologies 309
Vref
KC
SL
SL
number
SL
C SL
SL
number = KVb - Va
SL Vref
By linking the results to the constraint effects created by the differential pressure
of the membrane, we get:
⎛ δR ⎞
N ⋅ number = K ⋅ N ⎜ ⎟ = K ⋅ N ⋅ ⋅ ∆P [7.25]
⎝ 4R ⎠
to the quantification noise and to close to the charge injection errors of the clocks.
It is notable that in equation [7.24], as in equation [7.25], only the factor “K”
remains; the values of Vref and Cref have no effect from a metrological point of
view. The improvement of the modulator the proceeds by introducing two
functioning modes, “direct” and reverse”, which, as in the previous application,
310 Fundamentals of Instrumentation and Measurement
eliminate the effect of injection. And by using a reconfigured mode (shown by the
dotted line in the schema), the scale factor K can be measured precisely (the exact
value of other capacities are without importance from a metrological point of view).
Then K can be stored for subsequent measurements. This functioning mode can be
activated at any time; the modulator is able to performing autocalibration. By
reducing the impact of interference effects, integration, in the form of symmetrical
architecture, leads to very good performances (16 bits of resolution for a pass band
of 1,000 Hz, an oversampling frequency of 2 MHz for a first order modulator, to cite
two examples).
Accelerators are being used more and more inside of vehicles. Recently, airbags
have been equipped with similar microsystems; these are the basis of the inflation
that occurs in case of accidents. As well, accelerometers functioning as
inclinometers are used for controlling the suspension of some vehicles. These
mechanisms also function in automatic shock absorption systems, used more or less
often according to the state of the road or the car.
m
d 2 − d1 = γ [7.26]
k
where is the acceleration and m the seismic mass. The distances d2 and d1 measure,
are, respectively, the gaps (polegaps) that separate the seismic mass from the lower
and higher counter-electrodes.
The Contribution of Microtechnologies 311
Counter-electrode γ
Suspension arm
SLn Ø Counter-electrode
d1 C1 SL = SLn
Ø
Ø
d2
Seismic
mass
γ ∫ SL = SLp
C2
Vref
SLp Ø
N.number= n - p = N ( C1–C2
C1+C2 )
Figure 7.23. Schema principle of a modulator for an accelerator
ε S ε S
C1 = o and C 2 = o [7.27]
d1 d2
where S represents the surface of the seismic mass as relates to the counter-
electrodes, and i0 the permittivity of the void. These two capacities are not
independent and are linked by the relation:
1 1 d + d2 d 2
+ = 1 =2 o = [7.28]
C1 C 2 εoS ε o S Co
where d0 represents the distance of the seismic mass to one or the other of the
counter-electrodes in the absence of acceleration, and C0 represents the
corresponding capacity. The introduction of relations [7.2] and [7.28] to the interior
312 Fundamentals of Instrumentation and Measurement
d 2 − d1 m d −d C − C2 2d ⋅ k ⎛ C1 − C 2 ⎞
= γ = 2 1 = 1 ⇒γ = o ⎜⎜ ⎟⎟ [7.29]
2d o 2d o ⋅ k d 2 + d1 C 2 + C1 m ⎝ C 2 + C1 ⎠
n+ p = N
N N
nC1 ⋅ Vref − pC 2 ⋅ Vref = 0 with
∑ ∑
[7.30]
n= SLn (i )..... p = SL p (i )
0 0
to the quantification noise and to near to the clocks’ charge injection errors.
Also:
⎛ C − C2 ⎞ ⎛ d − d1 ⎞ ⎛ m ⎞
n − p = N ⎜⎜ 1 ⎟⎟ = N ⎜⎜ 2 ⎟⎟ = N ⎜⎜ ⎟γ
⎟ [7.31]
⎝ C1 + C 2 ⎠ ⎝ d1 + d 2 ⎠ ⎝ 2 kd o ⎠
However, when we must try to find linearity and a very good signal-to-noise
ratio in order to conserve the functioning of the open loop of the capacitive sensitive
cells, we must use other architectures that in principle eliminate this problem.
ε ⋅ L2 ⋅ V 2
Fe = ∞ V 2 ⇔ Fm = ρ ⋅ L3 ⋅ γ ∞ L3 ⋅ γ
2⋅d 2
ρ = massic density
These forces are of the same order of magnitude for the potential value
compatible with the microelectronics in the ranges of acceleration measurements
usual when the pole gap has an order of several micrometers for mass dimensions of
the seismic mass, which are millimetric. These values are typical of cells that can be
produced through current silicon microtechnologies.
f f
Fe1 + Fe 2 = 0
f Q2 f f Q2 f
Fe1 = 1 ⋅ n 21 ⇔ Fe 2 = 2 ⋅ n12 [7.32]
2εS 2εS
where Q1 and Q2 are the shared charges on the surfaces of the seismic mass in
relation to the two counter-electrodes. The equation is fulfilled under two specific
conditions:
Q1 = Q 2 or Q1 = −Q2 ⇒ Q = 0 [7.33]
314 Fundamentals of Instrumentation and Measurement
Only the second condition can lead to a relevant measurement of the technique
of the switching capacities. We will know if the verification of the sum of the
charges contained by the mass is actually zero.
+ Q + Q
Fe2 Fe2
Vin-x Vin-x
Seismic mass Seismic mass
n21 Fe1 -Vin-x n21 Fe1 -Vin-x
+ +
r r r r r r r
Fe1 + Fe2 = 0 Q=0 Fe1 + Fe2 + Fm = 0
C1
Measurement Measurement
of r of
∫ ∫
γ ∆V x Fm + ∆V x
∆Q A m ∆C A
r −
Fe
C2
dimensional factors
εS εS
Q = Q1 + Q 2 = ( x − Vin) + ( x + Vin) = C1 ( x − Vin) + C 2 ( x + Vin)
d1 d1
[7.34]
⎛ C − C2 ⎞
⇒ Q = 0 ⇔ x = ⎜⎜ 1 ⎟⎟Vin
⎝ C1 + C 2 ⎠
Since Vin is stable, the potential x of the seismic mass of the effects of the
electrostatic forces constitutes an acceleration measurement, since from equation
[7.31], we get:
⎛ C − C2 ⎞ ⎛ d − d1 ⎞ ⎛ m ⎞
x = Vin ⋅ ⎜⎜ 1 ⎟⎟ = Vin ⋅ ⎜⎜ 2 ⎟⎟ = Vin .⎜⎜ ⎟γ
⎟ [7.35]
⎝ C1 + C 2 ⎠ ⎝ d1 + d 2 ⎠ ⎝ 2 kd o ⎠
The Contribution of Microtechnologies 315
The circuit shown in Figure 7.25 ensures the required functionality [LEU 90].
+ Vin
C
f
C1 Vcc
Vcc
m - -
C C
2 4
+ +
-Vin
C Vee x
30 Vee
During the intervals ] nT, (n + 1/2)T [, the switches switched by the clock f are
closed. All the nodes of the charge amplifier, the counter-electrodes, and the seismic
mass are brought to the same potential. The seismic mass is subject only to
mechanical forces.
During the intervals ](n + 1/2)T, (n + 1)T [, the node “+” of the charge amplifier
does not change in potential, but the output voltage varies according to the
disequilibrium measurement of the shared charges on the surface of the seismic
mass, since:
(C1 + C 2 ) x n + (C 2 − C1 )Vin
V I ( n +1 2) − x n = [7.36]
C3
⎛ C − C2 ⎞ ⎛ m ⎞
x n → ∞ ( nT ) = V in ⋅ ⎜ 1 ⎟ = V in . ⎜ ⎟ static [7.37]
⎝ 1
C + C 2 ⎠n→ ∞ ⎝ 2 kd o ⎠
316 Fundamentals of Instrumentation and Measurement
C4 ⎛ C1 + C 2 ⎞
0< ⎜ ⎟<2
⎜ ⎟
Cf ⎝ C3 ⎠
Here we again see the inherent advantage of feedback structures. These help us
obtain a measurement quality (apart from response times) that is not sensitive to
component variability, with the exception of the sensitive cell itself.
m /(do ⋅ k )
The value of the stiffness “k” is a parameter that is highly sensitive to dispersive
effects related to dimension reduction and to certain thermal steps of manufacturing
processes. In addition, this value can vary according to conditions of usage.
Although the small pole gap “d0” is not identical to a sample taken from the same
series, but its value remains stable. The seismic mass is, however, a well-identified
physical object of some size even in the most miniaturized sensitive cells and is
time-invariant. “d0” and “m” can thus be memory-stored objects.
The theoretic static scale factor now depends on only two parameters if the
measurement method is based on the balancing of the mechanical force by the
electrostatic forces.
expresses at equilibrium:
f f f
Fm + Fe1 + Fe 2 = 0 ⇔ C1 + C 2 = C 0
C 0 (Vin − x) 2 f C (V + x) 2 f f [7.38]
n 21 + 0 in n12 + m ⋅ γ stat ⋅ n 21 = 0
2d 0 2d 0
In addition:
m ⋅d0 m ⋅ d 02
x th e o r e tic a l = sta t = s ta t [7.39]
2 C 0 ⋅ V in 2 i ⋅ S ⋅ V in
The switching capacities circuit (Figure 7.26) fulfills the desired functionality; the
potential difference “b” brings about the gap between the two capacities C1 and C2.
Under the effect of constant acceleration, the signal this circuit transmits is
expressed as:
m ⋅ d0
x rea l = sta t
2 C 0 V1 C ⋅ d 02 k
1+ [7.40]
2 A
2 b ⋅ V1⋅ C 0 sta t
V1 = V in + b 2
318 Fundamentals of Instrumentation and Measurement
This relation shows the dependence between the static scale factor and the stiffness
that occurs when the loop gain is not infinite. However, to further develop this
analysis, the dynamic characteristics of the sensitive cell must be taken into account.
The seismic mass is subject to the linking of four forces: the effects of suspension
stiffness; acceleration; viscous absorption; and the result of electrostatic forces.
In the absence of “small signal” electrostatic forces, the function of the harmonic
transfer of the sensitive cell is expressed as:
Y (ω ) m 1
= H (ω ) = [7.41]
Γ (ω ) k λ m 2
1+ i ω− ω
k k
X rea l (ω ) m ⋅d0 1
=
Γ (ω ) 2 C 0 ⋅V1 ⎛ C ⋅ d 02 ⎞⎛ m ⎞ [7.42]
1+ ⎜ 2 ⎟ ⎜ ⎟
⎝ 2 b ⋅V1 ⋅ C 0 ⎠ ⎝ H (ω ) A (ω ) ⎠
V 1 = V in + b 2
This expression shows that the error analysis operator, A( ) cannot be chosen
independently of the sensitive cell. To clarify this point, we will look at two cases:
the cell with dominant absorption and the cell with optimum absorption.
m
H (ω ) ≅
m
iλω (1 + i ω)
λ
To avoid any instability in the looped system, the error analysis operator is a
limited-gain amplifier. Expression [7.42] then becomes:
X rea l ( ) m ⋅ d0 1
=
d( ) 2 C 0 ⋅ V1 ⎛ C ⋅ d 02 ⎞⎛ (1 + i ( m λ ) ) ⎞
1+ i⎜ ⎟⎜ ⎟
⎜ 2b ⋅V ⋅ C 2 ⎟⎝ A ⎠
⎝ 1 0 ⎠
V1 = V in + b 2 [7.43]
C ⋅ d 02 2
A<
2 b ⋅ V1 ⋅ C 02 m
The signal to mechanical noise ratio of this type of sensitive cell is potentially
the best. But these cells are also the most fragile because of the low stiffness of the
seismic mass suspension; and they also withstand miniaturization poorly.
m
H (ω ) ≅
k
k
0 < ω < ΩR =
m
The error analysis operator can then be an integrator with a limited transition
pulsation “ T”. Expression [7.42] then becomes:
X rea l ( ) m ⋅ d0 1
=
d( ) 2 C 0 ⋅ V1 ⎛ C ⋅ d 02 ⎞⎛ k ⎞
1+ i⎜ ⎟⎜
⎜ 2b ⋅V ⋅ C 2 ⎟ ⎝ T ⎟⎠
⎝ 1 0 ⎠
V1 = V in + b 2 [7.44]
⎛ C ⋅ d 02 ⎞
<⎜ ⎟ kΩ R
T ⎜ 2b ⋅V ⋅ C 2 ⎟
⎝ 1 0 ⎠
320 Fundamentals of Instrumentation and Measurement
C 0V 2 ⎛ T − τ ⎞ p − n
γ= ⎜ ⎟( ) …… [7.45]
2d 0 ⋅ m ⎝ T ⎠ N
The Contribution of Microtechnologies 321
with:
p+n = N
N −1 N −1
n ( kT ) = ∑ SL (( k − i )T ).......... p ( kT ) = ∑ SL (( k − i )T )
i=0 i=0
N ⋅ numberγ ( kT ) = p ( kT ) − n ( kT )
C 0 ⋅ V 2 ⎛⎜ τ n − τ n' ⎞
⎟
γ =
2 ⋅ d 0 ⋅ m ⎜⎝ T ⎟
⎠
[7.46]
Accelerometers, and more generally inert systems, have been the subjects of
much research; the final performances of these microsystems are very sensitive to
using electronics within technological parameters. However, it would be outside the
scope of this chapter to treat this topic in depth, since here we are limiting our
discussion to initiation technologies
322 Fundamentals of Instrumentation and Measurement
V
Øm
Ø2 VC
Øm 1
Øm SL
+
∆C = C2 – C1
b b b
γ γ γ=0
τn τ’n+1
VC2 –b –b –b
VC1
τ’n b τ’n+1 b b
γ –b γ –b γ=0 –b
VC2 T
Figure 7.29. Digital architecture and chronogram of a sensitive cell with optimum absorption
Much research has been undertaken on electronics with low consumption levels,
microsources (electrothermal, electrodynamic and photovoltaic), Hz and optical
telesupplying, among other areas.
Up to now, this research has not had much impact in the area of automotive
technology devoted to the battery. However, telesupplying has already appeared in
mobile parts microsystems responsible for chassis-road connections (such as tire
pressure), remote control, and engine immobilization.
7.4. Bibliography
In this section, we will discuss standard measurement devices and their recent,
digital, developments. Recently digitization is done throughout the system before the
overall measurement process begins. This means it is carried out increasingly with
sensors.
8.1.1. Multimeters
ADC VX idVm
VX
Displays are carried out with a certain number of numerals or digits. We find
devices called “3 ½ digits” or “6 ½ digits”. In this designation, the ½ digit is the first
character to be displayed, with 0 or 1 or even 2 (it is a half digit because it does not
take all the possible values). The whole number (here a 3 or a 6) represents the
following digits. Thus, the displayed number (without taking into account possible
decimal points) can have values of between 0 and the maximum displayable or
number of points. A 3 ½ multimeter can display from 0 to 1,999 and is called a 2
million point device. The HP34410A multimeter made by Agilent TechnologiesTM is
a 6 ½ digit or 2 million point multimeter.
RS
VX RIn iVm
100 R S
δm = [8.1]
R S + Rin
The input resistance is typically 10 M but can reach much higher values. For
example, on the HP34401 multimeter, we can choose on the ranges of 0.1 V, 1 V
and 10 V as an input impedance of 10 M or even above G , whereas on the 100
and 1,000 V ranges it is worth 10 M .
R sh R in idVm
I0
RX RIn idVm
A problem occurs when measuring low resistances. In this case, the resistance of
the measurement wires is of the same order of magnitude and introduces a
significant error (drop of potential in the wires). Here we use the “4 point” method.
I0
RX RIn idVm
One basic feature to consider is the passband of the device, which for
multimeters is a fairly general base. For the HP34401A multimeter, the passband has
100 kHz of voltage and 5 kHz of intensity (for an accuracy below 1%).
The measurement of the effective value takes into account the effective value of
the alternating current Vf-ac and that of the direct current Vf-cc. We especially need to
Instruments and Measurement Chains 329
separate the two in cases of an end face alternative value of a component remaining
high. In this case, we get the true effective value.
V f = V f −cc 2 + V f − ac 2 [8.2]
Signals having high comb factors (relation between the comb value and the
effective value) set at relatively high frequencies pose error risks, due to the
presence of fairly high harmonics.
Of course, the fact that a digital value is displayed is not a sign of absolute
accuracy. Measurement accuracy is not directly accessible with a device, as was the
case with magnetoelectric instruments with moving coils indicating the class of the
instrument. It is necessary now to refer to the device’s instructions for information
concerning accuracy, taking into account a percentage of the measurement, of the
range, and even the number of digits. Accuracy is expressed as:
Still using the example of the HP34401A multimeter, to measure the direct
voltage on a caliber 1 V, the accuracy for a year is given as 0.0040 + 0.0007 or
0.0047% full-scale maximum. This means an accuracy of 4.7 · 10-5 V for a 1 V
measurement.
Impulses
Port Counter Display
Control
Whether frequency or periods are being measured, one synoptic can be given for
the instrument being used. This synoptic is shown in Figure 8.6. In this figure, the
synoptic counts the number of impulses during the high state period of the control
signal.
TCK .N CK
T= [8.4]
Np
Instruments and Measurement Chains 331
8.1.3. Oscilloscopes
8.1.3.1. Introduction
Today, since oscilloscopes are usually digital oscilloscopes, we will only discuss
this type. The synoptic of a oscilloscope is shown in Figure 8.7.
After selecting the coupling mode as input (DC for direct current and AC for the
part of the signal without a direct current), the principle chain is connected to a
preamplifier (G), a sample and hold (Samp.B), an analog-to-digital converter and a
memory storage device. The converter usually has an 8-bit accuracy.
The criteria which fix the frequency of the oscilloscope are the analog pass band
of the input stages and the sampling frequency. Rise time is also a possible criterion
but is seldom used anymore.
Sampling can be carried out either in real time (or in single-cycle mode) or
repetitively (in sequential or random mode). In the first case, in order to respect
Shannon’s theorem, we must have fs > 2.BW where fs represents the sampling
frequency and BW the length of the signal band. In the second case, we must
implement a reconstruction procedure and we have f s > k R .BW where kR is equal to
2.5 for a sinusoid, around 4 for a transient regime, or even 10 when there is no
reconstruction. We find oscilloscopes with sampling frequencies of up to 20 MEch/s
(as with series HP54600 made by Agilent TechnologiesTM) to 2 GEch/s (the
HP54615 and Infinium).
332 Fundamentals of Instrumentation and Measurement
AC
CH1
G Samp. B CAN Mem.
DC
AC
CH2
G Samp. B CAN Mem.
DC
Clock Clock/
Samp.
External
trigger
Time µPr.
base
After the analog-to-digital conversion, the data are stored in a memory unit that
currently can be up to 1 Mega-samplings. Recent instruments almost always have an
interface (serial or IEEE488) that facilitates data transfer to a calculator.
There are several visualization modes for signals: by points or vectorial (the
points obtained are linked). We can also carry out an averaging of the data to
improve the signal-to-noise ratio on in general 8, 64 or even 256 acquisitions. The
visualization of modulated signals (amplitude or angular modulation) can cause
some problems with display. This is due to the fact that multiple frequencies occur
during reconstruction. With such cases we must be careful to avoid taking a
vectorial display so as not to place too much importance on falsely correlated
phenomena. A single-cycle sampling can be most practical here, since the passband
is strictly limited by Shannon’s theorem.
Rin Cin
The input capacitor has a capacity with an order of magnitude of 15 pF, but this
value is not normalized (and varies from around 10 to 20 pF). If we carry out the
measurements with a standard coaxial cable (with a line capacity of 100 pF/m), the
capacitor brought to the measurement point is therefore of 15 pF + d*100 pF where
d is the cable length. For a coaxial cable of 1 m this can come to 115 pF, which can
prove to be counter-productive. To resolve this problem, we can use measurement
probes.
The model of the probe used with the input stage of an oscilloscope is shown in
Figure 8.9. For this model, we can establish the corresponding transfer function that
links the voltage analyzed by the oscilloscope (written as Vin) to the voltage to be
measured (written Vm).
Rso
Oscilloscope
Regulating the probe to obtain an all-pass filter means adjusting the capacitor Cso
so that:
R so C so = Rin C in [8.6]
334 Fundamentals of Instrumentation and Measurement
R so + Rin
Z eq = [8.7]
1 + jRin C in ω
So, with Rin = 1 M ; Rso= 9 M ; Cin = 12 pF. In this case, Cso must be adjusted
to 1.33pF, which gives an input impedance probe + oscilloscope equivalent to 10
M in parallel with 1.2 pF. The capacity of the cable being used for the probe is
often around 7 to 8 pF, which gives the ensemble a capacity below 10 pF! But of
course, even the smallest disturbance brought to the assembly is obtained to the
detriment of an attenuation by 10 of the voltage to be measured. There are also
active probes that present very high input impedances and very low disturbance
capacities (the measurement element is a MOS transistor).
There are mainly two methods of carrying out a frequency analysis of a signal.
With low frequencies (up to a few 100 of kHz), we used an analyzer based on the
calculation of the Fourier transform of the signal, which has been digitized
beforehand. For radiofrequencies, high frequencies and microwaves, we use a
sweeping analyzer (a technique using several 100 kHz up to 100 GHz).
In
Att. G det
IF
LO
Ramp Local
generator osc.
We see that the most important unit in this process is the mixer, the signal to be
studied having been applied after formatting on the RF input and the sweeping
signal having been formatted on the LO input. Since the signals we apply on the RF
input of a mixer must come up to certain levels, the input has an attenuation stage
and an amplifier. It can have several successive conversion stages before optimally
adapting the frequency range of the signal to be studied to the frequency of the
selective filter. Each stage has a mixer, a local oscillator and an intermediary
frequency filter. With recent instruments, we use a digital filter for the last IF stage
that facilitates high stability in the filter, even at very low resolutions like 1 Hz. In
this case, the final section is digital, including the peak detection and the display
control.
After the selective filter, there is a peak detector for finding the amplitude of
selected lines, then a filtering before beginning the visualization process. Recently,
in some microprocessing variations, the signal is digitized after the video filtering
and a microprocessor controls the local oscillator and display functions.
The main parameters to regulate are the central frequency, the frequency span,
the resolution (length of the selective filter or resolution bandwidth, written as
BWres) and the sweep rate, written as SWr. However, we should remember that the
higher the response, the more selective the passband filter. We have to “wait” for the
output signal to go through transient regime before being able to correctly carry out
peak detection. So, if we want to improve resolution (that is, separate the close
lines), we must increase the sweep time so that the span parameters, sweep rate and
resolution cannot be regulated independently. The resolution values usually can be
336 Fundamentals of Instrumentation and Measurement
regulated by sequences 1, 3 (100 Hz, 300 Hz, 1 kHz, etc.). We can give an
approximate formula between sweep rate and resolution:
BW res 2
SW r = [8.8]
kF
where kF is a kind of form factor dependent on the type of filter used (for a Gaussian
filter we have kF ≈ 2).
For the adjustments concerning detected amplitude, we find the reference, that is
the maximum value displayed at the top of the screen, the unit chosen for the display
and the scale (quantity displayed by division). The unit of amplitude can be in volts
(or its sub-multiples), in dB (with different variants dB, dBv, dBm, and so on,
according to the reference chosen for the calculation of decibels). We see that
display in dB (logarithmic) is the most widely used, taking into account the very
wide gaps possible between amplitudes of different peaks.
Sweeping can be done by a analog VCO, but today it is more often carried out by
frequency synthesis devices with frequencies that are calculated according to the
span and number of measurement points. With these devices it is possible to average
the measurements carried out at each of the frequencies; this increases the signal-to-
noise ratio.
In
Att. ADC FFT
analog num.
FS
One important advantage of the FFT analyzer is its rapidity, since it establishes
all the components of the spectrum in frequency in one time; the measurement speed
is, at equal resolution, well above that of a sweep analyzer. Another advantage is
that it also allows us to obtain a good resolution even at low frequencies (as low as
Hz), which would be impossible with a sweep method.
The windows proposed by FFT analyzers are the following: uniform; Barlett (or
triangular); Hanning (in cosine); Hamming; Blackman; Kaiser; and Flattop. Each
type of window has certain advantages (such as fewer secondary lobes, a good
respect for the maximum value, among others), but these are to the detriment of the
length at half-maximum or to the measurement accuracy of the amplitude. Here, we
give the expressions of some of these windows, by the function w(n), defined as
0 ≤ n ≤ N – 1, and which is zero outside, n representing the number of the sample
[OPP 74].
2n N −1
w(n) = for 0 ≤ n ≤
N −1 2
Bartlett: [8.10b]
2n N −1
w(n) = 2 − for ≤ n ≤ N −1
N −1 2
338 Fundamentals of Instrumentation and Measurement
1⎡ ⎛ 2πn ⎞⎤
Hanning: w(n) = ⎢1 − cos⎜ ⎟⎥ [8.10c]
2⎣ ⎝ N − 1 ⎠⎦
⎛ 2πn ⎞
Hamming: w(n) = 0.54 − 0.46 cos ⎜ ⎟ [8.10d]
⎝ N −1 ⎠
⎛ 2πn ⎞ ⎛ 4πn ⎞
Blackmann: w(n) = 0.42 − 0.5cos ⎜ ⎟ + 0.08cos ⎜ ⎟ [8.10e]
⎝ N −1 ⎠ ⎝ N −1 ⎠
⎡ 2 2⎤
⎛ N −1 ⎞ ⎡ ⎛ N − 1 ⎞⎤ ⎥
I 0 ⎢ω a ⎜ ⎟ − ⎢n − ⎜ ⎟ ⎥
⎢ ⎝ 2 ⎠ ⎣ ⎝ 2 ⎠⎦ ⎥⎥
⎣⎢ ⎦
Kaiser: w(n) = [8.10f]
⎡ ⎛ N − 1 ⎞⎤
I 0 ⎢ω a ⎜ ⎟⎥
⎣ ⎝ 2 ⎠⎦
The Kaiser is defined from the Bessel function of order zero, of the first type.
The parameter a allows us to adjust the compromise between width of the central
lobe and amplitude of the secondary lobes.
Of course, we must mention direct spectral studies of a signal, for example the
signal delivered by an oscillator or modulated signals of type AM or FM. These help
us establish the spectral dimension around the carrier. Some spectrum analyzers
even include demodulation functions (one example is the ESA1500 made by Agilent
TechnologiesTM).
When we study the output signal of an amplifier, the spectrum analyzer allows a
certain number of measurements. The measurement of harmonic distortion rate is
worth mentioning here. This is an indicator of the relation of the energy contained in
all the harmonics and the energy contained in the fundamental; the results are
usually shown in percentages. The measurement is limited to a few harmonics, with
the analyzer offering the possibility of choosing the number (here we cite the
example of the FFT analyzer SR760 made by Stanford Research SystemsTM).
∑ Vi 2
2,..i..
THD = [8.11]
V1
Instruments and Measurement Chains 339
A very important measurement concerns the power spectral density (PSD). This
is the amplitude normalized to 1 Hz of passband (expressed as V / Hz or in
dB / Hz ). This measurement gives us a result independent of the span (an example
of this type of signal analyzer is the HP89410 made by Agilent TechnologiesTM).
The study of oscillators requires another type of measurement carried out with a
spectrum analyzer: the measurement of phase noise. This measurement lets us
encode the spectral purity of an oscillator, or the sharpness of the line corresponding
to the oscillation frequency. If the resolution of the spectrum analyzer filter is
written as BW and if we carry out the measurement at a distance fx of the central line
or carrier fc, the phase noise is then:
Figure 8.12 shows the principle of the measurement of phase noise done with a
spectrum analyzer. The measurement is carried out with a gap f in frequency in
relation to the central frequency of the oscillator (or of the carrier in a transmission
system).
f
B Bϕ (dBc / Hz ) = B + 10 log BW
8.1.5.1. S parameters
With high frequencies (radio frequencies or hyper frequencies), it is imperative
to take into account the propagation effects on the transmission lines, and even, in
some cases, on the components. In the case of lines (bifilar, coaxial, microstrip, and
waveguide), we define the incident and reflected waves as a line plane (V1, Vr) the
v
reflection coefficient ρ = r , the characteristic impedance RC, the reduced
vi
Z
impedance z = .
RC
340 Fundamentals of Instrumentation and Measurement
νi
We define the incident traveling wave by a = and the reflected traveling
RC
vr
wave by b = .
Rc
z −1 1+ ρ
ρ= and z = [8.13]
z +1 1− ρ
(b ) = [S ](a ) [8.14]
or also:
bi = ∑ Sij a j [8.15]
j
⎛b ⎞
S ii = ⎜⎜ i ⎟
⎟ is therefore the reflection coefficient of the port (i) when the
⎝ ai ⎠ a j =0
ports (j) are matched to Rc .
⎛b ⎞
Sij = ⎜ i ⎟⎟ is the transmission coefficient of the port (j) towards the port (i)
⎜a
⎝ j ⎠ ak = 0
all the ports (k ≠ j) being matched to Rc .
Att. ∆Φ R
S
Att. (j) Q (i) A
At first, an all pass replaces Q. The attenuators and the phase shifter are adjusted
to regulate R = A. Then we place Q, R is unchanged and A’ = Sij*A so that we get
A'
S ij = , and from this the module and the phase of the transmission coefficient.
R'
⎛ 1 ⎞
⎜0 0⎟
⎜ 2 ⎟
S =⎜
1 1⎟
0 [8.16]
⎜2 2⎟
⎜ 1 ⎟
⎜0 0⎟
⎝ 2 ⎠
Att. ∆Φ R
S
Att. (1) (3) A
(2)
(i)
Q
1
quadrupole at port (2) of the reflectometer. R = R’ and A' = S ii a1 and so
4
A'
S ii = , from which we get the module and the coefficient relection phase.
R'
The S-parameter test set is an essential part of a network analyzer and can be
integrated either to the device itself or in casing. This unit ensures all successive
connections of the source, of the quadrupole, of the reflectometer to ports (i) and (j)
alternately, and allows us to determine the four parameters S ii , Sij , S ji and S jj .
The results obtained can be shown on the analyzer screen unit in the form of a
Smith chart, mainly for the reflection coefficients or in the module and phase form
for the transmission coefficients. From the transmission coefficients Sii we establish
the input impedance at the level of port (i).
ZX
V1
V2
V
Z X = −R 1 [8.17]
V2
There is also a four-point method for very low impedances, which is quite
similar to that used for multimeters.
DUT
R
A
V1
V2
V2
V1 DUT
V2
Z=R [8.18]
V1
1 − S11
Z X = RC × [8.19]
1 + S11
P1 P2
DUT
With all these methods it is important to closely observe the quality of the fixture
systems of the components being tested (in particular for CMS components) and
then proceed to a calibration step of the measurement system. This last step controls
all stages leading up to and including measurement planning.
synchronous amplifier
In fact, with the useful signal spectrum x(t) concentrated close to f = 0, the
modulation of x(t) by the reference signal of the frequency fr leads to the creation of
a signal s(t) whose spectrum is that of x(t) but is shifted from f = 0 to f = ± fr. The
signal s(t) is then filtered by an amplifier tuned to the frequency fr, which eliminates
all the noise found outside the passband of the tuned amplifier. A synchronous
detection using a multiplier, a phase shifter and a passband filter restores the useful
signal x(t).
To illustrate this method, let us consider a signal x(t) provided with a white
noise. We suppose that the reference signal is purely sinusoidal. Figure 8.20 shows
the spectrums obtained at the output of the different analysis units shown in Figure
8.19.
346 Fundamentals of Instrumentation and Measurement
We see that this type of detection helps us measure the phase shifting between
two signals of the same frequency, as well as the separation of real and imaginary
components from a signal in the complex ensemble.
One of the important parameters of the lock-in amplifier is the time constant,
which is directly linked to the cut-off frequency of the passband filters. These filters
eliminate the 2 fr component, and also help reduce noise by diminishing the
bandwidth. For example, the SR510/SR530 amplifiers made by Stanford ResearchTM
have two passband filters. The first has time constants that are adjusted from 1 ms to
100 s for the first stage, and from 1 s to 0.1 s for the second stage. The bandwidth of
the amplifier is of 100 kHz.
To improve measurement accuracy, we find the input of band reject filters set to
the frequency of the sector. We then double this frequency.
There are also digital lock-in amplifiers (one example is the SR810/830 made by
Stanford ResearchTM). With these instruments, the synchronous detection is carried
out digitally. The signal is sampled (after antialiasing filtering) at the maximum
frequency of 256 kHz, then a DSP carries out the demodulation, that is, the “digital”
multiplication of the signal sampled by the reference signal. The signal is then
filtered. As an example, the DSP can carry out 16 million multiplications and
additions per second on 24 bits. The time constant is adjustable from 10 µs to 30 ks,
the input passband going from 1 mHz to 102 kHz.
Instruments and Measurement Chains 347
8.2.1. Introduction
Conditioning (SCXI)
Card E/S
internal
VXI
ISA,PCI,PCMCIA Sensor
or
Actuator
IEEE488
Serial
Depending on the type of measurement and the environment in which the system
functions, the designer can choose between different options. There is no one
solution to any given problem, since today there are so many possibilities from
which to choose. The new communication interfaces available to the public are
reflected in the field of instrumentation, and so offer new possibilities. The serial
bus USB is one of the latest examples.
348 Fundamentals of Instrumentation and Measurement
There are many devices (those previously mentioned) and almost 250
manufacturers (among them Agilent TechnologiesTM, TektronicsTM, National
InstrumentTM) that offer interfaces and programs for this bus.
Transmitter Receptor
Controller and/or and/or
receptor transmitter
EOI
REM Control line
SQR
ATN of the bus
IFC
NDAC Handshake
NRFD
DAV lines
The signals used for transmission are divided into three main levels, with
subdivisions according to functions.
– there are eight data lines. Each is a bidirectional bus ensuring word,
transmission, addresses and ASCII data;
350 Fundamentals of Instrumentation and Measurement
Electrical signals have levels compatible with TTL standards and work in
negative logic.
SRQ, NRFD and NDAC lines only use open collectors. For other lines, we find
both open collectors and three-state buffers that help us obtain transfer speeds above
250 Ko/s.
+5V
Input
14×3.3
+5V
=49.2mA
3kΩ
3.3mA max
0.4V
6.2kΩ
100pF
Output: max
level 0
Asynchronous mode
The ensemble of bits to be transmitted is organized as a frame of about 10 bits
maximum. Each bit has a duration Tb in time. This frame contains the following
features:
– a low-level start bit;
– a message containing N bits, usually a type ASCII code, with N having
between 5 to 8 bits;
– a control bit (Checksum) more or less equal to the transmitted message;
– one or two high-level stop bits.
Rest
Start message
Stop
Transmission P
t
Reception Activation
(sampling)
Tb
This very simple transmission mode is widely used. However, it does not allow
for significant flows (lower than 19,200 bauds, with a corresponding Tb period of
52 µs). In the majority of cases, the transmitter is faster than the receiver, usually
because the receiver takes longer to analyze data. However, the reception buffer is
generally very limited. This means the signal receiver used with the transmitter must
use a handshake system. This system can be material (one example is the Data
Terminal Ready (DTR) protocol) or it can consist of software (such as the Xon Xoff
protocol).
Synchronous mode
Here, the transmitter and the receiver are synchronized, which means their
transmission and reception clocks of identical frequencies and phases. We then can
obtain very long and high speed message transmissions.
We can speak of two structures for driver connections and line receivers:
– Unbalanced structures, in which one conductor wire is used for transmitting a
logical signal, as well as a ground conductor wire that can be shared when several
transmission lines are necessary.
– Balanced structures, a mode in which two conductor wires are needed to
transmit a logical signal. As with the unbalanced structure, one ground conductor
wire is used. The major advantage of this structure is that it is relatively insensitive
to environmental noise. For this reason it is widely used in industrial applications.
As well, it allows larger flows than unbalanced structures do.
Instruments and Measurement Chains 353
At this time, there are four important standards used in serial transmission
systems. There are drivers and receivers of connected lines corresponding to each
norm. Figure 8.25 shows the type of wiring for each norm needed for the
transmission of logical signals (apart from the TIA/EIA485 norm, which permits
bidirectional transmission).
E R TIA/EIA232
E
T R
TIA/EIA423
R R
E T R TIA/EIA422
R R
E T T E
R R
R
E
The voltages we find are basically due to the line capacities used for carrying
signals.
Bit/s speed
100M
10M Ethernet
1M
IEEE488 RS422
100k RS485
10k
I 2C
1k RS232 RS423
100
1 10 100 1k 10k Distance m
Their use is relatively simple and their costs are moderate, considering their
capabilities and performances. The major advantage of these cards is their good
transfer speed for data in terms of measurement and control. This means we can use
real-time analysis.
Instruments and Measurement Chains 355
In general, the number of inputs/outputs is fairly limited and can pose some
problems when a computer is not close to the process.
CH0 AGC
CH1
MUX ADC Memory
CH2
CH3
AR filter
Tim0
Timer Clock
Tim1
Interface PC
control
PIO
Logic
IN/OUT
DAC Synthesis
CH4
Bus PC
If signal conditioning is still hard to carry out, we can use external signal
conditioning modules, such as the SCXI (Signal Conditioning eXtension for
Instrumentation).
The cards developed for instrumentation use the standard principles of computer
extension cards.
– The first is the Industry Standard Architecture (ISA) bus. Although it is older,
it is still used for instrumentation purposes. This is an asynchronous 16 bit bus set to
a rhythm of 8.33 MHz, with a transmission rhythm that does not exceed 1 to 2 Mo/s
because of cycles and interruptions. However, there are cards in ISA format that
allow for sampling of signals at frequencies of the order of several tens of Mech/s
that must be integrated with memory. The transfer towards the PC is then done
according to a lower rhythm and does not allow for a real-time analysis.
– The second is the Peripheral Component Interconnect (PCI) bus, which was
developed by Intel in 1993 and has been widely used since 1995. This is a 32 bit bus
set at 33 MHz, allowing for a theoretical maximum flow of 132 Mo/s. This is higher
than the flows allowed by the ISA bus, which explains its popularity for users
Instruments and Measurement Chains 357
needing rapid acquisition cards. It also has “plug and play” features, eliminating the
cordless plugs needed for cards of the ISA formats. Again, differing from the ISA
system, the cards of PCI format integrate the DMA controller on the card itself, so
the bus can have autonomous control; thus, these are called master cards. There are
also “slave” cards that cannot integrate DMA functions in this way.
Specifications for the VXI bus completely describe the P2 and P3 connectors.
The P2 bus is made of several elements. It has a VME bus, a 10 MHz clock,
activation lines with ECL and TTL logical levels, an analog summation line, a
module identification line and a local bus. The P3 connector has supplementary lines
for a local bus structure, a clock line set at 100 MHz, and high performance
activation lines in ECL logic.
This design gives activation signals above 50 MHz and local data transfers of
100 Mb/s.
358 Fundamentals of Instrumentation and Measurement
VXI VME
P1 A
3.9’× 6.3’
P1 P1 VME Computer bus
16 bit transfer, 16 Mo addressing
P1 Multi-master Arbitration bus
B 9.2’× 6.3’ Priority Interrupt bus, Utilities bus
P2
P2 Center rows adds:
VME 32 bit data & 4Go addressing
P1 P2 Outer rows adds:
C P2 10 MHz Clock bus, TTL&ECL Trigger
P2 9.2’× 13.4’ 12 Pin Local bus, Analog Sum bus
Module Identification bus
Power Distribution bus
P1 P3 adds:
D
P2
14.4’× 13.4’ P3 100 MHz Clock bus, ECL Star Bus
P3 ECL Trigger bus, 24 pin local
Power Distribution bus
The VXI system is very complete and offers a standard of performance for the
field of instrumentation. However, it is still relatively expensive and therefore is
found only in top-of-the-line acquisition and automatic test systems.
8.3. Bibliography
[ASC 99] ASCH G., Acquisition de données, du capteur à l’ordinateur, Dunod, 1999.
[COM 96] COMBES P.F., Micro-ondes, volume 1, Lignes guides et cavités, Dunod, 1996.
[COM 97] COMBES P.F., Micro-ondes, volume 2, Circuits passifs, propagation, antennes,
Dunod, 1997.
[COO 95] COOMBS C.F., Electronic Instrument Handbook, McGraw-Hill, 1995.
[DIE 99] DE DIEULEVEULT F., Electronique appliquée aux hautes fréquences, Dunod,
1999.
[IMH 90] The Impedance Measurement Handbook, Hewlett-Packard, 1990.
[OPP 74] OPPENHEIM A.V., SCHAFER R.W., Digital Signal Processing, Prentice Hall,
1974.
[SEI 99] Scientific and Engineering Instruments, Stanford Research Systems, 1998-1999.
[TMC 99] Test and Measurement Catalog 2000, Agilent Technologies, Dec. 1999.
[WHI 93a] WHITE R.A., Electronic Test Instruments – Theory and Applications, Prentice
Hall, 1993.
[WHI 93B] WHITE R.A., Spectrum & Network Measurements, Prentice Hall, 1993.
This page intentionally left blank
Chapter 9
The focus of this chapter will be to describe the relation between a sensor’s
output variables and the physical variable applied to its input, called the measurand.
This relation can also take into account the role played by other variables, a
priori external, that can cause variations in the output signal (some examples are the
sensor’s feed tension and the temperature).
The best approach is first to analyze what a sensor does and then to understand
and completely describe the physical processes of transduction. We then convert
these into mathematical forms by using physical laws. The result is an equation that
links output variables to input variables (creating a knowledge model).
This is an experimental method that consists of collecting data. The values of the
output signals are compared to the values taken by the input variables under a set of
given conditions. The study of the structure of these data helps us collect results in
the form of mathematical relations (dependency models) that explain the
observations.
9.1.3. Adjustment
Sensor output
Sensor
Adjustment Deviation – +
algorithm
In both cases, the method uses the following steps (see Figure 9.1):
– using either physical laws or a close observation of data, we establish a model
(equation) that has a certain number of unknown parameters;
– by setting an optimization criterion (the least squares, for example), we look
for parameter values that “at least” adjust the observed deviations between the
sensor’s output signal and the output signal of the model;
– the study of deviations between the data and the adjusted variables allows us,
for the first time, to verify the adequacy of the model and then, a second time, to
estimate the limits of this adequacy in terms of variability.
Elaboration of Models for the Interaction Between the Sensor and its Environment 363
Let us assume Y, X1, X2 are the variables represented by the output signal of a
sensor according to the input variables (signal, limiting quantity and so on). After a
physical analysis, we know that there is a mathematical relation between these
quantities, written as:
We can also observe the measurements carried out with this sensor and make the
hypothesis that there is an expression such as the one in equation [9.1].
In both cases, the problem consists, apart from observations, of calculating the
values of the parameter model. For example, the relation between the input signal X
and the output signal Y can be a polynomial form of degree k:
Y = θ 0 + θ1 X + θ 2 X 2 + ... + θ k X k [9.2]
where we estimate the k+1 parameters from the n pairs of observations (x1, y1), (x2,
y2), …, (xn, yn).
The situation discussed here is when the number of points is strictly higher than
the number of parameters to be assessed. Under these conditions, the n equations
representing the n measurements cannot be resolved simultaneously. We then have
to analyze the system:
y 2 = θ 0 + θ 1 x 2 + θ 2 x 22 + .... + θ k x 2k + e 2 [9.4]
y n = θ 0 + θ 1 x n + θ 2 x n2 + .... + θ k x nk + e n [9.6]
where the quantities e1,e2,….,en represent the deviations between the supposed or
theoretical model and the effected measurements. The value of each of these
deviations varies according to the optimization criterion adapted to analyze this
problem.
The criterion used here will be the least squares method of Gauss, who described
it as follows: “The estimator of the parameters 0, 1,…, k are the specific values that
reduce to minimum the sum of the squared deviations between the experimental
observations and the corresponding values predicted by the adopted theoretic
model.”
n
Q(θ à , θ 1 ,..., θ k ) = ∑e
i =1
2
i [9.7]
Let x1, x2, …, xn be the results obtained during n independent repetitions of the
measurement of the same variable under the same experimental conditions. We
make the hypothesis that these results are n specific numerical values of an expected
variable µ that we want to estimate (the expectation represents the ideal value that
the variable has under ideal conditions, that is, without random disturbance). The
obtaining conditions being identical, we can formulate the hypothesis that these
results have the same variance j2. The n measurement results are translated by the
system of the following n equations:
x1 = µ + e1 [9.8]
x 2 = µ + e2 [9.9]
x i = µ + ei [9.10]
Elaboration of Models for the Interaction Between the Sensor and its Environment 365
x n = µ + en [9.11]
n n
Q( µ ) = ∑ i =1
ei2 = ∑ (x
i =1
i − µ)2 [9.12]
300
Sum of squared
250
deviations
200
150
100
50
0
0 2 4
Parameter 6 8 10
The minimum of this function is attained for the value µ̂ of µ , which is the
solution of the equation obtained by writing that:
dQ
=0 [9.13]
dµ
∑x
i =1
i
µ̂ = [9.14]
n
In other terms, the estimator of the central value of the measurement results, in
the sense of the least squares, is simply the average arithmetic x of the values
obtained. We know that this estimator is not biased; that is, its expectation is equal
to µ. The variance of this estimator is then:
366 Fundamentals of Instrumentation and Measurement
σ2
V ( µˆ ) = V ( x ) = [9.15]
n
Let us look again at the example of finding the estimator of the central value
parameter of a measurement series.
Let x1, x2, …, xn be the results obtained during n independent repetitions of the
measurement of the same value. We now make the hypothesis that certain
experimental conditions have varied during the measurements, so that the variance
associated with each result varies from one measurement to another. This leads to:
V ( xi ) = σ i2 [9.16]
However, the expectation of each result is the same; thus, we let µ be the
expectation of this variable. The n measurement results are expressed by the same
system as before:
x1 = µ + e1 [9.17]
x 2 = µ + e2 [9.18]
x i = µ + ei [9.19]
x n = µ + en [9.20]
To take in consideration the fact that the measurements do not have the same
variance, we weigh each square of the deviance with a weighting (or weight)
coefficient gi. We then look at the expression:
n n
Q( µ ) = ∑
i =1
g i. ei2 = ∑ g (x
i =1
i i − µ) 2 [9.21]
The minimum of this function is attained for the value µ̂ p of µ, which is the
solution of the equation obtained by writing that:
dQ
=0 [9.22]
dµ
Elaboration of Models for the Interaction Between the Sensor and its Environment 367
∑g x i i n
µ̂ p = i =1
n
= ∑w x i i [9.23]
∑g i =1
i
i =1
n
where wi is a normed weight such that ∑w
i =1
i = 1. [9.24]
V ( µˆ p ) = ∑ wi2σi2 [9.25]
We can try to determine the form of the weights that bring about the smallest
variance for ûp. Looking for the minimum of this function, taking into account the
constraint on the sum of the weights, gives us:
1
gi = [9.26]
σ i2
and
1
σ i2
wi = n
[9.27]
∑
i =1
1
σ i2
1
V ( µˆ p ) = n
[9.28]
∑σ
1
2
i =1 i
368 Fundamentals of Instrumentation and Measurement
With a sensor, the parameter factor of X is the sensitivity of the sensor, with the
scale gap being a constant parameter that is either optional when the instrument has
an expanded scale or is required in the case of an inopportune zero gap. We discuss
the two situations by presenting the following hypotheses:
– H1, with values xi that are perfectly known (E(xi) = xi and V(xi ) = 0);
– H2, with values of Y made without systematic errors;
– H3, with the variable Y measured with a constant variance, so V(yi) = 2
constant;
– H4, with independent measurements of Y, that is cov(yi, yj) = 0 when i ≠ j.
y1 = θ .x1 + e1 [9.29]
y 2 = θ .x 2 + e 2 [9.30]
y i = θ .x i + e i [9.31]
y n = θ .x n + e n [9.32]
The terms e1, e2, …, en express the gaps between the observed values of Y and
the values predicted by the chosen model. Hypothesis H2 immediately shows that
each of these gaps is, on average, zero. So E(ei) = 0 whatever the index i of the
measurement.
dQ
=0 [9.34]
dθ
∑
i =1
θˆ.x i2 = ∑x y
i =1
i i [9.35]
∑x y
i =1
i i
θˆ = n
[9.36]
∑
i =1
x i2
σ2
V (θˆ) = n [9.37]
∑i =1
x i2
For each measured point of the abscissa xi, we call the residuals ri the gap
existing between the measured value yi and Y of the corresponding value yˆ i = θˆ.x i
of the model:
ri = y i − yˆ i = y i − θˆ.x i [9.38]
n
If we look at the normal equation, we see that ∑ r .x
i =1
i i =0 [9.39]
370 Fundamentals of Instrumentation and Measurement
This property allows us, a posteriori, to verify the validity of the numerical
value of θˆ .
∑r
i =1
i
2
σˆ 2 = [9.40]
n −1
Using the least squares line equation helps us calculate the interposed value of
Y that corresponds to an undetermined abscissa of the field validity of the model:
yˆ = θˆ.x [9.41]
σ 2x2
V ( yˆ ) = x 2V (θˆ) = n [9.42]
∑
i =1
x i2
We will look at the previous schema again by adapting it to the case of an affine
model in order to explain the measurements. The system representing the measured
n points is written:
y1 = θ o + θ 1 .x1 + e1 [9.43]
y 2 = θ o + θ 1 .x 2 + e 2 [9.44]
Elaboration of Models for the Interaction Between the Sensor and its Environment 371
y i = θ o + θ 1 .x i + e i [9.45]
y n = θ o + θ 1 .x n + e n [9.46]
The terms e1, e2, …, en express the deviations between the observed values of Y
and the values predicted by the chosen model. Hypothesis H2 leads us immediately
to the conclusion that each of the gaps are on average zero, so E(ei) = 0 whatever the
index i of the measurement. By applying the Gaussian criterion, we get the quantity:
n n
Q(θ 0 , θ 1 ) = ∑
i =1
e i2 = ∑ (x
i =1
i − θ 0 − θ 1 .x i ) 2 [9.47]
Even though no relation exists between 0 and 1 (that is, they can vary
independently from one another), the minimum value of this surface is reached for
the value θˆ0 and 0 and the value θˆ1 of 1 that are the solutions of the usual
equation systems obtained when we write:
∂Q
=0 [9.48]
∂θ 0
∂Q
=0 [9.49]
∂θ 1
n n
nθˆ0 + θˆ1 ∑
i =1
xi = ∑y
i =1
i [9.50]
n n n
θˆ0 ∑
i =1
x i + θˆ1 ∑
i =1
x i2 = ∑x y
i =1
i i [9.51]
372 Fundamentals of Instrumentation and Measurement
n n n n ⎡ ⎤
∑ x ∑ y − ∑ x ∑ x .y
2
i i i i i n
⎢
⎢1
⎥
x (xi − x ) ⎥
θˆ0 = i =1 i =1 i =1 i =1
2
= ∑ yi ⎢ − n ⎥ [9.52]
⎛ n ⎞ ⎢n
∑ (xi − x ) ⎥
2⎥
n
i =1
n ∑
i =1
x i2 − ⎜
⎜
xi ⎟
⎝ i =1 ⎠
⎟ ∑ ⎢
⎣ i =1 ⎦
n n n n
n ∑
i =1
xi . y i − ∑ ∑
i =1
xi
i =1
yi ∑ y .(x
i =1
i i − x)
θˆ1 = 2
= n
[9.53]
⎛ n ⎞
∑ (x − x)
n
∑ ∑
2
n x i2 − ⎜ xi ⎟ i
⎜ ⎟ i =1
i =1 ⎝ i =1 ⎠
For each of the solutions, the first form, directly deduced from normal equations,
is the fastest for carrying out calculations. This is because the different sums that are
part of it are calculated as input data arrives. However, the second form, which
requires a priori calculation of the average of the values of X and Y, interposes the
differences into these averages and is less sensitive to rounding errors of the
calculation systems.
These estimators are expressed according to the random variables. These are
therefore random quantities. Taking into account hypothesis H2, these are unbiased
estimators, which means that E ( θˆ0 ) = θ 0 and that E( θˆ1 ) = θ 1 .
nσ 2 σ2
V (θˆ1 ) = 2
= n
[9.55]
⎛ n ⎞
∑ (x − x)
n
∑ ∑
2
n x i2 − ⎜ xi ⎟ i
⎜ ⎟ i =1
i =1 ⎝ i =1 ⎠
Elaboration of Models for the Interaction Between the Sensor and its Environment 373
Also, we see that the obtained estimators are usually correlated, so their
covariance expression is as follows:
n
−σ 2 ∑x
i =1
i
−σ 2 x
cov(θˆ0 , θˆ1 ) = 2
= n
[9.56]
⎛ n ⎞
∑ (x − x)
n
∑ ∑
2
n x i2 − ⎜ xi ⎟ i
⎜ ⎟ i =1
i =1 ⎝ i =1 ⎠
Later we will give an explanation for this correlation and why the covariance is
zero when the arithmetic mean of the values of X are zero.
For each measured point, we call residuals ri, the gap existing between the value
measured yi of Y and the corresponding value yˆ i = θˆ0 + θˆ1.xi of the model:
∑r
i =1
i =0 [9.58]
This property shows that the residuals are positive for some, negative for others,
and that overall, the sum cancels itself out: the line of the least squares goes from
“the middle” of the scatter of measured points and can be either above or below the
line.
∑ r .x
i =1
i i =0 [9.59]
The residuals also help to obtain an unbiased estimator σ̂ 2 of the variance. The
residuals make up the measurements of Y, on condition that the chosen model is
pertinent:
374 Fundamentals of Instrumentation and Measurement
∑r i =1
i
2
σˆ 2 = [9.60]
n−2
The first of the normal equations shows that the point whose coordinates are:
n n
∑
i =1
xi ∑y
i =1
i
x= , y= [9.61]
n n
belongs to the least squares line. This point is linked to the measurements made by
the system and not to the adjusted line.
is always satisfied.
This has two consequences. The first is that we can now calculate one of the
estimators (in practice, θˆ0 ) from knowing the other one (in practice, θˆ1 ), and the
values of x and y . The second is that it qualitatively explains the correlation
between θˆ0 and θˆ1 . Since the line must go through the point of the coordinates x ,
y being fixed for a data set, any attempt to modify the gradient, for example, can
only be done by rotation around this point.
Let us suppose that the average of X is strictly positive. Under these conditions,
the covariance between θˆ0 and θˆ1 is negative. We then see graphically that
augmenting the gradient means a diminishment of the ordinate source. This explains
the covariance sign. A completely similar conclusion can be obtained when the
average of X is negative. In the specific case when the average of X is zero, this
particular point is on the ordinates axis. The rotation of the line does not mean
source ordinate modification, and the covariance is zero.
Elaboration of Models for the Interaction Between the Sensor and its Environment 375
Using the least squares line equation allows us to calculate the interposed value
of Y corresponding to an indeterminate abscissa x of the validity domain of the
model:
By replacing the variances of the estimators and the covariance between the
estimators with their respective expressions, we get the forms:
⎡ ⎤ ⎡ ⎤
⎢ ⎥ ⎢ ⎥
⎢ n( x − x )2 ⎥ ⎢ ( x − x )2 ⎥
2 ⎢1 ⎥ = σ ⎢1 +
V ( yˆ ) = σ + 2
⎥ [9.65]
⎢n ⎛ n ⎞ ⎥
2
⎢n 2⎥
∑ (x i − x ) ⎥
n
⎢
⎢
n ∑
x i2 − ⎜
⎜ ∑ xi ⎟ ⎥
⎟ ⎥
⎝ i =1 ⎠ ⎦
⎢
⎢⎣ ⎥⎦
⎣ i =1
We see that the variance of varies according to the value of x (see Figure 9.3).
It presents a minimum of x = x , where its value is:
σ2
[V ( yˆ )]x = x = [9.66]
n
The estimated ordinate variance of the line of least squares increases according
to the lengthening function at x .
376 Fundamentals of Instrumentation and Measurement
All these expressions can be evaluated by replacing the variance j2 with its
estimator, which has been obtained from the residual sum of squares.
Y
0 5 10 15
Figure 9.3. Approximate gap envelope around the least squares line
We can also reverse the least squares line method by calculating the abscissa
x̂ that corresponds to an ordinate :
yˆ − θˆ0
xˆ = [9.68]
θˆ 1
Applying the variance composition law to this expression leads to the following
conclusion:
V ( yˆ )
V ( xˆ ) = [9.69]
θˆ 2
1
There is a direct correspondence between the gap type of the abscissa and the
gap type of the ordinate throughout the gradient of the least squares line.
We should remember that the least squares criterion does not, when used alone,
allow us to test the validity of the chosen linear model.
40
30
20
residuals
10
0
-10
-20
-30
x
Figure 9.4. Distribution of residuals according to the model being used
2
⎡ n ⎤
∑
⎢ ( x i − x ).( y i − y )⎥
⎣⎢ i =1 ⎦⎥
R= [9.70]
⎡ n ⎤⎡ n ⎤
∑
⎢ ( xi − x ) ⎥ ⎢ ( y i − y ) ⎥
⎢⎣ i =1
2
⎥⎦ ⎢⎣ i =1
2
⎥⎦
∑
n n
θˆ1 ∑
i =1
( xi − x ) 2 ∑r
i =1
1
2
R= n
= 1− n
[9.71]
∑(yi=
i − y) 2
∑(y
i=
i − y) 2
We can easily see that if the experimental points are perfectly aligned, that is, if
the dispersion of the values of Y are completely explained by the chosen theoretical
model, the residuals are zero, so R = 1. However, if the values of Y are independent
of the values of X, that is, if the gradient of the model is zero, then R = 0. Aside
from these two extreme cases (which are rarely found in practice), we must be
378 Fundamentals of Instrumentation and Measurement
careful to draw conclusions solely from the value of R, since different forms of point
scatters can lead to the same value of the correlation coefficient.
V ( y i ) = σ i2 [9.72]
This is the case when measurements are made with a type of constant relative
gap ji/yi. In these conditions, we use a weighting coefficient that converts the more
or less high proximity of the passage of the line near to the point according to the
uncertainty function that is being affected.
n n
Q(θ ) = ∑ g e = ∑ g (y
i =1
2
i i
i =1
i i − θ .x i ) 2 [9.73]
The minimum of this function is reached for the value θˆ p of , which is the
solution of the equation obtained by writing that:
dQ
=0 [9.74]
dθ
n n
∑i =1
θˆ p g i .x i2 = ∑g x y
i =1
i i i [9.75]
∑g x y
i =1
i i i
θˆ p = n
[9.76]
∑i =1
g i x i2
Taking into account the hypotheses that have already been formulated, the
estimator is always unbiased, which means that E( θˆ p ) = θ .
∑g
i =1
i xi σ i
2 2 2
V (θˆ p ) = 2
[9.77]
⎡ n ⎤
⎢ ∑ g i x i2 ⎥
⎣⎢ i =1 ⎦⎥
This is the function of values that we can give to the weighting coefficients.
We can find the weighting values that allow us to obtain a minimal value for
V (θˆ p ) , that is, for the solutions obtained by writing:
∂V (θˆ p )
=0 [9.78]
∂p j
This means that we find the fact that the weight that minimizes the variance is
inversely proportional to the variance of the quantity it weights:
gi = 1 2 [9.80]
σ i
In these conditions, the gradient estimator and its variance respectively take the
following expressions:
380 Fundamentals of Instrumentation and Measurement
∑σ
xi y i
2
i =1
θˆ p = n
i
[9.81]
x i2
∑i =1 σ i
2
1
V (θˆ p ) = n
[9.82]
x i2
∑σ
i =1
2
i
In the specific case of a variance being identical for each measurement, these
expressions result in the same relations as those shown for the non-weighted case.
For the measurement point of an abscissa xi, the estimated ordinate is yˆ i = θˆ.x i ,
and the residual is worth:
ri = y i − yˆ i = y i − θˆ.x i [9.83]
∑ g r .x
i =1
i i i =0 [9.84]
A posteriori, this property allows us to verify the validity of the numerical value
of θ̂ p .
n n
Q(θ o , θ 1 ) = ∑
i =1
g i ei2 = ∑ g (y
i =1
i i − θ 0 − θ 1 .x i ) 2 [9.85]
Elaboration of Models for the Interaction Between the Sensor and its Environment 381
The minimum of this function is attained for the value θˆ0 p of 0 and the value
θˆ
1 p of 1, solutions of the equation obtained by writing:
∂Q
=0 [9.86]
∂θ 0
∂Q
=0 [9.87]
∂θ 1
n n n
θˆ0 p ∑g
i =1
i + θˆ1 p ∑g x =∑g y
i =1
i i
i =1
i i [9.88]
n n n
θˆ0 p ∑
i =1
g i x i + θˆ1 p ∑i =1
g i x i2 = ∑g x y
i =1
i i i [9.89]
n n n n
∑ i =1
g i x i2 ∑
i =1
g i yi − ∑i =1
g i xi ∑g x y
i =1
i i i
θˆ0 p = 2
[9.90]
n n ⎛ n ⎞
∑g ∑ i g i x i2 − ⎜
⎜ ∑ g i xi ⎟
⎟
i =1 i =1 ⎝ i =1 ⎠
n n n n
∑ ∑
i =1
gi
i =1
g i xi y i − ∑
i =1
g i xi ∑g y
i =1
i i
θˆ1p = 2
[9.91]
n n ⎛ n ⎞
∑g ∑ i g i x i2 − ⎜
⎜ ∑ g i xi ⎟
⎟
i =1 i =1 ⎝ i =1 ⎠
Taking into account the previously formulated hypotheses, these estimators are
unbiased, that is:
E( θˆ0 p ) = θ 0 [9.92]
E( θˆ1 p ) = θ 1 [9.93]
382 Fundamentals of Instrumentation and Measurement
by positing:
n n
∑g x
i =1
i i ∑g y
i =1
i i
xp = n
and y p = n
[9.95]
∑g
i =1
i ∑g
i =1
i
which are respectively the weighted average of the values of X and Y. The point of
the coordinates x p , y p belongs to the least squares line. Consequently, it is possible
to obtain a new set of relations that give estimators by using a new axial system,
parallel to the initial axes but centered on the point of coordinates x p , y p . Here, the
line intersects the origin and we find a proportional form, with the same leading
coefficient, so:
∑ g y .(x )
n n
∑
i =1
g i ( x i − x p ).( y i − y p )
i =1
i i i − xp
θˆ1p = = [9.96]
∑ g (x )
n n
∑ g (x
i =1
i i − xp) 2
i =1
i i − xp 2
As was the case before, this second set of solutions, even though requiring a
prior calculation of the weighted averages of X and Y, produce values that are less
sensitive to calculation errors. This is because the expressions only relate to gap
values measured in relation to these averages.
These estimators are expressed according to random variables. This means they
are random quantities. Taking into consideration hypothesis H2, these are unbiased
estimators, so that E( θˆ0 ) = θ 0 and E( θˆ1 ) = θ 1 .
Elaboration of Models for the Interaction Between the Sensor and its Environment 383
The following expressions give the variances of each estimator, as well as their
covariances.
∑g x
i =1
2
i i
1 x p2
V (θˆ0 p ) = = + [9.98]
∑ g ∑ g (x )
2 n n
n n ⎛ n ⎞
∑g ∑ ∑ − xp 2
g i x i2 − ⎜ g i xi ⎟ i i i
i ⎜ ⎟ i =1 i =1
i =1 i =1 ⎝ i =1 ⎠
n
∑g
i =1
i
1
V (θˆ1 ) = = [9.99]
∑ g (x )
2 n
n n ⎛ ⎞n
∑g ∑ ∑ − xp 2
g i x i2 − ⎜ g i xi ⎟ i i
i ⎜ ⎟ i =1
i =1 i =1 ⎝ i =1 ⎠
and
n
− ∑g xi =1
i i
− xp
cov(θˆ0 , θˆ1 ) = = [9.100]
∑ g (x )
2 n
n n ⎛ n ⎞
∑g ∑ ∑ − xp 2
g i x i2 − ⎜ g i xi ⎟ i i
i ⎜ ⎟ i =1
i =1 i =1 ⎝ i =1 ⎠
The variance values of the obtained estimators are functions of the values
attributed to weighting coefficients.
As before, we can show that the weights that minimize these quantities are
inversely proportional to the gap variance, so:
gi = 1 [9.101]
σ i2
Using the least squares line equation allows us to calculate the interposed value
of Y corresponding to an indeterminate abscissa x of the validity field of the model:
By replacing the variance of the estimators and the covariance between the
estimators with their respective expressions, we get the form:
1 ( x − x p )2
v( yˆ ) = n + n [9.104]
∑ gi ∑ gi (xi − x p ) 2
i =1 i =1
9.3.4. The least measured-squares line: when two measured variables contain
uncertainties
The previous sections do not discuss the problem of determining line coefficients
when there are uncertainties only for variables represented as ordinates.
If this inequality is not resolved, we can resolve the problem by using the method
developed by Williamson.
y = θ 0 + θ1 x [9.106]
We measure n pairs of values (xi, yi), each of these measurements being seen as a
random variable, with the following variances:
V ( xi ) = pi [9.107]
V ( y i ) = qi [9.108]
Elaboration of Models for the Interaction Between the Sensor and its Environment 385
It is not possible to bring in covariance a priori, since the variables X and Y are
results of different experimental processes.
n ⎡⎛ x − X 2
⎞ ⎛ y i − Yi ⎞
2⎤
Q= ∑ ⎢⎜ i
⎢⎜⎝ p i
i
⎟ +⎜
⎟ ⎜ q
⎠ ⎝
⎟
⎟
⎠
⎥
⎥
[9.109]
i =1 ⎣ i
⎦
or again:
n ⎡⎛ x − X 2
⎞ ⎛ yi −θ 0 −θ1 X i ⎞
2⎤
Q= ∑ ⎢⎜ i
⎢⎜⎝ p i
i
⎟ +⎜
⎟ ⎜
⎠ ⎝ qi
⎟
⎟
⎠
⎥
⎥
[9.110]
i =1 ⎣ ⎦
n ⎡⎛ x − X 2
⎞ ⎛ vi ⎞
2⎤
Q= ∑ ⎢⎜ i
⎢⎜⎝ p i
i
⎟ +⎜
⎟ ⎜q
⎠ ⎝ i
⎟
⎟
⎠
⎥
⎥
[9.111]
i =1 ⎣ ⎦
by proposing:
vi = yi − θ 0 − θ 1 xi [9.112]
Williamson handles the problem by minimizing each quantity written inside the
bracket by proposing that:
∂Q
=0 [9.113]
∂X i
386 Fundamentals of Instrumentation and Measurement
We get:
n
Q= ∑ g .v
i =1
i
2
i [9.114]
by proposing:
1
gi = [9.115]
q i + θ 12 p i
We see that the proposed method means considering the squared gap, measured
parallel to the axis of the ordinates, between the experimental point and the
theoretical line. This quantity is modified by a weighted coefficient projected onto
the axis of the ordinates of the variance of yi, and then by the component projected
onto the axis of the ordinates of the variance of xi. This means that we come back to
the standard situation analyzed above: the calculation of vi as well as of gi requires
knowledge of the theoretical line that we are trying to measure. We thus
immediately know that resolving the problem requires going through an iterative
process or phase.
∂Q
n
∂v i
∂θ 0
= 2 g i vi
i =1
∑∂θ 0
=0 [9.116]
n
∂Q ⎛ ∂g ∂v ⎞
∂θ 1
= ∑ ⎜⎜⎝ vi2 ∂θ1i + 2 g i vi ∂θi1 ⎟⎟⎠ = 0 [9.117]
i =1
Taking into account the relations existing between vi and gi on the one hand, and
between 0 and 1 on the other, we get:
∂v i
= −1 [9.118]
∂θ 0
∂v i
= − xi [9.119]
∂θ 1
Elaboration of Models for the Interaction Between the Sensor and its Environment 387
∂g i
= −2 p iθ 1 g i2 [9.120]
∂θ 1
∑g v
i =1
i i =0 [9.121]
n
∑ g i vi ( g i vi piθˆ1 + xi ) = 0 [9.122]
i =1
By replacing gi and vi with their expressions in the first of the normal equations,
and by proposing:
n n
∑ g i xi ∑ g i yi
xπ = i =1 and yπ = i =1 [9.123]
n n
∑ gi ∑ gi
i =1 i =1
This easily shows that the estimated line intersects with the point of the
coordinates xπ , yπ , a result we have already seen with more restrictive hypotheses.
x i′ = x i − xπ [9.125]
y i′ = y i − yπ [9.126]
and:
z i = x i′ g i q i − g i p iθˆ1 y i′ [9.127]
388 Fundamentals of Instrumentation and Measurement
∑ g z ( y ′ θˆ .x ′ ) = 0
i =1
i i i− 1 i [9.128]
∑ g z y′
i =1
i i i
θˆ1 = n
[9.129]
∑ g z x′
i =1
i i i
This format is less satisfactory than it seems, since gi and also zi are functions of
θˆ1 and therefore of the solution!
1
0 gi = [9.130]
qi + 0 θˆ12 . pi
∑ i =1
0 g i .x i
0 xπ = n
[9.131]
∑i =1
0 gi
∑ i =1
0 g i .yi
0 yπ = n
[9.132]
∑ i =1
0 gi
0 x i′ = x i − 0 xπ [9.133]
0 y i′ = y i − 0 yπ [9.134]
Elaboration of Models for the Interaction Between the Sensor and its Environment 389
and:
0 z i = 0 x i′ . 0 g i .q i − 0 g i . p i . 0 θˆ1 . 0 y i′ [9.135]
∑ i =1
0 g i . 0 z i . y i′
1θ 1
ˆ = [9.136]
n
∑ i =1
0 g i . 0 z i .x i′
which helps us calculate new values for g i , xπ , yπ , x i′ , y i′ , z i and thus for θˆ1 , etc.,
continuing up to the convergence towards the gradient value.
As with all iterative problems, it is important to start from a value close to the
solution, both to minimize the number of iterations and to guarantee the
convergence towards the desired value, even though here this last point is not an
issue.
n
V (θˆ1 ) = T 2 ∑ g .( x ′ q
i =1
i i
2
i + y i′ 2 p i ) [9.138]
1
V (θˆ0 ) = n
+ 2.( xπ + 2 z π ) z π T + ( xπ + 2 z π ) 2 V (θˆ1 ) [9.139]
∑g
i =1
i
expressions in which
1
T= [9.140]
n ⎡ x′.y′ ⎤
∑i =1
g i .⎢ i i + 4.( z i − z π ).( z i − x i′ )⎥
⎣⎢ θ 1
ˆ
⎦⎥
390 Fundamentals of Instrumentation and Measurement
and:
n
∑ g .z
i =1
i i
zπ = n
[9.141]
∑g
i =1
i
All these relations restore values already seen when uncertainties only affect the
variable Y, whether in a constant way as a variable according to the measurement
index.
Now we come to a situation where the adjustment model is written in the form of
a polynomial development of X. We get:
Y = f ( X ) = θ 0 . f 0 ( X ) + θ 1 . f1 ( X ) + θ 2 . f 2 ( X ) + θ k . f k ( X ) [9.142]
We try to find the estimators of the parameters 0, 1, 2,…, k of the model from
the measurement of n pairs of values (xi , yi) when n > k + 1.
Y = f ( X ) = θ 0 + θ1 .X + θ 2 .X 2 + θ k .X k [9.143]
θ 0 . f 0 ( x1 ) + θ 1 . f 1 ( x1 ) + θ 2 . f 2 ( x1 ) + θ k . f k ( x1 ) + e1 = y1 [9.144]
θ 0 . f 0 ( x 2 ) + θ 1 . f1 ( x 2 ) + θ 2 . f 2 ( x 2 ) + θ k . f k ( x 2 ) + e2 = y 2 [9.145]
θ 0 . f 0 ( x n ) + θ 1 . f1 ( x n ) + θ 2 . f 2 ( x n ) + θ k . f k ( x n ) + en = y n [9.146]
⎡ y1 ⎤ ⎡ e1 ⎤ ⎡θ1 ⎤
⎢y ⎥ ⎢e ⎥ ⎢θ ⎥
y = ⎢ 2 ⎥ e = ⎢ 2⎥ θ = ⎢ 2⎥ [9.147]
⎢ ... ⎥ ⎢ ... ⎥ ⎢ ... ⎥
⎢ ⎥ ⎢ ⎥ ⎢ ⎥
⎣ yn ⎦ ⎣e n ⎦ ⎣θ n ⎦
A.θ + e = Y [9.148]
The matrix:
⎡ f 0 ( x1 ) f1 ( x1 ) ... f k ( x1 ) ⎤
⎢ f (x ) f1 ( x2 ) ... f k ( x 2 )⎥⎥
A=⎢ 0 2 [9.149]
⎢ ... ... ... ... ⎥
⎢ ⎥
⎣ f 0 ( xn ) f1 ( xn ) ... f k ( x n )⎦
n
Q= ∑ ei2 = e T .e [9.150]
i =1
This sum is minimum when the derivations of Q in relation to each of the model
parameters are simultaneously zero. This leads us to the system of normal equations:
n n n n
θˆ0 ∑ f 02 ( xi ) + θˆ1 ∑ f 0 ( xi ) f1 ( xi ) + .... + θˆk ∑ f 0 ( xi ) f k ( xi ) = ∑ yi f 0 ( xi ) [9.151]
i =1 i =1 i =1 i =1
n n n n
θˆ0 ∑ f 0 ( xi ) f1 ( xi ) + θˆ1 ∑ f12 ( xi ) + .... + θˆk ∑ f1 xi f k xi = ∑ yi f1 xi [9.152]
i =1 i =1 i =1 i =1
n n n n
θˆ0 ∑ f 0 ( xi ) f k ( xi ) + θˆ1 ∑ f1 ( xi ) f k ( xi ) + .... + θˆk ∑ f k2 xi = ∑ yi f k xi [9.153]
i =1 i =1 i =1 i =1
In this system, θˆ0 , θˆ1 , ..., θˆk , solutions of this system, are the estimators of the
parameters θ 0 , θ1 , ..., θ k in the sense of the least squares.
AT .A.θˆ = AT .y [9.155]
From this we get the solution to the problem:
(
θˆ = AT .A AT .y)
-1
[9.156]
We see that the matrix:
⎡ n n n ⎤
⎢
⎢ i =1
∑
f 02 ( x i ) ∑ f 0 ( x i ) f 1 ( x i ) .... ∑ f 0 ( xi ) f k ( xi )⎥⎥
i =1 i =1
⎢ n n n ⎥
⎢
A .A = ⎢
T ∑
f 0 ( xi ) f 1 ( xi ) ∑ f 12 ( x i ) .... ∑ f 1 ( x i ) f k ( x i ) ⎥⎥ [9.157]
⎢ i =1 i =1 i =1 ⎥
⎢ .... .... .... .... ⎥
⎢ n n n ⎥
⎢
⎢⎣ i =1
∑
f 0 ( xi ) f k ( xi ) ∑ f 1 ( xi ) f k ( xi ) .... ∑ f k2 ( x i ) ⎥
⎥⎦
i =1 i =1
Elaboration of Models for the Interaction Between the Sensor and its Environment 393
is a squared symmetrical matrix with lines and columns that are equal to the number
of estimated parameters.
We can calculate the matrix given the variances and the covariances of the
obtained estimators.
(
= AT .A ) -1 T
A .V(y).A. AT .A ( )-1
[9.158]
where V(y) is the variances-covariances matrix of y.
⎡σ 2 0 ... 0 ⎤ ⎡1 0 ... 0 ⎤
⎢ ⎥ ⎢0
σ 2 1 ... 0 ⎥⎥
V(y ) = ⎢
0 ⎥
=σ 2⎢
0 ...
[9.159]
⎢ ... ... ... ... ⎥ ⎢... ... ... ...⎥
⎢ ⎥ ⎢ ⎥
⎣⎢ 0 0 ... σ 2 ⎦⎥ ⎣0 0 ... 1 ⎦
and the variances-covariances matrix of the estimators takes the simpler form:
(
V (θˆ) = σ 2 AT .A
-1
) [9.160]
Near to the factor j2, this is simply the inverse matrix of AT · A. Even though it is
symmetrical, this matrix generally is expressed in non-zero terms outside the
diagonal principle. This becomes a general rule: the obtained estimators are
correlated.
∑r
i =1
i
2
σ2 = [9.161]
n− p
394 Fundamentals of Instrumentation and Measurement
We can use the coefficients obtained to calculate the value of the polynomial
corresponding to a given value of x, so:
⎡ f 0 ( x) ⎤
⎢ f ( x) ⎥
x=⎢ 1 ⎥ [9.163]
⎢ ... ⎥
⎢ ⎥
⎣ f k ( x)⎦
we also get:
yˆ = x T .θˆ [9.164]
This solution done with a least squares matrix contains the results already
obtained in the example of the line.
Here, we look once more at a situation in which the values of Y are not
correlated but are obtained with a different variance for each measurement. Under
these conditions, hypothesis H3 is written:
V(yi) = σ i2 [9.166]
⎡σ 12 0 ... 0 ⎤
⎢ ⎥
σ 22
V(y) = ⎢
0 ... 0 ⎥
[9.167]
⎢ ... ... ... ... ⎥
⎢ ⎥
⎣⎢ 0 0 ... σ n2 ⎦⎥
Elaboration of Models for the Interaction Between the Sensor and its Environment 395
A.θ + e = Y [9.168]
However, we need to know the sum of the squared gaps weighted by a weight
g i = 1 2 , so:
σi
n
Q= ∑g ei =1
2
i i [9.169]
Finding the minimum of this quantity leads us to the system of normal equations
which, in matrix form, is written:
(
θˆ = AT .g.A ) −1
AT .g.y [9.171]
g = [V(y)]−1 [9.172]
The matrix giving the variances and covariances of the obtained estimators is
written:
V (θˆ) = AT .g.A (
−1
) [9.173]
This matrix solution of the least squares contains the results already found in the
example of the line.
The matrix notation of the least squares provides the simpler calculation
solutions than the algebraic form.
396 Fundamentals of Instrumentation and Measurement
For example, let’s look at the following example: we are measuring pairs of
values (xi, yi) whose representation in a system of axes Ox, Oy gives points that are
fairly well-aligned. We then use the following representation model:
y = θ o + θ 1 .x [9.174]
In relation to the examples analyzed above, we make the hypothesis that the
measurements of y are made with the same variance and are correlated. This occurs
fairly often in practice, if only because of the uncertainties introduced by the
measurement instrument.
If we suppose that the covariance between the values of y, taken two by two,
remain the same, the variance-covariance matrix of the vector y takes the form:
Since the measurements have the same variance, it is not necessary to weight the
results; and the solution of the least squares retains the usual form:
(
θˆ = AT .A )
−1
AT .y [9.176]
(
V ( θˆ ) = AT .A )-1
(
AT .V(y).A. AT .A )
−1
[9.177]
⎡ n ⎤
⎢
⎢ ∑ 2
xi ⎥
⎥
V (θˆ0 ) = σ 2 ⎢ i =1
(1 − ρ ) + ρ ⎥ [9.178]
⎢ n ⎛ n ⎞
2 ⎥
⎢n
∑
⎢ i =1
x i2 − ⎜
⎜ ∑ xi ⎟
⎟
⎥
⎥
⎣ ⎝ i =1 ⎠ ⎦
Elaboration of Models for the Interaction Between the Sensor and its Environment 397
nσ 2
V (θˆ1 ) = 2
(1 − ρ ) [9.179]
n ⎛ n ⎞
n ∑ x i2 − ⎜
⎜ ∑xi ⎟
⎟
i =1 ⎝ i =1 ⎠
n
−σ 2 ∑x i
cov(θˆ0 , θˆ1 ) = i =1
2
(1 − ρ ) [9.180]
n ⎛ n ⎞
n ∑ x i2 − ⎜
⎜ ∑ xi ⎟
⎟
i =1 ⎝ i =1 ⎠
All these values obviously depend on the correlation coefficient. When this takes
the value 0, we find the same results as obtained before.
V (θˆ0 ) = σ 2 [9.181]
V (θˆ1 ) = 0 [9.182]
has a variance:
⎧ ⎡ ⎤⎫
⎪ ⎢ ⎥⎪
⎪ ⎢1 ⎥⎪
⎪ n( x − x ) 2
V ( yˆ ) = σ 2 ⎨ ρ + (1 − ρ )⎢ + ⎥⎪
2 ⎥⎬
[9.185]
⎪ ⎢ n n ⎛ n ⎞ ⎪
⎪
⎪⎩
⎢
⎢
n ∑ x i2 − ⎜
⎜∑ xi ⎟ ⎥⎪
⎟ ⎥
⎣ i =1 ⎝ i =1 ⎠ ⎦ ⎪⎭
V ( yˆ ) = σ 2 [9.186]
The least squares method also applies to situations when the model chosen to
represent the dependence between the variables X and Y results in a non-linear
normal equation system. Here, we can linearize the equation in several ways: by
changing the variable; by developing the function serially near to the representative
measurement points; or by working numerically and finding the function of the sum
of the squared gaps.
There are functions linking X and Y that lead to non-linear systems expressed in
parameters to be estimated, but which, by changing a variable, can be relevant to
this example. This means they can be analyzed by the standard methods. By way of
example, we cite the following cases:
– U = A.exp(BT) which refers to the linear example Y = 0 + 1 X by proposing
Y = In (U) and X = T, from which we derive 0 + In(A) and 1 = B.
– U = A + B.In(T), which refers to the linear example Y = 0 + 1X by proposing
that Y = U and X = In(T), from which we derive 0 = A and 1 = B.
– U = A.TB, which refers to the linear example Y = 0 + 1X by proposing that
Y = U and X = In(T), from which 0 = In(A) and 1 = B.
Elaboration of Models for the Interaction Between the Sensor and its Environment 399
T
– U= which refers to the linear example Y = 0 + 1X by proposing
A. X + B
that Y = 1 and X = 1 , from which we derive 0 = A and 1 = B.
U X
We should be aware of the fact that even if the values of the transformed variable
U have the same variance V(U), changing the variable usually involves unequal
variances for the values of the resulting variable Y, since the variance Y is expressed
as:
2
⎛ ∂Y ⎞
V (Y ) = ⎜ ⎟ V (U ) [9.187]
⎝ ∂U ⎠
Y = f ( X , θ 0 , θ 1 ,..., θ k ) [9.188]
The equation transforms the relations between the values of X and the
measurements of Y. It depends on the values of k + 1 parameters 0, 1,… k.
We measure n pairs of values (xi, yi) so that V(yi) = ji2. Applying the principle of
the least squares gives us the function:
n
Q= ∑ g [y
i =1
i i − f ( x i , θ 0 , θ 1 ,..., θ k )]2 [9.189]
The estimators we are trying to find are the solution of the normal equations
systems:
∂Q
[
∂f ( x i , θˆ0 , θˆ1 ,..., θˆk )
]
n
∂θ 0
= ∑ − 2. g
i =1
i
∂θˆ0
y i − f ( x i , θˆ0 , θˆ1 ,..., θˆk ) = 0 [9.191]
400 Fundamentals of Instrumentation and Measurement
∂θ 1
= ∑ − 2. g
i =1
i
∂θˆ 1
y i − f ( x i , θˆ0 , θˆ1 ,..., θˆk ) = 0 [9.192]
∂θ k
= ∑ − 2.g
i =1
i
∂θˆ k
y i − f ( x i , θˆ0 , θˆ1 ,..., θˆk ) = 0 [9.193]
In general, this system is not linear, which makes its resolution difficult. We can
resolve it from the initial values of the written solutions 0 θ 0 , 0 θ 1 , …, 0 θ k and
giving each of them an increase e 0 , e1 , …, e k , so that:
θˆ0 = 0 θ 0+ e0 [9.194]
θˆ1 = 0 θ 1+ e1 [9.195]
θˆk = 0 θ k +e k [9.196]
k
∂f ( x i , 0 θ 0 , 0 θ 1 , ,..., 0 θ k )
f ( x i , θˆ0 , θˆ1 ,..., θˆk ) = f ( x i , 0 θ 0 , 0 θ 1 , ,..., 0 θ k ) ∑ j =0
∂θ j
[9.197]
The normal equations are then of the type (what we have written here is only that
of the index parameter j):
n
∂f ( x i , 0 θ 0 , 0 θ 1 , ,..., 0 θ k ) ⎡ k
∂f ( x i , 0 θ 0 , 0 θ 1 , ,..., 0 θ k ) ⎤
∑g i
∂θ j
⎢ y i − f ( x i , 0 θ 0 , 0 θ 1 , ,..., 0 θ k ) −
⎢
∑ ∂θ j
ej⎥ = 0
⎥
i =0 ⎣ j =0 ⎦
[9.198]
We find a linear system as ej. From this solution, and from the initial values
given to the parameters, we get a new set of values that can help iterate the
calculation.
Elaboration of Models for the Interaction Between the Sensor and its Environment 401
9.5.2. Numerical search for the minimum of the function of the sum of the
squared gaps
n
Q= ∑ g [y
i =1
i i − f ( x i , θ 0 , θ 1 ,..., θ k )]2 [9.199]
As a general rule, the form of this surface can vary, but close to the solution
(corresponding to a surface summit), we find an elliptic paraboloid.
The concept is as follows: from the initial values, we calculate a first value of Q.
We then give an increase to each of the values of the parameters and observe the
variation of Q. If Q increases, we move away from the surface summit, so that the
increases are in the wrong direction. In the opposite situation, the sense of
displacement is correct, and we continue until converging on the solution. To put it
another way, the representative point of the parameter values is displaced on the
surface until it joins its summit again.
Carrying out the method can be somewhat difficult when the number of
parameters is high. There are methods that allow us to systematize finding the
solution and the speed of the convergence towards this solution (the Maquard
method, for example).
Certain precautions must be taken if we use graphic methods. When using these
methods, the following conditions must be met:
– The departure point must be sufficiently close to the solution in order to have a
quick maximum convergence.
– The variations given to the parameters must not be so significant that
oscillation from one part to another of the solution occurs and no solution is
achieved.
– Due to the local curvature of the surface, the rapidity of convergence may be
different depending on whether the solution is reached by larger or smaller values.
402 Fundamentals of Instrumentation and Measurement
The principle of the least squares also applies to finding the parameters of a
multivariable model, that is, to describing the development of an explained variable
Z according to the explicative variables.
This is an example of the least squares plane when the relation between Z and
the variables X and Y is expressed as:
z = θ 0 + θ1 x + θ 2 y [9.200]
The problem consists of finding the estimators θˆ0 , θˆ1, θˆ2 of θ 0 , θ1, θ 2 from the
measurement of n triplets of the paired values xi, yi, zi which, in an n plane brought
to an axial system (Ox, Oy, Oz), is represented by n points.
θ 0 + θ 1 x1 + θ 2 y1 + e1 = z1 [9.201]
θ 0 + θ1 x2 + θ 2 y 2 + e2 = z 2 [9.202]
θ 0 + θ1 xn + θ 2 y n + en = z n [9.203]
or again, by proposing:
⎡ z1 ⎤ ⎡ e1 ⎤
⎢z ⎥ ⎢e ⎥ ⎡θ 0 ⎤
z= ⎢ 2⎥
e= ⎢ 2⎥
θ = ⎢⎢θ 1 ⎥⎥ [9.204]
⎢ ... ⎥ ⎢ ... ⎥
⎢ ⎥ ⎢ ⎥ ⎢⎣θ 2 ⎥⎦
⎣zn ⎦ ⎣e n ⎦
Elaboration of Models for the Interaction Between the Sensor and its Environment 403
and:
⎡ 1 x1 y1 ⎤
⎢ ⎥
1 x2 y2 ⎥
A=⎢ [9.205]
⎢... ... ... ⎥
⎢ ⎥
⎣1 xn yn ⎦
A.θ + e = z [9.206]
Here we give as examples the expressions of the coefficients of the least squares
plane:
with:
2 2 2
n n n n n n ⎛ n ⎞ ⎛ n ⎞ n ⎛ n ⎞
den = n ∑ ∑
i =1
x i2
i =1
y i2 +2 ∑ x ∑ y ∑ x y − ∑ ⎝∑
i =1
i
i =1
i
i =1
i i
i =1
x i2 ⎜
⎜
i =1
yi ⎟ − ⎜
⎟
⎠
⎜
⎝ i =1 ⎠
∑
xi ⎟
⎟ ∑
i =1
y i2 − n⎜
⎜ ∑
⎝ i =1
xi y i ⎟
⎟
⎠
[9.210]
n n n n n n n n n
num(θˆ0 ) = ∑ ∑ ∑
i =1
x i2
i =1
y i2
i =1
zi + ∑ ∑
i =1
xi
i =1
xi y i ∑
i =1
yi zi + ∑ ∑
i =1
yi
i =1
xi y i ∑x z
i =1
i i
2
n n n n n n ⎛ n ⎞ n
− ∑ ∑ ∑ x i2 yi yi zi − ∑ ∑ ∑ xi y i2 xi z i − ⎜
⎜ ∑
xi y i ⎟
⎟ ∑z i [9.211]
i =1 i =1 i =1 i =1 i =1 i =1 ⎝ i =1 ⎠ i =1
404 Fundamentals of Instrumentation and Measurement
n n n n n n n n
num(θˆ1 ) = n ∑ y ∑ x z +∑ y ∑ z ∑ x y + ∑ x ∑ y ∑ y z
i =1
2
i
i =1
i i
i =1
i
i =1
i
i =1
i i
i =1
i
i =1
i
i =1
i i
2
n n n n n ⎛ n ⎞ n
−n ∑ xi y i ∑ yi zi − ∑ ∑ ∑ xi y i2 zi − ⎜
⎜
yi ⎟
⎟∑ ∑x z i i [9.212]
i =1 i =1 i =1 i =1 i =1 ⎝ i =1 ⎠ i =1
n n n n n n n n
num(θˆ2 ) = n ∑ ∑
i =1
x i2
i =1
yi zi + ∑ ∑ ∑
i =1
xi
i =1
yi
i =1
xi z i + ∑ ∑ ∑x y
i =1
xi
i =1
zi
i =1
i i
2
n n n n n ⎛ n ⎞ n
−n ∑x y ∑x z −∑ ∑y ∑
i i i i x i2 i zi − ⎜
⎜
xi ⎟
⎟∑ ∑y z i i [9.213]
i =1 i =1 i =1 i =1 i =1 ⎝ i =1 ⎠ i =1
The multivariable forms also help us in cases where sensors have sensitivities
that vary according to a variable; that it is external to measurement (influence
variables).
y = θ 0 + Sx [9.214]
S = θ1 + θ 2t [9.215]
1 being the sensitivity of the sensor when t = 0 and 2 describing the development
of the sensitivity according to t. We then get:
We recognize the plane equation, the variable y being described according to the
function of the variable x, and xt.
Elaboration of Models for the Interaction Between the Sensor and its Environment 405
The least squares method consists of finding the minimum of the parameter
function of the model Q( 0, 1…. k) formed by writing the sum of the squared gaps.
This minimum is reached when:
∂Q ∂Q ∂Q
dQ = .dθ 0 + .dθ 1 + .... + .dθ k = 0 [9.217]
∂θ 0 ∂θ 1 ∂θ k
∂Q
=0 [9.218]
∂θ 0
∂Q
=0 [9.219]
∂θ 1
∂Q
=0 [9.220]
∂θ k
g (θ 0 , θ1 , ..., θ k ) = C [9.221]
∂g ∂g ∂g
dg = .dθ 0 + .dθ 1 + .... + .dθ k = 0 [9.222]
∂θ 0 ∂θ 1 ∂θ k
406 Fundamentals of Instrumentation and Measurement
The result is that we can express one of the elements (for example, d k)
according to the others:
1 ⎛ ∂g ∂g ∂g ⎞
dθ k = − ⎜ .dθ k −1 ⎟⎟
∂g ⎜ ∂θ .dθ 0 + ∂θ .dθ 1 + .... + ∂θ [9.223]
⎝ 0 1 k −1 ⎠
∂θ k
By bringing back this value to equation [9.217], and by writing that the quantity
as a factor of each differential element is zero, we end up with:
∂Q ∂Q ∂g 1
− =0 [9.224]
∂θ 0 ∂θ k ∂θ 0 ∂g
∂θ k
∂Q ∂Q ∂g 1
− =0 [9.225]
∂θ 1 ∂θ k ∂θ 1 ∂g
∂θ k
∂Q ∂Q ∂g 1
− =0 [9.226]
∂θ k −1 ∂θ k ∂θ k −1 ∂g
∂θ k
The solution of these gives the estimators θˆ0 , θˆ1 , ..., θˆk −1 .
The last estimator θˆk is obtained by applying the constraint relation, so:
We can systematize our search for solutions by applying the method of Lagrange
multipliers.
Elaboration of Models for the Interaction Between the Sensor and its Environment 407
Let us look at the system formed by the sum of the squared gaps and the
constraint condition:
n
Q(θ 0 , θ 1 ,..., θ k ) = ∑e
i =1
2
i [9.228]
g (θ 0 , θ 1 ,..., θ k ) = C [9.229]
Q L (θ 0 , θ 1 ,..., θ k ) = ∑e 2
i + λ [g (θ 0 , θ 1 ,..., θ k ) − C ] [9.230]
where is the Lagrange multiplier. We thus add an unknown and the system giving
the solution of the problem is formed by writing the following k + 2 equations:
∂Q L ∂g
+λ =0 [9.231]
∂θ 0 ∂θ 0
∂Q L ∂g
+λ =0 [9.232]
∂θ 1 ∂θ 1
∂Q L ∂g
+λ =0 [9.233]
∂θ k ∂θ k
∂Q L
= g (θ 0 , θ 1 ,..., θ k − C ) = 0 [9.234]
∂λ
We can either resolve the system of unknown k + 2 equations or extract the value
of the multiplier of one of the equations and bring it to the others in order to
decrease the order of the system. This methodology is applicable to situations where
several constraint relations exist simultaneously.
Y = θ 0 . f 0 ( X ) + θ 1 . f 1 ( X ) + θ 2 . f 2 ( X ) + ... + θ k . f k ( X ) [9.235]
408 Fundamentals of Instrumentation and Measurement
We have seen that the resolution of the least squares method with the help of
matrix formalism works by means of matrix inversion:
⎡ n n n ⎤
⎢
⎢ i =1
∑ f 02 ( x i ) ∑
i =1
f 0 ( x i ) f 1 ( x i ) .... ∑f
i =1
0 ( xi ) f k ( x i )⎥
⎥
⎢ n n n ⎥
⎢
A .A = ⎢
T
i =1
∑
f 0 ( xi ) f 1 ( xi ) ∑f
i =1
2
1 ( xi ) .... ∑
i =1
f 1 ( xi ) f k ( xi ) ⎥
⎥ [9.236]
⎢ .... .... .... .... ⎥
⎢ n n n ⎥
⎢
⎢ ∑
f 0 ( xi ) f k ( xi ) ∑f 1 ( xi ) f k ( x i ) .... ∑ f k2 ( x i ) ⎥
⎥
⎣ i =1 i =1 i =1 ⎦
We see that the calculations are considerably simplified if all the terms outside
the principle diagonal are zero; that is, if:
∑f
i =1
j ( x i ). f h ( x i ) =0 [9.237]
Y = θ 0 . f 0 ( X ) + θ 1 . f 1 ( X ) + θ 2 . f 2 ( X ) + ... + θ k . f k ( X ) [9.238]
The least squares resolution remains unchanged, the only modification being the
matrix to be inversed, which takes the following form:
⎡ n n n ⎤
⎢
⎢ i =1
∑
P02 ( x i ) ∑i =1
P0 ( x i ) P1 ( x i ) .... ∑ P ( x ) P ( x )⎥⎥
i =1
0 i k i
⎢ n n n ⎥
⎢
B .B = ⎢
T
i =1
∑P0 ( x i ) P1 ( x i ) ∑
i =1
P12 ( x i ) .... ∑i =1
P1 ( x i ) Pk ( x i ) ⎥ [9.240]
⎥
⎢ .... .... .... .... ⎥
⎢ n n n ⎥
⎢ ∑
⎢ P ( x )P (x )
0 i k i ∑ P (x )P ( x )
1 i k i .... ∑ Pk2 ( x i ) ⎥
⎥
⎣ i =1 i =1 i =1 ⎦
or, taking into account the condition of orthogonality imposed on the polynomials:
⎡ n ⎤
⎢
⎢ i =1
∑P 2
0 ( xi ) 0 .... 0 ⎥
⎥
⎢ n ⎥
⎢
B .B = ⎢
T 0 ∑P
i =1
2
1 ( xi ) .... 0 ⎥
⎥ [9.241]
⎢ .... .... .... .... ⎥
⎢ n ⎥
⎢
⎢
0 0 .... ∑ Pk2 ( x i )⎥
⎥
⎣ i =1 ⎦
The matrix is diagonal and the resolution of the problem no longer poses a
problem. This is because the equations of the system are independent. We thus
obtain the general form of the equation:
∑ y P (x )
i =1
i j i
Φ
ˆ =
j n
[9.242]
∑i =1
P j2 ( x i )
σ2
V (Φ
ˆ )=
j n
[9.243]
∑i =1
P j2 ( x i )
410 Fundamentals of Instrumentation and Measurement
The orthogonality has the secondary effect of making the estimators non-
correlated, so that:
cov(Φ
ˆ ,Φ
j
ˆ ) = 0 ∀h ≠ j
h [9.244]
In books dealing with this subject, we find many polynomial forms that have
orthogonal characteristics. We can start here with trigonometric polynomials with
orthogonality features used to calculate the coefficients of the Fourier development
of a periodic function. There are also Lagrange and Legendre polynomials that
present constraints on the values situated on the axes of abscissas (these values are
between -1 and +1 and/or equidistant values of x).The polynomials used by Forsythe
do not have that constraint. They are written:
P0 ( x) = 1 [9.245]
P1 ( x) = ( x − α 1 ).P0 ( x) [9.246]
P2 ( x) = ( x − α 2 ).P1 ( x) − β 2 P0 ( x) [9.247]
P j ( x) = ( x − α j ).P j −1 ( x) − β j P j − 2 ( x) [9.248]
n n
∑
i =1
x i .P j2−1 ( x i ) ∑P
i =1
2
j −1 ( x i )
αj = n
and β j = n
[9.249]
∑
i =1
P j2−1 ( x i ) ∑
i =1
P j2− 2 ( x i )
Y = Φ 0 .P0 ( X ) + Φ 1 P1 ( X ) [9.250]
a) Calculation of f0
∑y i
ˆ )=σ
2
i =1
Consequently, Φ
ˆ =
0 = y and V (Φ j [9.251]
n n
b) Calculation of f1
n
∑x
i =1
i
We have P1 ( X ) = ( x − α 1 ) with α 1 = =x [9.252]
n
n
∑ y (x
i =1
i i − x)
σ2
Consequently, Φ
ˆ =
1 n
and V (Φ
ˆ )=
j n
[9.253]
∑ (x
i =1
i − x) 2
∑ (x
i =1
i − x) 2
y = y + Φ1 ( x − x ) [9.254]
This can be attained if we remember the fact that the least squares line intersects
the coordinate point ( x , y ) .
The recurrent nature of the Forsythe polynomials has the advantage of easily
increasing the degree of the adjustment polynomial.
412 Fundamentals of Instrumentation and Measurement
Generally, we can show that the sum of the squares of the residuals Rk after an
adjustment of the degree k, is expressed according to the sum of the squares of the
residuals Rk-1 that correspond to an adjustment by the polynomial of degree k – 1 by
the relation:
n
R k = R k −1 − Φ
ˆ2
k ∑P
i =1
2
k ( xi ) [9.255]
This quantity decreases as k increases. In other words, the fact of increasing the
degree of the adjustment polynomial acts to constrain the polynomial from coming
closer to the experimental points. This can carry the risk of being unnecessary
because these points have a variance that defines an uncertainty field. Therefore, it is
physically sufficient that the polynomial goes into the interior of the uncertainty
field without intersecting the experimental points. In addition, this constraint has no
effect on the measured points. Outside these points, no constraint applies, even if the
polynomial might oscillate with an amplitude well above that of the desired
adjustment gain.
There are two simple methods for determining the optimum degree of the
smoothing polynomial.
The first method calculates, after adjusting the degree k, the quantity:
Rk
[9.256]
n − (k + 1)
This is an estimator of the variance linked to the variable Y until this estimation
is coherent with a predetermined value.
250
9.9. Bibliography
[FOR 77] FORSYTHE G.E., “Generation and use of orthogonal polynomials for data fitting
with a digital computer”, J. Soc. Indus. Appl. Math., vol. 5, no. 2, 1957.
[JAF 96] JAFFARD P., Méthodes de la statistique et du calcul des probabilités, Masson
Editions, 1996.
[NIE 98] NIELSEI L., “Least-squares estimation using Lagrange multipliers”, Metrologia,
vol. 35, no. 2, 115-118, 1998.
[SAP 90] SAPORTA G., Probabilités, analyse de données statistiques, Technip Editions,
1990.
[WIL 45] WILLIAMSON J.H., “Least square fitting of a straight line”, Can. J., Phys; vol. 46,
1845-1847, 1968.
This page intentionally left blank
Chapter 10
10.1. Introduction
The goal of this chapter is twofold. We will first present some basic
mathematical tools necessary for analyzing analog and digital signals that are
present in instrumentation chains. Then we will provide some examples of signal
processing methods that can lead to adaptations, according to the applications being
considered. In this chapter we will especially emphasize mathematical tools for
digital signal processing and time-frequency representations that are useful for
extracting continuous information in signals that may not be stationary, a usual
situation in practice. Any reader wanting to gain a more in-depth knowledge of basic
signal representations may consult, for example, the following books: [CHA 90];
[COT 97]; [COU 84]; [DEL 91]; [GAS 90]; [MAX 89]; [MAX 96]; [PIC 89, 93, 94,
95]; and [ROD 78]. In addition, some methods of analog processing are discussed in
Chapters 4 and 5.
10.2.1. Introduction
Physical Shaping
Analog
phenomenon Sensor and amplification
to be observed processing
of the signal
In certain applications, for example, in control systems, the signal obtained after
processing can be reinjected into the process input in the form of a control law that
will eventually modify the behavior of the observed physical phenomenon. For
example, in active vision, the analyzed signal can serve to follow the trajectory of an
object; and, if necessary, correct the position, the direction and the settings of the
sensors that are following the scene.
Signals can be analyzed in the time domain or in the frequency domain. The
following section gives some definitions that are often used in the analysis of analog
signals.
time origin. Practically, a deterministic model is only partially known, and its
unknown parameters introduce random behaviors that are more or less
unpredictable. A deterministic signal has little importance in real situations, since it
does not carry information, apart from its presence or absence. However, this kind of
signal can act as an excitation signal that indirectly obtains information about a
physical variable through the interaction that this signal can have with the physical
system being analyzed. A signal carrying information of uncertain nature is a
random signal. Some statistical properties of this signal allow us to describe it
simply or even to evaluate our knowledge of this signal by relating it to a known
model.
In practice, a sensor is activated from a given time t0, very often chosen as a time
origin. The signals which are observed and processed are then considered as zero
signals up to the time t0 = 0. These are called causal signals. In addition, signals are
observed during a finite period T. By commodity, especially of calculation, we very
often represent an observed signal by a periodic signal; or we construct a periodic
auxiliary signal from the observed signal. The rest of this section will provide some
fundamental descriptions of signals and present some of their usual features.
An analog signal represented in the time domain by a scalar function x(t) of the
continuous variable t can be characterized in different ways. Subject to the existence
of integrals, we have the following definitions:
T
1
– mean value mx = x(t ) = lim
T → +∞ 2T ∫ x(t ) dt ;
−T
T
T → +∞ ∫
2
– energy E x = lim x(t ) dt ;
−T
2
– instantaneous power p x ( t ) = x ( t ) ;
T
1
∫ x(t )
2 2
– mean power Px = x (t ) = lim dt ;
T → +∞ 2T
−T
– mean power of the signal fluctuations around its mean
2 2 2
σ x2 = x(t ) − m x = x(t ) − m x .
The time representation of a signal in its form x(t) is the most natural. It directly
shows the magnitude variation of the signal according to time. However, there are
other representations. The remainder of this section will discuss the frequency
representation that indicates the variation frequency of the signal magnitude. This
418 Fundamentals of Instrumentation and Measurement
+∞
X(f ) = ∫ x(t ) exp(− j 2πft ) dt [10.1]
−∞
and:
+∞
X (ω ) = ∫ x(t ) exp(− jωt ) dt [10.2]
−∞
+∞ +∞
1
x(t ) = ∫ X ( f ) exp(j 2πft ) df =
2π ∫ X (ω ) exp( jωt ) dω [10.3]
−∞ −∞
If x(t) is a real signal, its Fourier transform is a complex function of even module
and odd argument: X(-f)= X*(f). In addition to the linearity of the Fourier transform,
we see several other properties: the time reversal property:
TF[x(-t)] = X(-f); the conjugation property TF[x*(t)] = X*(-f ); and the delay
theorem TF[x(t-k)] = exp(-j2ヾfk)X(f). The convolution product of two signals x(t)
and y(t), denoted by (x*y)(t) is defined by:
+∞
( x ∗ y )(t ) = ( y ∗ x)(t ) = ∫ x(u) y(t − u) du
−∞
and this expression may be written more easily in the frequency domain than in the
time domain. Indeed, we have TF[(x*y)(t)] = TF[x(t)]TF[y(t)]. This property, which
is also called the Plancherel theorem, is very useful for the linear filtering of signals
and for calculating correlation functions.
Representation and Analysis of Signals 419
+∞ +∞ +∞
1
∫ ∫ ∫ X (ω )
2 2 2
Ex = x(t ) dt = X(f ) df = dω
2π
−∞ −∞ −∞
This means the energy of a signal does not depend on the chosen representation.
2
The function Φ xx ( f ) = X ( f ) is called the energy spectral density of the signal
x(t) and is a frequency representation of the energy. This quantity is, by definition,
always real and positive. The inverse Fourier transform of fxx(f), written xx(k), is
called the energy autocorrelation function of the energy signal x(t). By writing that
2
X ( f ) = X ( f ) X * ( f ) and by using the property X * ( f ) = TF[ x * (−t )] we deduce
from this:
+∞ +∞
γ xx (τ ) = ∫ Φ xx ( f ) exp( j 2πfτ ) df = ∫ x(t ) x (t − τ ) dt
*
−∞ −∞
xx(k) expresses the resemblance between x(t) and x*(t-k) and produces the
continuous autosimilarities in the signal. If the signal x(t) is real, its autocorrelation
function is real, even, and maximum at the time origin.
These results can be applied to two signals x(t) and y(t). First of all, we have:
+∞ +∞
* *
∫ x(t ) y (t ) dt = ∫ X ( f )Y ( f ) df
−∞ −∞
+∞ +∞
γ xy (τ ) = ∫ Φ xy ( f ) exp( j 2πfτ ) df = ∫ x(t ) y * (t − τ ) dt
−∞ −∞
which constitutes the Wiener-Kintchine theorem. If the signal y(t) is also obtained
by a linear filtering of the signal x(t), then we have Φ yx ( f ) = G ( f )Φ xx ( f ) or
2
Φ yy ( f ) = G ( f ) Φ xx ( f ), G(f) being the frequency response of the filter. These
relations correspond to the interference formula.
+∞ 1T
xT (t ) = ∑ cn exp( j 2πn t T ) with c n = ∫ x(t ) exp(− j 2πn t T )dt
n =−∞ T 0
where 1/T is the fundamental frequency and cn the amplitude of the harmonic of
rank n. With this new representation, integrating the signal is done on a finite
interval, but we have to calculate an infinite number of coefficients. This is why
another representation will be defined later for the analysis of digital signals that are
defined by a finite number of harmonics (see section 10.6).
+∞
XT ( f ) = ∑ c nδ ( f − n T )
n =−∞
The mean power PT of the periodic signal xT(t) is written in the form:
+∞ +∞ +∞
1 T /2 2 2 2
PT = ∫ x(t ) dt = ∑ c n = ∫ ∑ c n δ ( f − n T ) df
T −T / 2 n =−∞ −∞ n = −∞
+∞ 2
The function ΦT ( f ) = ∑ cn δ ( f − n T ) is the power spectral density of the
n =−∞
periodic signal xT(t). The inverse Fourier transform of this quantity is the power
autocorrelation function of this signal. It is defined by:
1 T /2 +∞ 2
γ T (τ ) = ∫ xT (t ) xT* (t − τ ) dt = ∑ cn exp( j 2πnτ T )
T −T / 2 n =−∞
Representation and Analysis of Signals 421
Strictly speaking, a random signal is stationary if all its statistical properties are
invariant by changing the origin of time. If we limit this property only to the first
and second statistical times, we say that the signal is in a wide sense stationary. This
last feature is commonly accepted as the starting hypothesis for processing methods
that use the statistical properties of the signal to be analyzed. The random process
fluctuates and the observed signal corresponds only to a special realization of the
random process x(t, u) called trajectory. In absolute terms, in order to know the
statistical variables of the process, we must carry out the same experiment many
times. This is obviously not possible in most situations and we usually assume that
the nature of the information conveyed by the time behavior of the signal is the same
as that which is conveyed by carrying out the process a number of times. The
stationary random process is called ergodic if all the statistical means coincide
asymptomatically towards the time means, in particular that is if:
+∞
1 T n
E[ x n (t , u )] = ∫ x n f ( x) dx = lim ∫ x (t , u ) dt = x (t , u )
n
−∞ T →+∞ 2T −T
where f(x) is the probability density of the random signal x(t, u). The correlation
function of a stationary signal is defined by:
γ xx (τ ) = E[ x(t , u ) x * (t − τ , u )]
For this kind of signal, xy(k) does not depend on the time t. Moreover, if the
signal is ergodic, we get:
1 T
γ xx (τ ) = lim ∫ x(t , u ) x (t − τ , u ) dt
*
T →+∞ 2T −T
The power spectral density fxx(f) of a stationary signal is the Fourier transform
of the autocorrelation function (the Wiener-Khintchine theorem):
Φ xx ( f ) = TF[γ xx (τ )] . For stationary and ergodic signals, the time means and the
422 Fundamentals of Instrumentation and Measurement
statistical means coincide, the quadratic mean is directly related to the power
spectral density, and we have:
+∞
= E ⎡ x ( t , u ) ⎤ = γ xx ( 0 ) = ∫ Φ xx ( f ) d f
2 2
Px = x ( t , u )
⎢⎣ ⎥⎦
−∞
The cross-correlation function of two random signals x(t, u) and y(t, u), which
are jointly stationary and ergodic, is defined by:
1 T
γ xy (τ ) = lim ∫ x(t , u ) y (t − τ , u ) dt = E[ x(t , u ) y (t − τ , u )]
* *
T →+∞ 2T −T
and their cross-power spectral density is shown by fxy (f) = TF[ xy(k)]. The signals x(t,
u) and y(t, u) are said to be uncorrelated if their correlation is zero (whatever k may
be). If the signal y(t, u) is obtained by a linear filtering of the signal x(t, u), the
interference formula is written Φ yx ( f ) = G ( f )Φ xx ( f ) or
2
Φ yy ( f ) = G ( f ) Φ xx ( f ), G(f) being the filter’s frequency response.
10.3.1. Introduction
Digital functions process series of numbers that usually come from sampling an
analog signal, the amplitude being quantified. Figure 10.2 shows part of a synoptic
of a digital instrumentation chain. We note by xb(t) the sampled and hold signal and
xq[n] the digital signal.
Representation and Analysis of Signals 423
ADC
x (t ) x b (t ) x q [n] y[n] y (t )
E/B Q Digital
ADC
processing
or also:
Thus, the signals xe(t) and x[n] are equivalent. First, taking the Fourier transform
of equation [10.7], we establish that:
∑ δ ( f − n Te ) = δ 1 Te ( f )
1 1
TF[δ Te (t )] =
Te n Te
This tells us that the Fourier transform of an impulse train is still an impulse
train. The Fourier transform of equation [10.6] is then written:
1
Xe( f ) = ( X ∗ δ 1 Te )( f )
Te
The impulse function being the neutral element of the convolution product, we
have deduced from this fact that:
∑ X ( f − n Te )
1
X e ( f ) = X TD ( fTe ) =
Te n
The Fourier transform of the sampled signal is obtained from that of the analog
signal by periodic replication. Figure 10.3 illustrates this point for a bandlimited
Representation and Analysis of Signals 425
signal; that is, that B exists so that X ( f ) = 0 ∀ f ≥ B or that all the frequency
representation of x(t) is inside the band ]–B, B[. Two examples are then considered:
B < Fe 2 or B ≥ Fe 2 .
−B 0 B f − Fe 2 0 Fe 2 Fe f
cas B ≥ Fe 2 Te X e ( f )
−B 0 B Fe f
Figure 10.3. Sampling and aliasing
If BTe < 0.5 , X ( f ) and X ( f − n Te ) do not overlap for all n ≠ 0; X(f) can
then be obtained from Xe(f) by simple multiplication by the frequency response H(f)
of a ideal lowpass filter defined by:
⎧T if − Fe 2 ≤ f ≤ Fe 2,
H( f ) = ⎨ e
⎩0 otherwise.
The original signal x(t) can then be reconstructed from the ideal sampled signal
xe(t) or then from the sampled signal x[n]. After filtering the signal xe(t) by the
lowpass filter, the information is intact.
10.3.2.1.3. Interpolation
The interpolation formula is obtained by going back to the time domain of the
relation X ( f ) = H ( f ) X e ( f ) . We then have x(t ) = (h ∗ xe )(t ) , with:
+ Fe / 2
h(t ) = Te ∫ e
j 2πft
df = sinc(Fe t )
− FE / 2
H ( f ) (dB)
0 B Fe / 2 f
Passband Transition
of the filter band
10.3.2.2. Quantization
We are only looking at real-valued signals. Representing numbers on a calculator
requires approaching their values by whole numbers coded on a given number of
bits. The quantization models the operation that carries out this approximation. The
input-output non-linear characteristic represents the nature of the approximation
(Figure 10.5).
xq
xqi +1
xqi q
s0 xi xi +1 xb
We write x qi as the output values of the quantizer. With the example shown in
Figure 10.5, all the quantization steps are equal: quantization is then uniform and
x qi +1 = x qi + q , q being the quantization step. The quantity s0, shown in Figure
10.5, models a threshold that characterizes the nature of the quantization. Two
examples are current today: quantization by truncation (s0 = q) and quantization by
rounding off (s0 = q/2).
f (a)
x (t ) + x q (t ) 1/ q
+
ηq (t )
q q a
−
2 2
+∞
1q2 2 q2
σ η2 = ∫ a 2 f (a) da = ∫ a da =
−∞ q −q 2 12
If the input signal is a triangular signal that is not clipped, the quantization noise
is a sawtooth signal of period T with a power expressed by:
+T / 2 2
1 ⎛ − qt ⎞ q2
Pη = ∫ ⎜ ⎟ dt =
T −T / 2 ⎝ T ⎠ 12
We find the same result as before. If the input signal has any form, the power of
the quantization error must be calculated by considering both its probability density
and time representation. However, if the quantization step and sampling period are
sufficiently low, it is justifiable to conserve the uniform law hypothesis. With this
hypothesis, for a sinusoidal signal of maximum amplitude A and a quantization
operation on n bits with a dynamic [-A, A], the quantization step is q = 2A/2n and the
error variance σ η2 = A 2 (3 × 2 2 n ) .The power of the signal being equal to A2/2, the
signal-to-noise ratio is RSB = 3 × 22n–1. In decibels, this becomes a widely used
relation in technical documentation [AZI 96]:
We must keep in mind that the signal-to-noise ratio increases by 6 dB each time
we increase the converter capacity by one bit. In Chapter 6, we described the
principles of sigma-delta converters that can significantly increase the signal-to-
noise ratio with a looped system. In the remainder of this chapter, no more
distinction will be made between x[n] and xq[n].
Representation and Analysis of Signals 429
The concepts of correlation and spectral density relating to analog signals seen in
section 10.2 can easily be generalized to discrete time signals; the time integrals are
then replaced by sums.
We call the z transform of the sequence of x[n], the complex function defined by:
X ( z ) = TZ{x[n]} = ∑ x[n] z − n
n
We can show that this series converges if the complex number z belongs to a ring
of complex plane that is delimited by two concentric circles centered on the origin of
the radius R min and R max , that is, for z ∈ D x =]Rmin , Rmax [ (see Figure 10.7). If
the signal is causal as well, the convergence domain becomes D x = ]R min , +∞[ . For
certain signals, the z transform converges in z = Rmin ; for example, the z transform
of the unit impulsion unit x[n] = δ [n] where δ [n] is the Kronecker symbol
( δ [n] = 1 for n = 0 and δ [n] = 0 for n ≠ 0 , exists whatever the value of z and the
convergence ring is the entire complex plane: D x = [0,+∞[ ).
– delay: TZ{x[n − p ]} = z − p X z ( z ) ;
R max
R min
1 n −1
x[n] = ∫ X ( z ) z dz
j 2π C
We are going to apply certain properties of the z transform for digital filtering.
The kind of digital processing described here concerns linear time and invariant
systems, that is, linear filters, for which the input signal x[n] is related to that of the
output y[n] by the constant-coefficient difference equation:
p q
y[n] = ∑ ai x[n − i ] − ∑ b j y[n − j ] with p > 0 and q > 0 [10.8]
i =0 j =1
Representation and Analysis of Signals 431
This kind of filter is causal. This relation is written in the form of a discrete
convolution equation:
The causality of the filter means that h[n] = 0 for n < 0. When the coefficients bj
are not all zero, the filter is called recursive and its impulse response is infinite (IIR).
However, when the coefficients bj (j = 1, …q) are zero, the coefficients
ai (i = 0, …p) are the p + 1 non-zero coefficients of the impulse response of the filter
and we say that the filter has a finite impulse response (FIR).
By taking the z transform from the relation shown in equation [10.8], and by
using the delay property of the z transform, we deduce that:
⎛ q ⎞ p
Y ( z )⎜1 +
⎜ ∑ b j z − j ⎟⎟ = ∑ ai z − i X ( z )
⎝ j =1 ⎠ i =0
−1 −p
Y ( z ) a 0 + a1 z + … + a p z
H ( z) = =
X ( z) 1 + b1 z −1 + … + bq z −q
This transfer function is a rational fraction. The filter is then called a rational
filter. The roots of a 0 + a1 z −1 + … + a p z − p = 0 are called zeros of H(z) and the
roots of 1 + b1 z −1 +…+bq z − q = 0 poles of H(z). In the specific example when all the
coefficients of ai and bj are real, we have the property H ( z ∗ ) = H ∗ ( z ) . The poles
and zeros of H(z) are then real or complex and conjugate. If the unit circle
( z = exp( j 2πf ) ) belongs to the convergence domain, the transfer function of the
filter H(z) allows for the expression of the complex gain G(f). We then have
G ( f ) = H (exp( j 2πf ) ) . The unit circle can be either graduated according to angle or
frequency (between –1/2 and 1/2).
The stable rational and causal filters constitute a specific class of filters called
dynamic filters. Minimum phase filters are dynamic filters with zeros that are also
inside the unit circle. These filters are especially useful because their inverse is also
a minimum phase filter. The inverse filter is also stable.
G( f )
1+ δ1
1
1− δ1
δ2
0 f1 f2 f
We here present two synthesis methods for IIR filters. The first method involves
approaching the transfer function of an analog filter by a discrete time transfer
function. The second method calculates the coefficients of the filter by an
optimization process. Analog filters are usually Butterworth, Bessel, Tchebycheff or
elliptic filters.
2 1 − z −1
p=
Te 1 + z −1
( n+1)Te ( n +1)Te
y ((n + 1)Te ) = ∫ x(t ) dt = y (nTe ) + ∫ x(t ) dt
−∞ nTe
1 − z −1
p = k
1 + z −1
Several criteria can be chosen to carry out approximation in the best possible
way. In general, these criteria lead to resolving a non-linear optimization issue with
constraints. This means we need a method to determine the coefficients of the filter;
these coefficients will optimize the criterion with the goal of determining the
optimal filter in a given family. For example, we look for the coefficients of a filter
with a given structure that will minimize the distance between the frequency
response of this filter and the required frequency response for the frequencies
fi (i = 1, …N).
The least squares method, initially conceived to study the movement of planets,
is widely used and is basic to many estimation techniques. As we will see, its
popularity is due to the fact that the choice of a quadratic parameter criterion leads to
an explicit solution of the parameters we want to find. The principle of this method
will now be discussed.
Representation and Analysis of Signals 435
We try to minimize the mean-squares gap Q between the given coordinate points
( xi , y i ) (i = 1, … , N ) and the values ( x i , g i (θ)) given by a model that depends
on the parameters vector = [ 1, …, p]T. This gap is written:
N
Q= ∑ yi − g i ( 2
) = y − g( )
2
i =1
∂g
g (θ) ≈ c + G∆θ , with c = g(θˆ k ) , G = θˆ k
and ∆θ = (θ − θˆ k )
∂θ T
We then apply the previous results with this linearized function to obtain the new
estimated value of . After calculations, we get θˆ k +1 = θˆ k + ∆θˆ with
∆θ = (G T G ) −1 G T (y − c). This new recursive formula is also expressed according
to the gradient ∇ Q (θ) of the criterion Q and of its approximate Hessian formula
~
H Q (θ) (in this formula, the terms on which the second derivatives depend are
~
ignored). So we have ∇ Q (θˆ k ) = −G T (y − c ) and H Q (θˆ k ) = G T G ; the estimated
value of is updated according to the recursive formula
~ −1 ˆ
θˆ k +1 = θˆ k − H Q (θ k ) ∇ Q (θˆ k ). However, the decrease of the criterion Q is not
guaranteed. This is why we introduce a coefficient k that allows the algorithm
~ −1 ˆ
converge, at least to a local minimum: θˆ k +1 = θˆ k − λ k H Q (θ k ) ∇ Q (θˆ k ). This is
called a Gauss-Newton algorithm and can be obtained by a second order Taylor
approximation of the criterion. There are methods that avoid the inverse calculation
of the approximate Hessian formula. The algorithms are based on a first order
Taylor approximation of the criterion, only using a gradient but converging more
slowly. For more details on optimization methods, we direct the reader to [MIN 83]
and [WAL 94].
436 Fundamentals of Instrumentation and Measurement
The goal of this section is to use several concepts discussed earlier to solve
common signal processing problems that lead to finding an optimal filter. Even
though signals can be complex, in the remainder of this section, in order to simplify
formulae, we will suppose that they are real.
From observing a signal drowning in noise, we want to learn from linear filtering
if this effective (noiseless) signal, whose waveshape is known, is present or absent.
The matched filter that maximizes the signal-to-noise ratio at its outset at the time of
our decision. This problem is found, for example, in radar applications when we
need to determine the presence or absence of a target from measuring the received
signal.
Let us suppose that s(t) is the effective signal of known waveshape that we
assume is buried in an additive noise b(t). A digital version of an matched filter is
then introduced, and an analog filter can be used in a similar way. The impulse
response of the filter is noted h[n] and its frequency response H(f). The time we
decide if the effective signal is present (P) or absent (A) is written as t0 = n0Te, with
Te the sampling period. The process performed by the matched filter is shown in
Figure 10.9.
b[n ]
+
s[n ] y[n ]
x[n ] P
Filter y[ n 0 ]
h[n ]
+ ty=(t )t 0 = n 0 T e A
We also assume that the noise b[n] is white, of a power j2b. The mean power of
the output noise of the filter is then Phb = E[((h ∗ b)[n])2 ] = σ b2 ∑ h[k ] 2 . The
k
instantaneous power of the effective output signal of the filter is:
2
⎛ ⎞
p hs [n o ] = ((h ∗ s )[n 0 ]) 2
=⎜
⎜ ∑ h[k ]s[n o − k ] ⎟
⎟
⎝ k ⎠
Representation and Analysis of Signals 437
and the signal-to noise ratio of the filter at the time of detection no Te is
RSB[n0 ] = phs [n0 ] Phb . From the Schwartz inequality we get:
⎛ ⎞⎛ ⎞ ⎛ ⎞⎛ ⎞
p hs [n0 ] ≤ ⎜⎜ ∑ h[k ]2 ⎟⎟ ⎜⎜ ∑ s[n0 − k ]2 ⎟⎟ = ⎜⎜ ∑ h[k ]2 ⎟⎟ ⎜⎜ ∑ s[k ]2 ⎟⎟
⎝k ⎠⎝ k ⎠ ⎝k ⎠⎝ k ⎠
the inequality being satisfied if and only if h[k ] = c s [n0 − k ] , which leads to
1 γ ss [0]
RSB[n0 ] ≤
σ 2 ∑ s[k ]
k
2
=
σ b2
where γ ss [n] is the energy autocorrelation of
b
the effective signal s[n]. The maximum output signal-to-noise ratio therefore only
depends on the energy of the effective signal s[n] and on the noise power b[n]. The
impulse response of the optimal filter ( h[n] = c s [n0 − n] ) is a reversed and shifted
duplicate of the signal s[n]. This filter is said to be matched to the signal s[n]. Its
response to the effective signal s[n] is y s [n] = c ∑ s[n0 − k ]s[n − k ] and again, to a
k
close multiplicative coefficient, the energy autocorrelation of the effective signal:
y s [n] = cγ ss [n − n0 ] . The matched filter is a correlator. For example, the response
of the matched filter to a rectangular impulse is a triangular signal. The choice of n0
depends on the application and on a causality constraint related to a real-time
process. This causality constraint can be strictly fulfilled if the effective signal s[n]
has a finite time structure.
Now we will look at the problem of estimating a signal x(t) by causal linear
filtering from a signal y(t) that is correlated with x(t). We consider here a digital
filter. The signal y(t) is then sampled at the frequency Fe = 1 Te and we want to
know the estimation of x[n] = x(nTe ) according to samples y[k ] ( k ≤ n ). We write
xˆ[n] as the estimated value of x[n].
∂P[n]
= −2 E[ y[n − k ](x[n] − xˆ[n])] = 0 , ∀ k = 0, … , M − 1
∂h[k ]
If the signals are stationary, these relations are rewritten according to the
M −1
correlation functions of the signals x[n] and y[n]: γ xy [k ] − ∑ h[i]γ yy [k − i] = 0 ,
i =0
∀k = 0, …, M − 1 . The causal FIR optimal filter thus satisfies the equations:
γ xy [k ] = (h ∗ γ yy )[k ] ∀k = 0, … , M − 1 [10.9]
We can show that this relation can also apply to a causal filter that is not
necessarily of finite impulse response:
and that it is also validated with an analog filter [LIF 81]; [MAX 89]; [PIN 95].
Equation [10.10] is called the Wiener-Hopf equation.
In the specific case where y (t ) = (h0 ∗ x)(t ) + b(t ) with b(t), which is a white
noise independent of x(t), we get:
H 0∗ ( f )Φ xx ( f ) 1 1
H( f ) = =
2
H 0 ( f ) Φ xx ( f ) + Φ ww ( f ) H 0 ( f ) 1 + Φ ww ( f ) Φ xx ( f )
which means that 1/H0 (f) is weighted by a coefficient which is all the closer to 1
since the noise is low.
The causality constraint leads to an mean square error above that obtained with
an optimal filter without constraints. The resolution, shown in equation [10.10] (or
its equivalent with an analog filter), leads to the Wiener filter; and this is fairly
complicated [LIF 81]; [PIN 95]. This is why often we prefer using a causal linear
filter with finite impulse response, in order to solve equation [10.9].
Representation and Analysis of Signals 439
To introduce the Kalman filter, we still keep in mind the previous problem, but
add that the system is described by state space representation:
the noises v[n] and w[n] being white and uncorrelated to each other and to the initial
state x[0]. We respectively write Rv and Rw as the covariances of v[n] and w[n]. The
signal x[n] is called autoregressive of order one. We can generalize this to a filter of
a higher order by introducing a state vector of equal dimension to the order of the
filter.
The best causal linear estimator (in the sense of mean square error) of x[n]
according to all available observations y[1], …, y[n] is now written as xˆ[n n] and its
quadratic error P[n n] = E[( x[n] − xˆ[n n]) 2 ] . From what has been previously
shown, we now have: xˆ[n n] = h Tn n y n with h n n = Γ n−1γ n .
We now introduce a new observation y[n + 1] and look for the estimator. We
will establish that xˆ[n + 1 n + 1] can be obtained in a recursive manner in two steps.
For that, we write xˆ[n + 1 n] as the best linear estimator of x[n + 1] according to all
previous observations y[1], ···, y[n]: xˆ[n + 1 n] = h Tn +1 n y n . This quantity is called a
440 Fundamentals of Instrumentation and Measurement
one-step prediction. To establish the equations of the predictive filter, we write that
h n +1 n minimizes P[n + 1 n] = E[( x[n + 1] − xˆ[n + 1 n]) 2 ] , which leads to:
∂P[n + 1 n]
= E[y n (ax[n] + v[n] − h Tn +1 1y n )] = 0
∂h n +1 n
Taking into account the statistical properties of noises, this relation is rewritten
Γ nh n +1 n = aγ n . We deduce from this that h n +1 n = ah n n , then xˆ[n + 1 n] = axˆ[n] .
By expressing x[n + 1] according to x[n] in the expression P[n + 1/n] and in
considering the statistical properties of v[n], we deduce that
P[n + 1 n] = a 2 P [n n] + Rv .
an optimal estimator to take into account u[n] (we are then looking for xˆ [n n] in the
form xˆ [n n] = h Tn n y n + α n ), we obtain the recurrence relations that are those of the
Kalman filter. Finding the estimator is carried out in two steps. The first prediction
step leads to:
⎧⎪ xˆ [n + 1 n] = A n xˆ [n n] + B n u[n]
⎨
⎪⎩P[n + 1 n] = A n P[n n]A Tn + R v [n]
This algorithm must be initialized. If we have the first and second order of the
initial state x[0], we can choose xˆ [0 0] = E[x[0]] for initialization and
[ ]
P[0 0] = E (x[0] − xˆ [0])(x[0] − xˆ [0]) T ; we then show that this estimator is non-
biased. Otherwise, we can choose xˆ [0 0] = 0 and P[0 0] = αI with a being rather
high. There are other ways (least squares, Bayesian approach) to obtain these
equations and we direct the reader to [AND 79], [JAZ 70] and [SCH 91]. We note
here that if the state noises and observations are Gaussian, the obtained estimator is
the best estimator without imposing a linearity constraint. The Kalman filter is
relatively robust and is being used more and more frequently. For non-linear
systems, the previous equations can be applied after a first order approximation of
the system around the current estimated state.
+∞
Tx (t , ω ) = x,ψ t ,ω = ∫ x( s ) ψ t*,ω ( s ) d s
−∞
We can define the orders of magnitude of the dimensions of this atom from the
variance of the energy of the analysis function and from its Fourier transform so
+∞ 2 2
that, for an analysis function of normed energy, we have σ t 2 = ∫− ∞ s ψ t ,ω ( s ) d s
1 +∞ 2
and σ ω 2 = ∫− ∞ θ Ψ t ,ω (θ ) d θ . These variables are shown in Figure 10.10.
2
2π
ω
σω
σt
t
Figure 10.10. Tiling of the time-frequency plane
It is easy to demonstrate that the surface of the atom (for an analysis function
normalized to 1) cannot be below a limit given by the Heisenberg-Gabor inequality:
σ t σ ω ≥ 1 2 . This result fixes the resolution limits of the time-frequency
representations. It is important to note that this limit is attained for an analysis
function of the Gaussian model:
ψ t ,ω ( x ) = π −1 / 4 exp(− ( x − t ) 2 2)
Representation and Analysis of Signals 443
For the Fourier transform (frequency representation), the analysis function only
depends on the pulsation and the atom is a band of infinite length and of zero height
parallel to the time axis: ψ .,ω ( s ) = exp( jωs ) , Ψ .,ω (θ ) = δ (ω − θ ) , σ t = ∞ and
σ ω = 0. In addition, we can use the same expression for the time representation:
ψ t ,. ( s ) = δ (t − s) and Ψ t ,. (θ ) = exp(− jθ ). The atom being analyzed in this case is a
band of zero thickness parallel to the frequency axis σ t = 0 and σ ω = ∞.
These various representations can also be used with digital signals and a certain
number of efficient processing algorithms have been developed. In this regard we
emphasize the importance of the fast Fourier transform (FFT) and the discrete
wavelet transform.
Fourier transform. Among them, the Hartley transform stays within the domain of
real functions, since its nucleus is real: ψ .,ω ( s ) = cos ωs + sin ωs . It is defined by:
+∞
H (ω ) = ∫ x(t ) (cos ωt + sin ωt ) d t
−∞
The inversion is obtained the same way because the analysis function is real and
forms an orthogonal base:
1 +∞
x(t ) = ∫ H (t ) (cos ωt + sin ωt ) d ω
2π −∞
The Hartley transform can easily be linked to the Fourier transform by the
intermediary of its even part H p (ω ) = (H (ω ) + H (−ω ) ) 2 and of its odd part
H i (ω ) = (H (ω ) − H (−ω ) ) 2 which are respectively the cosine transform and the
sinus transform: so we have X (ω ) = H p (ω ) − jH i (ω ) . The Hartley transform is
usually deduced from the Fourier transform, its real part providing the even part, and
its imaginary part, the odd part. Although this transform is not much used today, it
can be advantageous because the analysis is real and the inversion formula is
identical to direct transform. The cosine transform, seldom used as a frequency
analysis tool, has a particular importance in another context; its discrete version is a
good approximation of the Karhunen-Loève transform [BEL 87].
Calculating a discrete Fourier transform (DFT) means not only sampling (of
period Te) but also windowing the signal. Windowing is a constraint due to the
material impossibility of carrying out an infinite number of calculations. A sampling
in the frequency domain becomes necessary.
Signal windowing is imposed by the choice of the transform length, that is, by
the number N of samples retained for the calculation. The width of the window is
then T = NTe . The truncated signal is written x f (t ) = x(t ) rect T (t ) where rect T (t )
is the carried function equal to 1 on the observation horizon and to 0 elsewhere. The
Representation and Analysis of Signals 445
1
Fourier transform of the truncated signal becomes X f (ω ) = ( X ∗ Rect T )(ω )
2π
with, if the observation horizon is centered at the origin in time,
Rect T (ω ) = T sinc ( ωT 2 ) . The frequency representation X ( f ) of the analog
signal is thus convoluted with a cardinal sinus.
N −1
X [ n] = ∑ x[k ] exp(− j 2πk n N ) [10.11]
k =0
and Z [n] = X [n] Y [n] . The Fourier transform of the circular correlation:
N −1
γ xy [n] = ∑ x[k ] y * [k − n]
k =0
The basic operation, called the butterfly operation, is simply a DFT on two
samples, one of even range and the other of odd range:
⎧
⎪ X [n] = X 1p [n] + exp(− j 2πn N )X 1i [n]
⎨
⎪⎩ X [n + N / 2] = X 1 [n] − exp(− j 2πn N )X 1i [n]
p
100
50
10
5
m
2 4 7 10
+∞
T fg x(t , ω ) = x, ψ t ,ω = ∫ x( s) g ( s − t ) exp(− jωs ) d s
*
−∞
We see that the reconstruction can, in a more general way, be conducted with the
+∞
help of a function ’ satisfying ∫−∞ψ (t ) ψ ′ * (t ) d t = 1 . This representation
conserves the energy; and we have a theorem analogous to that of Parseval’s:
+∞
2 1 +∞ +∞
∫ ∫ T fg x(t , ω ) T fg x(t , ω ) d t d ω
*
E x = ∫ x(t ) d t =
−∞ 2π −∞ −∞
σω
σt
50
pulse
100
150
200
250
50 100 150 200 250 300 350 400 450 500
time
X( f )
70
60
50
40
30
20
10
σ t = a∆t and σ ω = ∆ω a . This means that the surface of the atom being analyzed
remains constant throughout the scales but that the tiling of the plane respects the
conditions of an optimum analysis: a time spread that is inversely proportional to the
frequency. This is called a time-scale representation.
+∞
1 ⎛t − b⎞
Toc x(a, b) = x, ψ a , b = ∫ x(t )
a
ψ * ⎜⎜
⎝ a ⎠
⎟⎟ d t
−∞
1/a
a
t
+∞ +∞
dadb
∫ ∫ ψ ab (t ) ψ ab (t ) = δ (t − t ' )
*
−∞ −∞ a2
+∞ 2
Ψ (ω )
∫ ω dω = 1
−∞
This condition is not very restrictive and a function of normalized energy and of
zero mean (Ψ (0) = 0 ) suitably located around the source will generally be
Representation and Analysis of Signals 451
admissible. We see that, as was the case with SWFT, reconstruction is possible with
another function ψ ' (t ) if it verifies the condition:
+∞
Ψ (ω ) Ψ '* (ω )
∫ dω = 1
−∞ ω
+∞ +∞
dadb
x(t ) = ∫ ∫ Toc x(a, b)ψ a, b (t ) a2
−∞ −∞
+∞ +∞ +∞
dadb
∫ ∫ ∫
2 2
Ex = x(t ) d t = Toc x(a, b)
−∞ −∞ −∞ a2
We can cite two examples of functional analyses used for the continuous wavelet
transform. One is called the Mexican hat: ψ (t ) = 2(1 − t )2 exp(− t 2 2) ( 3π 1 4 ) .
This is simply a second derivation of the Gaussian and the Morlet wavelet formula
ψ (t ) = exp(− t 2 2) exp( jω o t ) π 1 4 which itself came from gaborets; however,
gaborets are not, strictly speaking, admissible (see Figure 10.16).
t
t
+∞
Tod x(i, n) = x,ψ i,n = ∫ x(t ) 2 −i / 2ψ * (2 −i x − n) d x
−∞
Y. Meyer [MEY 90] and S. Mallat [MAL 99] propose an especially efficient
calculation algorithm from this transform and its inverse in the general field of
multiresolution analysis (MRA). We will present a very general overview.
⎧
⎪... ⊂ V1 ⊂ V0 ⊂ ... ⊂ Vi ⊂ Vi −1 ⊂ ...,
⎪
∪ ∩
⎨ Vi = L ( R), Vi = {0}, Vi −1 = Vi ⊕ Wi , Vi ⊥Wi , ∀i ∈ Z ,
⎪i∈Z
2
i∈Z
⎪
⎩ x (t ) ∈ Vi ⇔ x ( 2 t ) ∈ Vi −1, ∀i ∈ Z , x(t ) ∈ V0 ⇔ x(t − k ) ∈ V0 , ∀k ∈ Z .
All Wi spaces are, by construction, 2 by 2 orthogonal, the direct sum of all these
subspaces is equal to L2 (R) . This means the ensemble of ψ i,n for i and n integers
forms an orthonormal basis of L2 (R) . We therefore have a discrete wavelet basis
and coefficients d ni of the projection of x(t) on the subspaces Wi constituting the
discrete wavelet transform of x(t): Ai x = ∑ x, ϕ i ,n ϕ i ,n , Di x = ∑ x,ψ i,n ψ i,n ,
n n
a ni = x, ϕ i , n and d ni = x,ψ i ,n = Tod x(i, n) .
A0x
A1 x D1 x
A2x
D2 x
ϕ ψ
algorithms of analysis and reconstruction use the filtering operations and processes
of over-and under-sampling presented in Figures 10.18 and 10.19.
a nj −1 anj
h 2
d nj
g 2
anj anj −1
2 h +
d nj
2 g
Cubic splines
0.4
0.2
Signal
0
-0.2
-0.4
50 100 150 200 250
0.2
D1p
-0.2
50 100 150 200 250
0.2
0
D2p
-0.2
-0.4
50 100 150 200 250
0.2
0
D3p
-0.2
-0.4
50 100 150 200 250
0.4
0.2
D4p
0
-0.2
0.2
D5p
0
-0.2
50 100 150 200 250
The second family has been developed by I. Daubechies [DAU 92] and is made
of an FIR filter. The analysis functions have a finite structure (2N), and the family is
parametered by N. As size increases, the analysis function becomes regular, and
456 Fundamentals of Instrumentation and Measurement
analysis becomes better localized in the time-scale plane. We see that the linear-
phase condition can be met in this context for a sufficiently large structure
( N > 10 ).
Lastly, we point out that the discrete wavelet transform is not invariant in
translation.
In some applications, the relevant variable is not the signal itself but its energy.
In such cases, we may want time-frequency representations of this variable. Energy
is by its nature a quadratic variable and transforms produce these representations
using bilinear combinations of the signal (when it is real). These transforms are
called bilinear transforms. In the following sections, we discuss several of these, the
best known being the group of transforms of the Cohen class: the spectrogram; the
scalogram; the Wigner-Ville transform; and the pseudo-Wigner-Ville transform.
2
+∞
T fg x(t , ω ) T fg * x(t , ω ) = ∫ x( s ) g * ( s − t ) exp(− jωs ) d s
−∞
We can see that, when a signal is of limited duration, the transform does not
preserve the width of the structure, which has been enlarged by the analysis window.
As well, the spectrogram is a non-reversible transform, so some loss of information
is at the source of this phenomenon. However, in spite of this problem, this
representation is one of the most widely used and the range of apodization windows
used for calculating the Fourier transform (such as Blackman, cosinuisoidal,
Gaussian, Hamming, Hanning and Kaiser) influence the choice of a window
function.
Representation and Analysis of Signals 457
2
+∞
ψ ((t − b) a ) d t
* 1 *
Toc x(a, b)Toc x ( a, b) = ∫ x(t )
−∞ a
+∞
W x (t , ω ) = ∫ x(t + τ 2) x * (t − τ 2) exp(− jωτ ) d τ
−∞
+∞ 2
∫−∞ W x (t , ω ) d t = X (ω )
and:
+∞ 2
1 2π ∫−∞ W x (t , ω ) d ω = x (t )
+∞ +∞ +∞ +∞
1 1
∫ ∫ ∫ ∫
2 2
Ex = W x (t , ω ) d t d ω = x(t ) d t = X (ω ) d ω
2π −∞ −∞ −∞
2π −∞
W x + y (t , ω ) = W x (t , ω ) + W y (t , ω ) + 2 Re[W xy (t , ω )]
oscillatory terms that are concentrated in the time-frequency plane in the middle of
the segment that joins the representation centers of x(t) and y(t). The phenomenon is
illustrated in the figure below that shows the Wigner-Ville transform of the sum of
the two signals localized in time (100 and 350) and in frequency. We will see that
the interferences also affect the negative frequency terms.
50
100
pulse
150
200
250
50 100 150 200 250 300 350 400 450 500
time
We can decrease the number of these disturbance terms in this way: we use the
analytic signal deduced from the real signal by presetting to zero the negative
frequency components of the Fourier transform of the signal. Another way to
Representation and Analysis of Signals 459
attenuate the interferences is to smooth the oscillations of the Fourier transform with
a lowpass filter.
+∞
PW x (t , ω ) = ∫ g (τ ) x(t + τ 2) x (t − τ 2) exp(− jωτ ) d τ
*
−∞
+∞
PW x (t , ω ) = ∫ x g (t + τ 2) x *g (t − τ 2) exp(− jωτ ) d τ
−∞
This transform is simply the WVT of a windowed signal, and so, contrary to the
spectrogram, the PWVT preserves the structure of a compact signal. The other side
of this advantage is that the sampling, as is the case with the WVT, must be at the
double frequency of the minimum required by Shannon for x(t) to avoid spectrum
folding.
Among the fundamental difference between the two types of signals, we must
first note the impossibility of good planning for data nD. This has obvious
consequences for transposing recursive algorithms; the causality constraints, strictly
speaking, disappear but implementing the algorithms must take into account the
order in which the operations are possible (see the problem of recursive 2D filters).
This absence of a natural order of data reading emphasizes the need to use
algorithms that respect this symmetry; the problem of phase linearity of the filters
used for imaging is a good example of this constraint.
Multidimensional signals are often intrinsically low. CCD sensors, for example,
furnish finite dimension images that are known in advance, with the signal to be
processed usually available in its whole in the image memory. The problems must
be taken into account and a processing can never be completely invariant for
translation. Image sensors often work through spatial sampling of data before they
are measured. As such, we can say, in many instances, that the primary signal is
digital and the means of processing the signal must be approached in this
perspective.
However, there are operations and techniques which specifically deal with
multidimensional processing [KUN 93]. There are non-separable filters, geometric
transforms, segmentation methods by active contour [BLA 97], mathematic
morphology [COS 89] and Markovian field modelization [GUY 93], among others.
Here, we briefly mention the specific case of signal sequencing processing; n spatial
dimensions more than time. Specific tools are used in this case, since the time
dimension has particular properties.
10.8. Bibliography
[AND 79] ANDERSON B.D.O., MOORE J.B., Optimal Filtering, Prentice-Hall, Englewood
Cliffs, 1979.
[AZI 96] AZIZ P.M., SORENSEN H.V. and SPIEGEL J.V.D., “An Overview of Sigma-Delta
Converters”, IEEE Signal Processing Magazine, January 1996.
[BEL 87] BELLANGER M., Traitement Numérique du Signal, Masson, Paris, 1987.
[BLA 87] BLAKE A., ZISSERMAN A., Visual Reconstruction, MIT Press, Cambridge-MA,
1987.
[CHA 90] CHARBIT M., Eléments de Théorie du Signal: les Signaux Aléatoires, Ellipses,
Paris, 1990.
[COC 95] COCQUEREREZ J.P., PHILIPP S., Analyse d’Images: Filtrage et Segmentation,
Masson, Paris, 1995.
[COS 89] COSTER M., CHERMANT J.L., Précis d’Analyse d’Images, CNRS, Paris, 1989.
[COT 97] COTTET F., Traitement des signaux et acquisition de données, Dunod, Paris, 1997.
[COU 84] DE COULON F., Théorie et Traitement des Signaux, Dunod, 1984.
[DAU 92] DAUBECHIES I., Ten Lectures on Wavelets, SIAM, Philadelphia, 1992.
[DEL 91] DELMAS J.P., Eléments de théorie du signal: les signaux déterministes, Ellipses,
Paris, 1991.
[DUV 91] DUVAUT P., Traitement du signal – Concepts et Applications, Hermès, Paris,
1991.
[FLA 93] FLANDRIN P., Temps-fréquence, Hermès, Paris, 1993.
[FON 81] FONDANECHE P., GILBERTAS P., Filtres Numériques, Principes et
Réalisations, Masson, Paris, 1981.
[GAS 90] GASQUET C., WITOMSKI P., Analyse de Fourier et Applications – Filtrage,
Calcul Numérique et Ondelettes, Masson, Paris, 1990.
[GON 92] GONZALES R.C., WOODS R.E., Digital Image Processing, Addison-Wesley,
1992.
[GUY 93] GUYON X., Champs Aléatoires sur un Réseau, Masson, 1993.
462 Fundamentals of Instrumentation and Measurement
[JAZ 70] JAZWINSKI A., Stochastic Processes and Filtering Theory, Academic Press, San
Diego, 1970.
[HOR 93] HORAUD R., MONGA O., Vision par Ordinateur, Hermès, Paris, 1993.
[KUN 84] KUNT M., Traitement Numérique des Signaux, Dunod, Paris, 1984.
[KUN 93] KUNT M., GRANLUND G., KOCHER M., Traitement Numérique des Images,
Presses Polytechniques et Universitaires Romandes, 1993.
[LAB 88] LABARRERE M., KRIEF. J.P., GIMONET B., Le Filtrage et ses Applications,
collection sup’aéro, 1988.
[LIF 81] LIFERMANN J., Les principes du traitement statistique du signal: les méthodes
classiques, Masson, Paris, 1981.
[MAL 99] MALLAT S., A Wavelet Tour of Signal Processing, 2nd ed., Academic Press, 1999.
[MAX 89] MAX J., Méthodes et Techniques de Traitement du Signal et Applications aux
Mesures Physiques, vol. 1 and 2, Masson, Paris, 1989.
[MAX 96] MAX J., LACOUME J.L., Méthodes et Techniques de Traitement du Signal et
Applications aux Mesures Physiques, 5th ed., Masson, Paris, 1996.
[MEY 90] MEYER Y., Ondelettes et Opérateurs I, Ondelettes, Hermann, Paris, 1990.
[MIN 83] MINOUX M., Programmation mathématique : théorie et algorithmes, vol. 1,
Dunod, Paris, 1983.
[OPP 75] OPPENHEIM A.V., SCHAFER R.W., Digital Signal Processing, Prentice-Hall
Inc., Englewood Cliffs, New Jersey, 1975.
[PIC 89] PICINBONO B., Théorie des signaux et des systèmes avec problèmes résolus,
Dunod, Paris, 1989.
[PIC 93] PICINBONO B., Signaux aléatoires, tome 1: Probabilités et variables aléatoires,
Dunod, Paris, 1993.
[PIC 94] PICINBONO B., Signaux aléatoires, tome 2: Fonctions aléatoires et modèles avec
problèmes résolus, Dunod, Paris, 1994.
[PIC 95] PICINBONO B., Signaux aléatoires, tome 3: Bases du traitement statistique du
signal avec problèmes résolus, Dunod, Paris, 1995.
[PRO 92] PROAKIS J.G., MANOLAKIS D.G., Digital Signal Processing, 2nd ed.,
Macmillan, New York, 1992.
[ROD 78] RODDIER F., Distributions et transform de Fourier, McGraw-Hill, Paris, 1978.
[SCH 91] SCHARF L.L., Statistical Signal Processing: Détection, Estimation, and Time
Series Analysis, Addison Wesley, 1991.
[TRU 97a] TRUCHETET F., Ondelettes pour le signal numérique, Hermès, Paris, 1997.
[TRU 97b] TRUCHETET F., Traitement linéaire du signal numérique, Hermès, Paris, 1997.
[WAL 94] WALTER E., PRONZATO L., Identification de modèles paramétriques, Masson,
Paris, 1994.
Chapter 11
11.1. Introduction
Choosing the right sensor is obviously of crucial importance; the quality of the
obtained cartography depends on it. The image of the environment must have
sufficient precision and reliability to be used in subsequent processing. To meet this
requirement, a network of sensors called a multisensor is the best way to increase
and improve the overall performance of an observation system.
On the other hand, increasing the relevance and the quality of the cartography
usually leads to linking sensors of different modalities (different observed
variables). For example, in a situation where we want to predict the presence of ice
on roads, knowing the temperature and the hygrometric degree of the best
predictions will not help us learn anything about the temperature.
Once defined by the transducer or transducers used for learning about the
physical phenomenon or, more generally, the environment, we must go on to
analyze the information they provide to extract the variables that are relevant to the
diagnostic task. The term “information” is used here in its broadest sense, since it
can refer to signals, images, numerical or symbolic content, to mention a few
usages. The work of parameterization generates and chooses variables, which then
help us decide:
– if a system is or is not functioning in degraded mode;
– if a transient is or is not present in a signal;
– if the perceived object belongs or does not belong to a specific class.
We cannot cover all existing techniques in depth, but rather offer a range of
solutions with comparisons between these solutions whenever possible. We direct
the interested reader to reference works for more detailed implementation
techniques.
11.2.1. Introduction
The work is thus divided into two stages: parameterization, then variable
selection, but these stages are in fact joined. After a brief summary of the
parameterization currently in use, the following sections will offer a range of
variable selection methods.
Unfortunately, this kind of expertise is not always possible – especially for a new
problem – and it is often difficult to predict in advance which variable will be
important. Compressing information a priori cannot really be done, at least in the
same terms. It is better to use more complete modelizations of initial information
which will help in processing unknown degraded modes or a new class of objects.
Processing signals and images can be done using many different techniques,
among them being: AR, MA and ARMA modelization; Fourier descriptors; time-
frequency analyses; polynomial approximations; splines; and Prony models. All
these techniques provide, for each observation, a variable set, such as coefficients of
ARMA filters, serial Fourier coefficients, or polynomial coefficients. The
dimensions of this variable set are usually significant, but before trying to select
from among these variables, certain expressions must be represented.
Multi-sensor Systems: Diagnostics and Fusion 467
variable
↓
⎡ x1 ⎤ ⎡x11 x1 p ⎤ ← observation
⎢ x ⎥ ⎢x ⎥
X=⎢ ⎥=⎢
2
⎢ ⎥ ⎢
21
[
⎥= X ,X
⎥ 1 2 Xp ] [11.1]
⎢ x ⎥ ⎢x x np ⎥⎦
⎣ n ⎦ ⎣ n1
For each of these observations of value in RP, we have one desired output yi of
the decision system. This output corresponds to a class designation and takes values
of between 0 and K – 1 for a problem general to K classes Ω0, Ω1, ..., ΩK–1:
First and second order statistics will later be utilized. The center of gravity in the
scatter plot is a vector of p components:
[ ] with [ ] 1 n
m = m1 , m2 ... m p mi = E X i = ∑ x ki [11.3]
n k =1
⎡σ1 2 σ 12 σ1 p ⎤
n
⎢σ 21 ⎥
∑
1
V=⎢ ⎥ with σ ij = ( xki − mi )( xkj − m j ) [11.4]
⎢σ n
2⎥ k =1
⎣ p1 σp ⎦
1 n 2
σ2 = ∑ x −m [11.5]
n k =1 k
We can also show that this total variance is equal to the trace of V [SAP 90]
p
σ 2 = ∑ σ k2 [11.6]
k =1
468 Fundamentals of Instrumentation and Measurement
We see that this quantity is independent of the coordinates system chosen and
that in the particular axial system made of vectors specific to V, this total variance is
expressed by:
p
σ 2 = ∑ λk2 [11.7]
k =1
In order to suppress irrelevant variables that may appear, or variables that are
over-coordinated between themselves, and also to limit the curse of dimensionality
(see section 11.2.1), we must choose a subset of a variable set X1, X2, ..., Xp. The
following sections will present in some detail several good selection methods in a
supervised context in which a database, which we assume to be exhaustive, is
available for designing a diagnostic system.
1 n 2
J=
n
∑ x i − proj(x i ) [11.8]
i=1
We can easily show [SAP 90] that the subspace we are looking for is generated
by the pr vectors proper to Ui of the variance-covariance matrix V (see equation
[11.4]) related to its first pr proper values i ranged in decreasing order. The axes
defined by these proper vectors are called inertia axes or principle axes.
xi
+ subspace
X +
+ +
++ proj( x i )
+ + +
+ +
+ +
From the example of the 2D illustration in Figure 11.2, we can easily see that
1D space that least deforms the initial scatterplot is the linear regression line that
merges with the first inertia axis. In any dimension space, the PCA determines
through linear combinations the initial p variables Xi of the new centered variables
Ui of the maximum and uncorrelated variance between them.
The PCA can be used for two purposes: to reduce the representation space of the
data by projecting the data in a space of reduced dimensions, or simply to aid in
visualizing the observation bases of large dimensions. In analyzing data, it is
important to visualize the correlations between the initial variables and the variables
coming from the PCA that “summarize” a large part of the basic inertia. The
correlation between the initial variable Xi and the principle component Ui is
calculated with the following expression [LEB 97]:
λ j U j (i)
R( X i ,U j ) = [11.9]
σi
(R( X i ,U j ), R( X i ,U k ))
This visualization helps us understand the links between each Xi variable with
the principle inertia axes (the principle plane j = 1, k = 2 is the most used). These
points are contained in a center circle 0 and of radius 1, whatever i and j are:
−1 ≤ R( X i ,U j ) ≤ 1
Xk Xi
1st principle axis
-1 0 1
Xj
-1
The PCA is a relatively powerful data analysis method, allowing many ways of
visualizing information [SAP 90]. With the PCA, we can verify the independence of
descriptive signal variables. It offers a new representation base made by a linear
combination of initial variables. The axial hierarchy is established by using an
inertia criterion that favors variables presenting the highest variances.
If we want to reduce the number of initial variables by using the PCA, we then
face the problem of choosing the pr dimension of the representation subspace.
λi
principal axis
1 2 3 ... p numbers
k
∑ λ i2
i =1 σ2
threshold
k
pr
Usually this choice is made by visualizing the range of values belonging to the
matrix of variance-covariance ranged in decreasing order. By using equation [11.7],
each value raised to the square “explains” part of the total variance of the cloud of
data points X. Choosing a threshold percentage of the curve represents the
accumulated sum of the squares of the values that help us obtain the value of pr (see
Figure 11.4). The existence of a break in the value set also helps us set the threshold.
As we can observe, this choice is made without taking into account the class Y
labeling that is available to us. This is doubtlessly the biggest drawback to using the
PCA as a method for selecting variables for diagnostic purposes. Actually, nothing
shows that the retained variables will be the most relevant ones for separating
classes. With the example shown in Figure 11.5, the basis of data X is divided in two
groups (corresponding, for example, to the class of correct modes and to the class of
faulty modes). The principle axis Il (the one possessing the highest inertia) is not
evidentially the most important axis for distinguishing between the two classes;
choosing the axis Id seems much better.
I1 density
X
Ω1
Ω0
+ + I1
+ o
+ ++ + oo density
+ + o
+ ++ o
o
+ + + ++ o o o
+ + o
+ + + + o o
+ + oo
oo Id
Id
The work of Fischer and Mahalanobis (1936) first explored this statistical
method. Discriminate factorial analysis (DFA) is both a descriptive method and a
method of classifying data. In this section, we will only discuss the first method.
472 Fundamentals of Instrumentation and Measurement
nΩ 0 X
Ω0
mΩ
0 + + o Ω1
VΩ 0 + + oo o o
+ + o
mΩ0 + o
+
o m Ω1
+ + o o
+ m o o o
+ + o
+ + o nΩ 1
nΩ 2 x x x
x x mΩ mΩ
mΩ x 2
x 1
x x
Ω2
2
VΩ 1
VΩ 2
To present the basic principles of DFA, some equations must be represented. For
each i of X, it is possible to define a center of gravity and a variance-covariance
matrix by using equations [11.3] and [11.4]. These are expressed, respectively, mΩi
and VΩi. nΩi is the number of observations of X belonging to the class i (see Figure
11.6).
1 K -1
Vin = ∑ nΩ VΩ [11.10]
n i=0 i i
The interclass variance matrix produces the average of the matrices proper to
each class. It can be interpreted as an overall measurement of the concentration of
classes around their center of gravity. If all the classes were reduced to their center
of gravity, their matrix trace of interclass variance would be zero.
Inversely, the interclass variance matrix measures the dispersion of the centers of
gravity of classes:
1 K -1
( )( )
t
Vex = ∑ nΩi mΩ i − m mΩ i − m [11.11]
n i=0
The closer the K centers of gravity are to each other, the weaker the interclass
matrix traces are.
Multi-sensor Systems: Diagnostics and Fusion 473
We can easily understand the importance of these two matrices for diagnostic
problems. The DFA tries to find a projection into the subspace of Rp that minimizes
Vin (concentrated classes) while maximizing Vex (classes far from each other). The
main result of this method [SAP 90] is that the subspace we are looking for is
generated by the pr vectors belonging to the matrix Vin-1Vex linked to its first values
ranged in decreasing order. The choice of pr is made in exactly the same way as with
the PCA method. However, we can show that the ranking of Vex is at most K-1 [GAI
83]; this means we must limit pr to pr≤ K-1.
A more detailed analysis also shows that the DFA is nothing less than a PCA on
the K centers of gravity of classes having a non-Euclidian metric (the Nahalanobis
distance: ||u||2 = ut Vin-1u [GAI 83]).
Xj
σ Ω 0 ( j)
o Ω0
oo o o
o o m
m Ω0 ( j) +
+ + o Ω0
o o o
m Ω + ++ o o o
1
m Ω1 ( j) o
++ ++
+ ++
+ + Ω1
σ Ω1 ( j) Xi
σ Ω i (i) σ Ω 0 (i)
m Ω1 (i) m Ω0 (i)
but on the source axes by making the hypothesis that they are orthogonal. For a
problem of two classes, for example, we calculate the quantities:
(m Ω (i) − m Ω (i ))
2
F (X ) =
0 1
i [11.13]
n σ +n σ
2 2
Ω0 Ω (i) Ω1 Ω (i)
0 1
The larger this quantity, the better we can discriminate it from the axis i. For
example, in Figure 11.7, the variable Xi is more discriminate than the variable Xj.
This sub-optimal method does not allow us to combine the initial variables to
obtain new, more relevant variables. Selecting variables using this criterion is
therefore very simple to do; a version of equation [11.13] can be extended to a
situation in which K classes also exist [DOC 81]. We should avoid using this simple
method when the complete variance-covariance matrix structure is not close to being
a diagonal structure.
DFA is often more useful than the PCA for diagnostic problems and for formula
recognition. This is because it takes into account the fact that elements of the
observation base appear in classes. However, the interclass variance matrix used is
an averaged matrix that only imperfectly represents the internal variance matrices of
each class, especially if the classes are very disparate. Added to this theoretical
difficulty is a practical one that occurs during the estimation of matrices Vin and Vex,
mainly if the observation basis contains few examples.
The selection method described in this section uses the linear regression
formalism. With the representations proposed in section 11.2.1, choosing a linear
regression model imposes the following matrix relation between the input and
output variables:
Y = XP + ε [11.14]
Figure 11.8. Illustration of the OFR method for n=3 and p=5
During the first iteration, we choose the variable Xi1, the most colinear to Y:
⎡
( )
t 2 ⎤
⎢ X k .Y ⎥
( )
2
cos X i , Y = max ⎢ ⎥
1
1≤ k ≤ p ⎢ X 2 2
k Y ⎥
⎣ ⎦
During the second iteration, we first orthogonalize all the remaining variables, as
well as the output so it is perpendicular to Xi1
Xi tY Xi t X k
Y (2 ) = Y − 1
and X k (2 ) = X k − 1
Xi t Xi Xi t Xi
1 1 1 1
with 1 ≤ k ≤ p and k ≠ i1
476 Fundamentals of Instrumentation and Measurement
The procedure ends at the iteration p, when all the variables have been classed.
( )
j 2
o( j) = ∑ cos X (k)
ik , Y
( k)
k =1
The reader will find very complete developments in the variable selections in
texts such as [DUD 73], [FUK 72], [KIT 86], and [KRI 82].
j
0
1 2 3 ... p
choice of p r
11.3.1. Introduction
The last two methods to be presented come from the domain of Bayesian
classification; they are first of all decision trees and neuron networks, of either
multilayered perceptron types or radial base functions.
p ( x Ωi ) p ( Ωi )
p ( Ωi x ) = (Bayes formula) [11.15]
( )
∑ p x Ω j p (Ω j )
K-1
j =0
Applying the Bayes formula assumes that the range of classes is complete, or
that – since we are dealing with classification – each observation only belongs to
one class and that the group of classes entirely covers the representation space Rp.
Example: what is the probability that every person measuring 1.60 meters in height
is a woman?
Responding to this question brings us back to estimating the following
conditional probability: p( F。x = 1.60). If the person is chosen randomly from a
population, and if we suppose that this population is half women and half men,
applying the Bayes formula gives us:
p (1.60 Ω F ) × 0.5
p ( Ω F 1.60 ) =
p (1.60 Ω F ) × 0.5 + p (1.60 Ω H ) × 0.5
478 Fundamentals of Instrumentation and Measurement
With this simplified model of the distribution of the heights of men in France, a
Gaussian centered at 1.75 meters, and a deviation of 0.15 meters, and for women a
Gaussian centered at 1.65 meters and a deviation of 0.15 meters, applying the Bayes
formula finally gives us:
1.61 × 0.5
p ( Ω F 1.60 ) = = 60.9%
1.61 × 0.5 + 2.52 × 0.5
If now, we choose a person from among, say, a national legislative body, and not
from among the French population, the probabilities change, if we assume that this
legislative body is composed of 10% women and 90% men. Applying the Bayesian
rule in this case gives us:
1.61 × 0.1
p ( Ω F 1.60 ) = = 14.8%
1.61 × 0.1 + 2.52 × 0.9
x ∈ Ωi such that
j =0...K-1 ⎣
(
Max ⎡ p Ω j x ⎤
⎦ )
x ∈ Ωi such that
j =0...K-1 ⎢⎣
( ⎥⎦ )( )
Max ⎡p x Ω j p Ω j ⎤ (Bayes rule) [11.16]
Multi-sensor Systems: Diagnostics and Fusion 479
Non-optimum
threshold
A1 ’ A0 ’
x
Figure 11.10. Illustration of the Bayesian decision rule for two classes
Figure 11.10 shows the Bayesian rule in one dimension. On the first curve, we
see that the decision threshold is placed at the equality point of probabilities a
posteriori, as the Bayes rule shows. The line of this threshold (respective to the left),
the observation will be modified to class 1 (respective to class 0). The poorly
modified observations of class 0 (respective to class 1) are regrouped in the area A0
(respectively in the area A1). With a threshold differently arranged, as in the second
curve in Figure 11.10, the sum of these two areas can only increase (see areas A’0
and A’1). This explains that the Bayes rule must also be called a minimum cost rule.
The Bayesian classifier cannot always be used directly; it requires knowing the
probabilities of belonging to classes, as well as the internal probabilities densities of
classes. To resolve these difficulties, many classification methods have been
developed. Some of these, including derivation methods, parametric and non-
parametric methods, will be discussed at the end of this chapter. The Bayesian
decision rule still has great theoretic importance; it provides a standard for
comparison for all these methods.
480 Fundamentals of Instrumentation and Measurement
⎛ 1 ⎞
p ( x Ωi ) =
1
( ) ( )
t
exp ⎜ − x − mΩi VΩi −1 x − mΩi ⎟
( 2π ) ⎝ 2 ⎠
p/2 1/ 2
VΩi
The parameters mΩi and VΩi are determined on the learning base. Lacking
complementary information, the probabilities will be chosen equal:
nΩi
p ( Ωi ) = [11.17]
n
In this specific case of a parametric model, applying the rule of Bayes’ decision
(see equation [11.16]) will lead to [DUB 90]:
⎡
( ) V ( x − m ) + Log ( V ) − 2Log ( p (Ω ))⎥⎦⎤
t
−1
x ∈ Ωi such that Min ⎢ x − mΩ Ωj Ωj Ωj j
j=0...K-1 ⎣ j
The first term expresses the removal of the observation to the center of the class j
with a Mahalanobis metric. The second term is a corrective term linked to the
dispersion of of the class j. The final term takes into account the a priori probability
of the class j in the decision rule. If the classes are equiprobable and of the same
dispersion, the Bayes rule is reduced to an attempt to find the minimum distance
between the classes’ centers.
K -1
pˆ ( x Ωi ) =
ki
i = 0...K–1 with ∑ ki = k
nΩi × v ( x) i =0
Multi-sensor Systems: Diagnostics and Fusion 481
Choosing a specific metric influences the form of the volume v; the volume is a
sphere with a Euclidian distance, a cube with a Manhattan distance (||u|| = ∑|ui|) and
an ellipsoid with a Mahalanobis distance.
If we choose, a priori, a probability like the one in [11.17], the Bayes decision
rule then becomes very simple:
The observation is allocated to the class that is most represented among the
closest k neighbors. We can demonstrate that the error rate of the method tends
towards that of the Bayesian classifier if k tends to infinity (the error is two times
higher if k = 1) [KRI 82].
Unfortunately, this method, which is very easy to use, takes a long time to
calculate; with each new observation to be classed, we must calculate the distances
between each observation and all the observations of the learning base.
pˆ ( x Ωi ) =
ki ( x )
i = 0...K − 1 [11.18]
nΩi × h p
Applying this method often leads to obtaining very noisy probability densities,
the results of densities that are too low in the learning base in certain zones of the
representation space. Parzen proposed smoothing these densities by using nuclei
[PAR 62]; these “gently” modify the term ki(x) shown in equation [11.18], instead of
an all-or-nothing counting:
ˆp(x Ω i ) =
1 ⎛ x − xk ⎞
∑ ϕ⎜
⎝ h ⎠
⎟ [11.19]
nΩ × h p x k ∈ Ωi
i
Finally, by the bias of this function, each observation of the class I intervenes in
estimating p(x|Ωi), not those situated in the immediate proximity of x. Gaussian
nuclei are most often used. The adjustment parameter h then plays the role of the
gap type of the nuclei; the more h increases, the more the density estimation will be
smoothed.
The first work on this non-Bayesian method dates from the 1960s. According to
experts, tree classification brings about a series of interleaved tests with a learning
phase that helps define the structure. These tests operate successively on each
descriptive variable and divide the representation space into the most homogeneous
regions possible relative to the classes. Figure 11.1 gives an example of a tree
structure in a situation with two dimensions and three classes.
observation x2
x =[x1 ,x 2 ] intermediary terminal
segment segment
x 2 >0.5 + + + o
yes n + + o
0.5 + o
+
x 1 <1 x 1 >0.3 + +
o o + +
0.2
x
x1
class Ω1 class Ω0 x 2 >0.2
o 0.3 x 1 x
x x
class Ω1 class Ω2
There are many ways to construct this type of tree [GUE 88]. Their construction
principles are often the same. Using the complete learning group, we must look for
all the variables, with the best thresholds separating the base into two groups that are
as homogenous as possible relative the classes. We then have the threshold – and its
linked variable – which produces the best among the best separations. We then
develop the two branches obtained. With each of them, we apply the same procedure
Multi-sensor Systems: Diagnostics and Fusion 483
as before, but applied to the part of the learning base that corresponds to the
threshold as defined above.
Each branch develops this way up to the point when the observations verify all
the tests that already belong to the same class. This sub-group is then termed “pure”.
I(E) = ∑ ∑ p ( Ωi E ) × p ( Ω j E ) [11.20]
i j ,i ≠ j
For classification, the procedure is very simple; we carry out a new observation
on the majority class represented in the extremity of the branch where the
observation is located.
The main importance of this method is that the order of the tests allows us to
take a decision in an optimized way according to the zones of the space where the
observation is located. Modifying a particular class only requires a low number of
tests, and only a few variables need be introduced; but for other classes, the set of
tests can be longer. Overall, the calculation time required is low.
What is more, knowing all the descriptive variables is not necessary if the
observation is in a zone where the modification is simple. Here, the parametrization
time is less (but in industrial control settings, the cost of certain controls is quite
high).
The problems with implementing decision trees lie in choosing the development
level of each branch. If the sub-groups obtained are not pure, the tree structure can
be developed up to the point where each extremity only contains a single
observation! The correct classification rate is then 100% for the learning base but of
course, the generalization capacity of the tree is poor.
484 Fundamentals of Instrumentation and Measurement
Neural networks (NN) have undergone important developments since the mid-
1980s. The work of Rumelhart and Le Cun [RUM 86] is fundamental to this new
trend, since with it we can implement efficient and fast algorithms for learning
multilayered NNs. However, as early as the 1940s, McCulloch and Pitts had begun
working on elementary NNs.
The “biological analogy” is the source of the name, but today, NNs are no longer
seen as simple mathematical operators. Rather, their popularity is due more to their
universal approximate and inexpensive qualities [BIS 95] than to their potential to
reproduce the functioning of the human brain!
Several families exist. In this chapter, we will discuss the most widely used,
supervised-mode structures: the multilayer perceptron (MLP) and the Radial Basis
functions network (RBF). We direct the reader to other texts dealing with other, less
widely used networks [HAY 94] [HEY 94].
⎧1 Scalar Non-linear
⎪ product activation
⎪⎪ x 1 Output
x ⎨ x2 x, w φ
⎪ S = φ ( x, w )
⎪
⎪⎩ x p w weight
This neuron carries out two operations. The first operation is the scalar product
between the input vector and a vector w called the weight (we see that the input
vector includes a constant so that an adjustable bias may be introduced).
Multi-sensor Systems: Diagnostics and Fusion 485
This “scalar product” neuron carries out a linear separation of the input space in
two regions, whether or not the activation function being used is linear or non-linear.
An example is given in Figure 11.13 in two dimensions. Here we see the line of
the plane for which S = 0.5 and which defines two half-planes. The neuron output is
indicated in gray. We can understand the importance of using this type of operator in
a classification problem of two classes; for example:
x2 w 0 + x 1w1 + x 2 w2 = 0
φ (u )
1
u
0
x1
g
1 S=0
S=
1 + e ( 0 1 1 2 2)
− w +x w +x w SS=0,5
= 0.5
The second type of neuron is called a basic “distance” neuron. This time, the
neuron (or nucleus) calculates the distance, with a metric A, between the input vector
and a center C, before injecting the result into the activation function, which is
usually Gaussian. The possible partition of the space is then quadratic. An example
is given in Figure 11.15 in two dimensions.
486 Fundamentals of Instrumentation and Measurement
Non-linear
⎧ x1 Distance
activation
⎪ x
⎪ Output
x ⎨
2
x − C A φ
⎪
⎪ x p
⎩
(
S =φ x−C A )
C center
x2
x1
x2 x1
x2
x1
x2
x1
x2
x1
x2
x1
Regulating the weights of all the layers is done in supervised mode, thanks to the
retropropagation algorithm of the gradient which uses the derivation of the
prediction error of the outputs calculated on the group of the learning base, and
propagated in layers from the output to the input [HEY 94].
The parameters of the network’s architecture that require adjustment are the
number of layers and the number of neurons per layer. This means that the space
488 Fundamentals of Instrumentation and Measurement
partitions brought about by using these networks are no longer simple hyperplanes
(2D lines), but much more complex forms.
Figure 11.17 gives examples in two dimensions of possible partition speeds with
two or three layers and one output.
Even with few nuclei, the partitions of the space are of widely varying speeds,
depending both on the type of distance being used and the weights of the output
layer. Figure 11.19 gives several examples of partitions that are possible with three
nuclei.
Multi-sensor Systems: Diagnostics and Fusion 489
Figure 11.19. Partition in two dimensions with radial base function networks
For situations in which there are more than two classes, two approaches are
possible. Even if the network has as many outputs as there are classes, each output
in the end represents the a posteriori probability of belonging to the class i knowing
x: p(Ωi|x). This approach is called global classification. The retained class will be the
one with the highest associated output.
Let us assume that we subdivide the global problem of K classes into two classes
and construct sub-classifiers dedicated to each sub-problem. Several subdivision
techniques are possible [PRI 94]; the simplest is shown in Figure 11.20, where K
sub-classifiers are represented by the separation of each class among the
K – 1 others. In this simple subdivision instance, the retained class will also be the
one with the highest output of the associated sub-classifier.
The partition approach is very useful in practice. Following the aphorism “divide
and rule”, the complete problem is divided and an individualized adjustment of each
sub-classifier is possible according to the relative difficulty of its sub-problem.
Learning sub-classifiers are independent of each other, which often improves
convergence speeds. This approach even allows personalized selection of input
variables for each sub-classifier, since some variables prove to be relevant for the
separation of a specific class [OUK 98].
We should remember that the partition approach can be used for all types of
classifiers and not only for neural networks.
11.4.1. Introduction
The first objective of data fusion is to help manipulate this kind of knowledge,
always taking into account the changing nature of a dynamic environment. Data
fusion may be seen as a process which helps us integrate information coming from
multiple sources so as to produce more specific and relevant data about an entity, an
activity or an occurrence.
Finally, these are performances of an overall system that matter. Among them,
reactivity plays a basic role in “real time” systems when the reaction time of the
system to a change in the environment is a very significant parameter. Thus, with the
car, a system that detects obstacles must be able to react to a pedestrian in its field of
vision; that is, to provide information to the driver in less than 100 ms. The
reliability of the system is also a factor of utmost importance. In our example we
give an example of limiting or cancelling the false alarm rate.
Whatever its level, the issue is to know what should be fused. What are the
sensors or information sources? How these should be fused, and which techniques
should be implemented?
Among data fusion techniques, we can very generally distinguish three large
classes:
492 Fundamentals of Instrumentation and Measurement
We point out two stages in fusional processing of data: the fusion itself and the
decision. As with our discussion in earlier sections, here we will only speak of the
decision or the diagnostic procedures concerning the hypotheses (class homologues)
constructed on the data or measurements (signal homologues).
We will only discuss in detail three techniques of fusion that present a gradation
in possible modelization of the imprecision and uncertainty [HEN 88]. These
techniques also can be used in low level data processing as well as in symbolic
processing.
Data fusion very often uses the Bayesian method of decision (see section 11.3.2).
Here we use a well-established formalism that has theoretical and experimental
advantages [PEA 88].
p ( x Ωi ) p ( Ωi )
p ( Ωi x ) =
∑ p ( x Ω ) p (Ω )
K-1
j j
j =0
In this approach, we modelize the sensor by the probability p(x|Ωi) that the
observation equals x, knowing that the hypothesis i is verified. We then speak of
the likelihood that the measurement x is conditional to the hypothesis i. A
Gaussian distribution is often used to modelize the sensor’s measurement.
For example, let us assume a detection system using an ultrasonic sensor with
binary response, is triggered in the presence of obstacles in its field of vision: x = 0
(no triggering) or x = 1 (triggering). Assuming that we have two hypotheses:
Ω0 = no obstacles, Ω1 = presence of an obstacle. So, let us lastly assume that the
sensor has a false alarm rate of 1% (p(1|Ω0) = 0.01) and a non-detection rate of 5%
(p(0|Ω1) = 0.05). The sensor will then be modelized by the matrix of the conditional
probabilities shown in Table 11.1. The importance of choosing an example with two
hypotheses with a measurement that can take only two states is that it allows us to
calculate the group of related conditional probabilities.
p(x|Ωi) Ω0 Ω1
x=0 0.99 0.05
x=1 0.01 0.95
Let us now assume that an a priori knowledge of the environment in which the
vehicle moves helps us to estimate the probability of the presence of an obstacle up
to 0.1% (p( l) = 0.001, thus p( 0) = 0.999). Constructing the conditional probability
matrix using the Bayes formula gives us Table 11.2.
p(Ωi|x) Ω0 Ω1
x=0 0.999 5.10-5
x=1 0.913 0.087
We then observe that the decision rule (equation [11.16]) which chooses the
hypothesis giving the highest value of p(Ωi|x) leads to retaining, whatever the
observation, the “non-obstacle” hypothesis! We see how Bayesian information
fusion can modify this result.
p ( x1 , x2 Ωi ) = p ( x1 Ωi ) × p ( x2 Ωi )
The decision is made this time by choosing the hypothesis that a posteriori
maximizes the conditional probability. We know that x1 and x2 are calculated with
equation [11.22]. Under these conditions, the two sensors both have an information
fusion coming from them both:
p ( x1 Ωi ) × p ( x2 Ωi ) × p ( Ωi )
p ( Ωi x1 , x2 ) = [11.22]
∑p(x ) p(x ) p (Ω )
K-1
1 Ωj × 2 Ωj × j
j =0
Now we look at the above example but use two identical ultrasonic sensors. This
time, we will obtain the conditional probability matrix shown in Table 11.3.
p(Ωi|x1,x2) Ω0 Ω1
(x1,x2) = (0,0) 0.999 0.001
(x1,x2) = (0,1) 0.995 0.005
(x1,x2) = (1,0) 0.995 0.005
(x1,x2) = (1,1) 0.1 0.9
The decision rule shown in equation [11.16] chooses the same conclusion as
before (no obstacle), except when there is concomitant triggering in the two sensors
(x1, x2) = (1, 1). The effective fusion of information of the two sensors modifies the
result significantly.
This poses a problem. The measurement of these probability laws is on one hand
far from being easy, and on the other hand can lead to different results in a learning
context than in a reality context. It also assumes that we have an exhaustive
knowledge of the group of possible hypotheses.
If Bayesian fusion analyzes sensor imprecisions, it does not allow us to take into
account their reliability or, more generally, the uncertainties in their model.
This theory is based on the work of A.P. Dempster and was formalized by G.
Shafer [SHA 76]. It uses as a starting point the group E made of K hypotheses i
considered as exclusive and exhaustive. It also assumes that one of these hypotheses
corresponds to each observed situation. The group E is called the frame of
discernment.
A6=E A3
A00=ž
A =Ω0 A11=ž
A =Ω11
A4 A5
A22=Ω
A =ž 22
In the previous example, m(A3) will represent the likelihood that can be
attributed to the hypothesis Ω0∪Ω1, without discernment possible between Ω0 and
Ω1.
We call elements focal, Ai having a non-zero mass. We say that if all the focal
elements are singletons, then the function m(.) corresponds to a probability.
Practically, the mass sets generated each time the sensor is tested easily allows
us to express the indiscernability between hypotheses. When there are three
hypotheses, if m(E) = 1, then m(A0) = m(A1) = m(A2) = 0. This extreme situation
expresses total uncertainty: the sensor is incapable of discerning between the
different hypotheses of the discernment frame.
Multi-sensor Systems: Diagnostics and Fusion 497
The major advantage of this theory is that it allows for conjoint evaluation of
hypotheses, modelization of uncertainties and the indiscernability between
hypotheses – all factors that the Bayesian theory cannot handle well. Uncertainty
cannot be expressed in Bayesian theory, except by an equidistribution of
probabilities for the different hypotheses.
The value mC2(Ω0 ∪ Ω1) = 0.1 characterizes the uncertainty of the sensor’s
functioning. The distribution of the two other masses mC2(Ω0) mC2(Ω1) remains, in
our example, in the same proportions as before.
It is important to remember that the mass sets are not fixed entities and they can
be changed at each new measurement.
The fusion itself of the mass sets coming from several sensors will be discussed
in section 11.4.3.4.
498 Fundamentals of Instrumentation and Measurement
Cr ( A ) = ∑ m ( B)
B⊆ A
[11.24]
Pl ( A ) = 1 − Cr ( ¬A ) = ∑
B ∩ A ≠∅
m ( B) [11.25]
where ¬A represents the complementary group of A for the group 2E . We see that if
the only elements Ai having a non-zero mass (focal elements) are the singletons,
then:
Cr ( Ωi ) = Pl ( Ωi ) = m ( Ωi ) = p ( Ωi )
we have two sources, and thus the mass sets mC1 and mC2, the fusion is expressed by
the following rule:
1
m( Ak ) = m C1 ( Ak ) ⊕ mC 2 ( Ak ) = ∑ mC ( Ai ) m C 2 ( Aj ) [11.26]
1 − H Ai ∩ A j = Ak 1
H= ∑ m C1 ( Ai ) mC 2 ( A j )
Ai ∩ Aj =∅
H = 0 shows that the two sources are never in contradiction. This normalization
allows the mass set obtained by fusion to verify the definition [11.23]. This rule
generalizes when there are several or many sources and/or sensors, always
respecting the properties of commutability and associability that are indispensable to
all fusion mechanisms (see [JAN 96] and [SHA 76]).
Below we will define the decision rule constructed from the fused mass set.
11.4.3.6. Example
Let us look again at the example of the two binary response sensors working in
the framework of two hypotheses 0 = no obstacles; 1 = an obstacle. Assuming
that they independently analyze the same scene with different modalities, one of the
sensors is ultrasonic, the other infrared. The ultrasonic sensor is assumed to have
reliable functioning. The infrared sensor’s functioning is assumed to be less well
known ( ≠ 1). Once the mass set and response functions of the sensors are set up,
and taking into account the value attributed to for the infrared sensor, it is possible
to fuse the sets coming from the two sensors and to make a decision by retaining the
most possible hypothesis. Assuming that the ultrasonic sensor provides the response
xus = 1 (an obstacle is present). We then construct the mass set:
We assume that the infrared sensor provides the response xir = 0 (no obstacle)
and that the reliability of its functioning is characterized by g = 0.9. We then
construct the following mass set:
mir (Ω0) = 0.72 mir (Ω1) = 0.18 mir (Ω0 ∪ Ω1) = 0.1
The two sensors give contradictory results but the mass fusion allows us to
decide. Applying the rule in [11.26] and the representation [11.25] provides the
result given in Table 11.4 that chooses the “presence of an obstacle” hypothesis.
Ω0 Ω1 Ω0 ∪ Ω1
US xus = 1 0.2 0.8 0
IR xir = 0 0.72 0.18 0.1
Fusion 0.42 0.58 0
Plausibility 0.42 0.58 0
We see that the mass set of the infrared sensor with uncertainties (and which
responds to “no obstacle”) cannot counterbalance the response of the ultrasonic
sensor. We also see that, in this simple situation, plausibility is directly equal to the
mass set resulting from fusion.
Tables 11.5 and 11.6 show the results of the fusion obtained for the two previous
sensors in all possible response configurations.
mus / mir Ω0 Ω1 Ω0 ∪ Ω1
x=0 0.75 / 0.72 0.25 / 0.18 0 / 0.1
x=1 0.2 / 0.2 0.8 / 0.7 0 / 0.1
mus ⊕ mir = Pl Ω0 Ω1 Ω0 ∪ Ω1
We notice that in cases where there is contradiction, the ultrasonic sensor does
not impose its choice because of the distributions of the masses on singletons.
The theory of possibilities was introduced by D. Dubois and H. Prade [DUB 87].
It is based on the fuzzy subgroups of L.A. Zadeh [ZAD 65] [ZAD 78]. It allows us
to leave the probabilistic framework of many fusion approaches. It is particularly
well-adapted to a situation where we know very little about the information sources
(sensors, experts, etc.).
µ Ω1 (x) µ Ω2 (x)
1
x
0
The imperfections of a sensor and how well the person performing the experiment
knows the sensors can also be modelized by functions of the same type. Here, we
speak of a distribution of possibilities. We assume that a sensor delivers the
measurement xm of a real variable xr. The information we have about the value of the
measured variable is expressed by a distribution of possibilities written as πm(x): that
the variable is equal to x, knowing that the measurement is xm (see Figure 11.23).
π m( x )
x
xm
µ A∪ B ( x ) = max ( µ A ( x ) , µ B ( x ) )
µ A∩ B ( x ) = min ( µ A ( x ) , µ B ( x ) ) [11.28]
µ¬ A ( x ) = 1 − µ A ( x )
Multi-sensor Systems: Diagnostics and Fusion 503
If these sensors are not completely reliable, especially if the information they
provide is discordant, we can take this discordance and uncertainty of this information
into account by using the conjunction of distributions using the operator ∪.
Figure 11.24 illustrates these two types of fusion when there are two sensors.
πC1(x) and πC2(x) are the distributions of possibilities of the values of the variable x
observed by two sensors C1 and C2.
All intermediary forms of fusion are certainly possible [BLO 96], which justifies
sophisticated supervision techniques, according to a priori knowledge.
π C 1 ∩C 2 (x)
conjunctive
fusion
πC 1 (x)
x
πC 2 ( x)
x π C 1 ∪C 2 (x)
disjunctive
fusion x
As for hypotheses, the same types of fusion can be used. This means that the
ownership function µΩ1 ∪ Ω2(x) characterizes the veracity of the hypothesis
Ω1 ∪ Ω2.
Π ( S ) = Sup ( π ( x ) )
x∈ S
[11.29]
N ( S ) = Sup (1 − π ( x ) ) = 1 − Π ( ¬ S )
x∉S
These functions evaluate the confidence we have in the statement: “The variable
has a value belonging to the group S.”
( i ) = Sup { min ( µ i
( x), ヾ m ( x) )} [11.30]
x
We can evaluate the degree of possibility of each hypothesis (truck or car) and
finally decide by using the rule of maximum possibilities (equation [11.31]). Figure
11.25 shows this mechanism.
Multi-sensor Systems: Diagnostics and Fusion 505
π m (x)
µ car(x) µ truck (x)
πm (car)
π m (truck)
µ truck(xm)
µ car (xm)
x
x m
We should observe that, on the figure, the maximum possibility degree rule gives
a “car” result different from the “truck” result which would have been obtained from
the raw measurement xm by using as a decision rule the maximum of the ownership
function. So:
11.4.5. Conclusion
Data fusion can take various forms. The reader will find a more thorough
coverage of different fusion techniques in [ABI 92], [TS 94] and [TS 97].
On the other hand, if the data are numeric, we choose possibilistic fusion; in
particular if we do not have a reliable probabilistic modelization of the sensors.
506 Fundamentals of Instrumentation and Measurement
This chapter has presented some of the aspects of diagnostic problems with the
help of multisensor systems: choosing a representation space of signals and in
particular the reduction of it dimension; Bayesian and non-Bayesian classification
techniques; and fusion of probabilistic and possibilistic data.
All these topics have been extensively researched and developed. The recent
refinement of connectionist techniques, logical flow and the theory of evidence, to
name several, have all made their contribution.
11.6. Bibliography
[ABI 92] ABIDI M.A., GONZALEZ R.C., Data Fusion in Robotics and Machine
Intelligences, Academic Press, 1992.
[APR 91] APPRIOU A., “Probabilités et incertitudes en fusion de données multi-senseurs”,
Revue Scientifique et Technique de la Défense, no. 11, p. 27-40, 1991.
[BEL 61] BELLMAN R., Adaptive Control Process: A Guided Tour, Princeton University
Press, 1961.
[BIS 95] BISCHOP C.M., Neural Networks for Pattern Recognition, Clarendon Press, 1995.
Multi-sensor Systems: Diagnostics and Fusion 507
[BLO 96] BLOCH I., “Information Combination Operators for Data Fusion: A comparative
Review with Classification”, IEEE Trans on SMC, vol. 26, no. 1, 1996.
[CHE 89] CHEN S., BILLINGS S.A., LUO W., “Orthogonal least squares methods and their
application to non-linear system identification”, Int J. Control, vol. 50, no. 5,
p. 1873-1896, 1989.
[DEV 94] DEVEUGHELE S., DUBUISSON B., “Adaptabilité et combinaison possibiliste:
application à la vision multicaméra”, Revue Traitement du Signal, vol. 11, no. 6,
p. 559-568, 1994.
[DOC 81] DOCTOR P.J., HARRINGTON T.P., DAVIS T.J., MORRIS C.J., FRALEY D.W.,
“Pattern recognition methods for classifying and sizing flaws using eddy current data”,
Eddy Current Characterization of Materials and Structures, ASTM, p. 464-483, 1981.
[DUB 87] DUBOIS D., PRADE., “Une approche ensembliste de la combinaison
d’informations imprécises ou incertaines”, Revue d’intelligence artificielle, vol. 1, no. 4,
p. 23-42, 1987.
[DUB 90] DUBUISSON B., Diagnostic et reconnaissance des formes, Hermès, 1990.
[DUB 97] DUBOIS D., PRADE H., YAGER R.R., et al., Fuzzy Information Engineering,
Weisley, 1997.
[DUD 73] DUDA R., HART P., Pattern Recognition and Scene Analysis, Wiley & sons,
1973.
[FUK 72] FUKUNAGA K., Introduction to Statistical Pattern Recognition, Academic Press,
1972.
[GAI 83] GAILLAT G., Méthodes statistiques de reconnaissance de formes, Cours ENSTA,
1983.
[GUE 88] GUEGEN A., NAKACHE J.P., “Méthode de discrimination basée sur la
construction d’un arbre de décision binaire”, Revue Stat. Appl., 36 (1), p. 19-38, 1988.
[HAY 94] HAYKIN S., Neural networks, A Comprehensive Foundation, Prentice Hall, 1994.
[HEN 88] HENKIND S.J., HARISSON M.C., “An Analysis of Four Uncertainty Calculi”,
IEEE Trans SMC, vol .18, no. 5, p. 700-714, 1988.
[HEY 94] HEYRAULT J., JUTTEN C., Réseaux neuronaux et traitement du signal, Hermès,
1994.
[JAN 96] JANEZ F., APPRIOU A., “Théorie de l’évidence et cadres de discernement non
exhaustifs”, Revue Traitement du Signal, vol. 13, no. 3, 1996.
[KIT 86] KITTLER J., Feature Selection and Extraction. Handbook of Pattern Recognition
and Image Processing, Academic Press, 1986.
[KRI 82] KRIHNAIAH et al., “Classification, pattern recognition and reduction of
dimensionality”, Handbook of Statistics, vol. 2, North-Holland, 1982.
[LEB 97] LEBART L., MORINEAU A., PIRON M., Statistique exploratoire
multidimensionelle, Dunod, 1997.
508 Fundamentals of Instrumentation and Measurement
[NIF 97] NIFLE A., REYNAUD R., “Classification des comportements fondée sur
l’occurrence d’événements en théorie des possibilités”, Revue Traitement du Signal, vol.
14, no. 5, p. 523-533, 1997.
[OUK 98] OUKHELLOU L., AKNIN P., “Optimisation de l’espace de représentation dans un
problème de classification par réseaux de neurons”, Journal Européen des Systèmes
Automatisés, vol. 32, no. 7-8, p. 915-938, 1998.
[OUK 99] OUKHELLOU L., AKNIN P., “Hybrid training of radial basis function networks
in a partitioning context of classification”, Neurocomputing, vol. 28, no. 1-3, p. 165-175,
1999.
[PAR 62] PARZEN E., “On the estimation of a probability density function and mode”, Ann.
Math. Stat, vol. 33, p. 1065-1076, 1962.
[PEA 88] PEARL J., Probabilistic Reasoning in Intelligent Systems, Morgan Kaufmann,
1988.
[PRI 94] PRICE D., KNEER S., PERSONNAZ L., DREYFUS G., Pairwise neural network
classifiers with probabilistic outputs, Neural Information Processing Systems, 1994.
[RUM 86] RUMELHART D.E., HINTON G.E., WILLIAMS R.J., “Learning internal
representations by error propagation”, Parallel Distributed Processing: Explorations in
Microstructure of Cognition, vol. 1, p. 318-362, MIT Press, 1986.
[SAP 90] SAPORTA G., Probabilités, Analyse des données et Statistique, Technip, 1990.
[SHA 76] SHAFER G., Mathematical Theory of Evidence, Princeton University Press, 1976.
[STO 97] STOPPIGLIA H., Méthodes statistiques de sélection de modèles neuronaux, PhD
Thesis, University of Paris VI, 1997.
[THE 99] THEODORIDIS S., KOUTROUMBAS K., Pattern Recognition, Academic Press,
1999.
[TS 94] Revue Traitement du Signal, numéro spécial Fusion de données, vol. 11, no. 6, 1994.
[TS 97] Revue Traitement du Signal, numéro spécial Fusion de données, vol. 14, no. 5, 1997.
[ZAD 65] ZADEH L.A., “Fuzzy sets”, Information and Control, vol. 8, p. 338-353, 1965.
[ZAD 78] ZADEH L.A., “Fuzzy sets as a basis for a theory of possibility”, Fuzzy sets and
systems, vol. 1, p. 3-28, North-Holland Publishing, 1978.
[ZAD 92] ZADEH L.A., KACPRZYK J., Fuzzy Logic for the management of uncertainty,
Wiley, 1992.
[ZWI 95] ZWINGELSTEIN G., Diagnostic des défaillances, Hermès, 1995.
Chapter 12
Intelligent Sensors
12.1. Introduction
Since the end of the 1980s, many articles have appeared in scientific [BER 87],
[GIA 86] and technical [BLA 87], [JOR 87] books using the term “intelligent”, a
word often associated with sensors, transmitters, actuators and instrumentation. In
this chapter, we will define an intelligent sensor as the linkage of one or several
measurement chains to a computing “machine”. Its main function is providing
reliable, useful information. However, sometimes we read that the sensor is the
weakest link in the measurement chain.
– This action/reaction loop could also be closed by a human operator. Still in the
automotive context, it is the driver who usually observes the speedometer and either
brakes, accelerates or reduces gear to adjust the speed of the car to the speed limits.
Obviously in this context the speed sensor must be precise and reliable.
Goals
Analyze,
Make decision
Observe Act
Process
Whatever the context, sensor users want products that perform well; that is,
sensors that are reliable from the first performance or, to be more technologically
exact, that are accurate1. In complementing this accuracy, it is more particularly the
credibility2 of the information which is required by the users of the systems. This is
especially true in the automotive context, with users who tend to have absolute
confidence in their vehicles and would be angered by erroneous warnings generated
by a faulty surveillance system.
Together with this need for accurate measurements, integration strategies require
mutual exchanges of information between the automation units. The development of
these techniques and automation methods has led to less centralized processing,
made possible by new communication networks. These networks are called
fieldbuses [CIA 99] and the appearance, by the 1980s, of “smart” or “intelligent”
equipment.
Hierarchy
Automation
P&T Management
system
Safety Functional
Maintenance
Control
Sensors
Actuators
PROCESSING Topology
3 We will come back to this example of “intelligent distance sensors” after discussing the
concept of “intelligent sensors”.
Intelligent Sensors 513
Figure 12.3. Schema principle for the direct injection system with an injection pump
We see that the car can be described as a system and also can be broken down
into subsystems; the motor control and anti-lock brake systems are themselves
systems. In looking at the example of Figure 12.3, taken from [GUY 99], we see
there are no less than seven sensors that provide information to the injection system,
which, in this particular case, leads to an increase of 21.3% of the engine torque to
1,900 tr/mn.
514 Fundamentals of Instrumentation and Measurement
We are basing our ideas on those in [KLE 91] in distinguishing between a smart
sensor and an intelligent sensor (see Figure 12.4):
– a smart sensor has functionalities that improve its metrological performances
by using numerical processing;
– an intelligent sensor integrates functions that allow it to fully participate in the
goal of an automated system, which then becomes a distributed automated system.
Mechanisms are implanted in this system and exchange information through the
dedicated communication system. This system is the backbone of a true real-time
database [BAY 95].
Intelligent sensor
Automation
"Smart" sensor
Metrology
Integration
Signal
"Standard" processing
sensor
Compensation
Fieldbus
Point-to-point Point-to-point
connection connection
Although these names are now in common usage, we prefer the term “digital
sensor with processing and communication capacities”, which specifies that the
system is a measurement device, that it is created by digital technology, that it has
bidirectional communication means, and processing capacities. An intelligent sensor
is then seen as a fully functional system with its own processing abilities, one that
can take part in more complex systems.
This architecture clearly is more complex that that of a standard sensor. It links
one or several measurement chains and the equivalent of a computer machine [BRI
96]. The related processing possibilities improve metrological performances in terms
of reliability and sensor availability.
Logical
Measurand Influence variables
Transducer Transducer quantitie(s)
Conditioner Conditioner
P
O
W
E Multiplexer Ampli. S&H
R
Interfaces
U
N
I Multiplexer Gain S&H
T ADC
Command Command Command
Internal bus
Component
Micro... ROM RAM Fieldbus
Comm.
Among the essential qualities of sensors [AFN 94] are the following: freedom
from bias error, fidelity, accuracy, rangeability, sensitivity, linearity, sharpness or
keenness, rapidity, resolution, traceability, repeatability and reproducibility.
Accuracy is the attribute which is most often considered by users.
This improvement of accuracy can make the sensor more adaptable and improve
its range, so it can be used in a variety of situations and applications.
For example, the same type of temperature sensor can acquire information about
the interior temperature of a car and about the interior of a cylinder, always retaining
its precision. Obviously, modelizing the transfer function signal output =
f(measurand) feature goes beyond simple linear transformations. In addition, the
linearity requirements need no longer be respected, since the information provided is
quantified, digitized and transmitted according to a range of different codings. These
codings allows us to link the corresponding physical unity, freeing us from fixed
rule of linearity imposed by using an intermediary variable such as the 4-20 mA, or
0-10 V for reasons of interoperability between sensors, actuators, regulators or
recorders.
Intelligent Sensors 517
Thus, the overriding need for reliability was behind the early development of
intelligent sensors. An intelligent sensor must be able to provide valid information
leading to the dependability of the application, even if the application itself is not of
optimal reliability, availability, safety, maintainability or durability.
These goals are met through validation procedures that are completed by:
– auto-tests and auto-diagnostics;
– stored memory of the last delivered values;
– alarm systems used when failures are detected;
– configuration re-readings;
518 Fundamentals of Instrumentation and Measurement
– network reconfigurations.
These procedures are described in detail in [BRI 94] and [CIA 87].
In the context of cars, the need for reliable, certain and validated information is
obvious. No consumer will accept an airbag that doesn’t inflate; no driver will
accept false warnings about engine problems.
Technological
Acquisition, Measurement Metrological
Correction, Operational
Compensation,
Signal
processing, Validation
Operational
measurement
Communication Configuration
Technological
Functional
Synchronisation,
Operational
Date
There is a crucial need for intelligent sensors which can provide data and
generate validated information. Indeed, the issue of validation is crucial to
intelligent sensors.
Functio nal
measurement(s)
Qualified
me asure me nt(s)
Operational
me asureme nt
User(s)
In the following paragraphs we will illustrate these concepts by showing how the
level of gasoline in a tank is converted to relevant information about how far a car
can travel with this gasoline.
The principle quantity here is obviously the level of fuel in the gas tank.
The temperature inside the gas tank can be seen as an influence quantity that
through the laws of liquid dilation influences the quantity of energy actually
available. The influence quantities are intrinsically linked to metrological and
physico-chemical characteristics of the proof body, allowing us to carry out the first
measurement. We see that the chemical characteristics of the fuel (the octane rate
can vary according to where the gasoline was bought) can be taken into account to
make the prediction more exact.
520 Fundamentals of Instrumentation and Measurement
The autocontrol variables are, for example, the pressure and temperature in the
gas tank. These help to validate the nominal functioning of the sensor of the level
used and provide information relating to safety. We see that the variable
“temperature inside the gas tank” appears as both an influence quantity and an
autocontrol variable; it can be used in many validation steps. On another level, the
temperature of the processing unit, the supply voltage of the electronic module, and
the supply voltage of any sensor can also be variables requiring surveillance.
The primary measurement is an electronic signal coming from the height sensor,
a signal which must be validated technologically. Then it must be verified that the
frequency and amplitude of the supply voltage conditioner linked to the capacitive
transducer being used are close to nominal values, that the output of this conditioner
is coherent with the geometric characteristics of the gas tank.
The auxiliary measurement is the tension delivered by a thermistor that has been
traversed by a known or measured current that represents the temperature of the
tank. This measurement is obtained though a model specific to the thermistor that
describes the temperature relation of the range of potential utilization of this proof
body. So the temperature = f(tension, current). This model must be constructed and
stored in the intelligent sensor. We point out here that this auxiliary measurement
can be compensated by taking into account auto-heating phenomena, for example, or
validated by measuring the current going through the thermistor; so validation has a
recursive or returning aspect.
technological measurements requiring that the nominal conditions for primary and
auxiliary measurements be combined at the moment a measurement is made. If this
is not the case, the processing can be done following several different strategies:
– if the producer of the primary measurement is faulty, then an estimation of the
measurement can be obtained from measurements previously made that have been
stored on a specific temporal horizon in the “dynamic database” by taking into
account the distance traveled since the problem began;
– if the metrological conditions are not combined by the faulty temperature
sensor, for example, the estimation of the distance traveled will be less precise and
this estimation error may be transmitted to the user;
– in other circumstances, a foldback value that has previously been
parameterized according to the application may be produced. Typically, the foldback
value can be the last produced operational measurement;
– in addition, alarm systems can be produced; the receiving systems or units
must have be able to make decisions to regulate these unusual situations which,
however, have been foreseen in the design step of the system.
The physical variable or quantity “time” can also be measured or more or less
produced within the system itself, in order to represent the refresh period of the
operational measurement. This physical variable “time” can be considered as an
implicit physical quantity.
This last example underlines the need for integration and precise representations
of relevant data, specifications for related processing, and efficient exchanges of
data within a vehicle. It brings up the potential problems of optimal data distribution
and the related processing of a system.
For production systems, more than 50 types of fieldbuses are currently on the
market, many of which are described in [CIA 99], [FAG 96], [PET 96] and [TER
97].
In the automotive domain, the Controller Area Network (CAN) [PAR 96, 99] is
currently the reference fieldbus, as well as another automotive multiplexing
network, the Vehicle Area Network (VAN), which has mainly been used by PSA
and its subsidiaries [ABO 97].
CAN was first developed by Bosch and Intel for automotive applications. It has
had good reliability and low costs. Many car manufacturers use or are preparing to
use the CAN network. The association of utility vehicles of the USA have also
adopted the fieldbus as a standard, as have most industrial manufacturers, mostly
because of the buses’ wide availability and competitive costs. At this time,
according to [MEN 99], sales for CAN components for cars have surpassed sales to
the industry.
Schematically, the CAN uses a bus topology and belongs to the class of
multimaster networks of the producer/consumer type. In these networks, different
levels of information are transmitted according to the diffusion principle, and
regulated by implanting the protocol CSMA/CR (Carrier Sense Multiple
Intelligent Sensors 523
Many other multiplexing systems or, more generally, exchange systems in cars
are now being developed by automotive manufacturers or other consortia. Among
these are the J1850 or the ITS (Intelligent Transport System), the Data Bus made by
the SAE, Society of Automotive Engineers, and the OSEK/VDE or the Open
systems and Interfaces for Distributed Electronics in Cars/Vehicle Distributed
Executive. The latter is a consortium of European manufacturers and scientists that
aims to establish operating system norms and communication protocols in the field
of electronics.
The previous sections show that, in the world of automotive electronics, the
current trend is towards developing interacting, multiplexing systems. This approach
is very important to the field of intelligent sensors, which are integrated into a
distributed automation architecture since, in addition to the characteristics discussed
above, an intelligent sensor must have the following features;
– it must be interoperable. This means it must cooperate with other automation
components in and with a specific application. This, at the very least, means using a
common standard of communication to allow the exchange of information and also
means that the two components conform to the same interpretation of data;
– its components must be interprocessible. This feature, which is not easily
differentiated from the first one, is especially important for the equipment and its
integration with a automation system with distributed architecture (ASDA)
dedicated to an industrial process;
– it must be interchangeable. This means that the equipment of one manufacturer
can be replaced with that of another manufacturer without changing the components.
The above concepts [STA 96] are both an advance and a check to any large-scale
diffusion of intelligent sensors in industry. They imply normalization procedures
which could appear to be constraints. However, some studies ([INC 93]; [LEE 96])
have proposed a standard, or at least an aid, to producing intelligent instruments that
might disregard interface communication. In the near future, these works might lead
to a total or partial modelization of the abilities of intelligent sensors and also lead to
524 Fundamentals of Instrumentation and Measurement
This is one of the objectives of the LARII project for “Software of Assistance to
the Creation of Intelligent Interfaces for Sensors and Actuators” which profited from
the financial support of the French authorities and which must make it possible to
diffuse near the French companies manufacturing of the sensors, mainly PMI-PME
of the software libraries, comprising functional blocks based on the standard IEC
1131 (typing of the data, definition of the functions, programming language, etc.)
directly usable in the design of intelligent sensors. These libraries will be integrated
in the strategy of design which should no longer remain on a solely technological
approach [ LEM 99].
In spite of industry’s stated enthusiasm for the idea of intelligent sensors, there
are clear drawbacks to their present widespread use:
– the wait-and-see policy of designers, manufacturers and potential users, who
must choose between the current wide range of communication networks and
fieldbuses (more than 50 fieldbuses are currently on the market ([TER 97, PET
98]));
– advances in industrial instrumentation are basically marketing tools of
manufacturers who develop new techniques before there is a real need for them;
– the integration of intelligent automation components is linked to the issue of an
optimal distribution of data and processing [CON 99]; [HER 98].
Especially in the automotive context, Figure 12.9 shows the variety of sensors
which can be integrated in a vehicle that is used on a daily basis.
An article from the end of the last century [GRA 99] mentioned the sale of
sensors used in vehicles: these sales are estimated at $5.18 billion in 1997, while
Intelligent Sensors 525
sales predictions for 2002 are estimated at $6.86 billion. Even though the technical
works often combine sensors, associated electronics, associated software and
mulitplexing, the sale of electronic chips for cars was estimated at $8.25 million in
1998; predictions for 2000-2001 have been estimated at $13.3 million [VER 99].
Experts’ predictions say that automotive electronics will, in the short term, represent
20% of the finished product, which may well surpass the cost of the mechanics.
According to [GRA 99], the “salability threshold” for a vehicle sensor may be
around $5-7, including the cost of signal processing. This means using digital
technologies of the MEMS (Microelectochemical systems) or MST (Microsystems
technologies) type.
Aside from the optimal running of the engine and the comfort of the driver and
passengers, the key word in the automotive field is safety. We can see this simply by
looking at acronyms such as ESP (Electronic Stability Program), EBS (Electronic
Braking System), TCS (Traction Control System) and ASR (Acceleration Slip
Regulation), which imposes calculation powers comparable to those of a 1982 A310
Airbus [VER 99].
526 Fundamentals of Instrumentation and Measurement
We also note that on the margins of automotive manufacturing, the demand for
sensors and associated systems is also important in the field of vehicle testing, both
in the design phase and in the test-run phases [ERW 99].
12.8. Bibliography
[ABO 97] ABOU B., MALVILLE J., Le Bus Van (Vehicle Area Network). Fondements du
protocole, Dunod, 1997.
[AFN 94] AFNOR, Vocabulaire fondamental des termes fondamentaux et généraux de
métrologie, Norme AFNOR NF X07-001, December 1994.
[AFN 96] Maintenance industrielle, Recueil de normes françaises, réf. 321 61 40, Paris, Ed.
AFNOR, 1996.
[BAY 95] BAYART M., SIMONOT F., Rapport Convention MESR 92-P-0239: Impact de
l’émergence des réseaux de terrain et de l’instrumentation intelligente dans la conception
des architectures des systèmes d’automatisation de processus. Rapport Final. Résultats et
Perspectives, Ministère de l’Enseignement Supérieur et de la Recherche, January 1995.
[BER 87] BERRY, “Distributed intelligence in Process Control”, Control Engineering,
May 87, p. 62-64, 1987.
[BLA 87] BLADOU, “Comment l’intelligence vient aux capteurs”, Electronique Industrielle,
no. 127, 1987.
[BRI 94] BRIGNELL J., TANER A., “Aspects of smart sensor reconfiguration”, Proceedings
of Eurosensors VIII conference, Toulouse, p. 525-529, 25-28 September 1994.
[BRI 96] BRIGNELL J., Intelligent Senor Systems, Institute of Physics Publishing, Sensors
Series, 1996.
[CIA 87] CIAME, Livre blanc: les Capteurs Intelligents, réflexions des utilisateurs, CIAME
AFCET, 1987.
[CIA 99] CIAME (Collectif), Réseaux de terrain. Description et critères de choix, Hermès,
1999.
[CON 99] CONRARD B., Contribution à l’évaluation quantitative de la sûreté de
fonctionnement des systèmes d’automatisation en phase de conception, PhD Thesis,
University of Henri Poincaré-Nancy 1, 24 September 1999.
[CRO 99] CROSSE J., “Smart Tyres, The shape of things to come”, Financial Times
Automotive World, p. 45, October 1999.
[DEL 90] DELCUVELLERIE J.-L., “L’impact de FIP sur la conception des architectures”,
Revue Générale de l’Electricité, p. 57-64, 7 July 1990.
[END 97] ENDRES H.E., “Signal evaluation of gas sensors with artificial neural nets”,
Proceedings of SICICA’97, p. 399-404, IFAC, 9-11 June 1997.
Intelligent Sensors 527
[ERW 99] ERWIN F., “Improvements in Vehicle-Testing Technology”, Sensors, vol. 16,
no. 12, December 1999.
[FAG 96] FAGES G., Les Bus de Terrain, Technical collection, Schneider, 1996.
[GIA 86] GIACHINO, “Smart Sensors”, Sensors & Actuators, no. 10, p. 239-248, Elsevier
Science S.A., 1986.
[GRA 99] GRACE R., “The Growing Presence of MEMS and MST in Automotive
Applications”, Sensors, vol. 16, no. 9, September 1999.
[GUY 99] GUYOT R., “Injecteur pompe Diesel”, Auto Concept, no. 25, April-May 99.
[HER 97] HERMANN F., THIRIET J. M., ROBERT M., “Task allocation for the design of
distributed architecture in automation systems: application to a pilot thermal process using
a fieldbus network”, Proceedings of SICICA’97, Annecy, IFAC, p. 459-464, 9-11 June
1997.
[INC 93] Projet Eureka EU 666 INCA Interface Normalisée pour Capteurs et Actionneurs,
Tâche 0B, Rapport final des travaux, July 1993.
[JOR 87] JORDAN, “Sensor technologies of the future”, GEC Review, vol. 3, no. 1, p. 23-32,
1987.
[KLE 91] KLEINSCHMIDT P., SCHMIDT F., “How many sensors does a car need?”,
Proceedings of Eurosensors V Conference, Rome, p. 1-13, 2 October 1991.
[LEE 96] LEE K.B., SCHNEEMAN R., “A standardized approach for transducer interfacing:
implementing IEEE-P1451 Smart transducer interface standards”, Proceedings of Sensors
Conference, Philadelphia, p. 87-100, 22-24 October 1996.
[LEM 99] LEMAIRE E., Spécification et Vérification Fonctionnelle et comportementale d’un
Equipement Intelligent, PhD Thesis, University of Sciences and Technologies, Lille,
14 December 1999.
[MEN 99] MENARD C., “L’électronique embarquée se cherche une architecture et une
fiabilité”, l’Usine Nouvelle, no. 2703, p. 76-83, 30 September 1999.
[MUL 95] MULLINS M.A., VAN PUTTEN A.F.P., BAYFORD R., BUTCHER J.B.,
“Potential for a smart sensor based on an integrated silicon anemometer”, Sensors and
Actuators (A), no. 46-47, p. 342-348, Elsevier Science S.A., 1995.
[NOI 97] NOIZETTE J.L., ROBERT M., RIVIERE J.M., “Intelligent sensor calibration
complexity, methodology and consequences”, Proceedings of IEEE Instrumentation and
Measurement Technology Conference, Ottawa, p. 948-952, 19-21 May 1997.
[PAR 96] PARET D., Le bus CAN, Description, De la théorie à la pratique, Dunod, 1996.
[PAR 99] PARET D., Le bus CAN, Applications CAL, CANopen, DeviceNet, OSEK, SDS,
Dunod, 1999.
528 Fundamentals of Instrumentation and Measurement
[PET 98] PETERSON G.C., “Selecting the right Industrial Network”, Control Engineering
International, p. 43-46, January 1998.
[ROB 93] ROBERT M.J., MARCHANDIAUX M., PORTE M., Capteurs Intelligents et
Méthodologie d’Evaluation, Hermès, 1993.
[ROB 99] ROBERT M.R., Contribution au développement d’une méthodologie d’étalonnage
de capteurs : application à un transmetteur intelligent de température, Doctorat de
l’Université Henri Poincaré-Nancy 1, 18 February 1999.
[RUT 91] RUTLEDGE D.N., DUCAUZE C.J., “An iterative method for determining the
linear zone of a detector response”, Chemometrics and Intelligent Laboratory Systems,
no. 12, p. 5-19, 1991.
[STA 96] STAROSWIECKI M., BAYART M., “Models and languages for the
interoperability of smart instruments”, Automatica, vol. 32, no. 6, June 1996.
[TER 97] Terrain, “Le marché des bus de terrain”, no. 13, p. 18-22, May-June 1997.
[VER 99] VERNAY J.P., “La voiture branchée est électronique”, l’Usine Nouvelle,
no. 2703 30, p. 72-75, September 1999.
[VIL 98] VILLEMEUR A., Sûreté de fonctionnement des systèmes industriels, Eyrolles,
1998.
List of Authors
Patrice AKNIN
INRETS, Arcueil
François BAILLIEU
ESIEE, Noisy-le-Grand
Paul BILDSTEIN
ESIEE, Noisy-le-Grand
Cécile DURIEU
Ecole Normale Supérieure de Cachan
Bernard JOURNET
Ecole Normale Supérieure de Cachan
Michel LECOLLINET
CNAM, Paris
François LEPOUTRE
CNAM, Paris
ONERA/DMSE
Thierry MAURIN
Ecole Normale Supérieure de Cachan
IEF, Paris-Sud University
Dominique MILLER
Ecole Normale Supérieure de Cachan
530 Fundamentals of Instrumentation and Measurement
Mustapha NADI
LIEN
Henri Poincaré University
Nancy
Dominique PLACKO
Ecole Normale Supérieure de Cachan
Stéphane POUJOULY
IUT Cachan
Denis PRÉMEL
Ecole Normale Supérieure de Cachan
Michel ROBERT
Henri Poincaré University
Nancy
Eduardo SANTANDER
Ecole Normale Supérieure de Cachan
Frédéric TRUCHETET
Le2i
University of Bourgogne
Le Creusot
Olivier VANCAUWENBERGHE
ESIEE, Noisy-le-Grand
Index
A E
amplifier 66, 147 effect
instrumentation 160 load 12, 18
isolation 162 Peltier 131
logarithmic 163 photoelectric 94
operational 147, 153, 192 Seeback 129
axis Thomson 130, 131
inertia 468
principle 468 F
C filter 223
active 168, 191, 199
cell analog 167, 169
capacitive 291 anti-folding 215, 228
Fleisher-Tow 198 corrective 170
Friend 194 dynamic 432
piezoresistive 307 FIR 258, 431
photoconductor 94 half-band 215, 241
Sallen-Key 193 IIR 260, 431
Tow-Thomas 196 Kalman 439
converter low pass ladder 179
analog-to-digital 229 passive 168, 177
current Wiener 437
darkness 82
gap 156
polarization 156
532 Fundamentals of Instrumentation and Measurement
M S
sensor
measurement intelligent 509
auxiliary 520 signal
functional 520 analog 29, 416
operational 519, 521 causal 417
primary 520 deterministic 416
technological 520 ergodic 421
periodic 420
N quantization of 423, 427
sampled 423
noise stationary 421
electronic 138 signal-to-noise ratio 145, 306
thermal 138 synthesis
white 422 cascade 175
Darlington analytic 181
P
T
processing
analog 137, 416 theorem
digital 422, 436 Parseval 419
signal 464, 506 Plancherel 418
Wiener-Kintchine 419, 421
transform
Q bilinear 456
Fourier 418, 420, 423, 443, 444,
quantities 446, 447, 449
influence 11, 61, 518, 519 Gabor 448
interfering 11 Hartley 444
modifying 11 Wigner-Ville 456, 457, 459