SIGNAL PROCESSING & SIMULATION
NEWSLETTER
TUTORIAL 1 - Basic concepts
in signal processing
Power and Energy
Energy
of a pulse g(t) is defined as
1
This
expression is very similar to expression 1 except that the resistance term is
missing. By Parseval’s Therorem, the
Energy is defined in the frequency domain as
2
where
G(f) is the Fourier transform of pulse g(t).
A signal
is composed of a finite number of pulses, each of which these pulse have
well-defined energy. However for signals, we generally speak in terms of power
and not energy. Power is a more
meaningful term since it can actually be measured. When we speak of Power, we
may be talking about the following two things
1.
Avg
power of the signal and
2.
Power
Spectrum
Power
is defined as energy over a specific time interval and is given by
3
By
evaluation of above integral we note, that Power is not a function of time.
(-true only for wide-sense stationery processes.)
4
Example 1: Avg Power in a
sine wave
For
a sine wave of amplitude 1.0, its average amplitude is .707 (its RMS value). So
the power of a sine wave is .5, as calculated below.
5
Example 2: Power in a pulse
A
rectangular pulse of amplitude A, power
in the signal is A2.
Note
that Power is interpreted as the Energy in one pulse of a periodic signal. Also
the above definition of Power must be properly normalized for the impedance
value of the signal to convert it to
watts. This is one reason we typically use 1 ohm as our impedance. For an
impedance other than 1 ohm, the power numbers calculated need to be adjusted.
For
random processes, we define the power slightly differently as
6
which
is simply the mean squared value of the instantaneous amplitude. At any point
in the signal, the power of the signal is simply equal to its amplitude
squared.
For
a stationery process, the average power is however not a function of time and
is simply the second moment or the variance of the signal x(t).
7
A
signal with a large variance has large power. This is intuitively obvious. We
use the above equation to model white noise processes of a given power value.
Noting
the equivalence between variance and auto-correlation of a function, we
introduce this relationship
8
What
is auto-correlation of a signal? It is
a series of values obtained by the summing the individual values of the signal
multiplied by a shifted version of itself. At zero shift, the time domain
values coincide so we get the maximum sum. We say that that the signal x(t) and
its zero shifted version x(t-0) are perfectly correlated. Depending on the
signal, any other value of the autocorrelation is likely to have a lower value.
The
process of doing auto-correlation is exactly the same as computing the variance
of the signal. The variance in fact is defined as the maximum value of the
auto-correlation function and earlier we said that the variance of a signal is
the measure of the power in the signal.
Expression
. It states that the average power in a signal is equal to not only the
variance but also the value of its auto-correlation function at time 0 which in
turn are same.
Power or Energy per Bit
Often
we talk of Energy per bit (Eb), when talking about signals. This is
the average energy per bit of information transmitted. We define it as,
9
where
Rb is the bit rate.
Example 1: Eb of
a sine wave
For
a sine wave of amplitude 1, the average amplitude is .707. The energy per bit is
calculated by
10
Example 2: Eb of
a rectangular wave
For
a rectangular pulse of amplitude 1, the average amplitude is 1. The energy per
bit is 1/Rb.
Power Spectrum
and Power Spectral Density
Another
way to look at the Power is to look its spectrum. The spectrum of a signal
shows how much power is contained in each of its harmonic or spectral
components or the frequency spectrum of the signal. A sine wave PSD has only a
delta function at the carrier frequency since the signal contains just one
spectral component namely the carrier frequency. A random signal on the other
hand has a rich PSD with power occurring in many different components. Power
spectrum of white noise is flat to show that power occurs in all frequencies.
A
plot of the frequency components on the x-axis and attendant Power in that
frequency on the y-axis is called the Power Spectrum of the signal. Its units
are watts per Hertz. The Power Spectrum
is also referred to as the Power Spectral Density. The two terms refer to the
same thing. The Power spectrum does not directly give us the total or average
power in the signal, only power in a particular spectral component. To obtain
the total power in the signal or in a particular range, we must integrate the
Power Spectrum over the range of frequencies of interest and include both
negative and positive frequencies.
We
can define total power now as the integral of the Power Spectral Density.
11
and
12
where
Sx is called the two-sided spectral density. We multiplied it by 2, because our
integration limits were only the positive frequencies.
Properties of PSD
and Power Spectrum
1.
The
total area under the Power Spectrum or PSD is equal to the total avg. power of
the signal.
2.
The
PSD is always positive.
3.
The
PSD is an even function of frequency or in other words, it is symmetrical.
4.
The
auto-correlation function and PSD are a Fourier transform pair. To compute PSD,
we compute the auto-correlation of the signal and then take its FFT. (Another
estimation method called “periodogram” uses sampled FFT to compute the PSD.)
5.
The
value of the auto-correlation function at zero-time equals the total power in
the signal.
Keep
in mind that total or the average power in a signal is often not of as great an
interest. We are most often interested in the PSD or the Power Spectrum. We
often want to see is how the input power has been redistributed by the channel and in this frequency-based
redistribution of power is where most of the interesting information lies.
Example 1 - Computing PSD of
a Sine wave with random phase
Lets
take a random sine wave
15
where
is the random phase
from 0 to 360 degrees.
We
compute the auto-correlation of this wave by the following relationship
16
Now
taking Fourier Transform of above, we get the PSD by the following relationship
as shown below
17
Sx(f)
-fc 0 fc
The
PSD of the sine wave shows that power is concentrated at the carrier frequency
and that the total power is the sum of the powers in both the negative and
positive terms.
Now
we can compute the total power from both of these relationships. From the
auto-correlation, we see that by setting t = 0, we get
18
and
by integrating the PSD (a trivial case here: 2 x A2/4) , we also get
the same quantity. But unlike the signal power computation, the PSD also tells us
that the power in this signal lies only at one frequency and is equal to twice
the magnitude of the two-sided PSD. When interpreting PSD, it is important to
note whether the given result is one or two-sided or the results will be off by
3 dB.
Most
simulation programs do not use two-sided PSD but an alternate representation
called the complex envelope. The complex envelope shifts the all energy in the
positive axis and the PSD of a complex envelope signal is always one-sided and
contains only the positive frequencies.
Example 2 - PSD of a Random
binary wave
Now
lets take a random binary wave where the 0 and 1’s are defined by x(t) and
-x(t) of amplitude A = 2 as shown in Figure above. The auto-correlation
function of this wave is given by
19
Note
that for an amplitude of 2, we should get the signal power as 4. In figure
above, we see that the value of the auto-correlation function at its peak is
exactly this number.
Now
we compute the Power Spectral Density by taking the Fourier Transform of above.
We get
20
The
examination of the PSD tells us something interesting. It tells us that the maximum
power occurs in the spectral component at zero and the pattern of power
drop-off follows the sinc function, a well-known result.
Now
let’s repeat the above calculations for a sine wave.
The auto-correlation function of a
sine wave is also periodic. Above we calculated the power of a sine wave of
amplitude 1 = .5. From Figure below we see that the peak value of the
auto-correlation function is indeed .5. The Power spectrum consists of a single
spectral component at the wave frequency of 10 and is of magnitude 24.56 dBm
(the dBm equivalent of the .707 rms amplitude).
But
since a sine wave can collerate to itself every one perid, its auot-correlation
is periodic, however, each of the subsequent perfect correlation points are
fewer in number so the total value of the auto-correlation, that is the sum,
continues to decrease.
Periodogram
Periodogram
is a computationally economically way of estimating the Power Spectrum.
Mathematically we see that one needs to compute the auto-correlation of the whole
sequence in order to get an accurate PSD. But for large sequences, this takes
too long and a averaged PSD is computed
instead. This averaged PSD is referred to as the Periodogram.
Periordogram
method of computing the power spectrum also makes sense when the signal FFT is
very noisy and a desired signal level can not be. In such cases, the inherent
averaging of the Periodogram can help extract the signal.
We
will discuss this method of assessing power in signals in much more detail
later.
Exercise:
Compute the auto-correlation function and the Power Spectrum of a sine wave
with added noise.
Copyright 2000 Charan Langton
All Rights Reserved
mntcastle@earthlink.net