Phasor (sine waves)

From Wikipedia, the free encyclopedia

In physics and engineering, a phasor is a representation of a sine wave whose amplitude (A), phase (θ), and frequency (ω) are time-invariant. It is a subset of a more general concept called analytic representation. Phasors reduce the dependencies on these parameters to three independent factors, thereby simplifying certain kinds of calculations. In particular, the frequency factor, which also includes the time-dependence of the sine wave, is often common to all the components of a linear combination of sine waves. Using phasors, it can be factored out, leaving just the static amplitude and phase information to be combined algebraically (rather than trigonometrically). Similarly, linear differential equations can be reduced to algebraic ones. The term phasor therefore often refers to just those two factors. In older texts, a phasor is also referred to as a sinor.

Contents

[edit] Definition

Euler's formula indicates that sine waves can be represented mathematically as the sum of two complex-valued functions:

A\cdot \cos(\omega t + \theta) = A/2\cdot e^{j(\omega t + \theta)} + A/2\cdot e^{-j(\omega t + \theta)},    [1]

or as the real part of one of the functions:


\begin{align}
A\cdot \cos(\omega t + \theta) &= \operatorname{Re} \left\{ A\cdot e^{j(\omega t + \theta)}\right\} \\
&= \operatorname{Re} \left\{ A e^{j\theta} \cdot e^{j\omega t}\right\}.
\end{align}

As indicated above, phasor can refer to either  A e^{j\theta} e^{j\omega t}\, or just the complex constant,  A e^{j\theta}\,  . In the latter case, it is understood to be a shorthand notation, encoding the amplitude and phase of an underlying sinusoid.

An even more compact shorthand is angle notation:  A \angle \theta.\,

[edit] Phasor arithmetic

[edit] Multiplication by a constant (scalar)

Multiplication of the phasor Ae j θ by a complex constant, Be j φ,  produces another phasor. That means its only effect is to change the amplitude and phase of the underlying sinusoid:


\begin{align}
\operatorname{Re}\{(A e^{j\theta} \cdot B e^{j\phi})\cdot e^{j\omega t} \}
&= \operatorname{Re}\{(AB e^{j(\theta+\phi)})\cdot e^{j\omega t} \} \\
&= AB \cos(\omega t +(\theta+\phi))
\end{align}

Although Be j φ has the form of the shorthand notation for a phasor, it is not a phasor. In electronics, it is an impedance, and the phase shift is actually caused by the time-delay associated with a reactive circuit element. The product of two phasors (or squaring a phasor) would represent the product of two sine waves, which is a non-linear operation and does not produce another phasor.

[edit] Differentiation and integration

The time derivative or integral of a phasor produces another phasor[2]. For example:


\begin{align}
\operatorname{Re}\left\{\frac{d}{dt}(A e^{j\theta} \cdot e^{j\omega t})\right\}
&= \operatorname{Re}\{A e^{j\theta} \cdot j\omega e^{j\omega t}\} \\
&= \operatorname{Re}\{A e^{j\theta} \cdot e^{j\pi/2} \omega e^{j\omega t}\} \\
&= \operatorname{Re}\{\omega A e^{j(\theta + \pi/2)} \cdot e^{j\omega t}\} \\
&= \omega A\cdot \cos(\omega t + \theta + \pi/2)
\end{align}

Therefore, in phasor representation, the time derivative of a sinusoid becomes just multiplication by the constant, j \omega = (e^{j\pi/2} \cdot \omega).\,  Similarly, integrating a phasor corresponds to multiplication by \frac{1}{j\omega} = \frac{e^{-j\pi/2}}{\omega}.\,  The time-dependent factor,  e^{j\omega t}\,,  is unaffected. When we solve a linear differential equation with phasor arithmetic, we are merely factoring  e^{j\omega t}\,  out of all terms of the equation, and reinserting it into the answer. For example, consider the following differential equation for the voltage across the capacitor in an RC circuit:

\frac{d\ v_C(t)}{dt} + \frac{1}{RC}v_C(t) = \frac{1}{RC}v_S(t)

When the voltage source in this circuit is sinusoidal:

v_S(t) = V_P\cdot \cos(\omega t + \theta),\,

we may substitute:


\begin{align}
v_S(t) &= \operatorname{Re} \{V_s \cdot e^{j\omega t}\} \\
\end{align}
v_C(t) = \operatorname{Re} \{V_c \cdot e^{j\omega t}\},

where phasor  V_s = V_P e^{j\theta},\,  and phasor V_c\, is the unknown quantity to be determined.

In the phasor shorthand notation, the differential equation reduces to[3]:

j \omega V_c + \frac{1}{RC} V_c = \frac{1}{RC}V_s

Solving for the phasor capacitor voltage gives:


V_c = \frac{1}{1 + j \omega RC} \cdot (V_s) = \frac{1-j\omega R C}{1+(\omega R C)^2} \cdot (V_P e^{j\theta})\,

As we have see, the complex constant factor represents differences of the amplitude and phase of v_C(t)\,  relative to V_P\,  and \theta.\,

In polar coordinate form, the factor is:

\frac{1}{\sqrt{1 + (\omega RC)^2}}\cdot e^{-j \phi(\omega)},\,    where  \phi(\omega) = \arctan(\omega RC).\,

Therefore:

v_C(t) = \frac{1}{\sqrt{1 + (\omega RC)^2}}\cdot V_P \cos(\omega t + \theta- \phi(\omega))

[edit] Addition

The sum of multiple phasors produces another phasor. That is because the sum of sine waves of one frequency is also a sine wave:


\begin{align}
A_1 \cos(\omega t + \theta_1) + A_2 \cos(\omega t + \theta_2)
&= \operatorname{Re} \{A_1 e^{j\theta_1}e^{j\omega t}\} + \operatorname{Re} \{A_2 e^{j\theta_2}e^{j\omega t}\} \\
&= \operatorname{Re} \{A_1 e^{j\theta_1}e^{j\omega t} + A_2 e^{j\theta_2}e^{j\omega t}\} \\
&= \operatorname{Re} \{(A_1 e^{j\theta_1} + A_2 e^{j\theta_2})e^{j\omega t}\} \\
&= \operatorname{Re} \{(A_3 e^{j\theta_3})e^{j\omega t}\} \\
&= A_3 \cos(\omega t + \theta_3),
\end{align}

where:


A_3^2 = (A_1 \cos(\theta_1)+A_2 \cos(\theta_2))^2 + (A_1 \sin(\theta_1)+A_2 \sin(\theta_2))^2

\tan(\theta_3) = \frac{A_1 \sin(\theta_1)+A_2 \sin(\theta_2)}{A_1 \cos(\theta_1)+A_2 \cos(\theta_2)}.

In physics, this sort of addition occurs when sine waves "interfere" with each other, constructively or destructively. Another way to view the calculations above is that two vectors with coordinates [A_1 \cos(\theta_1), A_1 \sin(\theta_1)]\, and [A_2 \cos(\theta_2), A_2 \sin(\theta_2)]\, are added (see vector addition) to produce a resultant vector with coordinates [A_3 \cos(\theta_3), A_3 \sin(\theta_3)].\,

Phasor diagram of three waves in perfect destructive interference
Phasor diagram of three waves in perfect destructive interference

The vector concept provides useful insight into questions like this: "What phase difference would be required between three identical waves for perfect cancellation?" In this case, simply imagine taking three vectors of equal length and placing them head to tail such that the last head matches up with the first tail. Clearly, the shape which satisfies these conditions is an equilateral triangle, and the angle between each phasor to the next is 120° (2π/3 radians), or one third of a wavelength λ / 3.  So the phase difference between each wave must also be 120°.

In other words, what this shows is:

\cos(\omega t) + \cos(\omega t + 2\pi/3) + \cos(\omega t +4\pi/3) = 0.\,

In the example of three waves, the phase difference between the first and the last wave was 240 degrees, while for two waves destructive interference happens at 180 degrees. In the limit of many waves, the phasors must form a circle for destructive interference, so that the first phasor is nearly parallel with the last. This means that for many sources, destructive interference happens when the first and last wave differ by 360 degrees, a full wavelength λ. This is why in single slit diffraction, the minima occurs when light from the far edge travels a full wavelength further than the light from the near edge.


[edit] Phasor diagrams

Electrical engineers, electronics engineers, and electronic engineering technicians use phasor diagrams to visualize complex constants and variables (phasors). Like vectors, arrows drawn on graph paper or computer displays represent phasors. Cartesian and polar representations each have advantages.


Notes

  1. ^
    • j is the Imaginary unit (j2 = − 1).
    • The frequency of the wave, in Hz, is given by ω / 2π.
  2. ^ This results from:  \frac{d}{dt}(e^{j \omega t}) = j \omega e^{j \omega t} which means that the complex exponential is the eigenfunction of the derivative operation.
  3. ^ Proof:

    \frac{d\ \operatorname{Re} \{V_c \cdot e^{j\omega t}\}}{dt} + \frac{1}{RC}\operatorname{Re} \{V_c \cdot e^{j\omega t}\} = \frac{1}{RC}\operatorname{Re} \{V_s \cdot e^{j\omega t}\}

     

     

     (Eq.1)

     

    Since this must hold for all t\,, specifically:  t-\frac{\pi}{2\omega },\,  it follows that:

    \frac{d\ \operatorname{Im} \{V_c \cdot e^{j\omega t}\}}{dt} + \frac{1}{RC}\operatorname{Im} \{V_c \cdot e^{j\omega t}\} = \frac{1}{RC}\operatorname{Im} \{V_s \cdot e^{j\omega t}\}

     

     

     (Eq.2)

     

    It is also readily seen that:

    \frac{d\ \operatorname{Re} \{V_c \cdot e^{j\omega t}\}}{dt} 
= \operatorname{Re} \left\{ \frac{d\left( V_c \cdot e^{j\omega t}\right)}{dt} \right\}
= \operatorname{Re} \left\{ j\omega V_c \cdot e^{j\omega t} \right\}
    \frac{d\ \operatorname{Im} \{V_c \cdot e^{j\omega t}\}}{dt} 
= \operatorname{Im} \left\{ \frac{d\left( V_c \cdot e^{j\omega t}\right)}{dt} \right\}
= \operatorname{Im} \left\{ j\omega V_c \cdot e^{j\omega t} \right\}

    Substituting these into (1) and (2), and multiplying (2) by j\,, and adding both equations gives:

    j\omega V_c \cdot e^{j\omega t} + \frac{1}{RC}V_c \cdot e^{j\omega t} = \frac{1}{RC}V_s \cdot e^{j\omega t}
    \left(j\omega V_c + \frac{1}{RC}V_c = \frac{1}{RC}V_s\right) \cdot e^{j\omega t}
    j\omega V_c + \frac{1}{RC}V_c = \frac{1}{RC}V_s \quad\quad(QED)

[edit] Circuit laws

With phasors, the techniques for solving DC circuits can be applied to solve AC circuits. A list of the basic laws is given below.

  • Ohm's law for resistors: a resistor has no time delays and therefore doesn't change the phase of a signal therefore V=IR remains valid.
  • Ohm's law for resistors, inductors, and capacitors: V=IZ where Z is the complex impedance.
  • In an AC circuit we have real power (P) which is a representation of the average power into the circuit and reactive power (Q) which indicates power flowing back and forward. We can also define the complex power S=P+jQ and the apparent power which is the magnitude of S. The power law for an AC circuit expressed in phasors is then S=VI* (where I* is the complex conjugate of I).
  • Kirchhoff's circuit laws work with phasors in complex form

Given this we can apply the techniques of analysis of resistive circuits with phasors to analyze single frequency AC circuits containing resistors, capacitors, and inductors. Multiple frequency linear AC circuits and AC circuits with different waveforms can be analyzed to find voltages and currents by transforming all waveforms to sine wave components with magnitude and phase then analyzing each frequency separately.

[edit] Power engineering

In analysis of three phase AC power systems, usually a set of phasors is defined as the three complex cube roots of unity, graphically represented as unit magnitudes at angles of 0, 120 and 240 degrees. By treating polyphase AC circuit quantities as phasors, balanced circuits can be simplified and unbalanced circuits can be treated as an algebraic combination of symmetrical circuits. This approach greatly simplifies the work required in electrical calculations of voltage drop, power flow, and short-circuit currents. In the context of power systems analysis, the phase angle is often given in degrees, and the magnitude in rms value rather than the peak amplitude of the sinusoid.

[edit] References

  • Douglas C. Giancoli (1989). Physics for Scientists and Engineers. Prentice Hall. ISBN 0-13-666322-2. 

[edit] External links