Overlap–save method

Overlap–save is the traditional name for an efficient way to evaluate the discrete convolution between a very long signal x[n] and a finite impulse response (FIR) filter h[n]:

y[n] = x[n] * h[n] \ \stackrel{\mathrm{def}}{=} \ \sum_{m=-\infty}^{\infty} h[m] \cdot x[n-m] = \sum_{m=1}^{M} h[m] \cdot x[n-m],\,

 

 

 

 

(Eq.1)

where h[m]=0 for m outside the region [1, M].

A sequence of 4 plots depicts one cycle of the Overlap-save convolution algorithm. The 1st plot is a long sequence of data to be processed with a lowpass FIR filter. The 2nd plot is one segment of the data to be processed in piecewise fashion. The 3rd plot is the filtered segment, with the usable portion colored red. The 4th plot shows the filtered segment appended to the output stream. The FIR filter is a boxcar lowpass with M=16 samples, the length of the segments is L=100 samples and the overlap is 15 samples.

The concept is to compute short segments of y[n] of an arbitrary length L, and concatenate the segments together. Consider a segment that begins at n = kL + M, for any integer k, and define:

x_k[n]  \ \stackrel{\mathrm{def}}{=}
\begin{cases}
x[n+kL] & 1 \le n \le L+M-1\\
0 & \textrm{otherwise}.
\end{cases}
y_k[n] \ \stackrel{\mathrm{def}}{=} \ x_k[n]*h[n]\,

Then, for kL + M    n    kL + L + M  1, and equivalently M    n  kL    L + M  1, we can write:


\begin{align}
y[n] = \sum_{m=1}^{M} h[m] \cdot x_k[n-kL-m]
&= x_k[n-kL] * h[n] \\
&\stackrel{\mathrm{def}}{=} \ y_k[n-kL].
\end{align}

The task is thereby reduced to computing yk[n], for M    n    L + M  1. The process described above is illustrated in the accompanying figure.

Now note that if we periodically extend xk[n] with period N    L + M  1, according to:

x_{k,N}[n] \ \stackrel{\mathrm{def}}{=} \ \sum_{\ell=-\infty}^{\infty} x_k[n - \ell N],

the convolutions  (x_{k,N})*h\,  and  x_k*h\,  are equivalent in the region M    n    L + M  1. So it is sufficient to compute the N-point circular (or cyclic) convolution of x_k[n]\, with h[n]\,  in the region [1, N].  The subregion [M, L + M  1] is appended to the output stream, and the other values are discarded.

The advantage is that the circular convolution can be computed very efficiently as follows, according to the circular convolution theorem:

y_k[n] = \scriptstyle \text{DFT}^{-1} \displaystyle  (\ \scriptstyle \text{DFT} \displaystyle (x_k[n])\cdot \scriptstyle \text{DFT} \displaystyle (h[n])\ ),

where:

Pseudocode

 (Overlap–save algorithm for linear convolution)
 h = FIR_impulse_response
 M = length(h)
 overlap = M-1
 N = 4*overlap    (or a nearby power-of-2)
 step_size = N-overlap
 H = DFT(h, N)
 position = 0
 while position+N <= length(x)
     yt = IDFT( DFT( x(1+position : N+position), N ) * H, N )
     y(1+position : step_size+position) = yt(M : N)    #discard M-1 y-values
     position = position + step_size
 end

Efficiency

When the DFT and its inverse is implemented by the FFT algorithm, the pseudocode above requires about N log2(N) + N complex multiplications for the FFT, product of arrays, and IFFT.[1] Each iteration produces N-M+1 output samples, so the number of complex multiplications per output sample is about:

\frac{N \log_2(N) + N}{N-M+1}.\,

 

 

 

 

(Eq.2)

For example, when M=201 and N=1024, Eq.2 equals 13.67, whereas direct evaluation of Eq.1 would require up to 201 complex multiplications per output sample, the worst case being when both x and h are complex-valued. Also note that for any given M, Eq.2 has a minimum with respect to N. It diverges for both small and large block sizes.

Overlap–discard

Overlap–discard[2] and Overlap–scrap[3] are less commonly used labels for the same method described here. However, these labels are actually better (than overlap–save) to distinguish from overlap–add, because both methods "save", but only one discards. "Save" merely refers to the fact that M  1 input (or output) samples from segment k are needed to process segment k + 1.

Extending overlap–save

The overlap-save algorithm may be extended to include other common operations of a system:[4][5]

Notes

  1. Cooley-Tukey FFT algorithm for N=2k needs (N/2) log2(N) - see FFT - Definition and speed
  2. Harris 1987.
  3. Frerking 1994.
  4. Carlin et al. 1999, p 31, col 20.
  5. Borgerding 2006, pp 158–161.

See also

References

This article is issued from Wikipedia - version of the Thursday, April 02, 2015. The text is available under the Creative Commons Attribution/Share Alike but additional terms may apply for the media files.