<< Chapter < Page | Chapter >> Page > |
An important consideration in the implementation of any practical numerical algorithm is numerical accuracy: how quickly dofloating-point roundoff errors accumulate in the course of the computation? Fortunately, FFT algorithms for the most part haveremarkably good accuracy characteristics. In particular, for a DFT of length computed by a Cooley-Tukey algorithm with finite-precision floating-point arithmetic, the worst-case error growth is [link] , [link] and the mean error growth for random inputs is only [link] , [link] . This is so good that, in practical applications, a properly implemented FFT will rarely be asignificant contributor to the numerical error.
The amazingly small roundoff errors of FFT algorithms are sometimes explained incorrectly as simply a consequence of the reduced number of operations: since there are fewer operations compared to a naive algorithm, the argument goes, there is less accumulation of roundoff error. The real reason, however, is more subtle than that, and has to do with the ordering of the operations rather than their number. For example, consider the computation of only the output in the radix-2 algorithm of [link] , ignoring all of the other outputs of the FFT. is the sum of all of the inputs, requiring additions. The FFT does not change this requirement, it merely changes the order of the additions so as to re-use some of them for other outputs. In particular, this radix-2 DIT FFT computes as follows: it first sums the even-indexed inputs, then sums the odd-indexed inputs, then adds the two sums; the even- and odd-indexed inputs are summed recursively by the same procedure. This process is sometimes called cascade summation , and even though it still requires total additions to compute by itself, its roundoff error grows much more slowly than simply adding , , and so on in sequence. Specifically, the roundoff error when adding up floating-point numbers in sequence grows as in the worst case, or as on average for random inputs (where the errors grow according to a random walk), but simply reordering these n-1 additions into a cascade summation yields worst-case and average-case error growth [link] .
However, these encouraging error-growth rates only apply if the trigonometric “twiddle” factors in the FFT algorithm are computedvery accurately. Many FFT implementations, including FFTW and common manufacturer-optimized libraries, therefore useprecomputed tables of twiddle factors calculated by means of standard library functions (which compute trigonometric constants to roughlymachine precision). The other common method to compute twiddle factors is to use a trigonometric recurrence formula—this savesmemory (and cache), but almost all recurrences have errors that grow as , , or even [link] , which lead to corresponding errors in the FFT. For example, one simplerecurrence is , multiplying repeatedly by to obtain a sequence of equally spaced angles, but the errors when using this process grow as [link] . A common improved recurrence is , where the small quantity In an FFT, the twiddle factors are powers of , so is a small angle proportional to and is close to 1. is computed using [link] ; unfortunately, the error using this method still grows as [link] , far worse than logarithmic.
Notification Switch
Would you like to follow the 'Fast fourier transforms' conversation and receive update notifications?