<?xml version="1.0" encoding="UTF-8"?><rss version="2.0"
	xmlns:content="http://purl.org/rss/1.0/modules/content/"
	xmlns:wfw="http://wellformedweb.org/CommentAPI/"
	xmlns:dc="http://purl.org/dc/elements/1.1/"
	xmlns:atom="http://www.w3.org/2005/Atom"
	xmlns:sy="http://purl.org/rss/1.0/modules/syndication/"
	xmlns:slash="http://purl.org/rss/1.0/modules/slash/"
	
	xmlns:georss="http://www.georss.org/georss"
	xmlns:geo="http://www.w3.org/2003/01/geo/wgs84_pos#"
	>

<channel>
	<title>Multicore FFT Archives - CenterSpace</title>
	<atom:link href="https://www.centerspace.net/tag/multicore-fft/feed" rel="self" type="application/rss+xml" />
	<link>https://www.centerspace.net/tag/multicore-fft</link>
	<description>.NET numerical class libraries</description>
	<lastBuildDate>Tue, 07 Feb 2023 21:48:41 +0000</lastBuildDate>
	<language>en-US</language>
	<sy:updatePeriod>
	hourly	</sy:updatePeriod>
	<sy:updateFrequency>
	1	</sy:updateFrequency>
	<generator>https://wordpress.org/?v=6.1.1</generator>
<site xmlns="com-wordpress:feed-additions:1">104092929</site>	<item>
		<title>FFT Performance Benchmarks in .NET</title>
		<link>https://www.centerspace.net/fft-performance-benchmarks-in-net</link>
					<comments>https://www.centerspace.net/fft-performance-benchmarks-in-net#respond</comments>
		
		<dc:creator><![CDATA[Paul Shirkey]]></dc:creator>
		<pubDate>Wed, 05 Jan 2011 20:07:46 +0000</pubDate>
				<category><![CDATA[.NET]]></category>
		<category><![CDATA[NMath]]></category>
		<category><![CDATA[FFT]]></category>
		<category><![CDATA[FFT .NET benchmarks]]></category>
		<category><![CDATA[FFT benchmarks]]></category>
		<category><![CDATA[fft C#]]></category>
		<category><![CDATA[FFT in .NET]]></category>
		<category><![CDATA[Multicore FFT]]></category>
		<category><![CDATA[NMATH FFT and FFTW]]></category>
		<category><![CDATA[Non power of 2 FFT]]></category>
		<guid isPermaLink="false">http://www.centerspace.net/blog/?p=2942</guid>

					<description><![CDATA[<p>We've had a number of inquires about the CenterSpace FFT benchmarks, so I thought I would code up a few tests and run them on my machine.  I've included our FFT performance numbers and the code that generated those numbers so you can try them on your machine.  (If you don't have NMath, you'll need to download the <a href="https://www.centerspace.net/downloads/trial-versions/">eval version</a>).  I also did a head-to-head comparison with FFTW, one of the fastest desktop FFT implementations.</p>
<p>The post <a rel="nofollow" href="https://www.centerspace.net/fft-performance-benchmarks-in-net">FFT Performance Benchmarks in .NET</a> appeared first on <a rel="nofollow" href="https://www.centerspace.net">CenterSpace</a>.</p>
]]></description>
										<content:encoded><![CDATA[<p>We&#8217;ve had a number of inquires about the CenterSpace FFT benchmarks, so I thought I would code up a few tests and run them on my machine.  I&#8217;ve included our FFT performance numbers and the code that generated those numbers so you can try them on your machine.  (If you don&#8217;t have NMath, you&#8217;ll need to download the <a href="/trial-version/">eval version</a>).  I also did a comparison of 1 dimensional real DFTs, with FFTW, one of the fastest desktop FFT implementations available.</p>
<h3> Benchmarks </h3>
<p>These benchmarks were run on a 2.80 Ghz, Intel Core i7 CPU, with 4Gb of memory installed. </p>
<pre class="code">
The clock resolution is 0.003 ns
1024 point, forward, real FFT required 4361.364 ns, Mflops 4069
1000 point, forward, real FFT required 5338.785 ns, Mflops 3235
4096 point, forward, real FFT required 21708.565 ns, Mflops 3924
4095 point, forward, real FFT required 43012.010 ns, Mflops 1980
1024 * 1024 point, forward, real FFT required 15.635 ms, Mflops 2324
</pre>
<p>I&#8217;m estimating the megaflop performance during the FFT using:<br />
<center><br />
<img decoding="async" src="http://latex.codecogs.com/gif.latex?MFlops \approx {2.5*n \ ln (n)) \over{ \textit{time in} \ \mu s} }" title="MFlops \approx {2.5*n \ ln (n)) \over{ \textit{time in} \ \mu s} }" /><br />
</center></p>
<p>This is the asymptotic number of floating point operations for the radix-2 Cooley-Tukey FFT algorithm. This FFT MFlop estimate is used in a number of FFT benchmark reports and serves as a good basis for comparing algorithm efficiency.</p>
<p>As expected we take a performance hit for non-power of 2 lengths, but due to various optimizations for processing prime length FFT kernels (3, 5, 7 &#038; 11), the performance hit is minimal in many cases. The 1000-point FFT has prime factors <code>(2)(2)(2)(5)(5)(5)</code>, and the 4095-point FFT has prime factors <code>(3)(3)(5)(7)(13)</code>, so those larger prime factors in the 4095-point FFT cost us some performance.  Typically, user&#8217;s zero pad their data vectors to a power-of-two length to get optimal performance.</p>
<h3> Side by side comparison with FFTW </h3>
<p>FFTW claims to be the &#8220;Fastest Fourier Transform in the West&#8221;, and is a clever, high performance implementation of the discrete Fourier transform.  This algorithm is shipped with all copies of MATLAB.  FFTW is implemented in C and has the reputation as being one of the fastest desktop FFT algorithm.  </p>
<p>Both the NMath FFT and the FFTW have a pre-computation setup that establishes the best algorithmic approach for the DFT at hand, before computing any FFT&#8217;s.  This pre-computational phase is not included in the times below.   In the case of the NMath FFT classes, this pre-computational phase in done in the class constructor; Therefore users must avoid constructing NMath FFT classes in tight loops for best performance (as shown in the benchmark code below).  Below is a small side-by-side comparison between FFTW and NMath&#8217;s FFT (using the numbers from above).</p>
<pre class="code">
<table><tbody>
<tr>
<th colspan="3"> Comparison of a forward, real, out-of-place FFT. </th>
</tr>
<tr> 
<th> FFT length</th> <th> FFTW </th> <th> NMATH FFT </th>
</tr>
<tr> 
<td> 1024 </td> <td> 4.14 &mu;s</td> <td> 4.36 &mu;s </td> 
</tr>
<tr> 
<td> 1000</td> <td> 5.98 &mu;s </td> <td> 5.33 &mu;s </td> 
</tr>
<tr> 
<td> 4096</td> <td> 20.31 &mu;s </td> <td> 21.71 &mu;s </td> 
</tr>
<tr> 
<td> 4095</td> <td> 49.90 &mu;s </td> <td> 43.01 &mu;s </td> 
</tr>
<tr> 
<td> 1024^2 </td> <td> 17.16 ms </td> <td> 15.63 ms </td> 
</tr>
</tbody>
</table>
</pre>
<p>Clearly NMATH is very competitive with, and at times out-performs FFTW for real FFT&#8217;s of both power-of-2 length signals and otherwise.  I chose 1D real signals as a test case because this is one of the most frequent use cases of our NMATH FFT library. </p>
<p>On a subjective scale, running a 1024-point FFT on a desktop commodity machine at around (an algorithm normalized) 4 GFlops is amazing.  That means that in a real time measurement situation, users can compute 1024-point FFT&#8217;s at around 220kHz &#8211; all with just a couple of lines of code.</p>
<p>Happy Computing,<br />
<em> Paul </em></p>
<h3> Benchmark Code </h3>
<pre lang="csharp">
 public void BenchMarks()
    {
      Double numberTrials = 10000;
      Double flops;

      Stopwatch timer = new System.Diagnostics.Stopwatch();
      Console.WriteLine( String.Format("The clock resolution is {0:0.000} ns", Stopwatch.Frequency / 1000000000.0 ) );

      // Snip one - power of two
      RandGenUniform rand = new RandGenUniform();
      DoubleForward1DFFT fft = new DoubleForward1DFFT( 1024 );
      DoubleVector realsignal = new DoubleVector( 1024, rand );

      DoubleVector result = new DoubleVector( 1024 * 1024 );

      timer.Reset();
      for( int i = 0; i < numberTrials; i++ )
      {
        timer.Start();
        fft.FFT( realsignal, ref result );
        timer.Stop();
      }
      flops = (2.5 * 1024 * NMathFunctions.Log(1024)) / (((timer.ElapsedTicks / numberTrials) / Stopwatch.Frequency) * 1000000.0 );
      Console.WriteLine( String.Format( "1024 point, forward, real FFT required {0:0.000} ns, Mflops {1:0}", ( ( timer.ElapsedTicks / numberTrials ) / Stopwatch.Frequency ) * 1000000000.0, flops ) );

      // length 1000
      fft = new DoubleForward1DFFT( 1000 );
      realsignal = new DoubleVector( 1000, rand );

      timer.Reset();
      for( int i = 0; i < numberTrials; i++ )
      {
        timer.Start();
        fft.FFT( realsignal, ref result );
        timer.Stop();
      }
      flops = ( 2.5 * 1000 * NMathFunctions.Log( 1000 ) ) / ( ( ( timer.ElapsedTicks / numberTrials ) / Stopwatch.Frequency ) * 1000000.0 );
      Console.WriteLine( String.Format( "1000 point, forward, real FFT required {0:0.000} ns, Mflops {1:0}", ( ( timer.ElapsedTicks / numberTrials ) / Stopwatch.Frequency ) * 1000000000.0, flops ) );

      // length 4096
      fft = new DoubleForward1DFFT( 4096 );
      realsignal = new DoubleVector( 4096, rand );

      timer.Reset();
      for( int i = 0; i < numberTrials; i++ )
      {
        timer.Start();
        fft.FFT( realsignal, ref result );
        timer.Stop();
      }
      flops = ( 2.5 * 4096 * NMathFunctions.Log( 4096 ) ) / ( ( ( timer.ElapsedTicks / numberTrials ) / Stopwatch.Frequency ) * 1000000.0 );
      Console.WriteLine( String.Format( "4096 point, forward, real FFT required {0:0.000} ns, Mflops {1:0}", ( ( timer.ElapsedTicks / numberTrials ) / Stopwatch.Frequency ) * 1000000000.0, flops ) );

      // length 4095
      fft = new DoubleForward1DFFT( 4095 );
      realsignal = new DoubleVector( 4095, rand );

      timer.Reset();
      for( int i = 0; i < numberTrials; i++ )
      {
        timer.Start();
        fft.FFT( realsignal, ref result );
        timer.Stop();
      }
      flops = ( 2.5 * 4095 * NMathFunctions.Log( 4095 ) ) / ( ( ( timer.ElapsedTicks / numberTrials ) / Stopwatch.Frequency ) * 1000000.0 );
      Console.WriteLine( String.Format( "4095 point, forward, real FFT required {0:0.000} ns, Mflops {1:0}", ( ( timer.ElapsedTicks / numberTrials ) / Stopwatch.Frequency ) * 1000000000.0, flops ) );


      // length 1M
      fft = new DoubleForward1DFFT( 1024 * 1024 );
      realsignal = new DoubleVector( 1024 * 1024, rand );

      timer.Reset();
      for( int i = 0; i < 100; i++ )
      {
        timer.Start();
        fft.FFT( realsignal, ref result );
        timer.Stop();
      }
      flops = ( 2.5 * 1024 * 1024 * NMathFunctions.Log( 1024 * 1024 ) ) / ( ( ( timer.ElapsedTicks / 100.0 ) / Stopwatch.Frequency ) * 1000000.0 );
      Console.WriteLine( String.Format( "Million point (1024 * 1024), forward, real point FFT required {0:0.000} ms, Mflops {1:0}", ( ( timer.ElapsedTicks / 100.0 ) / Stopwatch.Frequency ) * 1000.0, flops ) );

    }
</pre>
<p>The post <a rel="nofollow" href="https://www.centerspace.net/fft-performance-benchmarks-in-net">FFT Performance Benchmarks in .NET</a> appeared first on <a rel="nofollow" href="https://www.centerspace.net">CenterSpace</a>.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://www.centerspace.net/fft-performance-benchmarks-in-net/feed</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
		<post-id xmlns="com-wordpress:feed-additions:1">2942</post-id>	</item>
		<item>
		<title>Modern Fast Fourier Transform</title>
		<link>https://www.centerspace.net/modern-fast-fourier-transform</link>
					<comments>https://www.centerspace.net/modern-fast-fourier-transform#comments</comments>
		
		<dc:creator><![CDATA[Paul Shirkey]]></dc:creator>
		<pubDate>Tue, 29 Sep 2009 05:32:46 +0000</pubDate>
				<category><![CDATA[NMath]]></category>
		<category><![CDATA[Performance]]></category>
		<category><![CDATA[FFT]]></category>
		<category><![CDATA[FFT performance]]></category>
		<category><![CDATA[High performance FFT]]></category>
		<category><![CDATA[Multicore FFT]]></category>
		<guid isPermaLink="false">http://www.centerspace.net/blog/?p=226</guid>

					<description><![CDATA[<p>All variants of the original Cooley-Tukey O(n log n) fast Fourier transform fundamentally exploit different ways to factor the discrete Fourier summation of length N. For example, the split-radix FFT algorithm divides the Fourier summation of length N into three new Fourier summations: one of length N/2 and two of length N/4. The prime factor [&#8230;]</p>
<p>The post <a rel="nofollow" href="https://www.centerspace.net/modern-fast-fourier-transform">Modern Fast Fourier Transform</a> appeared first on <a rel="nofollow" href="https://www.centerspace.net">CenterSpace</a>.</p>
]]></description>
										<content:encoded><![CDATA[<p>All variants of the original Cooley-Tukey O(n log n) fast Fourier transform fundamentally exploit different ways to factor the discrete Fourier summation of length N.</p>
<p><center><br />
<a href="http://www.codecogs.com/eqnedit.php?latex=X_k = \sum_{n=0}^{N-1} x_n e^{(-2 \pi i / N) kn} \ \ \ \ \ k = 0, ... ,N-1" target="_blank" rel="noopener"><img decoding="async" title="X_k = \sum_{n=0}^{N-1} x_n e^{(-2 \pi i / N) kn} \ \ \ \ \ k = 0, ... ,N-1" src="http://latex.codecogs.com/gif.latex?X_k = \sum_{n=0}^{N-1} x_n e^{(-2 \pi i / N) kn} \ \ \ \ \ k = 0, ... ,N-1" alt="" /></a></center><br />
For example, the <em>split-radix FFT</em> algorithm divides the Fourier summation of length N into three new Fourier summations: one of length N/2 and two of length N/4.</p>
<p><center><br />
<a href="http://www.codecogs.com/eqnedit.php?latex=X_{k_N} = X_{k_{N/2}} @plus; X_{k_{N/4}} @plus; X_{k_{N/4}}" target="_blank" rel="noopener"><img decoding="async" title="X_{k_N} = X_{k_{N/2}} + X_{k_{N/4}} + X_{k_{N/4}}" src="http://latex.codecogs.com/gif.latex?X_{k_N} = X_{k_{N/2}} + X_{k_{N/4}} + X_{k_{N/4}}" alt="" /></a></center><br />
The <em>prime factor FFT</em>, divides the Fourier summation of length N, into two (if they exist) summations of length N1 and N2, where N1 and N2 must be relatively prime.</p>
<p><center><br />
<a href="http://www.codecogs.com/eqnedit.php?latex=X_{k_N} = X_{k_{N1}} ( X_{k_{N2}} ) \ \ where \ N1 \perp N2" target="_blank" rel="noopener"><img decoding="async" title="X_{k_N} = X_{k_{N1}} ( X_{k_{N2}} ) \ \ where \ N1 \perp N2" src="http://latex.codecogs.com/gif.latex?X_{k_N} = X_{k_{N1}} ( X_{k_{N2}} ) \ \ where \ N1 \perp N2" alt="" /></a></center><br />
These algorithms are typically applied recursively, and in combination with one another (or with still other factorizations) to maximize performance for a particular N.</p>
<p>In modern implementations there really isn&#8217;t a single static FFT algorithm, but more a dynamic collection of FFT algorithms and tools that are cleverly collated for the Fourier transform type at hand. Major algorithmic changes occur in the underlying implementation as the length and forward domain (real or complex) of the problem vary. Sophisticated FFT implementations insulate the end-user programmer from all of this background machinery.</p>
<h5>DFT length is fundamental to performance</h5>
<p>The days of power-of-2-only FFT algorithms are dead. Users of modern FFT libraries should not need to worry about the large complexities involved in finding the optimal algorithm for the FFT computation at hand; the library should look at the FFT length, problem domain (real or complex), number of machine cores, and machine architecture, and find and compute with the best hybridized FFT algorithm available. However, it is still helpful to understand that your realized performance will depend fundamentally on the various factorization of the length of your FFT. Most know that the best FFT performance will be had when N is a power of 2. If this stringent length requirement cannot be met, then it is best to use a length that be factored into small primes. CenterSpace&#8217;s FFT algorithms contain optimized kernels for prime factor lengths of 2, 3, 5, 7 and 11. The table below demonstrates the FFT performance sensitivity to FFT length.</p>
<table border="0" cellpadding="4">
<caption>Forward real 1D FFT performance at various lengths.</caption>
<tbody>
<tr align="center">
<td><em> DFT Length </em></td>
<td><em>Factors </em></td>
<td><em>MFLOP approximation </em></td>
</tr>
<tr>
<td>512</td>
<td>2 x 2 x 2 x 2 x 2 x 2 x 2 x 2 x 2</td>
<td>5324.5</td>
</tr>
<tr>
<td>511</td>
<td>7 x 73</td>
<td>1327.8</td>
</tr>
<tr>
<td>510</td>
<td>2 x 3 x 5 x 17</td>
<td>3879.4</td>
</tr>
<tr>
<td>509</td>
<td>509 (prime)</td>
<td>1762.4</td>
</tr>
<tr>
<td>508</td>
<td>2 x 2 x 127</td>
<td>2637.6</td>
</tr>
<tr>
<td>507</td>
<td>3 x 13 x 13</td>
<td>2631.5</td>
</tr>
<tr>
<td>506</td>
<td>2 x 11 x 23</td>
<td>3938.3</td>
</tr>
<tr>
<td>505</td>
<td>5 x 101</td>
<td>1122.6</td>
</tr>
<tr>
<td>504</td>
<td>2 x 2 x 2 x 3 x 3 x 7</td>
<td>5227</td>
</tr>
</tbody>
</table>
<p>Clearly the fastest FFT&#8217;s are for lengths that can be factored into small primes (512, 510, 507, 506, 504), and especially small primes that have optimized kernels (512 and 504). The more kernel optimized primes your FFT length contains the faster it will run. This is a universal fact that all FFT implementations confront and holds true for higher dimension FFT&#8217;s as well. <em> Slight changes in length can have a profound impact on FFT performance</em>.</p>
<p>You can factor your FFT length using an online service to assess how your FFT will perform.</p>
<h5>Multi-core Scalability</h5>
<p>The ability to factor a particular FFT into a set independent computations makes it fundamentally suitable for parallelization. All modern desktop and many laptop computers today contain at least two processor cores and any modern math library should be exploiting this fact where possible. CenterSpace&#8217;s complex domain FFT&#8217;s (and related convolutions) are multi-core aware, and automatically expand to fully utilize the available processor cores. Small problems are run on a single core, but once the computational advantages of algorithm parallelization overcome the overhead costs of multi-core parallelization, the computation is spread across all available cores. This automatic parallelization is gained simply by using CenterSpace&#8217;s NMath class libraries. No end-user programming effort is involved.</p>
<table border="0" cellpadding="6">
<caption>Forward complex 1D FFT performance on 1 and 8 cores.</caption>
<tbody>
<tr align="center">
<th><em> FFT Length </em></th>
<th><em> Machine Cores </em></th>
<th><em> Time (seconds) </em></th>
<th><em> MFLOP approximation </em></th>
</tr>
<tr>
<td>2^20</td>
<td>One</td>
<td>56.7</td>
<td>6405.9</td>
</tr>
<tr>
<td>2^20 + 1</td>
<td>One</td>
<td>554.6</td>
<td>655.3</td>
</tr>
<tr>
<td>2^20</td>
<td>Eight</td>
<td>53.3</td>
<td>6813.7</td>
</tr>
<tr>
<td>2^20 + 1</td>
<td>Eight</td>
<td>124.2</td>
<td>2925.3</td>
</tr>
</tbody>
</table>
<p>The power of two FFT&#8217;s are so computationally efficient on modern processors that the gain between one and eight cores is only about 3 seconds on a 2^20-point FFT. However, for the non-power-of-two case we get a 4.5 times speed improvement going from one core to eight. Looked at another way, with multi-core scalability of the FFT, we suffered only a 2X loss in performance going from a 2^20 length FFT to a 2^20+1 length FFT, instead of a 10X loss in performance. In other words, the multi-core scalability of CenterSpace&#8217;s NMath FFT algorithms mitigate the performance loss in using non-power-of-2 lengths, and this simplifies the end-user programmer&#8217;s job.</p>
<p><em> -Paul </em></p>
<p>See our <a href="/topic-fast-fourier-transforms/">FFT landing page </a> for complete documentation and code examples.</p>
<p>The post <a rel="nofollow" href="https://www.centerspace.net/modern-fast-fourier-transform">Modern Fast Fourier Transform</a> appeared first on <a rel="nofollow" href="https://www.centerspace.net">CenterSpace</a>.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://www.centerspace.net/modern-fast-fourier-transform/feed</wfw:commentRss>
			<slash:comments>2</slash:comments>
		
		
		<post-id xmlns="com-wordpress:feed-additions:1">226</post-id>	</item>
	</channel>
</rss>
