SUMMARY
We demonstrate that radically differing implementations of finite element methods (FEMs) are needed on multi‐core (CPU) and many‐core (GPU) architectures, if their respective performance potential is to be realised. Our numerical investigations using a finite element advection–diffusion solver show that increased performance on each architecture can only be achieved by committing to specific and diverse algorithmic choices that cut across the high‐level structure of the implementation. Making these commitments to achieve high performance for a single architecture leads to a loss of performance portability. Data structures that include redundant data but enable coalesced memory accesses are faster on many‐core architectures, whereas redundancy‐free data structures that are accessed indirectly are faster on multi‐core architectures. The Addto algorithm for global assembly is optimal on multi‐core architectures, whereas the Local Matrix Approach is optimal on many‐core architectures despite requiring more computation than the Addto algorithm. These results demonstrate the value in making the correct choice of algorithm and data structure when implementing FEMs, spectral element methods and low‐order discontinuous Galerkin methods on modern high‐performance architectures. Copyright © 2012 John Wiley & Sons, Ltd.