University Distinguished Professor
Jack Dongarra received his Ph.D. in Applied Mathematics from the University of New Mexico in 1980. He is a University Distinguished Professor at UTK, Distinguished Research Staff at ORNL, Turing Fellow at Manchester University, Adjunct Professor at Rice University, and director of the ICL at UTK. He specializes in numerical algorithms in linear algebra, parallel computing, use of advanced-computer architectures, programming methodology, and tools for parallel computers. His research includes the development, testing and documentation of high-quality mathematical software. He has contributed to the following: EISPACK, LINPACK, the BLAS, LAPACK, ScaLAPACK, Netlib, PVM, MPI, NetSolve, Top500, ATLAS, and PAPI.
The objective of the ICPP at UTK is the development and optimization of numerical linear algebra libraries and technologies for applications, while tackling current challenges in heterogeneous Intel® Xeon Phi™ coprocessor-based High Performance Computing. The developments will be disseminated through the MAGMA MIC library, designed as a replacement of the popular LAPACK for heterogeneous systems with Intel Xeon Phi Coprocessors.
Over the first year we developed the main dense linear algebra routines to solve dense linear systems and eigenvalue problems on heterogeneous Intel Xeon Phi coprocessor-based platforms. The developments were disseminated through two major software releases. Further, we developed benchmarks and two APIs, and evaluated programming models for the Intel Xeon Phi coprocessor architectures. We taught a graduate Scientific Computing for Engineers class with the use of Intel Xeon Phi coprocessors, organized tutorials and gave presentations at HPC conferences like SC13, IPDPS14, ISC14, and VECPAR14.
Solving linear systems of equations and eigenvalue problems is fundamental to scientific computing. Our developments are thus likely to enable Intel Xeon Phi coprocessor architectures for high-performance computing by providing an effortless migration of LAPACK-relying existing scientific and engineering codes to coprocessor-accelerated architectures. The developments will further help to explore the full potential of the Intel Xeon Phi coprocessor architecture and related programming models on the scientific computing community.
Our plan for the next year includes the development of new new algorithms and software tools in four main research and software development trusts:
- Dense linear algebra
Algorithmic improvements and new methods will be developed, e.g., in the area of eigensolvers and SVD, we will develop two-stage reductions to tridiagonal and bidiogonal forms. These algorithms remove the memory-bound limitations of the LAPACK algorithms, and depending on hardware can be an order of magnitude faster. Another direction will be the development of batched linear algebra operations to provide support for various applications. Batched LU, QR, and Cholesky will be developed for the simultaneous factorization of many very small dense matrices. This will include the development of batched BLAS as needed in the solvers and basic applications.
- Sparse linear algebra (SLA)
While extremely important for applications, SLA is notorious for running only at a fraction of the peak of modern architectures. We will first develop a highly optimized MAGMA MIC Sparse package, including the standard CG, BiCGSTAB, GMRES, and preconditioned versions. Second, we will develop communication-avoiding algorithms that significantly exceed in performance the standard memory and latency bound algorithms. This will include new s-step methods like the CA-GMRES, and blocked eigensolvers like the LOBPCG.
- Mixed-precision methods
We will develop numerical algorithms that recognize and exploit the presence of mixed-precision mathematics. This will include mixed-precision iterative refinement solvers for dense problems and mixed-precision orthogonalization schemes with applications to sparse iterative linear system and eigenproblem solvers.
We will develop a set of benchmarks, including the newly proposed HPCG, and optimize them for Intel Xeon Phi coprocessor architectures. The benchmarks will show essential communication and computation patterns in various applications, with the goal to encourage the focus of both hardware and software developers on architecture features and application needs.
- Asim YarKhan, Jakub Kurzak, Piotr Luszzczek, Jack Dongarra, 5/31/2016, Porting the PLASMA Numerical Library to the OpenMP Standard, netlib
- Asim YarKhan, Jakub Kurzak, Piotr Luszzczek, Jack Dongarra, 8/30/2016, Accelerating Tensor Contractions for High-Order FEM on CPUs, GPUs, and KNLs, utk.edu
- Azzam Haidar, Ahmad Abdelfattah, Veselin Dobrev, Ian Karlin, Tzanio Kolev, Stanimire Tomov, and Jack Dongarra, 11/12/2017, Introduction - 8th Workshop on Latest Advances in Scalable Algorithms for Large-Scale Systems, SC17
- Jakub Kurzak, Jack Dongarra, Michael Heroux, James Demmel, 11/12/2017, Linear Algebra Libraries for High-Performance Computing: Scientific Computing with Multicore and Accelerators, SC17
- Khairul Kabir, Azzam Haidar, Stanimire Tomov, and Jack Dongarra, 7/1/2015, On the Design, Development, and Analysis of Optimized Matrix-Vector Multiplication Routines for Coprocessors, UTK
- Khairul Kabir, Azzam Haidar, Stanimire Tomov, and Jack Dongarra, 6/1/2015, Performance Analysis and Optimisation of Two-Sided Factorization Algorithms for Heterogeneous Platform, UTK
- Jack Dongarra, Mark Gates, Azzam Haidar, Yulu Jia, Khairul Kabir, Piotr Luszczek and Stanimire Tomov, 9/16/2014, HPC programming on Intel Many-Integrated-Core hardware with MAGMA Port to Intel Xeon Phi, UTK
- Khairul Kabir, Azzam Haidar, Stanimire Tomov, and Jack Dongarra, 4/1/2015, Performance Analysis and Design of a Hessenberg Reduction using Stabilized Blocked Elementary Transformations for New Architectures, UTK
- Azzam Haidar, Konstantin Arturov, Murat Guney, Shane Story, Jack Dongarra, 1/1/2016, LU, QR, and Cholesky factorizations: Programming Model, Performance Analysis and Optimization Techniques for the Intel Xeon Phi Processor, icl.utk.edu
- Azzam Haidar, Piotr Luszczek, Stanimire Tomov, and Jack Dongarra, 6/1/2014, Heterogeneous Acceleration for Linear Algebra in Multi-Coprocessor Environments, UTK
- Piotr Luszczek, 11/19/2014, MAGMA MIC: HPC Linear Algebra for the Intel Xeon Phi Processor IXPUG