2217154
Project Grant
Overview
Grant Description
Collaborative Research: PPOSS: Large: A Comprehensive Framework for Efficient, Scalable, and Performance-Portable Tensor Applications
Computations on tensors are fundamental to many large-scale parallel software applications in scientific computing and machine learning, and their efficient implementation has been crucial for the significant advances they have enabled. However, with the end of Moore's Law, two critical challenges now threaten continued progress: (1) with transistors becoming a bounded resource, hardware customization is critical to sustaining improved performance and energy efficiency, requiring advances in algorithm-architecture co-design methodology; (2) increasing customization and heterogeneity of hardware architectures aggravates the already daunting challenges of application-developer productivity and performance-portability of software.
This project brings together researchers with expertise spanning the algorithm/software/hardware stack to address these challenges. The project's impacts include (1) improved performance and energy efficiency of hardware architectures through algorithm-architecture co-design; (2) increased developer productivity for software applications and the performance achieved on a variety of target platforms, which enhances the benefits of computing technology in science and industry; (3) advances in scalable machine-learning and scientific computing applications.
The project makes contributions along multiple directions:
1. Compiler Optimization: Powerful unified methodology for automated optimization of dense tensor computations, based on non-linear cost models for multi-level hyper-rectangular tiled execution on a range of target computing platforms.
2. Scalability with Sparsity: Multi-level blocking methodology to enhance scalability of sparse-tensor computations, based on analysis of the intrinsic sparsity patterns of the data and the corresponding data-reuse patterns.
3. Algorithm-Architecture Co-Design: By leveraging new cost models, development of powerful and general new approaches for hardware-software co-design of accelerators for dense- and sparse-tensor computations.
4. Correctness and Accuracy: Development of techniques to ensure correctness and floating-point accuracy with compiler transformations and compiler/hardware design-space exploration.
5. Applications: Use of the developed methodology and tools to advance cutting-edge applications in machine learning and scientific computing, including PDE solvers, quantum many-body simulation, tensor networks in machine learning, and large-scale image analysis.
This award reflects NSF's statutory mission and has been deemed worthy of support through evaluation using the foundation's intellectual merit and broader impacts review criteria.
Computations on tensors are fundamental to many large-scale parallel software applications in scientific computing and machine learning, and their efficient implementation has been crucial for the significant advances they have enabled. However, with the end of Moore's Law, two critical challenges now threaten continued progress: (1) with transistors becoming a bounded resource, hardware customization is critical to sustaining improved performance and energy efficiency, requiring advances in algorithm-architecture co-design methodology; (2) increasing customization and heterogeneity of hardware architectures aggravates the already daunting challenges of application-developer productivity and performance-portability of software.
This project brings together researchers with expertise spanning the algorithm/software/hardware stack to address these challenges. The project's impacts include (1) improved performance and energy efficiency of hardware architectures through algorithm-architecture co-design; (2) increased developer productivity for software applications and the performance achieved on a variety of target platforms, which enhances the benefits of computing technology in science and industry; (3) advances in scalable machine-learning and scientific computing applications.
The project makes contributions along multiple directions:
1. Compiler Optimization: Powerful unified methodology for automated optimization of dense tensor computations, based on non-linear cost models for multi-level hyper-rectangular tiled execution on a range of target computing platforms.
2. Scalability with Sparsity: Multi-level blocking methodology to enhance scalability of sparse-tensor computations, based on analysis of the intrinsic sparsity patterns of the data and the corresponding data-reuse patterns.
3. Algorithm-Architecture Co-Design: By leveraging new cost models, development of powerful and general new approaches for hardware-software co-design of accelerators for dense- and sparse-tensor computations.
4. Correctness and Accuracy: Development of techniques to ensure correctness and floating-point accuracy with compiler transformations and compiler/hardware design-space exploration.
5. Applications: Use of the developed methodology and tools to advance cutting-edge applications in machine learning and scientific computing, including PDE solvers, quantum many-body simulation, tensor networks in machine learning, and large-scale image analysis.
This award reflects NSF's statutory mission and has been deemed worthy of support through evaluation using the foundation's intellectual merit and broader impacts review criteria.
Awardee
Grant Program (CFDA)
Awarding Agency
Funding Agency
Place of Performance
Salt Lake City,
Utah
84112-9249
United States
Geographic Scope
Single Zip Code
Related Opportunity
None
University Of Utah was awarded
Efficient Tensor Applications: PPOSS Framework
Project Grant 2217154
worth $3,649,636
from the NSF Office of Advanced Cyberinfrastructure in July 2022 with work to be completed primarily in Salt Lake City Utah United States.
The grant
has a duration of 5 years and
was awarded through assistance program 47.070 Computer and Information Science and Engineering.
Status
(Ongoing)
Last Modified 7/6/22
Period of Performance
7/1/22
Start Date
6/30/27
End Date
Funding Split
$3.6M
Federal Obligation
$0.0
Non-Federal Obligation
$3.6M
Total Obligated
Activity Timeline
Additional Detail
Award ID FAIN
2217154
SAI Number
None
Award ID URI
SAI EXEMPT
Awardee Classifications
Public/State Controlled Institution Of Higher Education
Awarding Office
490501 DIV OF COMPUTER COMM FOUNDATIONS
Funding Office
490509 OFC OF ADV CYBERINFRASTRUCTURE
Awardee UEI
LL8GLEVH6MG3
Awardee CAGE
3T624
Performance District
02
Senators
Mike Lee
Mitt Romney
Mitt Romney
Representative
Chris Stewart
Budget Funding
| Federal Account | Budget Subfunction | Object Class | Total | Percentage |
|---|---|---|---|---|
| Research and Related Activities, National Science Foundation (049-0100) | General science and basic research | Grants, subsidies, and contributions (41.0) | $3,649,636 | 100% |
Modified: 7/6/22