2413244
Cooperative Agreement
Overview
Grant Description
Research infrastructure: NSF Mid-Scale RI-2: Open multimodal AI infrastructure to accelerate science.
Language models with billions of possible adjustments and trained on trillions of words are now powering the fastest-growing computing applications in history.
Large language models (LLM) are built using massive amounts of text, usually obtained by pulling data from multiple sources on the internet.
Recent advances enable these models to process other kinds of data, including images, graphs and tables.
Models with these abilities are known as multimodal LLMs.
The best-performing LLMs currently deployed are proprietary, so their parameters, training data, code and documentation are not openly available.
Thus, most artificial intelligence (AI) scientists cannot study, experiment directly with, or improve these state-of-the-art models.
This project – Open, Multimodal Artificial Intelligence (OMAI) – will provide infrastructure in the form of a suite of powerful, well-documented, up-to-date, open models, and open-source interfaces designed for scientific work.
Scientists will be able to access the models, use discipline-specific data and optimize the models.
The project empowers researchers, provides documentation to accelerate research and education, and has an active program in early-career training to advance US economic and scientific competitiveness.
In addition, opportunities provided through partnerships with a range of institutions and programs will enhance training.
The long-term plan is to make the infrastructure available as a low- or zero-cost service to the research community in a manner like open-source code repositories and science-focused digital libraries, to maximize usage.
The OMAI research infrastructure consists of a series of open, multimodal language models kept up to date with recent scientific publications and open-source application programming interfaces that enable scientists to use, expand, and modify those models.
It addresses priorities set forth in the White House AI Action Plan (https://www.whitehouse.gov/wp-content/uploads/2025/07/Americas-AI-Action-Plan.pdf) to accelerate AI-enabled science and ensure the United States is producing the leading open models.
The infrastructure aims to accelerate scientific discovery across disciplines ranging from materials to protein function prediction and weather models.
It will also enable new understanding and improvement of future LLMs while contributing to the development of a well-trained workforce capable of building, customizing, and maintaining such models.
In allowing researchers to fine-tune the models, researchers can optimize performance and understand design decisions that influence training speed and stability, which can impact the short-term and long-term economic costs of LLM development.
The project emphasizes reproducibility, transparency, open data, open and evolving evaluations, multimodality, and scientific use-cases.
It will enable a broad population of scientist-users across all disciplines to use and adapt artificial intelligence models to their own needs and lays the foundation for future research in AI for science.
By supporting work in these novel, critical research areas, OMAI can ultimately benefit both science and the public.
This award reflects NSF's statutory mission and has been deemed worthy of support through evaluation using the foundation's intellectual merit and broader impacts review criteria.
Subawards are planned for this award.
Language models with billions of possible adjustments and trained on trillions of words are now powering the fastest-growing computing applications in history.
Large language models (LLM) are built using massive amounts of text, usually obtained by pulling data from multiple sources on the internet.
Recent advances enable these models to process other kinds of data, including images, graphs and tables.
Models with these abilities are known as multimodal LLMs.
The best-performing LLMs currently deployed are proprietary, so their parameters, training data, code and documentation are not openly available.
Thus, most artificial intelligence (AI) scientists cannot study, experiment directly with, or improve these state-of-the-art models.
This project – Open, Multimodal Artificial Intelligence (OMAI) – will provide infrastructure in the form of a suite of powerful, well-documented, up-to-date, open models, and open-source interfaces designed for scientific work.
Scientists will be able to access the models, use discipline-specific data and optimize the models.
The project empowers researchers, provides documentation to accelerate research and education, and has an active program in early-career training to advance US economic and scientific competitiveness.
In addition, opportunities provided through partnerships with a range of institutions and programs will enhance training.
The long-term plan is to make the infrastructure available as a low- or zero-cost service to the research community in a manner like open-source code repositories and science-focused digital libraries, to maximize usage.
The OMAI research infrastructure consists of a series of open, multimodal language models kept up to date with recent scientific publications and open-source application programming interfaces that enable scientists to use, expand, and modify those models.
It addresses priorities set forth in the White House AI Action Plan (https://www.whitehouse.gov/wp-content/uploads/2025/07/Americas-AI-Action-Plan.pdf) to accelerate AI-enabled science and ensure the United States is producing the leading open models.
The infrastructure aims to accelerate scientific discovery across disciplines ranging from materials to protein function prediction and weather models.
It will also enable new understanding and improvement of future LLMs while contributing to the development of a well-trained workforce capable of building, customizing, and maintaining such models.
In allowing researchers to fine-tune the models, researchers can optimize performance and understand design decisions that influence training speed and stability, which can impact the short-term and long-term economic costs of LLM development.
The project emphasizes reproducibility, transparency, open data, open and evolving evaluations, multimodality, and scientific use-cases.
It will enable a broad population of scientist-users across all disciplines to use and adapt artificial intelligence models to their own needs and lays the foundation for future research in AI for science.
By supporting work in these novel, critical research areas, OMAI can ultimately benefit both science and the public.
This award reflects NSF's statutory mission and has been deemed worthy of support through evaluation using the foundation's intellectual merit and broader impacts review criteria.
Subawards are planned for this award.
Funding Goals
THE GOAL OF THIS FUNDING OPPORTUNITY, "MID-SCALE RESEARCH INFRASTRUCTURE-2", IS IDENTIFIED IN THE LINK: HTTPS://WWW.NSF.GOV/PUBLICATIONS/PUB_SUMM.JSP?ODS_KEY=NSF23570
Grant Program (CFDA)
Awarding / Funding Agency
Place of Performance
Seattle,
Washington
98103-9185
United States
Geographic Scope
Single Zip Code
Related Opportunity
The Allen Institute For Artificial Intelligence was awarded
Open Multimodal AI Infrastructure for Accelerating Science
Cooperative Agreement 2413244
worth $18,113,645
from the Division of Information and Intelligent Systems in August 2025 with work to be completed primarily in Seattle Washington United States.
The grant
has a duration of 5 years and
was awarded through assistance program 47.070 Computer and Information Science and Engineering.
The Cooperative Agreement was awarded through grant opportunity Mid-scale Research Infrastructure-2.
Status
(Ongoing)
Last Modified 8/21/25
Period of Performance
8/15/25
Start Date
7/31/30
End Date
Funding Split
$18.1M
Federal Obligation
$0.0
Non-Federal Obligation
$18.1M
Total Obligated
Activity Timeline
Additional Detail
Award ID FAIN
2413244
SAI Number
None
Award ID URI
SAI EXEMPT
Awardee Classifications
Small Business
Awarding Office
490502 DIV OF INFOR INTELLIGENT SYSTEMS
Funding Office
490502 DIV OF INFOR INTELLIGENT SYSTEMS
Awardee UEI
H2VSQE9MCUU5
Awardee CAGE
9JQD5
Performance District
WA-07
Senators
Maria Cantwell
Patty Murray
Patty Murray
Modified: 8/21/25