NSF requires disclosure of AI tool usage in proposal preparation. Ensure you disclose the use of FindGrants' AI drafting in your application.
NSF
Large Language Models (LLMs) show great promise for generating source code and automating programming tasks. But these models are error-prone and can produce code with subtle bugs. This poses a risk for deploying LLMs in industrial settings for software engineering tasks - the subtly erroneous code generated by LLMs can expose vulnerabilities that compromise system security. It has been shown that the weakness of LLMs for code generation primarily stems from not accounting for the semantic properties of programs when training, using, and evaluating these models. This project aims to improve LLMs’ ability to generate high-quality code by deeply integrating program analyses with all the stages in the life cycle of LLMs: training, code generation, and evaluation. This project develops novel quantitative program analyses techniques to provide feedback to LLMs during training and decoding. First, the project leverages symbolic execution and Bayesian program analyses to design meaningful metrics to evaluate LLM-generated code. This project then uses program scores to train a differentiable reward model that can assess the quality of partial or complete generated code. At training time, inspired by Reinforcement Learning with Human Feedback (RLHF), this project uses the reward model for fine-tuning LLMs to generate high-quality code. To improve code generation at decoding time, this project leverages the reward model and similarity-based program ranking techniques to constrain and prune the decoding tree. Finally, this project develops semantics-guided metrics and collects new benchmarks consisting of realistic coding tasks for training and evaluating code LLMs. This award reflects NSF's statutory mission and has been deemed worthy of support through evaluation using the Foundation's intellectual merit and broader impacts review criteria.
Up to $225K
2028-09-30
Detailed requirements not yet analyzed
Have the NOFO? Paste it below for AI-powered requirement analysis.
One-time $749 fee · Includes AI drafting + templates + PDF export
Category I: CloudBank 2: Accelerating Science and Engineering Research in the Commercial Cloud
NSF — up to $24M
Category I: Nexus: A Confluence of High-Performance AI and Scientific Computing with Seamless Scaling from Local to National Resources
NSF — up to $24.0M
Research Infrastructure: Mid-scale RI-1 (MI:IP): Dual-Doppler 3D Mobile Ka-band Rapid-Scanning Volume Imaging Radar for Earth System Science
NSF — up to $20.0M
A Scientific Ocean Drilling Coordinating Office for the US Community
NSF — up to $17.6M
Category I: AMA27: Sustainable Cyber-infrastructure for Expanding Participation
NSF — up to $13.8M
Graduate Research Fellowship Program (GRFP)
NSF — up to $9.0M