Skip directly to content

Minimize RSR Award Detail

Research Spending & Results

Award Detail

Doing Business As Name:University of Minnesota-Twin Cities
  • Nikolaos Papanikolopoulos
  • (612) 625-0163
  • Daniel F Keefe
  • Gail Bernstein
  • Christine A Conelea
  • Victoria L Interrante
Award Date:09/18/2019
Estimated Total Award Amount: $ 1,831,323
Funds Obligated to Date: $ 1,831,323
  • FY 2019=$1,831,323
Start Date:10/01/2019
End Date:09/30/2024
Transaction Type:Grant
Awarding Agency Code:4900
Funding Agency Code:4900
CFDA Number:47.070
Primary Program Source:040100 NSF RESEARCH & RELATED ACTIVIT
Award Title or Description:MRI: Development of an Instrument that Performs Behavioral Analysis for Neuropsychiatric Disorders like Tourette Syndrome
Federal Award ID Number:1919631
DUNS ID:555917996
Parent DUNS ID:117178941
Program:Major Research Instrumentation
Program Officer:
  • Rita Rodriguez
  • (703) 292-8950

Awardee Location

Street:200 OAK ST SE
Awardee Cong. District:05

Primary Place of Performance

Organization Name:University of Minnesota-Twin Cities
Street:312 Church Street
Cong. District:05

Abstract at Time of Award

This project, developing an instrument to enable a pioneering way to canonically represent and systematically quantify behavioral manifestations for a suite of neuropsychiatric disorders, integrates a powerful computer cluster with various sensor modalities and associated immersive technology components that facilitate data analysis and perception of physical and virtual experiences. The instrument will bring together neuropsychiatrists and engineering scientists seeking the development and deployment of tools that promote discovery and new knowledge in their respective domains, and innovation with immediate societal impact. It aims to enable medical experts to explore clinical hypotheses during diagnosis and treatment. Concurrently, scientists/engineers will explore the implementation of new computational tools and algorithms that satisfy new requirements utilizing a set of clinical hypotheses. Challenging the current understanding of a spectrum of neuropsychiatric conditions and triggering the explorations of new ways to affect them, this instrument should contribute to our understanding of the presence and evolution of mental states. Its unique characteristics will help in integrating a range of informational cues (e.g., visual, acoustic, haptic) along with virtual simulated information. This interplay between physical and virtual experiences and stimuli and their perception in an immersive environment constitutes the ultimate objective which will unleash creativity in better understanding and treat mental illnesses. The tight coupling of hardware and software components will enhance the various types of interactions such as expert-patient, and patient-virtual objects. Specifically, the instrument will support the following research efforts: - Comprehensive compilation of behavioral data, - Development of visualization tools, - Delivery of haptic information feedback, - Creation and interaction of virtual objects, and - Enhanced sentiment understanding from audio/visual signals, facial expression representation and analysis. The instrument enables the creation of communication channels with clinics across the country. Moreover, it contributes in forming a strong foundation to explore and understand a broad variety of neuropsychiatric disorders manifested via verbal and non-verbal communication pathways, aims to form a foundation that promotes strong collaboration targeting real-world challenges with immediate societal and economic impact, and offers an environment for discovery via the strong coupling of data visualization, User Interfaces-User eXperience (UI-UX) design, immersive perception, data exploration, and machine learning. This award reflects NSF's statutory mission and has been deemed worthy of support through evaluation using the Foundation's intellectual merit and broader impacts review criteria.

For specific questions or comments about this information including the NSF Project Outcomes Report, contact us.