A postdoctoral position is available. Individuals with experience in multi-electrode neural recordings and/or fMRI in behaving animals are particularly encouraged to apply.
We are looking for graduate students with previous STEM training. Interested students can apply through the following programs:
Additional competitive training and funding opportunities exist through:
- LUCID (learning, understanding, cognition, intelligence, data science)
- Vision Research Training Program
A technical position is available for a programmer interested in developing systems for real-time dynamic (closed-loop) control of neuroscience experiments involving 3D visualizations. Demonstrated previous experience with OpenGL and the software/hardware requirements of real-time control with millisecond precision is highly desired.
“MRI compatible, customizable, and 3D-printable microdrive for neuroscience research” is out today at eNeuro. Congrats Hoon, Ray, et al.!
Congratulations Ray on being accepted into the Vision Research Training Curriculum (VRTC T32) Program!
Congratulations Lowell on receiving the McPherson Eye Research Institute Graduate Student Support Initiative Award!
“Functional links between sensory representations, choice activity, and sensorimotor associations in parietal cortex” is out today at eLife. Congrats Ting-Yu, Ray, et al.!
“Cue-dependent effects of VR experience on motion-in-depth sensitivity” is out today at PLoS ONE. Congrats Jackie et al.!
Ari received the Vilas Associate Award.
- Older News
How do we perceive the three-dimensional (3D) structure of the world when our eyes only sense 2D projections like a movie on a screen?
Our visual system first encodes the environment in egocentric coordinates defined by our eyes. Such representations are inherently unstable in that they shift and rotate as we move our eyes or head. How then do we perceive the world as stable?
Neuro-computational underpinnings of autism
Autism spectrum disorder (ASD) manifests heterogeneously across individuals. We study the neural basis of this heterogeneity using multifaceted approaches including learning studies with adolescents, neural imaging, and computational modeling.
Real-time Experimental Control with Graphical User Interface
The Real-Time Experimental Control with Graphical User Interface (REC-GUI) framework is an open-source network-based parallel processing solution for performing behavioral control, high precision stimulus presentation, and high-density neurophysiological measurements.