Hello there!

I’m a PhD student advised by Chris Rozell at Georgia Tech, developing advanced closed-loop optogenetic control techniques for neuroscience. Specifically, I am applying optimal, model-based control to precisely perturb population dynamics.

I’m fascinated by the applications of biological computational principles for next-generation AI/ML technologies. And vice-versa: how we can apply modern ML to neurotech/neuroscience, especially for building life- and time-saving brain-computer interfaces. I am looking for an industry research scientist role post-graduation.

Education

PhD Biomedical Engineering

2019
present

Georgia Institute of Technology/Emory University

BS Bioinformatics

2015
2019

Brigham Young University

Interests

Brain-computer interfaces Brain-inspired machine learning High-throughput neuroscience
Recent Publications
Featured Projects

PhD dissertation proposal featured image

PhD dissertation proposal

Towards Optogenetic Feedback Control of Neural Population Dynamics

'Step the Brain along a Path' Lobby Installation featured image

'Step the Brain along a Path' Lobby Installation

A foray into neuroscience-based music and art

Cleo: Closed-Loop, Electrophysiology, and Optophysiology experiment simulation testbed featured image

Cleo: Closed-Loop, Electrophysiology, and Optophysiology experiment simulation testbed

A Python package built around Brian 2 designed as a testbed for bridging computational models and experiments for mesoscale neuroscience. Specifically, it allows for convenient …

Featured Posts
Sonification of neural data: my experience and outlook featured image

Sonification of neural data: my experience and outlook

As explained here, in 2022 I had the opportunity to participate in a project bringing together neuroscience and fine arts to engage the public. In this post I want to share my …

Recent Posts
Mean field variational inference featured image

Mean field variational inference

In this problem, you will investigate mean field approximate inference algorithms (Koller & Friedman1 11.5). Consider the Markov network in the above figure. Define edge potentials …

Approximate inference via Gibbs sampling

Consider a setting in which there are $D$ diseases and a patient either has ($d_i=1$) or does not have ($d_i=0$) each disease. The hospital can measure $S$ symptoms, where $s_j=1$ …

Markov chain Monte Carlo sampling

Inverse CDF sampling A simple sampling method adopted by many of the standard math libraries is the inverse probability transform: draw $u \sim \text{Unif}(0, 1)$, then draw $x\sim …

Parameter learning in probabilistic graphical models

Parameter learning in Bayesian networks and Markov random fields Cost of learning CRF parameters Consider the process of gradient-ascent training for a conditional random field …

Learning maximum likelihood tree structure with the Chow-Liu algorithm

Write a function ChowLiu(X) -> A where X is a D by N data matrix containing a multivariate data point on each column that returns a Chow-Liu maximum likelihood tree for X. The tree …

Expectation-maximization for a Markov chain mixture model

Assume that a sequence $v_1,\ldots,v_T \in \{1,\dots,V\}$ is generated by a Markov chain. For a single chain of length $T$, we have $$ p(v_1,\dots,v_T) = p(v_1)\prod_{t=1}^{T-1} …

Recent & Upcoming Presentations
NAISys 2024 poster featured image

NAISys 2024 poster

A foundation for causal characterization of latent neural dynamics with limited observational and interventional capacity

Optogenetics GRC 2024 poster featured image

Optogenetics GRC 2024 poster

Bridging model and experiment with Cleo: the Closed-Loop, Electrophysiology, and Optophysiology experiment simulation testbed

SNUFA 2023 poster

Enhancing the Cleo experiment simulation testbed to support all-optical control, multi-channel optogenetics, and easier integration into data analysis pipelines