Turning multimodal brain data into actionable insight.
I am a final‑year Ph.D. candidate in Electrical Engineering at UC San Diego, advised by Prof. Gal Mishne, designing interpretable machine‑learning frameworks that weave multimodal biomedical signals into detailed maps of brain connectivity. My work has appeared in NeurIPS and ICML and builds on deep generative models—Transformers, VAEs, and coupled autoencoders—combined with rigorous statistics. Earlier, I graduated top of my class at IIT Hyderabad, earning the Institute Silver Medal and a Research Excellence Award for my EMBC’17 machine‑learning wearable‑sensing study. I mentor students and co‑lead brain‑computer‑interface workshops, aiming to transform algorithmic advances into real‑world biomedical impact.
Context‑aware feature selection
c‑STG learns which input features matter for each context, boosting accuracy and interpretability across domains.
Paper ↗Dissecting cell‑type‑specific striatal dynamics
Large‑scale recordings reveal how cortex and thalamus jointly shape direct‑ and indirect‑pathway MSNs during skilled locomotion.
Pre‑print ↗Finding feature clusters that separate conditions
DiSC spots groups of pixels/genes/voxels whose relationships shift between states, outperforming baselines on imaging and genomics.
Paper ↗Disentangling shared vs. private neural dynamics
CTAE captures long‑range, non‑linear interactions across brain regions with explicit shared/private latent spaces.
Under review (ICLR 2026)See the complete list on Google Scholar ↗. (Automatic import coming soon!)