Diffusion Models & Generative AI

Microscopy Denoising Diffusion with Poisson-Aware Physical Guidance

Work in Progress

This work aims to computationally overcome the inherent trade-offs between different fluorescence microscopy techniques. Widefield (WF) microscopy offers a large field-of-view and high frame rates but suffers from significant image degradation due to out-of-focus light contamination. Conversely, 2-photon (2P) microscopy provides excellent optical sectioning, deeper tissue penetration, and reduced scattering by selectively exciting a diffraction-limited focal volume, but at the cost of speed and field-of-view. This project initially explored using standard CNNs (like U-Net and VGG16) to transform blurry WF images into sharp, 2P-quality images. However, the methodology evolved to address the fundamental noise properties of microscopy more directly. The final implementation uses a more advanced Denoising Diffusion Probabilistic Model (DDPM), which was enhanced with a novel, physically-motivated Poisson-Kullback-Leibler (PKL) guidance mechanism. This corrects the erroneous assumption of Gaussian noise made by standard models and instead enforces data consistency based on the physically accurate Poisson distribution of photon arrivals. The impact of this more sophisticated approach is a state-of-the-art image restoration pipeline that is more robust and has a lower tendency to hallucinate artifacts, enabling higher-fidelity biological imaging from faster, more accessible WF systems.

Computer Vision & Image Processing

The Pupil Tracker: Fast and Efficient Pupil Segmentation and Tracking [pdf]

Spring 2022 - CS289 Final Project

To address a significant data processing bottleneck in neuroscience, this project developed a computationally efficient alternative to resource-intensive deep learning models for tracking pupil dynamics for behavioral study. Pupil size and gaze direction are critical, non-invasive proxies for an animal's cognitive state—including arousal, attention, and decision-making—making accurate, high-throughput tracking essential for correlating neural activity with behavior. Existing CNN-based methods like DeepLabCut require days of training and specialized GPUs, creating significant delays in post-processing analysis. This work implemented a "Pupil Tracker" using a classical K-nearest neighbor (KNN) algorithm on binarized image data. KNN was selected for computational efficiency in neuroscience research environments with limited computing resources. Benchmarking against Detectron2 and DeepLabCut revealed massive training overhead (10 days CPU for DeepLabCut, 4 days on 8 V100 GPUs for Detectron2), specialized hardware requirements, and labor-intensive data labeling. KNN's "training-free" implementation, simplicity on binarized data, and CPU-centric performance prioritized speed and accessibility over marginal accuracy gains. The impact was a dramatic increase in experimental throughput: the method eliminated the training phase entirely and achieved an inference time of 0.075 seconds per frame, making it ~20% faster than DeepLabCut and 60 times faster than Detectron2, while running on a standard CPU. This efficiency allows researchers to process vastly more behavioral data, accelerating the pace of discovery in systems neuroscience.

Deep RL

MARL: Learned Memory Allocation in Heterogeneous Memory Systems [pdf]

Fall 2021 - CS285 Final Project

This project addressed the exascale computing challenge of the "memory wall," where the performance of memory-bound applications, such as deep learning inference, is throttled by the limited capacity of high-bandwidth DRAM. A novel reinforcement learning framework, MARL, was developed to automate data placement in a heterogeneous memory system composed of 16GB of DRAM and 128GB of slower, high-capacity Persistent Memory (PM). The implementation represented workloads as computational graphs for a Graph Neural Network (GNN) policy, which was trained via an Evolutionary Algorithm to navigate the immense action space ($2^{311}$ for ResNet-101) of possible memory allocations. Deployed on a commercial Intel server, MARL's impact was twofold: it achieved a 1.2x inference speedup over Intel's own optimized Memory Mode and, by intelligently utilizing the entire memory pool, it increased server throughput by enabling 85 parallel inference tasks compared to only 65 in Memory Mode and one in a DRAM-only configuration—a 30% improvement over the industry baseline.

ML Tools for Energy Systems

Spring 2021 - ME249

Project 1: Genetic Algorithm for Heat Transfer Correlation

To improve predictive accuracy for nucleate boiling heat transfer, this project sought to optimize the empirical constants within a five-parameter, dimensionless correlation based on the Rohsenow model. A genetic algorithm was implemented to perform a multi-variable search across a dataset encompassing varied gravity levels, pressures, and surface tensions. The algorithm successfully converged to a solution, demonstrating its efficacy in navigating a complex parameter space to establish a more robust and physically grounded heat transfer correlation.

Project 2: Neural Network Modeling of Spray Cooling and Gas Turbine Systems

This project aimed to create a high-fidelity surrogate model of a hybrid solar/fossil-fuel gas turbine power system, where traditional fluid dynamics models would be computationally prohibitive. A sequential neural network with three hidden layers and the `elu` activation function was implemented in Keras. By systematically evaluating different network architectures, this data-driven approach provided a computationally efficient tool to predict system performance and revealed that solar heat input was a more dominant parameter than air inlet temperature.

Project 3: Performance Modeling of Solar PV Power Systems

The objective was to develop a predictive framework for the non-linear behavior of multi-panel solar PV systems under varying electrical configurations (series, parallel, series-parallel). A key innovation was the use of two distinct neural networks: the first predicted the optimal configuration for a given condition, while the second, more accurate model predicted the power output for a specified configuration. The second model proved vastly superior for system control, achieving a 35-fold improvement in accuracy, providing a highly reliable tool for performance prediction.

Project 4: Solar Thermal Power Plant Boiler and Electronics Cooling

This work applied neural networks to two distinct thermal management problems: predicting steam exit quality in a solar thermal boiler and determining maximum surface temperatures in a passively cooled electronics system. For the boiler, a Keras model accurately predicted performance and was further developed into an inverse design tool that could calculate the required mass flow rate for a desired output. For electronics cooling, the resulting neural network served as the foundation for a design tool that generated a surface plot of safe operating conditions, effectively translating a predictive model into actionable engineering guidance.