Hi, I'm Qasim

LinkedInLinkLink

I am a Ph.D. student at Purdue University's School of Electrical and Computer Engineering, working under the supervision of Professor Murat Kocaoglu and Professor Mahsa Ghasemi. My research interests include reinforcement learning, and causal reasoning/inference. 

Selected work



Adaptive Online Experimental Design for Causal Discovery (ICML 2024 Spotlight Poster) (Click to read the Full Paper)

Abstract: Causal discovery aims to uncover cause-and-effect relationships encoded in causal graphs by leveraging observational, interventional data, or their combination. The majority of existing causal discovery methods are developed assuming infinite interventional data. We focus on data interventional efficiency and formalize causal discovery from the perspective of online learning, inspired by pure exploration in bandit problems. A graph separating system, consisting of interventions that cut every edge of the graph at least once, is sufficient for learning causal graphs when infinite interventional data is available, even in the worst case. We propose a track-and-stop causal discovery algorithm that adaptively selects interventions from the graph separating system via allocation matching and learns the causal graph based on sampling history. Given any desired confidence value, the algorithm determines a termination condition and runs until it is met. We analyze the algorithm to establish a problem-dependent upper bound on the expected number of required interventional samples. Our proposed algorithm outperforms existing methods in simulations across various randomly generated causal graphs. It achieves higher accuracy, measured by the structural hamming distance (SHD) between the learned causal graph and the ground truth, with significantly fewer samples. 


Approximate allocation matching for structural causal bandits with unobserved confounders (NeurIPS 2023 Poster) (Click to read the Full Paper)

Abstract: Structural causal bandit provides a framework for online decision-making problems when causal information is available. It models the stochastic environment with a structural causal model (SCM) that governs the causal relations between random variables. In each round, an agent applies an intervention (or no intervention) by setting certain variables to some constants and receives a stochastic reward from a non-manipulable variable. Though the causal structure is given, the observational and interventional distributions of these random variables are unknown beforehand, and they can only be learned through interactions with the environment. Therefore, to maximize the expected cumulative reward, it is critical to balance the explore-versus-exploit tradeoff. We assume each random variable takes a finite number of distinct values, and consider a semi-Markovian setting, where random variables are affected by unobserved confounders. Using the canonical SCM formulation to discretize the domains of unobserved variables, we efficiently integrate samples to reduce model uncertainty. This gives the decision maker a natural advantage over those in a classical multi-armed bandit setup. We provide a logarithmic asymptotic regret lower bound for the structural causal bandit problem. Inspired by the lower bound, we design an algorithm that can utilize the causal structure to accelerate the learning process and take informative and rewarding interventions. We establish that our algorithm achieves a logarithmic regret and demonstrate that it outperforms the existing methods via simulations. 

400V EV Traction System



The lab setup is designed to simulate a real-life electric vehicle system, using a 400V Lithium-ion battery bank, a DSP controlled Voltage Source Variable Frequency Drive, and an Induction Motor. The aim is to improve the efficiency and robustness of motor controllers and bridge the gap between theoretical work and real-life application. Various AC motor drive algorithms are tested, with high voltage battery parameters being taken into account. The lab setup closely monitors and analyzes temperature, current, voltage, and power draw values during and after each test run.