I am a Ph.D. student at the RLAI lab at University of Alberta advised by Prof. Martha White. My research focuses on building Reinforcement Learning agents capable of learning and adapting their behavior online, both in applied and simulated settings. Previously, I completed my M.Sc. with Prof. Martha White on Neural Network Pruning in the continual learning setting, worked at the LAVIS lab with Prof. Adrian Ulges on Knowledge Graph Completion and Semantic Code Search and at TUKL-SEECS lab with Prof. Faisal Shafait on Catastrophic Forgetting.
Ph.D. in Computer Science, September 2024-present
Supervisor: Martha White
M.Sc. in Computer Science, 2021-2023
CGPA: 4.00/4.00
Supervisor: Martha White
Thesis: Greedy Pruning for Continually Adapting Networks
Bachlors of Engineering in Software Engineering, 2015-2019
Supervisor: Faisal Shafait
Thesis: An Open-World Extension for Knowledge Graph Completion Models
K. Javed, H. Shah, R. Sutton, M. White
Journal of Machine Learning Research (JMLR), 2023
International Conference on Machine Learning (ICML), 2024
We show that by either decomposing the network into independent modules or learning a recurrent network incrementally, we can make RTRL scale linearly with the number of parameters. Unlike prior scalable gradient estimation algorithms, our algorithms do not add noise or bias to the gradient estimate.
K. Janjua, H. Shah, M. White, E. Miahi, M. C. Machado, A. White
Machine Learning Journal (MLJ), 2023
Reinforcement Learning Conference (RLC), 2024
We propose a framework for making accurate predictions on a real world water treatment plant based on the General Value Functions. This work is one of the first to motivate the importance of adapting predictions in real-time, for non-stationary high-volume systems in the real world.
H. Shah, J. Villmow, A. Ulges, U. Schwanecke, F. Shafait
Oral presentation, Association for the Advancement of Artificial Intelligence (AAAI), 2019
We propose an extension that enables any existing Knowledge Graph Completion model to predict facts about the open-world entities. This approach is more robust, more portable and has better performance than the published state of the art on most datasets. We also released a new dataset that overcomes the shortcomings of previous ones.
L. Zhu*, H. Shah*, H. Wang*, M. White | *equal contribution
Under Review - 2024
We explore the effectiveness of q-exponential policies in policy optimization methods, finding that heavy-tailed policies (q > 1) are generally more effective and can consistently outperform the Gaussian policy.
H. Shah, J. Villmow, A. Ulges
TextGraphs Workshop at International Conference on Computational Linguistics (COLING), 2020
We introduced relation specific transformations to substantially improve the performance of Open World Knowledge Graph Completion models. We also proposed an approach for clustering of relations to reduce the training time and memory footprint.
H. Shah, K. Javed, F. Shafait.
Preprint, 2018
Standard neural networks suffer from catastrophic forgetting when they are trained on incrementally arriving stream of i.i.d. data. To combat this forgetting, one approach is to train GANs on previously arrived data and feed it to the network again. In this paper, we highlighted that this method is biased and proposed an approach to mitigate this bias and reduce the effect of catastrophic forgetting.
Tea Time Talks, AMII, 2022
A significant proportion of the representations learned by the current generate & test methods consist of highly redundant features. This talk demonstrates how the feature ranking criteria utilized by these methods are highly ineffective in addressing this problem. In this talk, I present a new approach for decorrelating features in an online setting. I demonstrate that this decorrelator can effectively eliminate redundant features and produce a statistically significant performance improvement in the low-capacity function approximation setting.