Haseeb Shah

I am a graduate student at the RLAI lab at University of Alberta advised by Prof. Martha White. My research focuses on building intelligent agents capable of learning and adapting their behavior online. Previously, I completed my M.Sc. with Prof. Martha White on Neural Network Pruning in the continual learning setting, worked at the LAVIS lab with Prof. Adrian Ulges on Knowledge Graph Completion and Semantic Code Search and at TUKL-SEECS lab with Prof. Faisal Shafait on Catastrophic Forgetting.

Education

University of Alberta

Master of Science in Computer Science, 2021-2023

CGPA: 4.00/4.00

Supervisor: Martha White

Thesis: Greedy Pruning for Continually Adapting Networks

National University of Sciences and Technology

Bachlors of Engineering in Software Engineering, 2015-2019

Supervisor: Faisal Shafait

Thesis: An Open-World Extension for Knowledge Graph Completion Models

Publications

Scalable Real-Time Recurrent Learning Using Columnar-Constructive Networks

K. Javed, H. Shah, R. Sutton, M. White

Journal of Machine Learning Research (JMLR), 2023

We show that by either decomposing the network into independent modules or learning a recurrent network incrementally, we can make RTRL scale linearly with the number of parameters. Unlike prior scalable gradient estimation algorithms, our algorithms do not add noise or bias to the gradient estimate.

GVFs in the Real World: Making Predictions Online for Water Treatment

K. Janjua, H. Shah, M. White, E. Miahi, M. C. Machado, A. White

Machine Learning Journal (MLJ), 2023

We propose a framework for making accurate predictions on a real world water treatment plant based on the General Value Functions. This work is one of the first to motivate the importance of adapting predictions in real-time, for non-stationary high-volume systems in the real world.

An Open-World Extension for Knowledge Graph Completion Models

H. Shah, J. Villmow, A. Ulges, U. Schwanecke, F. Shafait

Oral presentation, Association for the Advancement of Artificial Intelligence (AAAI), 2019

We propose an extension that enables any existing Knowledge Graph Completion model to predict facts about the open-world entities. This approach is more robust, more portable and has better performance than the published state of the art on most datasets. We also released a new dataset that overcomes the shortcomings of previous ones.

Workshops & Preprints

Relation Specific Transformations for Open World Knowledge Graph Completion

H. Shah, J. Villmow, A. Ulges

TextGraphs Workshop at International Conference on Computational Linguistics (COLING), 2020

We introduced relation specific transformations to substantially improve the performance of Open World Knowledge Graph Completion models. We also proposed an approach for clustering of relations to reduce the training time and memory footprint.

Distillation Techniques for Pseudo-rehearsal Based Incremental Learning

H. Shah, K. Javed, F. Shafait.

Preprint, 2018

Standard neural networks suffer from catastrophic forgetting when they are trained on incrementally arriving stream of i.i.d. data. To combat this forgetting, one approach is to train GANs on previously arrived data and feed it to the network again. In this paper, we highlighted that this method is biased and proposed an approach to mitigate this bias and reduce the effect of catastrophic forgetting.

Public Talks

Online Feature Decorrelation

Tea Time Talks, AMII, 2022

A significant proportion of the representations learned by the current generate & test methods consist of highly redundant features. This talk demonstrates how the feature ranking criteria utilized by these methods are highly ineffective in addressing this problem. In this talk, I present a new approach for decorrelating features in an online setting. I demonstrate that this decorrelator can effectively eliminate redundant features and produce a statistically significant performance improvement in the low-capacity function approximation setting.