Sourav Pal

I am a sixth year PhD student of Computer Sciences at UW-Madison advised by Prof. Vikas Singh. My broad research interests are in Machine Learning and their applications to problems in Computer Vision and lately Large Language Models (LLMs). I am specifically interested in studying the interplay of concepts from applied math such as differential equations, operators, etc. with strong function approximation properties of neural networks prevalant in modern deep learning.

Before coming to Madison I was part of Adobe Acrobat Reader team. Even before I was a Research Intern at the BigData Experience Lab of Adobe Research where I worked with an amazing mentor, Dr. Ritwik Sinha.

As an undergrad I spent four amazing years, at the Indian Institute of Technology (IIT) Kharagpur from where I graduated with a B.Tech (Hons.) in Computer Science and Engineering. I was fortunate enough to be advised by Prof. Pabitra Mitra during my undergraduate studies.

Email  /  Twitter /  GitHub  /  Resume  /  Google Scholar  /  LinkedIn

profile photo
Recent Updates:

June 24: New paper on Implicit Representations got accepted to ICML 2024!

March 24: Our work on sampling of temporal trajectories is out on ArXiv.

April 23: Our work on Controlled Differential Equation got accepted to ICML 2023!

June 22: Our work on machine unlearning will also appear at UpML 2022 – Updatable Machine Learning Workshop at ICML

May 22: Our paper on multi resolution analysis for efficient self attention got accepted to ICML 2022

March 22: Our paper on machine unlearning got accepted to CVPR 2022

Publications
Implicit Representations via Operator Learning
Sourav Pal, Harshavardhan Adepu, Clinton Wang , Polina Golland, Vikas Singh
ICML 2024

[Code]

The idea of representing a signal as the weights of a neural network, called Implicit Neural Representations (INRs), has led to exciting implications for compression, view synthesis and 3D volumetric data understanding. One problem in this setting pertains to the use of INRs for downstream processing tasks. Despite some conceptual results, this remains challenging because the INR for a given image/signal often exists in isolation. What does the neighborhood around a given INR correspond to? Based on this question, we offer an operator theoretic reformulation of the INR model, which we call Operator INR (or O-INR). At a high level, instead of mapping positional encodings to a signal, O-INR maps one function space to another function space. A practical form of this general casting is obtained by appealing to Integral Transforms. The resultant model does not need multi-layer perceptrons (MLPs), used in most existing INR models - we show that convolutions are sufficient and offer benefits including numerically stable behavior. We show that O-INR can easily handle most problem settings in the literature, and offers a similar performance profile as baselines. These benefits come with minimal, if any, compromise.

Controlled Differential Equations on Long Sequences via Non-standard Wavelets
Sourav Pal, Zhanpeng Zeng, Sathya N. Ravi, Vikas Singh
ICML 2023

[Code]

Neural Controlled Differential equations (NCDE) are a powerful mechanism to model the dynamics in temporal sequences, e.g., applications involving physiological measures, where apart from the initial condition, the dynamics also depend on subsequent measures or even a different “control” sequence. But NCDEs do not scale well to longer sequences. Existing strategies adapt rough path theory, and instead model the dynamics over summaries known as log signatures. While rigorous and elegant, invertibility of these summaries is difficult, and limits the scope of problems where these ideas can offer strong benefits (reconstruction, generative modeling). For tasks where it is sensible to assume that the (long) sequences in the training data are a fixed length of temporal measurements – this assumption holds in most experiments tackled in the literature – we describe an efficient simplification. First, we recast the regression/classification task as an integral transform. We then show how restricting the class of operators (permissible in the integral transform), allows the use of a known algorithm that leverages non-standard Wavelets to decompose the operator. Thereby, our task (learning the operator) radically simplifies. A neural variant of this idea yields consistent improvements across a wide gamut of use cases tackled in existing works. We also describe a novel application on modeling tasks involving coupled differential equations.

Multi Resolution Analysis (MRA) for Approximate Self-Attention
Zhanpeng Zeng, Sourav Pal, Jeffery Kline, Glenn Fung, Vikas Singh
ICML 2022

[Code]

We revisit classical Multiresolution Analysis (MRA) concepts such as Wavelets, whose potential value in this setting remains underexplored thus far. We show that simple approximations based on empirical feedback and design choices informed by modern hardware and implementation challenges, eventually yield a MRA-based approach for self-attention with an excellent performance profile across most criteria of interest. We undertake an extensive set of experiments and demonstrate that this multi-resolution scheme outperforms most efficient self-attention proposals and is favorable for both short and long sequences

Deep Unlearning via Randomized Conditionally Independent Hessians
Ronak Mehta*, Sourav Pal*, Vikas Singh, Sathya N. Ravi
CVPR 2022
Also in UpML 2022 – Updatable Machine Learning Workshop at ICML 2022

[Code]

Machine Unlearning is the art of removing specific training samples from a predictive model as if they never existed in the training dataset. Recent ideas leveraging optimization-based updates scale poorly with the model dimension d, due to inverting the Hessian of the loss function. We use a variant of a new conditional independence coefficient, L-CODEC, to identify a subset of the model parameters with the most semantic overlap on an individual sample level. Our approach completely avoids the need to invert a (possibly) huge matrix. Our approach makes approximate unlearning possible in settings that would otherwise be infeasible, including vision models used for face recognition, person reidentification and transformer based NLP models.

D-FJ: Deep Neural Network Based Factuality Judgment
Ankan Mullick, Sourav Pal, Projjal Chanda, Arijit Panigrahy, Anurag Bharadwaj, Siddhant Singh, Tanmoy Dam
TrueFact, SIGKDD, 2019

Deep neural networks to detect facts and opinions from online news media. We have also shown how factuality, opinionatedness and sentiment fraction of different news articles changes over certain events in different time frames.

Visual Attention for Behavioral Cloning in Autonomous Driving
Sourav Pal*, Tharun Mohandoss*, Pabitra Mitra
ICMV, 2018

We present two methods of predicting visual attention maps. The first method is a supervised learning approach in which we collect eye-gaze data for the task of driving and use this to train a model for predicting the attention map. The second method is a novel unsupervised approach where we train a model to learn to predict attention as it learns to drive a car.

Saliency Prediction for Mobile User Interfaces
Prakhar Gupta, Sourav Pal*, Shubh Gupta* , Ajaykrishnan Jayagopal*, Ritwik Sinha
WACV, 2018

We introduce deep learning models for saliency prediction for mobile user interfaces at the element level to improve their usability.

Patents
Saliency Prediction for a Mobile User Interface

Saliency Prediction for Informational Documents
Reviewer: ICML 24, NeurIPS 23, ICML 22, CVPR 22, WACV 22
Class note: Measure Theory

Stack Overflow'ed from here