Repository logo
 

A Study of Techniques for Robustness to Out-of-Distribution Examples

dc.contributor.authorShama Sastry, Chandramouli
dc.contributor.copyright-releaseNot Applicable
dc.contributor.degreeDoctor of Philosophy
dc.contributor.departmentFaculty of Computer Science
dc.contributor.ethics-approvalNot Applicable
dc.contributor.external-examinerDr. Pawan Lingras
dc.contributor.manuscriptsNot Applicable
dc.contributor.thesis-readerDr. Evangelos Milios
dc.contributor.thesis-readerDr. Ga Wu
dc.contributor.thesis-supervisorDr. Sageev Oore
dc.date.accessioned2025-04-21T17:31:25Z
dc.date.available2025-04-21T17:31:25Z
dc.date.defence2025-04-16
dc.date.issued2025-04-17
dc.description.abstractDeep neural networks have achieved remarkable success and human-level performance in many tasks and yet, behave unpredictably when input examples are not guaranteed to be similar to the train distribution. In this thesis, we address the limitations of deep neural networks under distributional shifts, focusing on adversarial examples, covariate shifts, and out-of-distribution (OOD) samples. Ideally, we expect a robust neural network to withstand adversarial perturbations, adapt to covariate shifts, and gracefully refuse to operate on OOD examples. Recognising robustness as a critical challenge for safe and trustworthy deployment, we develop and evaluate train-time and post-training methods --- and their combination --- to address the aforementioned aspects of robustness. First, we introduce a novel post-training OOD detection technique based on Gram matrices of intermediate representations. Notably, this method achieves state-of-the-art performance on several benchmarks without requiring prior knowledge of OOD examples. Our method can also be combined with Outlier-Exposure (OE) to achieve improved robustness, especially on challenging near-distribution outliers. However, since OE relies upon extra data, we explore generative models for improved robustness as described below. Next, we introduce DiffAug, a diffusion-based augmentation method for enhancing robustness against covariate shifts, adversarial perturbations, and OOD inputs. Using DiffAug, we also improve classifier-guided diffusion by achieving improved perceptual alignment of gradients. We thus introduce a computationally efficient technique for training with improved robustness that does not require any additional data, and effectively complements existing augmentation approaches. Moving beyond image classification, we also explore robustness in time-series forecasting --- a domain inherently affected by non-stationary distribution shifts. Building on the DeepTime framework, we propose a theoretically motivated regularization term that improves forecast accuracy under challenging conditions, such as missing data, reduced training set sizes, or higher test-time frequencies. In summary, we present train-time and post-training techniques to enhance model robustness. Beyond their application to improve model robustness, we believe that the research findings offer new insights about the internal workings of a neural-network opening up several interesting future research directions.
dc.identifier.urihttps://hdl.handle.net/10222/85036
dc.language.isoen_US
dc.subjectRobustness
dc.subjectDiffusion Models
dc.subjectTest-time Adaptation
dc.subjectOOD Detection
dc.subjectTime-series Forecasting
dc.titleA Study of Techniques for Robustness to Out-of-Distribution Examples

Files

Original bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
ChandramouliShamaSastry2025.pdf
Size:
18.96 MB
Format:
Adobe Portable Document Format

License bundle

Now showing 1 - 1 of 1
No Thumbnail Available
Name:
license.txt
Size:
2.03 KB
Format:
Item-specific license agreed upon to submission
Description: