Inferential Machine Learning: Towards Human-collaborative Foundation Models
Neural network driven applications like ChatGPT suffer from hallucinations where they confidently provide inaccurate information. A fundamental reason for this inaccuracy is the feed-forward nature of inductive decisions taken by neural networks. Such decisions are a result of training schemes that do not allow networks to deviate from and creatively abduce reasons. With the advent of foundation models that are adapted across applications and data, humans can directly intervene and prompt foundation models. However, without understanding the operational limits of the underlying networks, human interventions often lead to unfair, inaccurate, hallucinated and unintelligible outputs. These outputs undermine the trust in foundation models, thereby causing roadblocks to their adoption in everyday lives. In this tutorial, we provide a systematic way to analyze and understand human interventions in neural network functionality. Specifically, our insights are the following: 1) decision theory must be abductive rather than deductive or inductive, 2) interventions must be analyzed as a function of the ‘not-taken’ residual interventions, 3) interventions are not always positive and networks must be equipped to detect adversarial interactions. The end goal is to promote a human-AI collaborative environment via inferential machine learning techniques.
Tutorial Location and Time
Webinar Link: Join the Webinar
Venue: IEEE/CVF Winter Conference on Applications of Computer Vision (WACV)
Room: Salon 2
Date and Time: FEB 28th, Mountain time (8 – 12 PM MST (GMT-7))
Duration: 2 Hours
Presenters

Goals
The goal of the tutorial is threefold: 1) place human intervenability in neural network applicability as a function of existing applications, 2) provide a mathematical framework to understand the effects of interventions in a neural network’s decision-making process, 3) imbue networks with the ability to differentiate between positive and negative interventions. Detailed subtopics are provided in the Outline section below.
Relevance to Audience
IEEE Computer Society has historically been at the forefront of advancing research in computational decision making. As neural network-based AI systems transition from academia to everyday life, their vulnerabilities must be understood before acceptance by the public. A key vulnerability is the lack of knowledge regarding a neural network’s operational limits. This vulnerability leads to lack of trust in neural networks by non-expert users. In the past, applications of robustness have served the research community to define trustworthy measures. However, users rarely understand these applications. Recently, prompt-based architectures that allow limited inputs from users during inference stage have gained prominence. Hence, users employ a ‘trial and error’ policy that may exceed the operational limits as designed by researchers during training. Hence, a larger emphasis must be placed regarding machine learning at inference. This is a timely tutorial that emphasizes the trustworthiness of neural network predictions in terms of human-centric measures which aid AI’s applicability. Since IEEE Big Data is attended by stakeholders at all levels, including students, researchers, entrepreneurs, policy makers, and other interested parties in academia and industry, it provides an ideal opportunity to emphasize inferential machine learning for foundation models.
Target Audience
This tutorial is intended for graduate students, researchers, engineers, and policy makers working in different topics related to trustworthy and regulatory machine learning.
Audience Prerequisites
Audience are expected to have a basic understanding of neural networks and vision applications including image recognition and detection.
Content Level
Time spent in the short course regarding content level will roughly be split as follows – 10% beginner, 65% intermediate, and 25% advanced.
Content Description
Neural networks provide generalizable and task independent representation spaces that have garnered widespread applicability in vision, language, and multimodal understanding applications. The complicated semantics of feature interactions within data has been broken down into a set of non-linear functions, convolution parameters, attention, as well as multi-modal inputs among others. The complexity of these operations has introduced multiple vulnerabilities within neural network architectures. These vulnerabilities include adversarial samples, confidence calibration issues, and catastrophic forgetting among others. Given that AI promises to herald the fourth industrial revolution, it is critical to understand and overcome these vulnerabilities. Doing so requires creating trustworthy neural networks that drive the AI systems. Defining trustworthiness, however, is not trivial. Traditionally, trust is a function of trustworthiness attributes including simple invariance to noise and perturbations. However, these measures are neither applicable nor communicable to casual users at inference. In this tutorial, we provide a human-interventionist approach to understanding neural network interactions that allow AI to function in society. Doing so allows us to state the following: 1) neural networks must know when and what they don’t know, 2) neural networks must be amenable to being intervened upon by humans at any stage of their decision-making process. These statements call for neural networks to be equipped with uncertainty quantification and be intervenable.
We provide a probabilistic post-hoc analysis of uncertainty, and intervenability. Post-hoc implies that a decision has already been made. In larger context, the goal of uncertainty quantification and interventionist amenability must be to satisfy multiple stakeholders at various levels of expertise. This includes researchers, engineers, policymakers, and everyday users among others. In this tutorial, we expound on a gradient-based methodology that provides a probabilistic framework of analysis without requiring any retraining. Once a neural network is trained, it acts as a knowledge base through which different types of gradients can be used to traverse adversarial, contrastive, and counterfactual representation spaces. We demonstrate the utility these gradients to define uncertainty and intervenability. Several image understanding and robustness applications including anomaly, novelty, adversarial, and out-of-distribution image detection, image quality assessment, and noise recognition experiments among others will be discussed. In this tutorial, we examine the types, visual meanings, and interpretations of robustness as a human-centric measure of the utility of large-scale neural networks.
Tutorial Outline
Part 1: Introduction to Inferential Machine Learning (20 mins)
-
Speaker: Ghassan AlRegib
- Case Study: Active Learning in initial training stages
- Case Study: Catastrophic Forgetting in later training stages
- Inferential Machine Learning: Definition, history, and modern adaptations including Test Time Augmentation (TTA), Self-supervised Learning (SSL), and Reinforcement Learning from Human Feedback (RLHF)
Part 2: Statistical and Computational Learning Theory and Inference (20 mins)
-
Speaker: Ghassan AlRegib
- Computational Learning: Traditional supervised learning under inductive reasoning
- Statistical Learning: Expands computational learning theory to include unsupervised and reinforcement learning
- Case study: Prompting in Segment Anything Model
Part 3: Foundation Models (20 mins)
-
Speaker: Ghassan AlRegib
- Basics of neural networks
- Brief history of neural networks: MLPs to CNNs to Transformers
- Vision and Language Foundation models
Part 4: Human-collaborative Interventions at Inference (1 hr)
-
Speaker: Mohit Prabhushankar
- Intervenability in existing applications of causality, privacy, interpretability, prompting, and benchmarking
- Positive and negative interventions
- Case study 1: Membership inference attacks
- Case study 2: Catastrophic and Anastrophic prompting in Segment Anything Model
- Mathematical frameworks to analyze intervenability
- Incomplete interventions under causality
- Uncertainty of incomplete interventions
- Case study: Intervenability in Interpretability
Previous tutorials
Some of the previous tutorials along with slides and video links can be accessed at https://alregib.ece.gatech.edu/courses-and-tutorials/
Relevant Topics and Publications
•Explainability [1, 2]
•Out-of-distribution Detection [3, 6]
•Adversarial Detection [4]
•Anomaly Detection [5]
•Corruption Detection [3]
•Misprediction Detection [6]
•Causal Analysis [7]
•Open-set Recognition [8]
•Noise Robustness [9]
•Uncertainty Visualization [10]
•Image Quality Assessment [11, 12]
•Saliency Detection [13]
•Novelty Detection [14]
•Disease Severity Detection [15]
[1] AlRegib, G., & Prabhushankar, M. (2022). Explanatory Paradigms in Neural Networks: Towards relevant and contextual explanations. IEEE Signal Processing Magazine, 39(4), 59-72.
[2] Prabhushankar, M., Kwon, G., Temel, D., & AlRegib, G. (2020, October). Contrastive explanations in neural networks. In 2020 IEEE International Conference on Image Processing (ICIP) (pp. 3289-3293). IEEE.
[3] J. Lee, C. Lehman, M. Prabhushankar, and G. AlRegib, “Probing the Purview of Neural Networks via Gradient Analysis,” in IEEE Access, Mar. 21 2023.
[4] J. Lee, M. Prabhushankar, and G. AlRegib, “Gradient-Based Adversarial and Out-of-Distribution Detection,” in International Conference on Machine Learning (ICML) Workshop on New Frontiers in Adversarial Machine Learning, Baltimore, MD, Jul. 2022.
[5] Kwon, G., Prabhushankar, M., Temel, D., & AlRegib, G. (2020, August). Backpropagated gradient representations for anomaly detection. In European Conference on Computer Vision (pp. 206-226). Springer, Cham.
[6] Prabhushankar, M., & AlRegib, G. (2024, August). Counterfactual Gradients-based Quantification of Prediction Trust in Neural Networks. In 2024 IEEE 7th International Conference on Multimedia Information Processing and Retrieval (MIPR) (pp. 529-535). IEEE.
[7] M. Prabhushankar, and G. AlRegib, ”Extracting Causal Visual Features for Limited Label Classification,” in IEEE International Conference on Image Processing (ICIP), Sept. 2021.
[8] Lee, Jinsol, and Ghassan AlRegib. “Open-Set Recognition With Gradient-Based Representations.” 2021 IEEE International Conference on Image Processing (ICIP). IEEE, 2021.
[9] M. Prabhushankar, and G. AlRegib, “Introspective Learning : A Two-Stage Approach for Inference in Neural Networks,” in Advances in Neural Information Processing Systems (NeurIPS), New Orleans, LA, Nov. 29 – Dec. 1 2022
[10] Prabhushankar, M., & AlRegib, G. (2024). Voice: Variance of induced contrastive explanations to quantify uncertainty in neural network interpretability. IEEE Journal of Selected Topics in Signal Processing.
[11] M. Prabhushankar and G. AlRegib, “Stochastic Surprisal: An Inferential Measurement of Free Energy in Neural Networks,” in Frontiers in Neuroscience, Perception Science, Volume 17, Feb. 09 2023.
[12] G. Kwon*, M. Prabhushankar*, D. Temel, and G. AlRegib, “Distorted Representation Space Characterization Through Backpropagated Gradients,” in IEEE International Conference on Image Processing (ICIP), Taipei, Taiwan, Sep. 2019.
[13] Y. Sun, M. Prabhushankar, and G. AlRegib, “Implicit Saliency in Deep Neural Networks,” in IEEE International Conference on Image Processing (ICIP), Abu Dhabi, United Arab Emirates, Oct. 2020.
[14] Kwon, G., Prabhushankar, M., Temel, D., & AlRegib, G. (2020, October). Novelty detection through model-based characterization of neural networks. In 2020 IEEE International Conference on Image Processing (ICIP) (pp. 3179-3183). IEEE.
[15] K. Kokilepersaud, M. Prabhushankar, G. AlRegib, S. Trejo Corona, C. Wykoff, “Gradient Based Labeling for Biomarker Classification in OCT,” in IEEE International Conference on Image Processing (ICIP), Bordeaux, France, Oct. 16-19 2022