Efficient Inference-time Control and Alignment

dc.contributor.authorRashid, Ahmad
dc.date.accessioned2026-04-30T19:54:36Z
dc.date.available2026-04-30T19:54:36Z
dc.date.issued2026-04-30
dc.date.submitted2026-04-06
dc.description.abstractModern foundation models are typically trained in three broad stages. First, large-scale pre-training is performed using self-supervised learning on massive corpora. Second, models are adapted through mid-training using supervised fine-tuning or instruction tuning on labeled datasets. Finally, a post-training stage is often applied using preference data and reinforcement learning in order to align the model and improve its safety, reliability, and usefulness. Although effective, post-training methods can be computationally expensive and inflexible once large models are deployed. This thesis explores an alternative paradigm: enforcing behavioral objectives at inference time rather than modifying model parameters during post-training. In this approach, smaller modular control models are combined with a base model to shape predictions during the decision process. Our aim is to design alignment mechanisms that are both mathematically grounded and empirically strong while remaining computationally efficient and easy to deploy. We apply this perspective of inference-time control to three problems. First, we address reliability in neural classifiers. We introduce PreLoad, an inference-time mechanism that mitigates arbitrarily high confidence on inputs that lie outside the training support while preserving accuracy and training efficiency. Second, we study reward-guided text generation (RGTG) in large language models as a form of inference-time alignment. We show that stable reward-guided decoding requires carefully designed token-level reward models and propose two algorithms, PARGS and FaRMA, that enable effective reward-guided generation. Third, we address the computational cost of RGTG and propose an efficient algorithm that adds only a minor overhead during inference while preserving the performance and benefits of reward-guided decoding. Together, these results demonstrate that inference-time control provides a flexible and computationally efficient framework for shaping the behavior of modern neural systems. By decoupling representation learning from the decision-time objectives, this work introduces new tools for improving the reliability, alignment, and efficiency of large-scale machine learning models without retraining them.
dc.identifier.urihttps://hdl.handle.net/10012/23136
dc.language.isoen
dc.pendingfalse
dc.publisherUniversity of Waterlooen
dc.subjectArtificial Intelligence
dc.subjectAlignment
dc.subjectLarge Language Models
dc.subjectDeep Learning
dc.subjectReliable AI
dc.subjectInference-time Control
dc.subjectMachine Learning
dc.subjectNatural Language Processing
dc.subjectReinforcement Learning
dc.subjectTest-time Compute
dc.subjectControlled Decoding
dc.subjectAlgorithms
dc.subjectEfficient AI
dc.subjectReward Models
dc.subjectValue Functions
dc.subjectReward Guided Text Generation
dc.subjectOOD Detection
dc.subjectOut-of-Distribution
dc.titleEfficient Inference-time Control and Alignment
dc.typeDoctoral Thesis
uws-etd.degreeDoctor of Philosophy
uws-etd.degree.departmentDavid R. Cheriton School of Computer Science
uws-etd.degree.disciplineComputer Science
uws-etd.degree.grantorUniversity of Waterlooen
uws-etd.embargo.terms0
uws.contributor.advisorPoupart, Pascal
uws.contributor.affiliation1Faculty of Mathematics
uws.peerReviewStatusUnrevieweden
uws.published.cityWaterlooen
uws.published.countryCanadaen
uws.published.provinceOntarioen
uws.scholarLevelGraduateen
uws.typeOfResourceTexten

Files

Original bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
Rashid_Ahmad.pdf
Size:
1017.3 KB
Format:
Adobe Portable Document Format

License bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
license.txt
Size:
6.4 KB
Format:
Item-specific license agreed upon to submission
Description:

Collections