- Library Home /
- Search Collections /
- Open Collections /
- Browse Collections /
- UBC Theses and Dissertations /
- ML-empowered combinatorial search
Open Collections
UBC Theses and Dissertations
UBC Theses and Dissertations
ML-empowered combinatorial search Cameron, Christopher
Abstract
Combinatorial search problems underlie a wide range of computational tasks, from resource allocation in 5G networks and spectrum auctions to verification of hardware and software systems. Despite the best-known techniques having worst-case exponential complexity, many real-world instances exhibit rich structural patterns that specialized solvers can exploit. No single solver excels across all problem distributions, and tuning solvers to specific distributions is crucial for best performance. When done by hand, this application-specific algorithm-design process is very tedious and limited. This thesis explores how machine learning (ML) can empower combinatorial search by integrating data-driven methods with symbolic solvers, maintaining correctness guarantees while adapting to specific applications. We begin by exploring evaluation and applications of two classic ML approaches for algorithm design: algorithm selection and configuration. We later relax two limitations of these classic approaches: (1) reliance on hand-crafted features to represent problem instances; and (2) a black-box view of one or more solvers. We relax the first constraint by learning an end-to-end neural representation of problem instances that captures the right invariances for reasoning about Boolean logic. We relax the second constraint by integrating ML into the solver’s internal decisions, proposing a new reinforcement learning approach, Monte Carlo Forest Search (MCFS), to learn improved branching policies in tree search. With these new advances, learning models for each new application typically demands extensive training data. We show it is possible to train foundation models that dramatically reduce the overhead of learning new tasks and distributions. We trained a single model on a large dataset across a broad range of combinatorial tasks that generalized both on held-out tasks and distributions. We found that models fine-tuned from the foundation models were substantially more sample efficient than models trained from scratch. Finally, we consider the predict-then-optimize paradigm, showing that coupling predictive models with the solver objective can prevent large degradations in solution quality. We formalize this in the setting where forecasts are inherently stochastic and multiple correlated forecasts drive downstream decisions.
Item Metadata
Title |
ML-empowered combinatorial search
|
Creator | |
Supervisor | |
Publisher |
University of British Columbia
|
Date Issued |
2025
|
Description |
Combinatorial search problems underlie a wide range of computational tasks, from resource allocation in 5G networks and spectrum auctions to verification of hardware and software systems. Despite the best-known techniques having worst-case exponential complexity, many real-world instances exhibit rich structural patterns that specialized solvers can exploit. No single solver excels across all problem distributions, and tuning solvers to specific distributions is crucial for best performance. When done by hand, this application-specific algorithm-design process is very tedious and limited. This thesis explores how machine learning (ML) can empower combinatorial search by integrating data-driven methods with symbolic solvers, maintaining correctness guarantees while adapting to specific applications.
We begin by exploring evaluation and applications of two classic ML approaches for algorithm design: algorithm selection and configuration. We later relax two limitations of these classic approaches: (1) reliance on hand-crafted features to represent problem instances; and (2) a black-box view of one or more solvers. We relax the first constraint by learning an end-to-end neural representation of problem instances that captures the right invariances for reasoning about Boolean logic. We relax the second constraint by integrating ML into the solver’s internal decisions, proposing a new reinforcement learning approach, Monte Carlo Forest Search (MCFS), to learn improved branching policies in tree search.
With these new advances, learning models for each new application typically demands extensive training data. We show it is possible to train foundation models that dramatically reduce the overhead of learning new tasks and distributions. We trained a single model on a large dataset across a broad range of combinatorial tasks that generalized both on held-out tasks and distributions. We found that models fine-tuned from the foundation models were substantially more sample efficient than models trained from scratch. Finally, we consider the predict-then-optimize paradigm, showing that coupling predictive models with the solver objective can prevent large degradations in solution quality. We formalize this in the setting where forecasts are inherently stochastic and multiple correlated forecasts drive downstream decisions.
|
Genre | |
Type | |
Language |
eng
|
Date Available |
2025-04-25
|
Provider |
Vancouver : University of British Columbia Library
|
Rights |
Attribution 4.0 International
|
DOI |
10.14288/1.0448579
|
URI | |
Degree | |
Program | |
Affiliation | |
Degree Grantor |
University of British Columbia
|
Graduation Date |
2025-05
|
Campus | |
Scholarly Level |
Graduate
|
Rights URI | |
Aggregated Source Repository |
DSpace
|
Item Media
Item Citations and Data
Rights
Attribution 4.0 International