Skip to content

Commit

Permalink
Merge pull request #67 from MSDLLCpapers/main
Browse files Browse the repository at this point in the history
Sync bugfix branch
  • Loading branch information
kstone40 authored Sep 19, 2024
2 parents 3095d6b + eff5652 commit 35ae52a
Show file tree
Hide file tree
Showing 11 changed files with 262 additions and 38 deletions.
7 changes: 7 additions & 0 deletions CHANGELOG.md
Original file line number Diff line number Diff line change
@@ -1,6 +1,13 @@
# Changelog

## [Untracked Changes]
### Added
- Improved methods for fitting PyTorch surrogates, including auto-stopping by parameter value norm

### Modified
- Greatly reduced the number of samples for DNN posterior, speeding up optimization
- Stabilized the mean estimate of ensemble surrogates by avoiding resampling
- Disabled root caching for ensemble surrogates during optimization

## [0.8.5]
### Added
Expand Down
7 changes: 5 additions & 2 deletions obsidian/acquisition/custom.py
Original file line number Diff line number Diff line change
Expand Up @@ -64,10 +64,11 @@ class qSpaceFill(MCAcquisitionFunction):
"""
def __init__(self,
model: Model,
X_baseline: Tensor,
sampler: MCSampler | None = None,
objective: MCAcquisitionObjective | None = None,
posterior_transform: PosteriorTransform | None = None,
X_pending: Tensor | None = None):
X_pending: Tensor | None = None,):

if sampler is None:
sampler = SobolQMCNormalSampler(sample_shape=torch.Size([512]))
Expand All @@ -80,6 +81,8 @@ def __init__(self,

super().__init__(model=model, sampler=sampler, objective=objective,
posterior_transform=posterior_transform, X_pending=X_pending)

self.register_buffer('X_baseline', X_baseline)

@t_batch_mode_transform()
def forward(self,
Expand All @@ -88,7 +91,7 @@ def forward(self,
Evaluate the acquisition function on the candidate set x
"""
# x dimensions: b * q * d
x_train = self.model.train_inputs[0][0] # train_inputs is a list of tuples
x_train = self.X_baseline

# For sequential mode, add pending data points to "train"
if self.X_pending is not None:
Expand Down
27 changes: 20 additions & 7 deletions obsidian/optimizer/bayesian.py
Original file line number Diff line number Diff line change
Expand Up @@ -3,7 +3,7 @@
from .base import Optimizer

from obsidian.parameters import ParamSpace, Target, Task
from obsidian.surrogates import SurrogateBoTorch, DNN
from obsidian.surrogates import SurrogateBoTorch, EnsembleModel
from obsidian.acquisition import aq_class_dict, aq_defaults, aq_hp_defaults, valid_aqs
from obsidian.surrogates import model_class_dict
from obsidian.objectives import Index_Objective, Objective_Sequence
Expand All @@ -18,7 +18,7 @@
from botorch.sampling.index_sampler import IndexSampler
from botorch.models.model_list_gp_regression import ModelListGP
from botorch.models.gpytorch import GPyTorchModel
from botorch.models.model import ModelList
from botorch.models.model import ModelList, Model
from botorch.utils.sampling import draw_sobol_samples
from botorch.utils.multi_objective.box_decompositions.non_dominated import NondominatedPartitioning

Expand Down Expand Up @@ -478,6 +478,7 @@ def _parse_aq_kwargs(self,
hps: dict,
m_batch: int,
target_locs: list[int],
model: Model,
X_t_pending: Tensor | None = None,
objective: MCAcquisitionObjective | None = None) -> dict:
"""
Expand Down Expand Up @@ -533,7 +534,9 @@ def _parse_aq_kwargs(self,
# Noisy aqs require X_train reference
if aq in ['NEI', 'NEHVI', 'NParEGO']:
aq_kwargs['X_baseline'] = X_baseline

if any(isinstance(m, EnsembleModel) for m in model.models):
aq_kwargs['cache_root'] = False

# Hypervolume requires reference point
if aq in ['EHVI', 'NEHVI']:

Expand Down Expand Up @@ -570,6 +573,9 @@ def _parse_aq_kwargs(self,
w = w/torch.sum(torch.abs(w))
aq_kwargs['scalarization_weights'] = w

if aq == 'SF':
aq_kwargs['X_baseline'] = X_baseline

return aq_kwargs

def suggest(self,
Expand Down Expand Up @@ -712,7 +718,7 @@ def suggest(self,
if not isinstance(model, ModelListGP):
samplers = []
for m in model.models:
if isinstance(m, DNN):
if isinstance(m, EnsembleModel):
sampler_i = IndexSampler(sample_shape=torch.Size([optim_samples]), seed=self.seed)
else:
sampler_i = SobolQMCNormalSampler(sample_shape=torch.Size([optim_samples]), seed=self.seed)
Expand Down Expand Up @@ -757,7 +763,9 @@ def suggest(self,
# Use aq_kwargs so that extra unnecessary ones in hps get removed for certain aq funcs
aq_kwargs = {'model': model, 'sampler': sampler, 'X_pending': X_t_pending}

aq_kwargs.update(self._parse_aq_kwargs(aq_str, aq_hps, m_batch, target_locs, X_t_pending, objective))
aq_kwargs.update(self._parse_aq_kwargs(aq_str, aq_hps, m_batch,
target_locs, model,
X_t_pending, objective))

# Raise errors related to certain constraints
if aq_str in ['UCB', 'Mean', 'TS', 'SF', 'SR', 'NIPV']:
Expand Down Expand Up @@ -812,7 +820,10 @@ def suggest(self,

# Hypervolume aqs fail with X_t_pending when optim_sequential=True
if aq_str in ['NEHVI', 'EHVI']:
optim_sequential = False
if optim_sequential and X_t_pending is not None:
warnings.warn('Hypervolume aqs with X_pending require joint optimization. \
Setting optim_sequential to False', UserWarning)
optim_sequential = False

# If it's random search, no need to do optimization; Otherwise, initialize the aq function and optimize
if aq_str == 'RS':
Expand Down Expand Up @@ -978,7 +989,9 @@ def evaluate(self,
# Use aq_kwargs so that extra unnecessary ones in hps get removed for certain aq funcs
aq_kwargs = {'model': model, 'sampler': None, 'X_pending': X_t_pending}

aq_kwargs.update(self._parse_aq_kwargs(aq_str, aq_hps, X_suggest.shape[0], target_locs, X_t_pending, objective))
aq_kwargs.update(self._parse_aq_kwargs(aq_str, aq_hps, X_suggest.shape[0],
target_locs, model,
X_t_pending, objective))

# If it's random search, no need to evaluate aq
if aq_str == 'RS':
Expand Down
42 changes: 38 additions & 4 deletions obsidian/plotting/plotly.py
Original file line number Diff line number Diff line change
Expand Up @@ -10,7 +10,6 @@
import plotly.graph_objects as go
from plotly.graph_objects import Figure
from plotly.subplots import make_subplots
from sklearn.manifold import MDS

import pandas as pd
import numpy as np
Expand Down Expand Up @@ -52,18 +51,34 @@ def visualize_inputs(campaign: Campaign) -> Figure:
+ ['Correlation Matrix']
+ [X.columns[i] for i in range(cols, n_dim)]
)
# if campaign.optimizer is fitted, then X_best_f_idx is identified
if 'X_best_f_idx' in dir(campaign.optimizer):
marker_shapes = ['diamond' if rowInd in [campaign.optimizer.X_best_f_idx] else 'circle' for rowInd in range(campaign.X.shape[0])]
else:
marker_shapes = ['circle']*campaign.X.shape[0]

for i, param in enumerate(X.columns):
row_i = i // cols + 1
col_i = i % cols + 1
fig.add_trace(go.Scatter(x=X.index, y=X[param],
mode='markers', name=param,
marker=dict(color=color_list[i]),
marker=dict(color=color_list[i], symbol=marker_shapes),
showlegend=False),
row=row_i, col=col_i)
fig.update_xaxes(tickvals=np.around(np.linspace(0, campaign.m_exp, 5)),
row=row_i, col=col_i)

# Add note to explain the shape of markers
if hasattr(campaign.optimizer, 'X_best_f_idx'):
fig.add_annotation(
text="Note: The diamond markers denote samples that achieve the best sum of targets.",
showarrow=False,
xref="paper", yref="paper",
x=0,
y=-0.2,
font=dict(style="italic")
)

# Calculate the correlation matrix
X_u = campaign.X_space.unit_map(X)
corr_matrix = X_u.corr()
Expand Down Expand Up @@ -99,6 +114,12 @@ def MDS_plot(campaign: Campaign) -> Figure:
Returns:
fig (Figure): The MDS plot
"""
try:
from sklearn.manifold import MDS
except ImportError:
raise ImportError('The `sklearn` package (>1.0) is required for the MDS plot. \
Please install it using `pip install scikit-learn`')

mds = MDS(n_components=2)
X_mds = mds.fit_transform(campaign.X_space.encode(campaign.X))

Expand Down Expand Up @@ -320,8 +341,9 @@ def factor_plot(optimizer: Optimizer,
Y_mu_ref = Y_pred_ref[y_name+('_t (pred)' if f_transform else ' (pred)')].values
fig.add_trace(go.Scatter(x=X_ref.iloc[:, feature_id].values, y=Y_mu_ref,
mode='markers',
marker=dict(symbol='diamond'),
line={'color': obsidian_colors.teal},
name='Ref'),
name='Reference'),
)
fig.update_xaxes(title_text=X_name)
fig.update_yaxes(title_text=y_name)
Expand Down Expand Up @@ -539,7 +561,19 @@ def optim_progress(campaign: Campaign,
marker=marker_dict,
customdata=campaign.data[X_names],
name='Data'))


# Highlight the best samples
if hasattr(campaign.optimizer, 'X_best_f_idx'):
fig.add_trace(go.Scatter(x=pd.Series(out_exp.iloc[campaign.optimizer.X_best_f_idx, 0]),
y=pd.Series(out_exp.iloc[campaign.optimizer.X_best_f_idx, 1]),
mode='markers',
marker=dict(symbol='diamond-open', size=14),
line={'color': 'black'},
legendgroup='marker_shape', showlegend=True,
name='Best')
)
fig.update_layout(showlegend=True)

template = ["<b>"+str(param.name)+"</b>: "+" %{customdata["+str(i)+"]"
+ (":.3G}"if isinstance(param, Param_Continuous) else "}") + "<br>"
for i, param in enumerate(campaign.X_space)]
Expand Down
1 change: 1 addition & 0 deletions obsidian/surrogates/__init__.py
Original file line number Diff line number Diff line change
Expand Up @@ -4,3 +4,4 @@
from .custom_GP import *
from .custom_torch import *
from .config import *
from .utils import *
27 changes: 11 additions & 16 deletions obsidian/surrogates/botorch.py
Original file line number Diff line number Diff line change
Expand Up @@ -2,6 +2,7 @@

from .base import SurrogateModel
from .config import model_class_dict
from .utils import fit_pytorch

from obsidian.utils import tensordict_to_dict, dict_to_tensordict
from obsidian.exceptions import SurrogateFitError
Expand All @@ -10,11 +11,11 @@
from botorch.fit import fit_gpytorch_mll
from botorch.optim.fit import fit_gpytorch_mll_torch, fit_gpytorch_mll_scipy
from botorch.models.gpytorch import GPyTorchModel
from botorch.models.ensemble import EnsembleModel
from gpytorch.mlls import ExactMarginalLogLikelihood

import torch
import torch.nn as nn
import torch.optim as optim
import numpy as np
import pandas as pd
import warnings
Expand Down Expand Up @@ -156,20 +157,7 @@ def fit(self,
raise SurrogateFitError('BoTorch model failed to fit')
else:
self.loss_fcn = nn.MSELoss()
self.optimizer = optim.Adam(self.torch_model.parameters(), lr=1e-2)

self.torch_model.train()
for epoch in range(200):
self.optimizer.zero_grad()
output = self.torch_model(X_p)
loss = self.loss_fcn(output, y_p)
loss.backward()
self.optimizer.step()

if (epoch % 50 == 0 and self.verbose):
print(f'Epoch {epoch}: Loss {loss.item()}')

self.torch_model.eval()
fit_pytorch(self.torch_model, X_p, y_p, loss_fcn=self.loss_fcn, verbose=self.verbose)

self.is_fit = True

Expand Down Expand Up @@ -229,7 +217,14 @@ def predict(self,
X_p = self._prepare(X)

pred_posterior = self.torch_model.posterior(X_p)
mu = pred_posterior.mean.detach().cpu().squeeze(-1)

# We would prefer to have stability in the mean of ensemble models,
# So, we will not re-sample for prediction but use forward methods
if isinstance(self.torch_model, EnsembleModel):
mu = self.torch_model.forward(X_p).detach()
else:
mu = pred_posterior.mean.detach().cpu().squeeze(-1)

if q is not None:
if (q < 0) or (q > 1):
raise ValueError('Quantile must be between 0 and 1')
Expand Down
Loading

0 comments on commit 35ae52a

Please sign in to comment.