-
-
Notifications
You must be signed in to change notification settings - Fork 233
/
Copy path3_ask_and_tell.py
81 lines (61 loc) · 2.49 KB
/
3_ask_and_tell.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
"""
Ask-and-Tell
^^^^^^^^^^^^
This examples show how to use the Ask-and-Tell interface.
"""
from ConfigSpace import Configuration, ConfigurationSpace, Float
from smac import HyperparameterOptimizationFacade, Scenario
from smac.runhistory.dataclasses import TrialValue
__copyright__ = "Copyright 2021, AutoML.org Freiburg-Hannover"
__license__ = "3-clause BSD"
class Rosenbrock2D:
@property
def configspace(self) -> ConfigurationSpace:
cs = ConfigurationSpace(seed=0)
x0 = Float("x0", (-5, 10), default=-3)
x1 = Float("x1", (-5, 10), default=-4)
cs.add([x0, x1])
return cs
def train(self, config: Configuration, seed: int = 0) -> float:
"""The 2-dimensional Rosenbrock function as a toy model.
The Rosenbrock function is well know in the optimization community and
often serves as a toy problem. It can be defined for arbitrary
dimensions. The minimium is always at x_i = 1 with a function value of
zero. All input parameters are continuous. The search domain for
all x's is the interval [-5, 10].
"""
x1 = config["x0"]
x2 = config["x1"]
cost = 100.0 * (x2 - x1**2.0) ** 2.0 + (1 - x1) ** 2.0
return cost
if __name__ == "__main__":
model = Rosenbrock2D()
# Scenario object
scenario = Scenario(model.configspace, deterministic=False, n_trials=100)
intensifier = HyperparameterOptimizationFacade.get_intensifier(
scenario,
max_config_calls=1, # We basically use one seed per config only
)
# Now we use SMAC to find the best hyperparameters
smac = HyperparameterOptimizationFacade(
scenario,
model.train,
intensifier=intensifier,
overwrite=True,
)
# We can ask SMAC which trials should be evaluated next
for _ in range(10):
info = smac.ask()
assert info.seed is not None
cost = model.train(info.config, seed=info.seed)
value = TrialValue(cost=cost, time=0.5)
smac.tell(info, value)
# After calling ask+tell, we can still optimize
# Note: SMAC will optimize the next 90 trials because 10 trials already have been evaluated
incumbent = smac.optimize()
# Get cost of default configuration
default_cost = smac.validate(model.configspace.get_default_configuration())
print(f"Default cost: {default_cost}")
# Let's calculate the cost of the incumbent
incumbent_cost = smac.validate(incumbent)
print(f"Incumbent cost: {incumbent_cost}")