1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
|
# Copyright (c) 2022 Cisco and/or its affiliates.
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at:
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""Module for numerical integration, tightly coupled to PLRsearch algorithm.
See log_plus for an explanation why None acts as a special case "float" number.
TODO: Separate optimizations specific to PLRsearch and distribute the rest
as a standalone package so other projects may reuse.
"""
import copy
import traceback
import dill
from numpy import random
# TODO: Teach FD.io CSIT to use multiple dirs in PYTHONPATH,
# then switch to absolute imports within PLRsearch package.
# Current usage of relative imports is just a short term workaround.
from . import stat_trackers
def try_estimate_nd(communication_pipe, scale_coeff=8.0, trace_enabled=False):
"""Call estimate_nd but catch any exception and send traceback.
This function does not return anything, computation result
is sent via the communication pipe instead.
TODO: Move scale_coeff to a field of data class
with constructor/factory hiding the default value,
and receive its instance via pipe, instead of argument.
:param communication_pipe: Endpoint for communication with parent process.
:param scale_coeff: Float number to tweak convergence speed with.
:param trace_enabled: Whether to emit trace level debugs.
Keeping trace disabled improves speed and saves memory.
Enable trace only when debugging the computation itself.
:type communication_pipe: multiprocessing.Connection
:type scale_coeff: float
:type trace_enabled: bool
:raises BaseException: Anything raised by interpreter or estimate_nd.
"""
try:
estimate_nd(communication_pipe, scale_coeff, trace_enabled)
except BaseException:
# Any subclass could have caused estimate_nd to stop before sending,
# so we have to catch them all.
traceback_string = traceback.format_exc()
communication_pipe.send(traceback_string)
# After sending, re-raise, so usages other than "one process per call"
# keep behaving correctly.
raise
def generate_sample(averages, covariance_matrix, dimension, scale_coeff):
"""Generate next sample for estimate_nd.
Arguments control the multivariate normal "focus".
Keep generating until the sample point fits into unit area.
:param averages: Coordinates of the focus center.
:param covariance_matrix: Matrix controlling the spread around the average.
:param dimension: If N is dimension, average is N vector and matrix is NxN.
:param scale_coeff: Coefficient to conformally multiply the spread.
:type averages: Indexable of N floats
:type covariance_matrix: Indexable of N indexables of N floats
:type dimension: int
:type scale_coeff: float
:returns: The generated sample point.
:rtype: N-tuple of float
"""
covariance_matrix = copy.deepcopy(covariance_matrix)
for first in range(dimension):
for second in range(dimension):
covariance_matrix[first][second] *= scale_coeff
while 1:
sample_point = random.multivariate_normal(
averages, covariance_matrix, 1
)[0].tolist()
# Multivariate Gauss can fall outside (-1, 1) interval
for first in range(dimension):
sample_coordinate = sample_point[first]
if sample_coordinate <= -1.0 or sample_coordinate >= 1.0:
break
else:
return sample_point
def estimate_nd(communication_pipe, scale_coeff=8.0, trace_enabled=False):
"""Use Bayesian inference from control queue, put result to result queue.
TODO: Use a logging framework that works in a user friendly way.
(Note that multiprocessing_logging does not work well with robot
and robotbackgroundlogger only works for threads, not processes.
Or, wait for https://github.com/robotframework/robotframework/pull/2182
Anyway, the current implementation with trace_enabled looks ugly.)
The result is average and standard deviation for posterior distribution
of a single dependent (scalar, float) value.
The prior is assumed to be uniform on (-1, 1) for every parameter.
Number of parameters and the function for computing
the dependent value and likelihood both come from input.
The likelihood is assumed to be extremely uneven (but never zero),
so the function should return the logarithm of the likelihood.
The integration method is basically a Monte Carlo
(TODO: Add links to notions used here.),
but importance sampling is used in order to focus
on the part of parameter space with (relatively) non-negligible likelihood.
Multivariate Gauss distribution is used for focusing,
so only unimodal posterior distributions are handled correctly.
Initial samples are mostly used for shaping (and shifting)
the Gaussian distribution, later samples will probably dominate.
Thus, initially the algorithm behavior resembles more "find the maximum",
as opposed to "reliably integrate". As for later iterations of PLRsearch,
it is assumed that the distribution position does not change rapidly;
thus integration algorithm returns also the distribution data,
to be used as initial focus in next iteration.
There are workarounds in place that allow old or default focus tracker
to be updated reasonably, even when initial samples
of new iteration have way smaller (or larger) weights.
During the "find the maximum" phase, the focus tracker frequently takes
a wrong shape (compared to observed samples in equilibrium).
Therefore scale_coeff argument is left for humans to tweak,
so the convergence is reliable and quick.
Until the distribution locates itself roughly around
the maximum likeligood point, the integration results are probably wrong.
That means some minimal time is needed for the result to become reliable.
TODO: The folowing is not currently implemented.
The reported standard distribution attempts to signal inconsistence
(when one sample has dominating weight compared to the rest of samples),
but some human supervision is strongly encouraged.
To facilitate running in worker processes, arguments and results
are communicated via a pipe. The computation does not start
until arguments appear in the pipe, the computation stops
when another item (stop object) is detected in the pipe
(and result is put to pipe).
TODO: Create classes for arguments and results,
so their fields are documented (and code perhaps more readable).
Input/argument object (received from pipe)
is a 4-tuple of the following fields:
- dimension: Integer, number of parameters to consider.
- dilled_function: Function (serialized using dill), which:
- - Takes the dimension number of float parameters from (-1, 1).
- - Returns float 2-tuple of dependent value and parameter log-likelihood.
- param_focus_tracker: VectorStatTracker to use for initial focus.
- max_samples: None or a limit for samples to use.
Output/result object (sent to pipe queue)
is a 5-tuple of the following fields:
- value_tracker: ScalarDualStatTracker estimate of value posterior.
- param_focus_tracker: VectorStatTracker to use for initial focus next.
- debug_list: List of debug strings to log at main process.
- trace_list: List of trace strings to pass to main process if enabled.
- samples: Number of samples used in computation (to make it reproducible).
Trace strings are very verbose, it is not recommended to enable them.
In they are not enabled, trace_list will be empty.
It is recommended to edit some lines manually to debug_list if needed.
:param communication_pipe: Endpoint for communication with parent process.
:param scale_coeff: Float number to tweak convergence speed with.
:param trace_enabled: Whether trace list should be populated at all.
:type communication_pipe: multiprocessing.Connection
:type scale_coeff: float
:type trace_enabled: bool
:raises OverflowError: If one sample dominates the rest too much.
Or if value_logweight_function does not handle
some part of parameter space carefully enough.
:raises numpy.linalg.LinAlgError: If the focus shape gets singular
(due to rounding errors). Try changing scale_coeff.
"""
debug_list = list()
trace_list = list()
# Block until input object appears.
dimension, dilled_function, param_focus_tracker, max_samples = (
communication_pipe.recv()
)
debug_list.append(
f"Called with param_focus_tracker {param_focus_tracker!r}"
)
def trace(name, value):
"""
Add a variable (name and value) to trace list (if enabled).
This is a closure (not a pure function),
as it accesses trace_list and trace_enabled
(without any of them being an explicit argument).
:param name: Any string identifying the value.
:param value: Any object to log repr of.
:type name: str
:type value: object
"""
if trace_enabled:
trace_list.append(f"{name} {value!r}")
value_logweight_function = dill.loads(dilled_function)
samples = 0
# Importance sampling produces samples of higher weight (important)
# more frequently, and corrects that by adding weight bonus
# for the less frequently (unimportant) samples.
# But "corrected_weight" is too close to "weight" to be readable,
# so "importance" is used instead, even if it runs contrary to what
# important region is.
value_tracker = stat_trackers.ScalarDualStatTracker()
param_sampled_tracker = stat_trackers.VectorStatTracker(dimension).reset()
if not param_focus_tracker:
# First call has None instead of a real (even empty) tracker.
param_focus_tracker = stat_trackers.VectorStatTracker(dimension)
param_focus_tracker.unit_reset()
else:
# Focus tracker has probably too high weight.
param_focus_tracker.log_sum_weight = None
random.seed(0)
while not communication_pipe.poll():
if max_samples and samples >= max_samples:
break
sample_point = generate_sample(
param_focus_tracker.averages, param_focus_tracker.covariance_matrix,
dimension, scale_coeff
)
trace(u"sample_point", sample_point)
samples += 1
trace(u"samples", samples)
value, log_weight = value_logweight_function(trace, *sample_point)
trace(u"value", value)
trace(u"log_weight", log_weight)
trace(u"focus tracker before adding", param_focus_tracker)
# Update focus related statistics.
param_distance = param_focus_tracker.add_without_dominance_get_distance(
sample_point, log_weight
)
# The code above looked at weight (not importance).
# The code below looks at importance (not weight).
log_rarity = param_distance / 2.0 / scale_coeff
trace(u"log_rarity", log_rarity)
log_importance = log_weight + log_rarity
trace(u"log_importance", log_importance)
value_tracker.add(value, log_importance)
# Update sampled statistics.
param_sampled_tracker.add_get_shift(sample_point, log_importance)
debug_list.append(f"integrator used {samples!s} samples")
debug_list.append(
u" ".join([
u"value_avg", str(value_tracker.average),
u"param_sampled_avg", repr(param_sampled_tracker.averages),
u"param_sampled_cov", repr(param_sampled_tracker.covariance_matrix),
u"value_log_variance", str(value_tracker.log_variance),
u"value_log_secondary_variance",
str(value_tracker.secondary.log_variance)
])
)
communication_pipe.send(
(value_tracker, param_focus_tracker, debug_list, trace_list, samples)
)
|