-
-
Notifications
You must be signed in to change notification settings - Fork 2.1k
/
Copy pathgibbs.py
105 lines (83 loc) · 3.1 KB
/
gibbs.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
# Copyright 2020 The PyMC Developers
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""
Created on May 12, 2012
@author: john
"""
from warnings import warn
import aesara.tensor as at
from aesara.graph.basic import graph_inputs
from numpy import arange, array, cumsum, empty, exp, max, nested_iters, searchsorted
from numpy.random import uniform
from pymc3.distributions import logpt
from pymc3.distributions.discrete import Categorical
from pymc3.model import modelcontext
from pymc3.step_methods.arraystep import ArrayStep, Competence
__all__ = ["ElemwiseCategorical"]
class ElemwiseCategorical(ArrayStep):
"""
Gibbs sampling for categorical variables that only have ElemwiseCategoricalise effects
the variable can't be indexed into or transposed or anything otherwise that will mess things up
"""
# TODO: It would be great to come up with a way to make
# ElemwiseCategorical more general (handling more complex elementwise
# variables)
def __init__(self, vars, values=None, model=None):
warn(
"ElemwiseCategorical is deprecated, switch to CategoricalGibbsMetropolis.",
DeprecationWarning,
stacklevel=2,
)
model = modelcontext(model)
self.var = vars[0]
# XXX: This needs to be refactored
self.sh = None # ones(self.var.dshape, self.var.dtype)
if values is None:
self.values = arange(self.var.distribution.k)
else:
self.values = values
super().__init__(vars, [elemwise_logp(model, self.var)])
def astep(self, q, logp):
p = array([logp(v * self.sh) for v in self.values])
# XXX: This needs to be refactored
shape = None # self.var.dshape
return categorical(p, shape)
@staticmethod
def competence(var, has_grad):
dist = getattr(var.owner, "op", None)
if isinstance(dist, Categorical):
return Competence.COMPATIBLE
return Competence.INCOMPATIBLE
def elemwise_logp(model, var):
terms = []
for v in model.basic_RVs:
v_logp = logpt(v)
if var in graph_inputs([v_logp]):
terms.append(v_logp)
return model.fn(at.add(*terms))
def categorical(prob, shape):
out = empty([1] + list(shape))
n = len(shape)
it0, it1 = nested_iters(
[prob, out],
[list(range(1, n + 1)), [0]],
op_flags=[["readonly"], ["readwrite"]],
flags=["reduce_ok"],
)
for _ in it0:
p, o = it1.itviews
p = cumsum(exp(p - max(p, axis=0)))
r = uniform() * p[-1]
o[0] = searchsorted(p, r)
return out[0, ...]