using Distributions
using FillArrays
using StatsPlots
using LinearAlgebra
using Random
# Set a random seed.
Random.seed!(3)
# Define Gaussian mixture model.
= [0.5, 0.5]
w = [-3.5, 0.5]
μ = MixtureModel([MvNormal(Fill(μₖ, 2), I) for μₖ in μ], w)
mixturemodel
# We draw the data points.
= 60
N = rand(mixturemodel, N); x
Gaussian Mixture Models
The following tutorial illustrates the use of Turing for an unsupervised task, namely, clustering data using a Bayesian mixture model. The aim of this task is to infer a latent grouping (hidden structure) from unlabelled data.
Synthetic Data
We generate a synthetic dataset of \(N = 60\) two-dimensional points \(x_i \in \mathbb{R}^2\) drawn from a Gaussian mixture model. For simplicity, we use \(K = 2\) clusters with
- equal weights, i.e., we use mixture weights \(w = [0.5, 0.5]\), and
- isotropic Gaussian distributions of the points in each cluster.
More concretely, we use the Gaussian distributions \(\mathcal{N}([\mu_k, \mu_k]^\mathsf{T}, I)\) with parameters \(\mu_1 = -3.5\) and \(\mu_2 = 0.5\).
The following plot shows the dataset.
scatter(x[1, :], x[2, :]; legend=false, title="Synthetic Dataset")
Gaussian Mixture Model in Turing
We are interested in recovering the grouping from the dataset. More precisely, we want to infer the mixture weights, the parameters \(\mu_1\) and \(\mu_2\), and the assignment of each datum to a cluster for the generative Gaussian mixture model.
In a Bayesian Gaussian mixture model with \(K\) components each data point \(x_i\) (\(i = 1,\ldots,N\)) is generated according to the following generative process. First we draw the model parameters, i.e., in our example we draw parameters \(\mu_k\) for the mean of the isotropic normal distributions and the mixture weights \(w\) of the \(K\) clusters. We use standard normal distributions as priors for \(\mu_k\) and a Dirichlet distribution with parameters \(\alpha_1 = \cdots = \alpha_K = 1\) as prior for \(w\): \[ \begin{aligned} \mu_k &\sim \mathcal{N}(0, 1) \qquad (k = 1,\ldots,K)\\ w &\sim \operatorname{Dirichlet}(\alpha_1, \ldots, \alpha_K) \end{aligned} \] After having constructed all the necessary model parameters, we can generate an observation by first selecting one of the clusters \[ z_i \sim \operatorname{Categorical}(w) \qquad (i = 1,\ldots,N), \] and then drawing the datum accordingly, i.e., in our example drawing \[ x_i \sim \mathcal{N}([\mu_{z_i}, \mu_{z_i}]^\mathsf{T}, I) \qquad (i=1,\ldots,N). \] For more details on Gaussian mixture models, refer to Chapter 9 of Christopher M. Bishop, Pattern Recognition and Machine Learning.
We specify the model in Turing:
using Turing
@model function gaussian_mixture_model(x)
# Draw the parameters for each of the K=2 clusters from a standard normal distribution.
= 2
K ~ MvNormal(Zeros(K), I)
μ
# Draw the weights for the K clusters from a Dirichlet distribution with parameters αₖ = 1.
~ Dirichlet(K, 1.0)
w # Alternatively, one could use a fixed set of weights.
# w = fill(1/K, K)
# Construct categorical distribution of assignments.
= Categorical(w)
distribution_assignments
# Construct multivariate normal distributions of each cluster.
= size(x)
D, N = [MvNormal(Fill(μₖ, D), I) for μₖ in μ]
distribution_clusters
# Draw assignments for each datum and generate it from the multivariate normal distribution.
= Vector{Int}(undef, N)
k for i in 1:N
~ distribution_assignments
k[i] :, i] ~ distribution_clusters[k[i]]
x[end
return k
end
= gaussian_mixture_model(x); model
We run a MCMC simulation to obtain an approximation of the posterior distribution of the parameters \(\mu\) and \(w\) and assignments \(k\). We use a Gibbs
sampler that combines a particle Gibbs sampler for the discrete parameters (assignments \(k\)) and a Hamiltonian Monte Carlo sampler for the continuous parameters (\(\mu\) and \(w\)). We generate multiple chains in parallel using multi-threading.
= Gibbs(:k => PG(100), (:μ, :w) => HMC(0.05, 10))
sampler = 150
nsamples = 4
nchains = 10
burn = sample(model, sampler, MCMCThreads(), nsamples, nchains, discard_initial = burn); chains
The sample()
call above assumes that you have at least two threads available in your Julia instance. If you do not, the multiple chains will run sequentially, and you may notice a warning. For more information, see the Turing documentation on sampling multiple chains.
Inferred Mixture Model
After sampling we can visualize the trace and density of the parameters of interest.
We consider the samples of the location parameters \(\mu_1\) and \(\mu_2\) for the two clusters.
plot(chains[["μ[1]", "μ[2]"]]; legend=true)
From the plots above, we can see that the chains have converged to seemingly different values for the parameters \(\mu_1\) and \(\mu_2\). However, these actually represent the same solution: it does not matter whether we assign \(\mu_1\) to the first cluster and \(\mu_2\) to the second, or vice versa, since the resulting sum is the same. (In principle it is also possible for the parameters to swap places within a single chain, although this does not happen in this example.) For more information see the Stan documentation, or Bishop’s book, where the concept of identifiability is discussed.
Having \(\mu_1\) and \(\mu_2\) swap can complicate the interpretation of the results, especially when different chains converge to different assignments. One solution here is to enforce an ordering on our \(\mu\) vector, requiring \(\mu_k \geq \mu_{k-1}\) for all \(k\). Bijectors.jl
provides a convenient function, ordered()
, which can be applied to a (continuous multivariate) distribution to enforce this:
using Bijectors: ordered
@model function gaussian_mixture_model_ordered(x)
# Draw the parameters for each of the K=2 clusters from a standard normal distribution.
= 2
K ~ ordered(MvNormal(Zeros(K), I))
μ # Draw the weights for the K clusters from a Dirichlet distribution with parameters αₖ = 1.
~ Dirichlet(K, 1.0)
w # Alternatively, one could use a fixed set of weights.
# w = fill(1/K, K)
# Construct categorical distribution of assignments.
= Categorical(w)
distribution_assignments # Construct multivariate normal distributions of each cluster.
= size(x)
D, N = [MvNormal(Fill(μₖ, D), I) for μₖ in μ]
distribution_clusters # Draw assignments for each datum and generate it from the multivariate normal distribution.
= Vector{Int}(undef, N)
k for i in 1:N
~ distribution_assignments
k[i] :, i] ~ distribution_clusters[k[i]]
x[end
return k
end
= gaussian_mixture_model_ordered(x); model
Now, re-running our model, we can see that the assigned means are consistent between chains:
= sample(model, sampler, MCMCThreads(), nsamples, nchains, discard_initial = burn); chains
plot(chains[["μ[1]", "μ[2]"]]; legend=true)
We also inspect the samples of the mixture weights \(w\).
plot(chains[["w[1]", "w[2]"]]; legend=true)
As the distributions of the samples for the parameters \(\mu_1\), \(\mu_2\), \(w_1\), and \(w_2\) are unimodal, we can safely visualize the density region of our model using the average values.
# Model with mean of samples as parameters.
= [mean(chains, "μ[$i]") for i in 1:2]
μ_mean = [mean(chains, "w[$i]") for i in 1:2]
w_mean = MixtureModel([MvNormal(Fill(μₖ, 2), I) for μₖ in μ_mean], w_mean)
mixturemodel_mean contour(
range(-7.5, 3; length=1_000),
range(-6.5, 3; length=1_000),
-> logpdf(mixturemodel_mean, [x, y]);
(x, y) =false,
widen
)scatter!(x[1, :], x[2, :]; legend=false, title="Synthetic Dataset")
Inferred Assignments
Finally, we can inspect the assignments of the data points inferred using Turing. As we can see, the dataset is partitioned into two distinct groups.
= [mean(chains, "k[$i]") for i in 1:N]
assignments scatter(
1, :],
x[2, :];
x[=false,
legend="Assignments on Synthetic Dataset",
title=assignments,
zcolor )
Marginalizing Out The Assignments
We can write out the marginal posterior of (continuous) \(w, \mu\) by summing out the influence of our (discrete) assignments \(z_i\) from our likelihood:
\[p(y \mid w, \mu ) = \sum_{k=1}^K w_k p_k(y \mid \mu_k)\]
In our case, this gives us:
\[p(y \mid w, \mu) = \sum_{k=1}^K w_k \cdot \operatorname{MvNormal}(y \mid \mu_k, I)\]
Marginalizing By Hand
We could implement the above version of the Gaussian mixture model in Turing as follows.
First, Turing uses log-probabilities, so the likelihood above must be converted into log-space:
\[\log \left( p(y \mid w, \mu) \right) = \text{logsumexp} \left[\log (w_k) + \log(\operatorname{MvNormal}(y \mid \mu_k, I)) \right]\]
Where we sum the components with logsumexp
from the LogExpFunctions.jl
package. The manually incremented likelihood can be added to the log-probability with @addlogprob!
, giving us the following model:
using LogExpFunctions
@model function gmm_marginalized(x)
= 2
K = size(x)
D, N ~ ordered(MvNormal(Zeros(K), I))
μ ~ Dirichlet(K, 1.0)
w = [MvNormal(Fill(μₖ, D), I) for μₖ in μ]
dists for i in 1:N
= Vector(undef, K)
lvec for k in 1:K
= (w[k] + logpdf(dists[k], x[:, i]))
lvec[k] end
@addlogprob! logsumexp(lvec)
end
end
When possible, use of @addlogprob!
should be avoided, as it exists outside the usual structure of a Turing model. In most cases, a custom distribution should be used instead.
The next section demonstrates the preferred method: using the MixtureModel
distribution we have seen already to perform the marginalization automatically.
Marginalizing For Free With Distribution.jl’s MixtureModel
Implementation
We can use Turing’s ~
syntax with anything that Distributions.jl
provides logpdf
and rand
methods for. It turns out that the MixtureModel
distribution it provides has, as its logpdf
method, logpdf(MixtureModel([Component_Distributions], weight_vector), Y)
, where Y
can be either a single observation or vector of observations.
In fact, Distributions.jl
provides many convenient constructors for mixture models, allowing further simplification in common special cases.
For example, when mixtures distributions are of the same type, one can write: ~ MixtureModel(Normal, [(μ1, σ1), (μ2, σ2)], w)
, or when the weight vector is known to allocate probability equally, it can be ommited.
The logpdf
implementation for a MixtureModel
distribution is exactly the marginalization defined above, and so our model can be simplified to:
@model function gmm_marginalized(x)
= 2
K = size(x)
D, _ ~ ordered(MvNormal(Zeros(K), I))
μ ~ Dirichlet(K, 1.0)
w ~ MixtureModel([MvNormal(Fill(μₖ, D), I) for μₖ in μ], w)
x end
= gmm_marginalized(x); model
As we have summed out the discrete components, we can perform inference using NUTS()
alone.
= NUTS()
sampler = sample(model, sampler, MCMCThreads(), nsamples, nchains; discard_initial = burn); chains
NUTS()
significantly outperforms our compositional Gibbs sampler, in large part because our model is now Rao-Blackwellized thanks to the marginalization of our assignment parameter.
plot(chains[["μ[1]", "μ[2]"]], legend=true)
Inferred Assignments With The Marginalized Model
As we have summed over possible assignments, the latent parameter representing the assignments is no longer available in our chain. This is not a problem, however, as given any fixed sample \((\mu, w)\), the assignment probability \(p(z_i \mid y_i)\) can be recovered using Bayes’s theorme:
\[p(z_i \mid y_i) = \frac{p(y_i \mid z_i) p(z_i)}{\sum_{k = 1}^K \left(p(y_i \mid z_i) p(z_i) \right)}\]
This quantity can be computed for every \(p(z = z_i \mid y_i)\), resulting in a probability vector, which is then used to sample posterior predictive assignments from a categorial distribution. For details on the mathematics here, see the Stan documentation on latent discrete parameters.
function sample_class(xi, dists, w)
= [(logpdf(d, xi) + log(w[i])) for (i, d) in enumerate(dists)]
lvec rand(Categorical(softmax(lvec)))
end
@model function gmm_recover(x)
= 2
K = size(x)
D, N ~ ordered(MvNormal(Zeros(K), I))
μ ~ Dirichlet(K, 1.0)
w = [MvNormal(Fill(μₖ, D), I) for μₖ in μ]
dists ~ MixtureModel(dists, w)
x # Return assignment draws for each datapoint.
return [sample_class(x[:, i], dists, w) for i in 1:N]
end
We sample from this model as before:
= gmm_recover(x)
model = sample(model, sampler, MCMCThreads(), nsamples, nchains, discard_initial = burn); chains
Given a sample from the marginalized posterior, these assignments can be recovered with:
= mean(returned(gmm_recover(x), chains)); assignments
scatter(
1, :],
x[2, :];
x[=false,
legend="Assignments on Synthetic Dataset - Recovered",
title=assignments,
zcolor )