# Copyright (C) 2019 Steven Reyes
# This program is free software; you can redistribute it and/or modify it
# under the terms of the GNU General Public License as published by the
# Free Software Foundation; either version 3 of the License, or (at your
# option) any later version.
#
# This program is distributed in the hope that it will be useful, but
# WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General
# Public License for more details.
#
# You should have received a copy of the GNU General Public License along
# with this program; if not, write to the Free Software Foundation, Inc.,
# 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA.
"""
This modules provides functions for estimating the marginal
likelihood or evidence of a model.
"""
import numpy
from scipy import integrate
[docs]
def arithmetic_mean_estimator(log_likelihood):
"""Returns the log evidence via the prior arithmetic mean estimator (AME).
The logarithm form of AME is used. This is the most basic
evidence estimator, and often requires O(billions) of samples
from the prior.
Parameters
----------
log_likelihood : 1d array of floats
The log likelihood of the data sampled from the prior
distribution.
Returns
-------
float :
Estimation of the log of the evidence.
"""
num_samples = len(log_likelihood)
logl_max = numpy.max(log_likelihood)
log_evidence = 0.
for i, _ in enumerate(log_likelihood):
log_evidence += numpy.exp(log_likelihood[i] - logl_max)
log_evidence = numpy.log(log_evidence)
log_evidence += logl_max - numpy.log(num_samples)
return log_evidence
[docs]
def harmonic_mean_estimator(log_likelihood):
"""Returns the log evidence via posterior harmonic mean estimator (HME).
The logarithm form of HME is used. This method is not
recommended for general use. It is very slow to converge,
formally, has infinite variance, and very error prone.
Not recommended for general use.
Parameters
----------
log_likelihood : 1d array of floats
The log likelihood of the data sampled from the posterior
distribution.
Returns
-------
float :
Estimation of the log of the evidence.
"""
num_samples = len(log_likelihood)
logl_max = numpy.max(-1.0*log_likelihood)
log_evidence = 0.
for i, _ in enumerate(log_likelihood):
log_evidence += numpy.exp(-1.0*log_likelihood[i] + logl_max)
log_evidence = -1.0*numpy.log(log_evidence)
log_evidence += logl_max
log_evidence += numpy.log(num_samples)
return log_evidence
[docs]
def thermodynamic_integration(log_likelihood, betas,
method="simpsons"):
"""Returns the log evidence of the model via thermodynamic integration.
Also returns an estimated standard deviation for the log evidence.
Current options are integration through the trapezoid rule, a
first-order corrected trapezoid rule, and Simpson's rule.
Parameters
----------
log_likelihood : 3d array of shape (betas, walker, iteration)
The log likelihood for each temperature separated by
temperature, walker, and iteration.
betas : 1d array
The inverse temperatures used in the MCMC.
method : {"trapzoid", "trapezoid_corrected", "simpsons"},
optional.
The numerical integration method to use for the
thermodynamic integration. Choices include: "trapezoid",
"trapezoid_corrected", "simpsons", for the trapezoid rule,
the first-order correction to the trapezoid rule, and
Simpson's rule. [Default = "simpsons"]
Returns
-------
log_evidence : float
Estimation of the log of the evidence.
mcmc_std : float
The standard deviation of the log evidence estimate from
Monte-Carlo spread.
"""
# Check if the method of integration is in the list of choices
method_list = ["trapezoid", "trapezoid_corrected", "simpsons"]
if method not in method_list:
raise ValueError("Method %s not supported. Expected %s"
% (method, method_list))
# Read in the data and ensure ordering of data.
# Ascending order sort
order = numpy.argsort(betas)
betas = betas[order]
log_likelihood = log_likelihood[order]
# Assume log likelihood is given in shape of beta, walker,
# and iteration.
log_likelihood = numpy.reshape(log_likelihood,
(len(betas),
len(log_likelihood[0].flatten())))
average_logl = numpy.average(log_likelihood, axis=1)
if method in ("trapezoid", "trapezoid_corrected"):
log_evidence = numpy.trapz(average_logl, betas)
if method == "trapezoid_corrected":
# var_correction holds the derivative correction terms
# See Friel et al. 2014 for expression and derivation.
# https://link.springer.com/article/10.1007/s11222-013-9397-1
var_correction = 0
for i in range(len(betas) - 1):
delta_beta = betas[i+1] - betas[i]
pre_fac_var = (1. / 12.) * (delta_beta ** 2.0)
var_diff = numpy.var(log_likelihood[i+1])
var_diff -= numpy.var(log_likelihood[i])
var_correction -= pre_fac_var * var_diff
# Add the derivative correction term back to the log_evidence
# from the first if statement.
log_evidence += var_correction
elif method == "simpsons":
# beta -> 0 tends to contribute the least to the integral
# so we can sacrifice precision there, rather than near
# beta -> 1. Option even="last" puts trapezoid rule at
# first few points.
log_evidence = integrate.simps(average_logl, betas,
even="last")
# Estimate the Monte Carlo variance of the evidence calculation
# See (Evans, Annis, 2019.)
# https://www.sciencedirect.com/science/article/pii/S0022249617302651
ti_vec = numpy.zeros(len(log_likelihood[0]))
# Get log likelihood chains by sample and not by temperature.
logl_per_samp = []
for i, _ in enumerate(log_likelihood[0]):
logl_per_samp.append([log_likelihood[x][i] for x in range(len(betas))])
if method in ("trapezoid", "trapezoid_corrected"):
for i, _ in enumerate(log_likelihood[0]):
ti_vec[i] = numpy.trapz(logl_per_samp[i], betas)
elif method == "simpsons":
for i, _ in enumerate(log_likelihood[0]):
ti_vec[i] = integrate.simps(logl_per_samp[i], betas,
even="last")
# Standard error is sample std / sqrt(number of samples)
mcmc_std = numpy.std(ti_vec) / numpy.sqrt(float(len(log_likelihood[0])))
return log_evidence, mcmc_std
[docs]
def stepping_stone_algorithm(log_likelihood, betas):
"""Returns the log evidence of the model via stepping stone algorithm.
Also returns an estimated standard deviation for the log evidence.
Parameters
----------
log_likelihood : 3d array of shape (betas, walker, iteration)
The log likelihood for each temperature separated by
temperature, walker, and iteration.
betas : 1d array
The inverse temperatures used in the MCMC.
Returns
-------
log_evidence : float
Estimation of the log of the evidence.
mcmc_std : float
The standard deviation of the log evidence estimate from
Monte-Carlo spread.
"""
# Reverse order sort
order = numpy.argsort(betas)[::-1]
betas = betas[order]
log_likelihood = log_likelihood[order]
# Assume log likelihood is given in shape of beta,
# walker, iteration.
log_likelihood = numpy.reshape(log_likelihood,
(len(betas),
len(log_likelihood[0].flatten())))
log_rk_pb = numpy.zeros(len(betas) - 1)
for i in range(len(betas) - 1):
delta_beta = betas[i] - betas[i+1]
# Max log likelihood for beta [i+1]
max_logl_pb = numpy.max(log_likelihood[i+1])
val_1 = delta_beta * max_logl_pb
val_2 = delta_beta * (log_likelihood[i+1] - max_logl_pb)
val_2 = numpy.log(numpy.average(numpy.exp(val_2)))
log_rk_pb[i] = val_1 + val_2
log_rk = numpy.sum(log_rk_pb)
log_evidence = log_rk
# Calculate the Monte Carlo variation
mcmc_std = 0
for i in range(len(betas) - 1):
delta_beta = betas[i] - betas[i+1]
pre_fact = (delta_beta * log_likelihood[i+1]) - log_rk_pb[i]
pre_fact = numpy.exp(pre_fact) - 1.0
val = numpy.sum(pre_fact ** 2)
mcmc_std += val
mcmc_std /= float(len(log_likelihood[0])) ** 2.0
mcmc_std = numpy.sqrt(mcmc_std)
return log_evidence, mcmc_std