Overview and Tutorial

Introduction

This module provides tools for representing, manipulating, and simulating Gaussian random variables numerically. It can deal with individual variables or arbitrarily large sets of variables, correlated or uncorrelated. It also supports complicated (Python) functions of Gaussian variables, automatically propagating uncertainties and correlations through the functions.

A Gaussian variable x represents a Gaussian probability distribution, and is therefore completely characterized by its mean x.mean and standard deviation x.sdev. They are used to represent quantities whose values are uncertain: for example, the mass, 125.7±0.4 GeV, of the recently discovered Higgs boson from particle physics. The following code illustrates a (very) simple application of gvar; it calculates the Higgs boson’s energy when it carries momentum 50±0.15 GeV.

>>> import gvar as gv
>>> m = gv.gvar(125.7, 0.4)             # Higgs boson mass
>>> p = gv.gvar(50, 0.15)               # Higgs boson momentum
>>> E = (p ** 2 +  m ** 2) ** 0.5       # Higgs boson energy
>>> print(m, E)
125.70(40) 135.28(38)
>>> print(E.mean, '+-', E.sdev)
135.279303665 +- 0.375787639425

Here method gvar.gvar() creates objects m and p of type gvar.GVar that represent Gaussian random variables for the Higgs mass and momentum, respectively. The energy E computed from the mass and momentum must, like them, be uncertain and so is also an object of type gvar.GVar — with mean E.mean=135.28 and standard deviation E.sdev=0.38. (Note that gvar uses the compact notation 135.28(38) to represent a Gaussian variable, where the number in parentheses is the uncertainty in the corresponding rightmost digits of the quoted mean value.)

A highly nontrivial feature of gvar.GVars is that they automatically track statistical correlations between different Gaussian variables. In the Higgs boson code above, for example, the uncertainty in the energy is due mostly to the initial uncertainty in the boson’s mass. Consequently statistical fluctuations in the energy are strongly correlated with those in the mass, and largely cancel, for example, in the ratio:

>>> print(E / m)
1.07621(64)

The ratio is 4–5 times more accurate than the either the mass or energy separately.

The correlation between m and E is obvious from their covariance and correlation matrices, both of which have large off-diagonal elements:

>>> print(gv.evalcov([m, E]))           # covariance matrix
[[ 0.16        0.14867019]
[ 0.14867019  0.14121635]]
>>> print(gv.evalcorr([m, E]))          # correlation matrix
[[ 1.          0.98905722]
 [ 0.98905722  1.        ]]

The correlation matrix shows that there is a 98.9% statistical correlation between the mass and energy.

A extreme example of correlation arises if we reconstruct the Higgs boson’s mass from its energy and momentum:

>>> print((E ** 2 - p ** 2) / m ** 2)
1 +- 1.4e-18

The numerator and denominator are completely correlated, indeed identical to machine precision, as they should be. This works only because gvar.GVar object E knows that its uncertainty comes from the uncertainties associated with variables m and p.

We can verify that the uncertainty in the Higgs boson’s energy comes mostly from its mass by creating an error budget for the Higgs energy (and for its energy to mass ratio):

>>> inputs = {'m':m, 'p':p}             # sources of uncertainty
>>> outputs = {'E':E, 'E/m':E/m}        # derived quantities
>>> print(gv.fmt_errorbudget(outputs=outputs, inputs=inputs))
Partial % Errors:
                   E       E/m
------------------------------
        p:      0.04      0.04
        m:      0.27      0.04
------------------------------
    total:      0.28      0.06

For each output (E and E/m), the error budget lists the contribution to the total uncertainty coming from each of the inputs (m and p). The total uncertainty in the energy is ±0.28%, and almost all of that comes from the mass — only ±0.04% comes from the uncertainty in the momentum. The two sources of uncertainty contribute equally, however, to the ratio E/m, which has a total uncertainty of only 0.06%.

This example is relatively simple. Module gvar, however, can easily handle thousands of Gaussian random variables and all of their correlations. These can be combined in arbitrary arithmetic expressions and/or fed through complicated (pure) Python functions, while the gvar.GVars automatically track uncertainties and correlations for and between all of these variables. The code for tracking correlations is the most complex part of the module’s design, particularly since this is done automatically, behind the scenes.

What follows is a tutorial showing how to create gvar.GVars and manipulate them to solve common problems in error propagation. Another way to learn about gvar is to look at the case studies later in the documentation. Each focuses on a single problem, and includes the full code and data, to allow for further experimentation.

gvar was originally written for use by the lsqfit module, which does multidimensional (Bayesian) least-squares fitting. It used to be distributed as part of lsqfit, but is now distributed separately because it is used by other modules (e.g., vegas for multidimensional Monte Carlo integration).

About Printing: The examples in this tutorial use the print function as it is used in Python 3. Drop the outermost parenthesis in each print statement if using Python 2; or add

from __future__ import print_function

at the start of your file.

Gaussian Random Variables

The Higgs boson mass (125.7±0.4 GeV) from the previous section is an example of a Gaussian random variable. As discussed above, such variables x represent Gaussian probability distributions, and therefore are completely characterized by their mean x.mean and standard deviation x.sdev. A mathematical function f(x) of a Gaussian variable is defined as the probability distribution of function values obtained by evaluating the function for random numbers drawn from the original distribution. The distribution of function values is itself approximately Gaussian provided the standard deviation x.sdev of the Gaussian variable is sufficiently small. Thus we can define a function f of a Gaussian variable x to be a Gaussian variable itself, with

f(x).mean = f(x.mean)
f(x).sdev = x.sdev |f'(x.mean)|,

which follows from linearizing the x dependence of f(x) about point x.mean. This formula, together with its multidimensional generalization, lead to a full calculus for Gaussian random variables that assigns Gaussian- variable values to arbitrary arithmetic expressions and functions involving Gaussian variables. This calculus, which is built into gvar, provides the rules for standard error propagation — an important application of Gaussian random variables and of the gvar module.

A multidimensional collection x[i] of Gaussian variables is characterized by the means x[i].mean for each variable, together with a covariance matrix cov[i, j]. Diagonal elements of cov specify the standard deviations of different variables: x[i].sdev = cov[i, i]**0.5. Nonzero off-diagonal elements imply correlations (or anti-correlations) between different variables:

cov[i, j] = <x[i]*x[j]>  -  <x[i]> * <x[j]>

where <y> denotes the expectation value or mean for a random variable y.

Creating Gaussian Variables

Objects of type gvar.GVar are of two types: 1) primary gvar.GVars that are created from means and covariances using gvar.gvar(); and 2) derived gvar.GVars that result from arithmetic expressions or functions involving gvar.GVars. The primary gvar.GVars are the primordial sources of all uncertainties in a gvar code. A single (primary) gvar.GVar is created from its mean xmean and standard deviation xsdev using:

x = gvar.gvar(xmean, xsdev).

This function can also be used to convert strings like "-72.374(22)" or "511.2 +- 0.3" into gvar.GVars: for example,

>>> import gvar
>>> x = gvar.gvar(3.1415, 0.0002)
>>> print(x)
3.14150(20)
>>> x = gvar.gvar("3.1415(2)")
>>> print(x)
3.14150(20)
>>> x = gvar.gvar("3.1415 +- 0.0002")
>>> print(x)
3.14150(20)

Note that x = gvar.gvar(x) is useful when you are unsure whether x is initially a gvar.GVar or a string representing a gvar.GVar.

gvar.GVars are usually more interesting when used to describe multidimensional distributions, especially if there are correlations between different variables. Such distributions are represented by collections of gvar.GVars in one of two standard formats: 1) numpy arrays of gvar.GVars (any shape); or, more flexibly, 2) Python dictionaries whose values are gvar.GVars or arrays of gvar.GVars. Most functions in gvar that handle multiple gvar.GVars work with either format, and if they return multidimensional results do so in the same format as the inputs (that is, arrays or dictionaries). Any dictionary is converted internally into a specialized (ordered) dictionary of type gvar.BufferDict, and dictionary-valued results are also gvar.BufferDicts.

To create an array of gvar.GVars with mean values specified by array xmean and covariance matrix xcov, use

x = gvar.gvar(xmean, xcov)

where array x has the same shape as xmean (and xcov.shape = xmean.shape+xmean.shape). Then each element x[i] of a one-dimensional array, for example, is a gvar.GVar where:

x[i].mean = xmean[i]         # mean of x[i]
x[i].val  = xmean[i]         # same as x[i].mean
x[i].sdev = xcov[i, i]**0.5  # std deviation of x[i]
x[i].var  = xcov[i, i]       # variance of x[i]

As an example,

>>> x, y = gvar.gvar([0.1, 10.], [[0.015625, 0.24], [0.24, 4.]])
>>> print('x =', x, '   y =', y)
x = 0.10(13)    y = 10.0(2.0)

makes x and y gvar.GVars with standard deviations sigma_x=0.125 and sigma_y=2, and a fairly strong statistical correlation:

>>> print(gvar.evalcov([x, y]))     # covariance matrix
[[ 0.015625  0.24    ]
 [ 0.24      4.      ]]
>>> print(gvar.evalcorr([x, y]))    # correlation matrix
[[ 1.    0.96]
 [ 0.96  1.  ]]

Here functions gvar.evalcov() and gvar.evalcorr() compute the covariance and correlation matrices, respectively, of the list of gvar.GVars in their arguments.

gvar.gvar() can also be used to convert strings or tuples stored in arrays or dictionaries into gvar.GVars: for example,

>>> garray = gvar.gvar(['2(1)', '10+-5', (99, 3), gvar.gvar(0, 2)])
>>> print(garray)
[2.0(1.0) 10.0(5.0) 99.0(3.0) 0.0(2.0)]
>>> gdict = gvar.gvar(dict(a='2(1)', b=['10+-5', (99, 3), gvar.gvar(0, 2)]))
>>> print(gdict)
{'a': 2.0(1.0),'b': array([10.0(5.0), 99.0(3.0), 0.0(2.0)], dtype=object)}

If the covariance matrix in gvar.gvar is diagonal, it can be replaced by an array of standard deviations (square roots of diagonal entries in cov). The example above without correlations, therefore, would be:

>>> x, y = gvar.gvar([0.1, 10.], [0.125, 2.])
>>> print('x =', x, '   y =', y)
x = 0.10(12)    y = 10.0(2.0)
>>> print(gvar.evalcov([x, y]))     # covariance matrix
[[ 0.015625  0.      ]
 [ 0.        4.      ]]
>>> print(gvar.evalcorr([x, y]))    # correlation matrix
[[ 1.  0.]
 [ 0.  1.]]

gvar.GVar Arithmetic and Functions

The gvar.GVars discussed in the previous section are all primary gvar.GVars since they were created by specifying their means and covariances explicitly, using gvar.gvar(). What makes gvar.GVars particularly useful is that they can be used in arithemtic expressions (and numeric pure-Python functions), just like Python floats. Such expressions result in new, derived gvar.GVars whose means, standard deviations, and correlations are determined from the covariance matrix of the primary gvar.GVars. The automatic propagation of correlations through arbitrarily complicated arithmetic is an especially useful feature of gvar.GVars.

As an example, again define

>>> x, y = gvar.gvar([0.1, 10.], [0.125, 2.])

and set

>>> f = x + y
>>> print('f =', f)
f = 10.1(2.0)

Then f is a (derived) gvar.GVar whose variance f.var equals

df/dx cov[0, 0] df/dx + 2 df/dx cov[0, 1] df/dy + ... = 2.0039**2

where cov is the original covariance matrix used to define x and y (in gvar.gvar). Note that while f and y separately have 20% uncertainties in this example, the ratio f/y has much smaller errors:

>>> print(f / y)
1.010(13)

This happens, of course, because the errors in f and y are highly correlated — the error in f comes mostly from y. gvar.GVars automatically track correlations even through complicated arithmetic expressions and functions: for example, the following more complicated ratio has a still smaller error, because of stronger correlations between numerator and denominator:

>>> print(gvar.sqrt(f**2 + y**2) / f)
1.4072(87)
>>> print(gvar.evalcorr([f, y]))
[[ 1.          0.99805258]
 [ 0.99805258  1.        ]]
>>> print(gvar.evalcorr([gvar.sqrt(f**2 + y**2), f]))
[[ 1.         0.9995188]
 [ 0.9995188  1.       ]]

The gvar module defines versions of the standard Python mathematical functions that work with gvar.GVar arguments. These include: exp, log, sqrt, sin, cos, tan, arcsin, arccos, arctan, arctan2, sinh, cosh, tanh, arcsinh, arccosh, arctanh, erf, fabs, abs. Numeric functions defined entirely in Python (i.e., pure-Python functions) will likely also work with gvar.GVars.

Numeric functions implemented by modules using low-level languages like C will not work with gvar.GVars. Such functions must be replaced by equivalent code written directly in Python. In some cases it is possible to construct a gvar.GVar-capable function from low-level code for the function and its derivative. For example, the following code defines a new version of the standard Python error function that accepts either floats or gvar.GVars as its argument:

import math
import gvar

def erf(x):
    if isinstance(x, gvar.GVar):
        f = math.erf(x.mean)
        dfdx = 2. * math.exp(- x.mean ** 2) / math.sqrt(math.pi)
        return gvar.gvar_function(x, f, dfdx)
    else:
        return math.erf(x)

Here function gvar.gvar_function() creates the gvar.GVar for a function with mean value f and derivative dfdx at point x. A more complete version of erf is included in gvar.

Some sample numerical analysis codes, adapted for use with gvar.GVars, are described in Numerical Analysis Modules in gvar.

Arithmetic operators + - * / ** == != <> += -= *= /= are all defined for gvar.GVars. Comparison operators are also supported: == != > >= < <=. They are applied to the mean values of gvar.GVars: for example, gvar.gvar(1,1) == gvar.var(1,2) is true, as is gvar.gvar(1,1) > 0. Logically x>y for gvar.GVars should evaluate to a boolean-valued random variable, but such variables are beyond the scope of this module. Comparison operators that act only on the mean values make it easier to implement pure-Python functions that work with either gvar.GVars or floats as arguments.

Implementation Notes: Each gvar.GVar keeps track of three pieces of information: 1) its mean value; 2) its derivatives with respect to the primary gvar.GVars (created by gvar.gvar()); and 3) the location of the covariance matrix for the primary gvar.GVars. The standard deviations and covariances for all gvar.GVars originate with the primary gvar.GVars: any gvar.GVar z satisfies

z = z.mean + sum_p (p - p.mean) * dz/dp

where the sum is over all primary gvar.GVarp. gvar uses this expression to calculate covariances from the derivatives, and the covariance matrix of the primary gvar.GVars. The derivatives for derived gvar.GVars are computed automatically, using automatic differentiation.

The derivative of a gvar.GVar f with respect to a primary gvar.GVar x is obtained from f.deriv(x). A list of derivatives with respect to all primary gvar.GVars is given by f.der, where the order of derivatives is the same as the order in which the primary gvar.GVars were created.

A gvar.GVar can be constructed at a very low level by supplying all the three essential pieces of information — for example,

f = gvar.gvar(fmean, fder, cov)

where fmean is the mean, fder is an array where fder[i] is the derivative of f with respect to the i-th primary gvar.GVar (numbered in the order in which they were created using gvar.gvar()), and cov is the covariance matrix for the primary gvar.GVars (easily obtained from gvar.gvar.cov).

Error Budgets from gvar.GVars

It is sometimes useful to know how much of the uncertainty in a derived quantity is due to a particular input uncertainty. Continuing the example above, for example, we might want to know how much of fs standard deviation is due to the standard deviation of x and how much comes from y. This is easily computed:

>>> x, y = gvar.gvar([0.1, 10.], [0.125, 2.])
>>> f = x + y
>>> print(f.partialsdev(x))        # uncertainty in f due to x
0.125
>>> print(f.partialsdev(y))        # uncertainty in f due to y
2.0
>>> print(f.partialsdev(x, y))     # uncertainty in f due to x and y
2.00390244274
>>> print(f.sdev)                  # should be the same
2.00390244274

This shows, for example, that most (2.0) of the uncertainty in f (2.0039) is from y.

gvar provides a useful tool for compiling an “error budget” for derived gvar.GVars relative to the primary gvar.GVars from which they were constructed: continuing the example above,

>>> outputs = {'f':f, 'f/y':f/y}
>>> inputs = {'x':x, 'y':y}
>>> print(gvar.fmt_values(outputs))
Values:
                f/y: 1.010(13)
                  f: 10.1(2.0)

>>> print(gvar.fmt_errorbudget(outputs=outputs, inputs=inputs))
Partial % Errors:
                 f/y         f
------------------------------
        y:      0.20     19.80
        x:      1.24      1.24
------------------------------
    total:      1.25     19.84

This shows y is responsible for 19.80% of the 19.84% uncertainty in f, but only 0.2% of the 1.25% uncertainty in f/y. The total uncertainty in each case is obtained by adding the x and y contributions in quadrature.

Storing gvar.GVars for Later Use; gvar.BufferDicts

Storing gvar.GVars in a file for later use is complicated by the need to capture the covariances between different gvar.GVars as well as their means. The easiest way to save an array or dictionary or other object g that contains gvar.GVars is to use gvar.dump(): for example,

>>> gvar.dump(g, 'gfile.pkl')

saves the data from g in a Python pickle file named 'gfile.pkl'. To reassemble the the data from g we use:

>>> g = gvar.load('gfile.pkl')

Functions gvar.dump() and gvar.load() are similar to the corresponding functions in Python’s pickle module but they preserve information about the correlations between different gvar.GVars in g. Correlations with gvar.GVars that are not in g are lost, so it is important to include all gvar.GVars of interest in g before saving them.

gvar.GVars can also be pickled easily if they are stored in a gvar.BufferDict since this data type has explicit support for pickling that preserves correlations. So if g is a gvar.BufferDict containing gvar.GVars (and/or arrays of gvar.GVars),

>>> import pickle
>>> pickle.dump(g, open('gfile.pkl', 'wb'))

saves the contents of g to a file named gfile.pkl, and the gvar.GVars are retrieved using

>>> g = pickle.load(open('gfile.pkl', 'rb'))

Of course, gvar.dump(g, 'gfile.pkl') and g = gvar.load('gfile.pkl') are simpler and achieve the same goal.

Using pickle to pickle an object containing gvar.GVars usually generates a warning about lost correlations. This can be ignored if correlations are unimportant. If pickle must be used and correlations matter, gvar.dumps/loads can sometimes be used to make this possible. Consider, for example, a class A that stores gvar.GVars a and b. We might be able to modify it so that pickle uses the gvar routines: for example, the code

class A:
    def __init__(self, a, b):
        if isinstance(a, bytes):
            self.__dict__ = gv.loads(a)
        else:
            self.a = a
            self.b = b
            ...

    def __reduce__(self):
        return A, (gv.dumps(self.__dict__), None)

    ...

has pickle convert the class’s dictionary into a byte stream using gv.dumps(self.__dict__) when pickling. This is reconstituted into a dictionary in A.__init__, using gv.loads(a), upon un-pickling.

Non-Gaussian Expectation Values

By default functions of gvar.GVars are also gvar.GVars, but there are cases where such functions cannot be represented accurately by Gaussian distributions. The product of 0.1(4) and 0.2(5), for example, is not very Gaussian because the standard deviations are large compared to the scale over which the product changes appreciably. In such cases one may want to use the true distribution of the function, instead of its Gaussian approximation, in an analysis.

Class vegas.PDFIntegrator evaluates integrals over multi-dimensional Gaussian probability density functions (PDFs) using the vegas module, which does adaptive multi-dimensional integration. This permits us, for example, to calculate the true mean and standard deviation of a function of Gaussian variables, or to test the extent to which the true distribution of the function is Gaussian. The following code analyzes the distribution of sin(p[0] * p[1]) where p = [0.1(4), 0.2(5)]:

import numpy as np
import gvar as gv
import vegas

p = gv.gvar(['0.1(4)', '0.2(5)'])

# function of interest
def f(p):
    return np.sin(p[0] * p[1])

# histogram for values of f(p)
fhist = gv.PDFHistogram(f(p), nbin=16)

# want expectation value of fstats(p)
def fstats(p):
    fp = f(p)
    return dict(
        moments=[fp, fp ** 2, fp ** 3, fp ** 4],
        histogram=fhist.count(fp),
        )

# evaluate expectation value of fstats in 3 steps
# 1 - create an integrator to evaluate expectation values of functions of p
p_expval = vegas.PDFIntegrator(p)
# 2 - adapt p_expval to the p's PDF (N.B., no function specified)
p_expval(neval=5000, nitn=10)
# 3 - evaluate expectation value of function(s) fhist(p)
results = p_expval(fstats, neval=5000, nitn=10, adapt=False)

# results from expectation value integration
print(results.summary())
print('moments:', results['moments'])
stats = gv.PDFStatistics(
    moments=results['moments'],
    histogram=(fhist.bins, results['histogram']),
    )
print('Statistics from Bayesian integrals:')
print(stats)
print('Gaussian approx:', f(p))

# plot histogram from integration (plt = matplotlib.pyplot)
plt = fhist.make_plot(results['histogram'])
plt.xlabel(r'$\sin(p_0 p_1)$')
plt.xlim(-1, 1)
# add extra curve corresponding to Gaussian with "correct" mean and sdev
correct_fp = gv.gvar(stats.mean.mean, stats.sdev.mean)
x = np.linspace(-1.,1.,50)
pdf = gv.PDF(correct_fp)
y = [pdf(xi) * fhist.widths[0] for xi in x]
plt.plot(x, y, 'k:' )
plt.show()

The key construct here is p_expval which is a vegas integrator designed so that p_expval(f) returns the expectation value of any function f(p) with respect to the probability distribution specified by p = gv.gvar(['0.1(4)', '0.2(5)']). The integrator is adaptive so it is called once without a function, to allow it to adapt to the probability density function (PDF). It is then applied to function fstats(p), which calculates various moments of f(p) as well as information for histogramming values of f(p) (using gvar.PDFHistogram). Parameters nitn and neval control the multidimensional integrator, telling it how many iterations of its adaptive algorithm to use and the maximum number of integrand evaluations to use in each iteration.

The output from this code is:

itn   integral        average         chi2/dof        Q
-------------------------------------------------------
  1   1.00032(90)     1.00032(90)         0.00     1.00
  2   0.9992(10)      0.99976(69)         1.10     0.33
  3   0.9987(10)      0.99942(57)         0.97     0.53
  4   1.00058(92)     0.99971(49)         0.89     0.74
  5   0.99992(99)     0.99975(44)         0.91     0.73
  6   1.00059(99)     0.99989(40)         0.92     0.71
  7   0.99830(96)     0.99966(37)         0.90     0.80
  8   1.00201(88)     0.99996(34)         0.91     0.77
  9   0.9977(12)      0.99971(33)         0.89     0.86
 10   0.9996(10)      0.99970(32)         0.84     0.95

moments: [0.01862(13) 0.043161(90) 0.004672(80) 0.011470(72)]
Statistics from Bayesian integrals:
   mean = 0.01862(13)   sdev = 0.20692(21)   skew = 0.2567(75)   ex_kurt = 3.116(20)
   median = 0.00017(14)   plus = 0.17397(49)   minus = 0.11705(43)
Gaussian approx: 0.020(94)

The table summarizes the integrator’s performance over the nitn=10 iterations it performed to obtain the final results; see the vegas documentation for further information. The expectation values for moments of f(p) are then listed, followed by the mean and standard deviation computed from these moments, as well as the skewness and excess kurtosis of the f(p) distribution. The median value for the distribution is estimated from the histogram, as are the intervals on either side of the median ((median-minus,median) and (median,median+plus)) containing 34% of the probability. Finally the mean and standard deviation in the Gaussian approximate are listed.

The exact mean of the f(p) distribution is 0.0186(1), which is somewhat lower than Gaussian approximation of 0.020. A more important difference is in the standard deviation which is 0.2069(3) for the real distribution, but less than half that size (0.094) in the Gaussian approximation. The real distribution is significantly broader than the Gaussian approximation suggests, though its mean is close. The real distribution also has nonzero skewness (0.26(1)) and excess kurtosis (3.12(2)), which suggest that it is not well described by any Gaussian. (Skewness and excess kurtosis vanish for Gaussian distributions.)

The code also displays a histogram showing the probability distribution for values of f(p):

_images/histogram.png

This shows the actual probability associated with each f(p) bin, together with the shape (red dashed line) expected from the Gaussian approximation (0.020(94)). It also shows the Gaussian distribution corresponding to correct mean and standard deviation (0.019(207)) of the distribution (black dotted line).

Neither Gaussian in this plot is quite right: the first is more accurate close to the maximimum, while the second does better further out. From the histogram we can estimate that 68% of the probability lies within ±0.14 of 0.03, which is probably the best succinct characterization of the uncertainty (0.03(14)).

This example is relatively simple since the underlying Gaussian distribution is only two dimensional. The vegas integrator used here is adaptive and so can function effectively even for high dimensions (10, 20, 50 … Gaussian variables). High dimensions usually cost more, requiring many more function evaluations (neval).

Random Number Generators and Simulations

gvar.GVars represent probability distributions. It is possible to use them to generate random numbers from those distributions. For example, in

>>> z = gvar.gvar(2.0, 0.5)
>>> print(z())
2.29895701465
>>> print(z())
3.00633184275
>>> print(z())
1.92649199321

calls to z() generate random numbers from a Gaussian random number generator with mean z.mean=2.0 and standard deviation z.sdev=0.5.

To obtain random arrays from an array g of gvar.GVars use giter=gvar.raniter(g) (see gvar.raniter()) to create a random array generator giter. Each call to next(giter) generates a new array of random numbers. The random number arrays have the same shape as the array g of gvar.GVars and have the distribution implied by those random variables (including correlations). For example,

>>> a = gvar.gvar(1.0, 1.0)
>>> da = gvar.gvar(0.0, 0.1)
>>> g = [a, a+da]
>>> giter = gvar.raniter(g)
>>> print(next(giter))
[ 1.51874589  1.59987422]
>>> print(next(giter))
[-1.39755111 -1.24780937]
>>> print(next(giter))
[ 0.49840244  0.50643312]

Note how the two random numbers separately vary over the region 1±1 (approximately), but the separation between the two is rarely more than 0±0.1. This is as expected given the strong correlation between a and a+da.

gvar.raniter(g) also works when g is a dictionary (or gvar.BufferDict) whose entries g[k] are gvar.GVars or arrays of gvar.GVars. In such cases the iterator returns a dictionary with the same layout:

>>> g = dict(a=gvar.gvar(0, 1), b=[gvar.gvar(0, 100), gvar.gvar(10, 1e-3)])
>>> print(g)
{'a': 0.0(1.0), 'b': [0(100), 10.0000(10)]}
>>> giter = gvar.raniter(g)
>>> print(next(giter))
{'a': -0.88986130981173306, 'b': array([-67.02994213,   9.99973707])}
>>> print(next(giter))
{'a': 0.21289976681277872, 'b': array([ 29.9351328 ,  10.00008606])}

One use for such random number generators is dealing with situations where the standard deviations are too large to justify the linearization assumed in defining functions of Gaussian variables. Consider, for example,

>>> x = gvar.gvar(1., 3.)
>>> print(cos(x))
0.5(2.5)

The standard deviation for cos(x) is obviously wrong since cos(x) can never be larger than one. We can estimate the the real mean and standard deviation using a simulation. To do this, we: 1) generate a large number of random numbers xi from x; 2) compute cos(xi) for each; and 3) compute the mean and standard deviation for the resulting distribution (or any other statistical quantity, particularly if the resulting distribution is not Gaussian):

# estimate mean,sdev from 1000 random x's
>>> ran_x = numpy.array([x() for _ in range(1000)])
>>> ran_cos = numpy.cos(ran_x)
>>> print('mean =', ran_cos.mean(), '  std dev =', ran_cos.std())
mean = 0.0350548954142   std dev = 0.718647118869

# check by doing more (and different) random numbers
>>> ran_x = numpy.array([x() for _ in range(100000)])
>>> ran_cos = numpy.cos(ran_x)
>>> print('mean =', ran_cos.mean(), '  std dev =', ran_cos.std())
mean = 0.00806276057656   std dev = 0.706357174056

This procedure generalizes trivially for multidimensional analyses, using arrays or dictionaries with gvar.raniter().

Note finally that bootstrap copies of gvar.GVars are easily created. A bootstrap copy of gvar.GVar x ± dx is another gvar.GVar with the same width but where the mean value is replaced by a random number drawn from the original distribution. Bootstrap copies of a data set, described by a collection of gvar.GVars, can be used as new (fake) data sets having the same statistical errors and correlations:

>>> g = gvar.gvar([1.1, 0.8], [[0.01, 0.005], [0.005, 0.01]])
>>> print(g)
[1.10(10) 0.80(10)]
>>> print(gvar.evalcov(g))                  # print covariance matrix
[[ 0.01   0.005]
 [ 0.005  0.01 ]]
>>> gbs_iter = gvar.bootstrap_iter(g)
>>> gbs = next(gbs_iter)                    # bootstrap copy of f
>>> print(gbs)
[1.14(10) 0.90(10)]                         # different means
>>> print(gvar.evalcov(gbs))
[[ 0.01   0.005]                            # same covariance matrix
 [ 0.005  0.01 ]]

Such fake data sets are useful for analyzing non-Gaussian behavior, for example, in nonlinear fits.

Limitations

The most fundamental limitation of this module is that the calculus of Gaussian variables that it assumes is only valid when standard deviations are small (compared to the distances over which the functions of interest change appreciably). One way of dealing with this limitation is to use simulations, as discussed in Random Number Generators and Simulations.

Another potential issue is roundoff error, which can become problematic if there is a wide range of standard deviations among correlated modes. For example, the following code works as expected:

>>> from gvar import gvar, evalcov
>>> tiny = 1e-4
>>> a = gvar(0., 1.)
>>> da = gvar(tiny, tiny)
>>> a, ada = gvar([a.mean, (a+da).mean], evalcov([a, a+da])) # = a,a+da
>>> print(ada-a)   # should be da again
0.00010(10)

Reducing tiny, however, leads to problems:

>>> from gvar import gvar, evalcov
>>> tiny = 1e-8
>>> a = gvar(0., 1.)
>>> da = gvar(tiny, tiny)
>>> a, ada = gvar([a.mean, (a+da).mean], evalcov([a, a+da])) # = a, a+da
>>> print(ada-a)   # should be da again
1(0)e-08

Here the call to gvar.evalcov() creates a new covariance matrix for a and ada = a+da, but the matrix does not have enough numerical precision to encode the size of da’s variance, which gets set, in effect, to zero. The problem arises here for values of tiny less than about 2e-8 (with 64-bit floating point numbers; tiny**2 is what appears in the covariance matrix).

Optimizations

When there are lots of primary gvar.GVars, the number of derivatives stored for each derived gvar.GVar can become rather large, potentially (though not necessarily) leading to slower calculations. One way to alleviate this problem, should it arise, is to separate the primary variables into groups that are never mixed in calculations and to use different gvar.gvar()s when generating the variables in different groups. New versions of gvar.gvar() are obtained using gvar.switch_gvar(): for example,

import gvar
...
x = gvar.gvar(...)
y = gvar.gvar(...)
z = f(x, y)
... other manipulations involving x and y ...
gvar.switch_gvar()
a = gvar.gvar(...)
b = gvar.gvar(...)
c = g(a, b)
... other manipulations involving a and b (but not x and y) ...

Here the gvar.gvar() used to create a and b is a different function than the one used to create x and y. A derived quantity, like c, knows about its derivatives with respect to a and b, and about their covariance matrix; but it carries no derivative information about x and y. Absent the switch_gvar line, c would have information about its derivatives with respect to x and y (zero derivative in both cases) and this would make calculations involving c slightly slower than with the switch_gvar line. Usually the difference is negligible — it used to be more important, in earlier implementations of gvar.GVar before sparse matrices were introduced to keep track of covariances. Note that the previous gvar.gvar() can be restored using gvar.restore_gvar(). Function gvar.gvar_factory() can also be used to create new versions of gvar.gvar().