Documentation ¶
Overview ¶
Package distmv provides multivariate random distribution types.
Index ¶
- func NormalLogProb(x, mu []float64, chol *mat.Cholesky) float64
- func NormalRand(dst, mean []float64, chol *mat.Cholesky, src rand.Source) []float64
- func NormalRandCov(dst, mean []float64, cov mat.Symmetric, src rand.Source) []float64
- type Bhattacharyya
- type CrossEntropy
- type Dirichlet
- type EigenSym
- type Hellinger
- type KullbackLeibler
- type LogProber
- type Normal
- func (n *Normal) ConditionNormal(observed []int, values []float64, src rand.Source) (*Normal, bool)
- func (n *Normal) CovarianceMatrix(dst *mat.SymDense)
- func (n *Normal) Dim() int
- func (n *Normal) Entropy() float64
- func (n *Normal) LogProb(x []float64) float64
- func (n *Normal) MarginalNormal(vars []int, src rand.Source) (*Normal, bool)
- func (n *Normal) MarginalNormalSingle(i int, src rand.Source) distuv.Normal
- func (n *Normal) Mean(dst []float64) []float64
- func (n *Normal) Prob(x []float64) float64
- func (n *Normal) Quantile(dst, p []float64) []float64
- func (n *Normal) Rand(dst []float64) []float64
- func (n *Normal) ScoreInput(dst, x []float64) []float64
- func (n *Normal) SetMean(mu []float64)
- func (n *Normal) TransformNormal(dst, x []float64) []float64
- type PositivePartEigenSym
- func (ed *PositivePartEigenSym) At(i, j int) float64
- func (ed *PositivePartEigenSym) Dims() (r, c int)
- func (ed *PositivePartEigenSym) RawQ() mat.Matrix
- func (ed *PositivePartEigenSym) RawValues() []float64
- func (ed *PositivePartEigenSym) SymmetricDim() int
- func (ed *PositivePartEigenSym) T() mat.Matrix
- type Quantiler
- type RandLogProber
- type Rander
- type Renyi
- type StudentsT
- func (s *StudentsT) ConditionStudentsT(observed []int, values []float64, src rand.Source) (dist *StudentsT, ok bool)
- func (st *StudentsT) CovarianceMatrix(dst *mat.SymDense)
- func (s *StudentsT) Dim() int
- func (s *StudentsT) LogProb(y []float64) float64
- func (s *StudentsT) MarginalStudentsT(vars []int, src rand.Source) (dist *StudentsT, ok bool)
- func (s *StudentsT) MarginalStudentsTSingle(i int, src rand.Source) distuv.StudentsT
- func (s *StudentsT) Mean(dst []float64) []float64
- func (s *StudentsT) Nu() float64
- func (s *StudentsT) Prob(y []float64) float64
- func (s *StudentsT) Rand(dst []float64) []float64
- type Uniform
- func (u *Uniform) Bounds(bounds []r1.Interval) []r1.Interval
- func (u *Uniform) CDF(dst, x []float64) []float64
- func (u *Uniform) Dim() int
- func (u *Uniform) Entropy() float64
- func (u *Uniform) LogProb(x []float64) float64
- func (u *Uniform) Mean(dst []float64) []float64
- func (u *Uniform) Prob(x []float64) float64
- func (u *Uniform) Quantile(dst, p []float64) []float64
- func (u *Uniform) Rand(dst []float64) []float64
- type Wasserstein
Constants ¶
This section is empty.
Variables ¶
This section is empty.
Functions ¶
func NormalLogProb ¶
NormalLogProb computes the log probability of the location x for a Normal distribution the given mean and Cholesky decomposition of the covariance matrix. NormalLogProb panics if len(x) is not equal to len(mu), or if len(mu) != chol.Size().
This function saves time and memory if the Cholesky decomposition is already available. Otherwise, the NewNormal function should be used.
func NormalRand ¶
NormalRand generates a random sample from a multivariate normal distributon given by the mean and the Cholesky factorization of the covariance matrix.
If dst is not nil, the sample will be stored in-place into dst and returned, otherwise a new slice will be allocated first. If dst is not nil, it must have length equal to the dimension of the distribution.
This function saves time and memory if the Cholesky factorization is already available. Otherwise, the NewNormal function should be used.
func NormalRandCov ¶ added in v0.15.0
NormalRandCov generates a random sample from a multivariate normal distribution given by the mean and the covariance matrix.
If dst is not nil, the sample will be stored in-place into dst and returned, otherwise a new slice will be allocated first. If dst is not nil, it must have length equal to the dimension of the distribution.
cov should be *mat.Cholesky, *mat.PivotedCholesky or EigenSym, otherwise NormalRandCov will be very inefficient because a pivoted Cholesky factorization of cov will be computed for every sample.
If cov is an EigenSym, all eigenvalues returned by RawValues must be non-negative, otherwise NormalRandCov will panic.
Types ¶
type Bhattacharyya ¶
type Bhattacharyya struct{}
Bhattacharyya is a type for computing the Bhattacharyya distance between probability distributions.
The Bhattacharyya distance is defined as
D_B = -ln(BC(l,r)) BC = \int_-∞^∞ (p(x)q(x))^(1/2) dx
Where BC is known as the Bhattacharyya coefficient. The Bhattacharyya distance is related to the Hellinger distance by
H(l,r) = sqrt(1-BC(l,r))
For more information, see
https://en.wikipedia.org/wiki/Bhattacharyya_distance
func (Bhattacharyya) DistNormal ¶
func (Bhattacharyya) DistNormal(l, r *Normal) float64
DistNormal computes the Bhattacharyya distance between normal distributions l and r. The dimensions of the input distributions must match or DistNormal will panic.
For Normal distributions, the Bhattacharyya distance is
Σ = (Σ_l + Σ_r)/2 D_B = (1/8)*(μ_l - μ_r)ᵀ*Σ^-1*(μ_l - μ_r) + (1/2)*ln(det(Σ)/(det(Σ_l)*det(Σ_r))^(1/2))
func (Bhattacharyya) DistUniform ¶
func (Bhattacharyya) DistUniform(l, r *Uniform) float64
DistUniform computes the Bhattacharyya distance between uniform distributions l and r. The dimensions of the input distributions must match or DistUniform will panic.
type CrossEntropy ¶
type CrossEntropy struct{}
CrossEntropy is a type for computing the cross-entropy between probability distributions.
The cross-entropy is defined as
- \int_x l(x) log(r(x)) dx = KL(l || r) + H(l)
where KL is the Kullback-Leibler divergence and H is the entropy. For more information, see
https://en.wikipedia.org/wiki/Cross_entropy
func (CrossEntropy) DistNormal ¶
func (CrossEntropy) DistNormal(l, r *Normal) float64
DistNormal returns the cross-entropy between normal distributions l and r. The dimensions of the input distributions must match or DistNormal will panic.
type Dirichlet ¶
type Dirichlet struct {
// contains filtered or unexported fields
}
Dirichlet implements the Dirichlet probability distribution.
The Dirichlet distribution is a continuous probability distribution that generates elements over the probability simplex, i.e. ||x||_1 = 1. The Dirichlet distribution is the conjugate prior to the categorical distribution and the multivariate version of the beta distribution. The probability of a point x is
1/Beta(α) \prod_i x_i^(α_i - 1)
where Beta(α) is the multivariate Beta function (see the mathext package).
For more information see https://en.wikipedia.org/wiki/Dirichlet_distribution
func NewDirichlet ¶
NewDirichlet creates a new dirichlet distribution with the given parameters alpha. NewDirichlet will panic if len(alpha) == 0, or if any alpha is <= 0.
func (*Dirichlet) CovarianceMatrix ¶
CovarianceMatrix calculates the covariance matrix of the distribution, storing the result in dst. Upon return, the value at element {i, j} of the covariance matrix is equal to the covariance of the i^th and j^th variables.
covariance(i, j) = E[(x_i - E[x_i])(x_j - E[x_j])]
If the dst matrix is empty it will be resized to the correct dimensions, otherwise dst must match the dimension of the receiver or CovarianceMatrix will panic.
func (*Dirichlet) LogProb ¶
LogProb computes the log of the pdf of the point x.
It does not check that ||x||_1 = 1.
func (*Dirichlet) Mean ¶
Mean returns the mean of the probability distribution.
If dst is not nil, the mean will be stored in-place into dst and returned, otherwise a new slice will be allocated first. If dst is not nil, it must have length equal to the dimension of the distribution.
type EigenSym ¶ added in v0.15.0
type EigenSym interface { mat.Symmetric // RawValues returns all eigenvalues in ascending order. The returned slice // must not be modified. RawValues() []float64 // RawQ returns an orthogonal matrix whose columns contain the eigenvectors. // The returned matrix must not be modified. RawQ() mat.Matrix }
EigenSym is an eigendecomposition of a symmetric matrix.
type Hellinger ¶
type Hellinger struct{}
Hellinger is a type for computing the Hellinger distance between probability distributions.
The Hellinger distance is defined as
H^2(l,r) = 1/2 * int_x (\sqrt(l(x)) - \sqrt(r(x)))^2 dx
and is bounded between 0 and 1. Note the above formula defines the squared Hellinger distance, while this returns the Hellinger distance itself. The Hellinger distance is related to the Bhattacharyya distance by
H^2 = 1 - exp(-D_B)
For more information, see
https://en.wikipedia.org/wiki/Hellinger_distance
func (Hellinger) DistNormal ¶
DistNormal returns the Hellinger distance between normal distributions l and r. The dimensions of the input distributions must match or DistNormal will panic.
See the documentation of Bhattacharyya.DistNormal for the formula for Normal distributions.
type KullbackLeibler ¶
type KullbackLeibler struct{}
KullbackLeibler is a type for computing the Kullback-Leibler divergence from l to r.
The Kullback-Leibler divergence is defined as
D_KL(l || r ) = \int_x p(x) log(p(x)/q(x)) dx
Note that the Kullback-Leibler divergence is not symmetric with respect to the order of the input arguments.
func (KullbackLeibler) DistDirichlet ¶
func (KullbackLeibler) DistDirichlet(l, r *Dirichlet) float64
DistDirichlet returns the Kullback-Leibler divergence between Dirichlet distributions l and r. The dimensions of the input distributions must match or DistDirichlet will panic.
For two Dirichlet distributions, the KL divergence is computed as
D_KL(l || r) = log Γ(α_0_l) - \sum_i log Γ(α_i_l) - log Γ(α_0_r) + \sum_i log Γ(α_i_r) + \sum_i (α_i_l - α_i_r)(ψ(α_i_l)- ψ(α_0_l))
Where Γ is the gamma function, ψ is the digamma function, and α_0 is the sum of the Dirichlet parameters.
func (KullbackLeibler) DistNormal ¶
func (KullbackLeibler) DistNormal(l, r *Normal) float64
DistNormal returns the KullbackLeibler divergence between normal distributions l and r. The dimensions of the input distributions must match or DistNormal will panic.
For two normal distributions, the KL divergence is computed as
D_KL(l || r) = 0.5*[ln(|Σ_r|) - ln(|Σ_l|) + (μ_l - μ_r)ᵀ*Σ_r^-1*(μ_l - μ_r) + tr(Σ_r^-1*Σ_l)-d]
func (KullbackLeibler) DistUniform ¶
func (KullbackLeibler) DistUniform(l, r *Uniform) float64
DistUniform returns the KullbackLeibler divergence between uniform distributions l and r. The dimensions of the input distributions must match or DistUniform will panic.
type Normal ¶
type Normal struct {
// contains filtered or unexported fields
}
Normal is a multivariate normal distribution (also known as the multivariate Gaussian distribution). Its pdf in k dimensions is given by
(2 π)^(-k/2) |Σ|^(-1/2) exp(-1/2 (x-μ)'Σ^-1(x-μ))
where μ is the mean vector and Σ the covariance matrix. Σ must be symmetric and positive definite. Use NewNormal to construct.
func NewNormal ¶
NewNormal creates a new Normal with the given mean and covariance matrix. NewNormal panics if len(mu) == 0, or if len(mu) != sigma.N. If the covariance matrix is not positive-definite, the returned boolean is false.
func NewNormalChol ¶
NewNormalChol creates a new Normal distribution with the given mean and covariance matrix represented by its Cholesky decomposition. NewNormalChol panics if len(mu) is not equal to chol.Size().
func NewNormalPrecision ¶
NewNormalPrecision creates a new Normal distribution with the given mean and precision matrix (inverse of the covariance matrix). NewNormalPrecision panics if len(mu) is not equal to prec.SymmetricDim(). If the precision matrix is not positive-definite, NewNormalPrecision returns nil for norm and false for ok.
func (*Normal) ConditionNormal ¶
ConditionNormal returns the Normal distribution that is the receiver conditioned on the input evidence. The returned multivariate normal has dimension n - len(observed), where n is the dimension of the original receiver. The updated mean and covariance are
mu = mu_un + sigma_{ob,un}ᵀ * sigma_{ob,ob}^-1 (v - mu_ob) sigma = sigma_{un,un} - sigma_{ob,un}ᵀ * sigma_{ob,ob}^-1 * sigma_{ob,un}
where mu_un and mu_ob are the original means of the unobserved and observed variables respectively, sigma_{un,un} is the unobserved subset of the covariance matrix, sigma_{ob,ob} is the observed subset of the covariance matrix, and sigma_{un,ob} are the cross terms. The elements of x_2 have been observed with values v. The dimension order is preserved during conditioning, so if the value of dimension 1 is observed, the returned normal represents dimensions {0, 2, ...} of the original Normal distribution.
ConditionNormal returns {nil, false} if there is a failure during the update. Mathematically this is impossible, but can occur with finite precision arithmetic.
func (*Normal) CovarianceMatrix ¶
CovarianceMatrix stores the covariance matrix of the distribution in dst. Upon return, the value at element {i, j} of the covariance matrix is equal to the covariance of the i^th and j^th variables.
covariance(i, j) = E[(x_i - E[x_i])(x_j - E[x_j])]
If the dst matrix is empty it will be resized to the correct dimensions, otherwise dst must match the dimension of the receiver or CovarianceMatrix will panic.
func (*Normal) MarginalNormal ¶
MarginalNormal returns the marginal distribution of the given input variables. That is, MarginalNormal returns
p(x_i) = \int_{x_o} p(x_i | x_o) p(x_o) dx_o
where x_i are the dimensions in the input, and x_o are the remaining dimensions. See https://en.wikipedia.org/wiki/Marginal_distribution for more information.
The input src is passed to the call to NewNormal.
func (*Normal) MarginalNormalSingle ¶
MarginalNormalSingle returns the marginal of the given input variable. That is, MarginalNormal returns
p(x_i) = \int_{x_¬i} p(x_i | x_¬i) p(x_¬i) dx_¬i
where i is the input index. See https://en.wikipedia.org/wiki/Marginal_distribution for more information.
The input src is passed to the constructed distuv.Normal.
func (*Normal) Mean ¶
Mean returns the mean of the probability distribution.
If dst is not nil, the mean will be stored in-place into dst and returned, otherwise a new slice will be allocated first. If dst is not nil, it must have length equal to the dimension of the distribution.
func (*Normal) Quantile ¶
Quantile returns the value of the multi-dimensional inverse cumulative distribution function at p.
If dst is not nil, the quantile will be stored in-place into dst and returned, otherwise a new slice will be allocated first. If dst is not nil, it must have length equal to the dimension of the distribution. Quantile will also panic if the length of p is not equal to the dimension of the distribution.
All of the values of p must be between 0 and 1, inclusive, or Quantile will panic.
func (*Normal) Rand ¶
Rand generates a random sample according to the distributon.
If dst is not nil, the sample will be stored in-place into dst and returned, otherwise a new slice will be allocated first. If dst is not nil, it must have length equal to the dimension of the distribution.
func (*Normal) ScoreInput ¶
ScoreInput returns the gradient of the log-probability with respect to the input x. That is, ScoreInput computes
∇_x log(p(x))
If dst is not nil, the score will be stored in-place into dst and returned, otherwise a new slice will be allocated first. If dst is not nil, it must have length equal to the dimension of the distribution.
func (*Normal) SetMean ¶
SetMean changes the mean of the normal distribution. SetMean panics if len(mu) does not equal the dimension of the normal distribution.
func (*Normal) TransformNormal ¶
TransformNormal transforms x generated from a standard multivariate normal into a vector that has been generated under the normal distribution of the receiver.
If dst is not nil, the result will be stored in-place into dst and returned, otherwise a new slice will be allocated first. If dst is not nil, it must have length equal to the dimension of the distribution. TransformNormal will also panic if the length of x is not equal to the dimension of the receiver.
type PositivePartEigenSym ¶ added in v0.15.0
type PositivePartEigenSym struct {
// contains filtered or unexported fields
}
PositivePartEigenSym is an EigenSym that sets any negative eigenvalues from the given eigendecomposition to zero but otherwise returns the values unchanged.
This is useful for filtering eigenvalues of positive semi-definite matrices that are almost zero but negative due to rounding errors.
func NewPositivePartEigenSym ¶ added in v0.15.0
func NewPositivePartEigenSym(ed *mat.EigenSym) *PositivePartEigenSym
NewPositivePartEigenSym returns a new PositivePartEigenSym, wrapping the given eigendecomposition.
func (*PositivePartEigenSym) At ¶ added in v0.15.0
func (ed *PositivePartEigenSym) At(i, j int) float64
At returns the value from the wrapped eigendecomposition.
func (*PositivePartEigenSym) Dims ¶ added in v0.15.0
func (ed *PositivePartEigenSym) Dims() (r, c int)
Dims returns the dimensions from the wrapped eigendecomposition.
func (*PositivePartEigenSym) RawQ ¶ added in v0.15.0
func (ed *PositivePartEigenSym) RawQ() mat.Matrix
RawQ returns the orthogonal matrix Q from the wrapped eigendecomposition. The returned matrix must not be modified.
func (*PositivePartEigenSym) RawValues ¶ added in v0.15.0
func (ed *PositivePartEigenSym) RawValues() []float64
RawValues returns the eigenvalues from the wrapped eigendecomposition in ascending order with any negative value replaced by zero. The returned slice must not be modified.
func (*PositivePartEigenSym) SymmetricDim ¶ added in v0.15.0
func (ed *PositivePartEigenSym) SymmetricDim() int
SymmetricDim returns the value from the wrapped eigendecomposition.
func (*PositivePartEigenSym) T ¶ added in v0.15.0
func (ed *PositivePartEigenSym) T() mat.Matrix
T returns the transpose from the wrapped eigendecomposition.
type Quantiler ¶
Quantiler returns the multi-dimensional inverse cumulative distribution function. len(x) must equal len(p), and if x is non-nil, len(x) must also equal len(p). If x is nil, a new slice will be allocated and returned, otherwise the quantile will be stored in-place into x. All of the values of p must be between 0 and 1, or Quantile will panic.
type RandLogProber ¶
RandLogProber is both a Rander and a LogProber.
type Rander ¶
Rander generates a random number according to the distributon. If the input is non-nil, len(x) must equal len(p) and the dimension of the distribution, otherwise Quantile will panic. If the input is nil, a new slice will be allocated and returned.
type Renyi ¶
type Renyi struct {
Alpha float64
}
Renyi is a type for computing the Rényi divergence of order α from l to r.
The Rényi divergence with α > 0, α ≠ 1 is defined as
D_α(l || r) = 1/(α-1) log(\int_-∞^∞ l(x)^α r(x)^(1-α)dx)
The Rényi divergence has special forms for α = 0 and α = 1. This type does not implement α = ∞. For α = 0,
D_0(l || r) = -log \int_-∞^∞ r(x)1{p(x)>0} dx
that is, the negative log probability under r(x) that l(x) > 0. When α = 1, the Rényi divergence is equal to the Kullback-Leibler divergence. The Rényi divergence is also equal to half the Bhattacharyya distance when α = 0.5.
The parameter α must be in 0 ≤ α < ∞ or the distance functions will panic.
func (Renyi) DistNormal ¶
DistNormal returns the Rényi divergence between normal distributions l and r. The dimensions of the input distributions must match or DistNormal will panic.
For two normal distributions, the Rényi divergence is computed as
Σ_α = (1-α) Σ_l + αΣ_r D_α(l||r) = α/2 * (μ_l - μ_r)'*Σ_α^-1*(μ_l - μ_r) + 1/(2(α-1))*ln(|Σ_λ|/(|Σ_l|^(1-α)*|Σ_r|^α))
For a more nicely formatted version of the formula, see Eq. 15 of
Kolchinsky, Artemy, and Brendan D. Tracey. "Estimating Mixture Entropy with Pairwise Distances." arXiv preprint arXiv:1706.02419 (2017).
Note that the this formula is for Chernoff divergence, which differs from Rényi divergence by a factor of 1-α. Also be aware that most sources in the literature report this formula incorrectly.
type StudentsT ¶
type StudentsT struct {
// contains filtered or unexported fields
}
StudentsT is a multivariate Student's T distribution. It is a distribution over ℝ^n with the probability density
p(y) = (Γ((ν+n)/2) / Γ(ν/2)) * (νπ)^(-n/2) * |Ʃ|^(-1/2) * (1 + 1/ν * (y-μ)ᵀ * Ʃ^-1 * (y-μ))^(-(ν+n)/2)
where ν is a scalar greater than 2, μ is a vector in ℝ^n, and Ʃ is an n×n symmetric positive definite matrix.
In this distribution, ν sets the spread of the distribution, similar to the degrees of freedom in a univariate Student's T distribution. As ν → ∞, the distribution approaches a multi-variate normal distribution. μ is the mean of the distribution, and the covariance is ν/(ν-2)*Ʃ.
See https://en.wikipedia.org/wiki/Student%27s_t-distribution and http://users.isy.liu.se/en/rt/roth/student.pdf for more information.
func NewStudentsT ¶
func NewStudentsT(mu []float64, sigma mat.Symmetric, nu float64, src rand.Source) (dist *StudentsT, ok bool)
NewStudentsT creates a new StudentsT with the given nu, mu, and sigma parameters.
NewStudentsT panics if len(mu) == 0, or if len(mu) != sigma.SymmetricDim(). If the covariance matrix is not positive-definite, nil is returned and ok is false.
func (*StudentsT) ConditionStudentsT ¶
func (s *StudentsT) ConditionStudentsT(observed []int, values []float64, src rand.Source) (dist *StudentsT, ok bool)
ConditionStudentsT returns the Student's T distribution that is the receiver conditioned on the input evidence, and the success of the operation. The returned Student's T has dimension n - len(observed), where n is the dimension of the original receiver. The dimension order is preserved during conditioning, so if the value of dimension 1 is observed, the returned normal represents dimensions {0, 2, ...} of the original Student's T distribution.
ok indicates whether there was a failure during the update. If ok is false the operation failed and dist is not usable. Mathematically this is impossible, but can occur with finite precision arithmetic.
func (*StudentsT) CovarianceMatrix ¶
CovarianceMatrix calculates the covariance matrix of the distribution, storing the result in dst. Upon return, the value at element {i, j} of the covariance matrix is equal to the covariance of the i^th and j^th variables.
covariance(i, j) = E[(x_i - E[x_i])(x_j - E[x_j])]
If the dst matrix is empty it will be resized to the correct dimensions, otherwise dst must match the dimension of the receiver or CovarianceMatrix will panic.
func (*StudentsT) MarginalStudentsT ¶
MarginalStudentsT returns the marginal distribution of the given input variables, and the success of the operation. That is, MarginalStudentsT returns
p(x_i) = \int_{x_o} p(x_i | x_o) p(x_o) dx_o
where x_i are the dimensions in the input, and x_o are the remaining dimensions. See https://en.wikipedia.org/wiki/Marginal_distribution for more information.
The input src is passed to the created StudentsT.
ok indicates whether there was a failure during the marginalization. If ok is false the operation failed and dist is not usable. Mathematically this is impossible, but can occur with finite precision arithmetic.
func (*StudentsT) MarginalStudentsTSingle ¶
MarginalStudentsTSingle returns the marginal distribution of the given input variable. That is, MarginalStudentsTSingle returns
p(x_i) = \int_{x_o} p(x_i | x_o) p(x_o) dx_o
where i is the input index, and x_o are the remaining dimensions. See https://en.wikipedia.org/wiki/Marginal_distribution for more information.
The input src is passed to the call to NewStudentsT.
func (*StudentsT) Mean ¶
Mean returns the mean of the probability distribution.
If dst is not nil, the mean will be stored in-place into dst and returned, otherwise a new slice will be allocated first. If dst is not nil, it must have length equal to the dimension of the distribution.
type Uniform ¶
type Uniform struct {
// contains filtered or unexported fields
}
Uniform represents a multivariate uniform distribution.
func NewUniform ¶
NewUniform creates a new uniform distribution with the given bounds.
func NewUnitUniform ¶
NewUnitUniform creates a new Uniform distribution over the dim-dimensional unit hypercube. That is, a uniform distribution where each dimension has Min = 0 and Max = 1.
func (*Uniform) Bounds ¶
Bounds returns the bounds on the variables of the distribution.
If dst is not nil, the bounds will be stored in-place into dst and returned, otherwise a new slice will be allocated first. If dst is not nil, it must have length equal to the dimension of the distribution.
func (*Uniform) CDF ¶
CDF returns the value of the multidimensional cumulative distribution function of the probability distribution at the point x.
If dst is not nil, the value will be stored in-place into dst and returned, otherwise a new slice will be allocated first. If dst is not nil, it must have length equal to the dimension of the distribution. CDF will also panic if the length of x is not equal to the dimension of the distribution.
func (*Uniform) Mean ¶
Mean returns the mean of the probability distribution.
If dst is not nil, the mean will be stored in-place into dst and returned, otherwise a new slice will be allocated first. If dst is not nil, it must have length equal to the dimension of the distribution.
func (*Uniform) Quantile ¶
Quantile returns the value of the multi-dimensional inverse cumulative distribution function at p.
If dst is not nil, the quantile will be stored in-place into dst and returned, otherwise a new slice will be allocated first. If dst is not nil, it must have length equal to the dimension of the distribution. Quantile will also panic if the length of p is not equal to the dimension of the distribution.
All of the values of p must be between 0 and 1, inclusive, or Quantile will panic.
type Wasserstein ¶
type Wasserstein struct{}
Wasserstein is a type for computing the Wasserstein distance between two probability distributions.
The Wasserstein distance is defined as
W(l,r) := inf 𝔼(||X-Y||_2^2)^1/2
For more information, see
https://en.wikipedia.org/wiki/Wasserstein_metric
func (Wasserstein) DistNormal ¶
func (Wasserstein) DistNormal(l, r *Normal) float64
DistNormal returns the Wasserstein distance between normal distributions l and r. The dimensions of the input distributions must match or DistNormal will panic.
The Wasserstein distance for Normal distributions is
d^2 = ||m_l - m_r||_2^2 + Tr(Σ_l + Σ_r - 2(Σ_l^(1/2)*Σ_r*Σ_l^(1/2))^(1/2))
For more information, see
http://djalil.chafai.net/blog/2010/04/30/wasserstein-distance-between-two-gaussians/