Today we want to delve deeper into the topic of Inverse Gaussian distribution, a topic that has gained relevance in recent years and that undoubtedly generates great interest among the population. Inverse Gaussian distribution is a topic that covers multiple aspects and has been the subject of constant debate and analysis. In this article, we will explore different perspectives and approaches related to Inverse Gaussian distribution, with the aim of providing a comprehensive view on this topic. From its origins to its impact today, Inverse Gaussian distribution has captured the attention of academics, experts and the general public, being the object of study and interest in various areas. Without a doubt, Inverse Gaussian distribution has become a relevant topic in contemporary society, which is why it is essential to deepen its understanding and scope.
for x > 0, where is the mean and is the shape parameter.[1]
The inverse Gaussian distribution has several properties analogous to a Gaussian distribution. The name can be misleading: it is an "inverse" only in that, while the Gaussian describes a Brownian motion's level at a fixed time, the inverse Gaussian describes the distribution of the time a Brownian motion with positive drift takes to reach a fixed positive level.
Its cumulant generating function (logarithm of the characteristic function)[contradictory] is the inverse of the cumulant generating function of a Gaussian random variable.
To indicate that a random variableX is inverse Gaussian-distributed with mean μ and shape parameter λ we write .
Properties
Single parameter form
The probability density function (pdf) of the inverse Gaussian distribution has a single parameter form given by
In this form, the mean and variance of the distribution are equal,
Also, the cumulative distribution function (cdf) of the single parameter inverse Gaussian distribution is related to the standard normal distribution by
where , and the is the cdf of standard normal distribution. The variables and are related to each other by the identity
In the single parameter form, the MGF simplifies to
An inverse Gaussian distribution in double parameter form can be transformed into a single parameter form by appropriate scaling where
The standard form of inverse Gaussian distribution is
Summation
If Xi has an distribution for i = 1, 2, ..., n
and all Xi are independent, then
Note that
is constant for all i. This is a necessary condition for the summation. Otherwise S would not be Inverse Gaussian distributed.
Suppose that we have a Brownian motion with drift defined by:
And suppose that we wish to find the probability density function for the time when the process first hits some barrier - known as the first passage time. The Fokker-Planck equation describing the evolution of the probability distribution is:
Define a point , such that . This will allow the original and mirror solutions to cancel out exactly at the barrier at each instant in time. This implies that the initial condition should be augmented to become:
where is a constant. Due to the linearity of the BVP, the solution to the Fokker-Planck equation with this initial condition is:
Now we must determine the value of . The fully absorbing boundary condition implies that:
At , we have that . Substituting this back into the above equation, we find that:
Therefore, the full solution to the BVP is:
Now that we have the full probability density function, we are ready to find the first passage time distribution . The simplest route is to first compute the survival function, which is defined as:
where is the cumulative distribution function of the standard normal distribution. The survival function gives us the probability that the Brownian motion process has not crossed the barrier at some time . Finally, the first passage time distribution is obtained from the identity:
Assuming that , the first passage time follows an inverse Gaussian distribution:
When drift is zero
A common special case of the above arises when the Brownian motion has no drift. In that case, parameter μ tends to infinity, and the first passage time for fixed level α has probability density function
publicdoubleinverseGaussian(doublemu,doublelambda){Randomrand=newRandom();doublev=rand.nextGaussian();// Sample from a normal distribution with a mean of 0 and 1 standard deviationdoubley=v*v;doublex=mu+(mu*mu*y)/(2*lambda)-(mu/(2*lambda))*Math.sqrt(4*mu*lambda*y+mu*mu*y*y);doubletest=rand.nextDouble();// Sample from a uniform distribution between 0 and 1if(test<=(mu)/(mu+x))returnx;elsereturn(mu*mu)/x;}
The convolution of an inverse Gaussian distribution (a Wald distribution) and an exponential (an ex-Wald distribution) is used as a model for response times in psychology,[10] with visual search as one example.[11]
History
This distribution appears to have been first derived in 1900 by Louis Bachelier[6][7] as the time a stock reaches a certain price for the first time. In 1915 it was used independently by Erwin Schrödinger[3] and Marian v. Smoluchowski[4] as the time to first passage of a Brownian motion. In the field of reproduction modeling it is known as the Hadwiger function, after Hugo Hadwiger who described it in 1940.[12]Abraham Wald re-derived this distribution in 1944[13] as the limiting form of a sample in a sequential probability ratio test. The name inverse Gaussian was proposed by Maurice Tweedie in 1945.[14] Tweedie investigated this distribution in 1956[15] and 1957[16][17] and established some of its statistical properties. The distribution was extensively reviewed by Folks and Chhikara in 1978.[5]
Rated Inverse Gaussian Distribution
Assuming that the time intervals between occurrences of a random phenomenon follow an inverse Gaussian distribution, the probability distribution for the number of occurrences of this event within a specified time window is referred to as rated inverse Gaussian.[18] While, first and second moment of this distribution are calculated, the derivation of the moment generating function remains an open problem.
Numeric computation and software
Despite the simple formula for the probability density function, numerical probability calculations for the inverse Gaussian distribution nevertheless require special care to achieve full machine accuracy in floating point arithmetic for all parameter values.[19] Functions for the inverse Gaussian distribution are provided for the R programming language by several packages including rmutil,[20][21] SuppDists,[22] STAR,[23] invGauss,[24] LaplacesDemon,[25] and statmod.[26]
^ ab
Chhikara, Raj S.; Folks, J. Leroy (1989), The Inverse Gaussian Distribution: Theory, Methodology and Applications, New York, NY, USA: Marcel Dekker, Inc, ISBN0-8247-7997-5
^
Seshadri, V. (1999), The Inverse Gaussian Distribution, Springer-Verlag, ISBN978-0-387-98618-0
^Shuster, J. (1968). "On the inverse Gaussian distribution function". Journal of the American Statistical Association. 63 (4): 1514–1516. doi:10.1080/01621459.1968.10480942.
^
Schwarz, Wolfgang (2001), "The ex-Wald distribution as a descriptive model of response times", Behavior Research Methods, Instruments, and Computers, 33 (4): 457–469, doi:10.3758/bf03195403, PMID11816448
^Capacity per unit cost-achieving input distribution of rated-inverse gaussian biological neuron M Nasiraee, HM Kordy, J Kazemitabar IEEE Transactions on Communications 70 (6), 3788-3803