Marginal Distributions of Continuous Random Variable
Transitive comparison of random variables
Bernard De Baets , ... Bart De Schuymer , in Logical, Algebraic, Analytic and Probabilistic Aspects of Triangular Norms, 2005
15.5.3 Continuous random variables
We first consider a continuous random vector ( X 1, X 2, …, Xm ) with arbitrary marginal distributions pairwisely coupled by T M . We have demonstrated that [8]:
15.5.1 Proposition
Let (X 1, X 2, …, Xm ) be a continuous random vector. Then the probabilistic relation QM = [qM ij ] can be computed as:
(15.13)
Note that if Xi and Xj are identically distributed, i.e. ,then , as expected. In Figure 15.1 we give a graphical interpretation of formula(15.13). The two curves correspond to the marginal cumulative distribution functions FX and FY .
According to (15.13) we have to distinguish between three domains: the domain where FX lies beneath FY , the domain where FX lies above FY , and the domain where FX and FY coincide. The value of qM XY is computed as the sum of the increment of FX over the first domain and half of the increment of FX (or FY ) over the third domain. With the notations shown on the figure, we obtain for the example of Figure 15.1:
This example also illustrates that the probabilistic relation associated to a random vector pairwisely coupled by T M , can still be regarded as a graded version of the concept of stochastic dominance.
Note that the case of coupling by means of T P , has been treated before as the case of independent random variables, and the computation of QP = [qP ij ] according to (15.9), can be concisely written as
where denotes the expected value of the function between square brackets w.r.t. the distribution of X.
Next, we consider a continuous random vector (X 1, X 2, … ,Xm ) with arbitrary marginal distributions pairwisely coupled by T L In [8], we have also shown that:
15.5.2 Proposition
Let (X 1, X 2, … ,Xm) be a continuous random vector. Then the probabilistic relation QL = [qL ij ] can be computed as:
or, equivalently:
(15.14)
Note that u in (15.14) may not be unique, in which case any u fulfilling the right equality may be considered. Then qL ij is simply the height of in u. This is illustrated on Figure 15.2, where qL X , Y = FY (u) = t 1, since t 1 + t 2 = 1.
One can again easily verify that when FX = FY . Also, the probabilistic relation associated to a random vector pairwisely coupled by T L , can again be regarded as a graded version of stochastic dominance.
Read full chapter
URL:
https://www.sciencedirect.com/science/article/pii/B978044451814950015X
Multivariate stochastic orders
Félix Belzunce , ... Julio Mulero , in An Introduction to Stochastic Orders, 2016
3.5 The multivariate likelihood ratio order
In a similar way to the univariate case, it is possible to check the multivariate stochastic and hazard rate orders in terms of a property of the joint density functions. This property leads to the definition of the multivariate likelihood ratio order, which is a generalization of the univariate likelihood ratio order. In this section, the results are given in the continuous case but they can be restated for the general case. The main intention of the multivariate likelihood ratio order is to provide a sufficient condition for the multivariate hazard rate order.
Definition 3.5.1
Given two continuous random vectors X = (X 1,…,X n ) and Y = (Y 1,…,Y n ) with joint density functions f and g, respectively, we say that X is smaller than Y in the multivariate likelihood ratio order, denoted by X ≤lr Y, if
Clearly, this is a generalization of the likelihood ratio order in the univariate case. However, the multivariate likelihood ratio order is not necessarily reflexive. When a random vector X satisfies X ≤lr X, we have the MTP2 property introduced in Section 1.3.
The following result provides a set of sufficient conditions for the multivariate likelihood ratio order.
Theorem 3.5.2
Let X = (X 1,…,X n ) and Y = (Y 1,…,Y n ) be two continuous random vectors with joint density functions f and g, respectively. If X or Y or both are MTP2, and
(3.13)
then
Proof
Let us assume that X is MTP2 (in the other case the proof is similar), then we have the following inequalities:
for all , such that x ≤y, where the first inequality follows from the MTP2 property and the second one from (3.13). Therefore, X ≤lr Y.
The multivariate likelihood ratio order is preserved under conditioning on sublattices, as we see next. Recall that a subset is called a sublattice if x,y ∈ A implies x ∧y ∈ A and x ∨y ∈ A. This result will be used to show the relationship among the multivariate likelihood ratio order and the multivariate dynamic hazard rate order. The proof is obvious from the definition.
Theorem 3.5.3
Let X = (X 1,…,X n ) and Y = (Y 1,…,Y n ) be two continuous random vectors. If X ≤lr Y, then
In particular, from the previous theorem, the multivariate likelihood ratio order is preserved under marginalization. This result is useful because, in some cases, it is easier to provide a result in the multivariate case rather that in the univariate case.
Theorem 3.5.4
Let X = (X 1,…,X n ) and Y = (Y 1,…,Y n ) be two continuous random vectors. If X ≤lr Y, then
Next, it is showed that the multivariate likelihood ratio order is stronger than the multivariate hazard rate order.
Theorem 3.5.5
Let X = (X 1,…,X n ) and Y = (Y 1,…,Y n ) be two continuous random vectors. If X ≤lr Y, then
Proof
Let us check the conditions for the definition of the multivariate dynamic hazard rate order. In particular, denoting by η and λ the multivariate dynamic hazard rates of X and Y, respectively, let us see if
where
and
whenever , 0 ≤x I ≤y I ≤ t e, 0 ≤x J ≤ t e, and for all .
The result will follow by proving that, as we shall see later,
(3.14)
Condition (3.14) will follow if
(3.15)
holds.Let us see that (3.15) follows if X ≤lr Y holds. Denoting by f and g the joint density functions of and , respectively, and by f I,J and g I,J the joint densities of X I,J and Y I,J , respectively, we see that the joint density function of is given by
and the joint density function of is given by
Given y J > t e, we see that x J < t e <y J and, analogously, x I ≤y I . Since X ≤lr Y, given , we see that
which, upon integration, yields
Therefore, from previous inequality, we see that (3.15) holds. Now, from Theorem 3.5.3, we see that (3.14) also holds. In particular, holds, for all . Now, since (2.25), we see that the hazard rates of and are ordered and, consequently, we see that
Finally, a result for the multivariate likelihood ratio order among random vectors with conditionally independent components is provided. In particular, we consider the same background as that in Theorem 3.3.8.
Theorem 3.5.6
Assume that
- (i)
-
X i (θ) =st Y i (θ), for all θ and for all i = 1,…,n,
- (ii)
-
X i (θ) ≤lr Y j (θ′), for all θ ≤ θ′ and for all 1 ≤ i ≤ j ≤ n, and
- (iii)
-
θ 1 ≤lr θ 2.
Then,
Proof
Let f i (x|θ) be the density function of X i (θ). From condition (ii), we see that
Furthermore, condition (iii) is equivalent to the fact that h i (θ 1,…,θ n ) is TP2 in (θ 1,…,θ n ,i) ∈ S ×{1,2}. Due to these facts and from Theorem 1.2.3, we see that is TP2 in (x 1,…,x n ,i), and we get the result.
Read full chapter
URL:
https://www.sciencedirect.com/science/article/pii/B978012803768300003X
Truth, Possibility and Probability
In North-Holland Mathematics Studies, 1991
Uniform distributions
Let a, b be real numbers with a < b. Then the standard density function
is called a uniform density over [a, b]. It is clear that f is a density, since
We now show a probability distribution that is approximately uniform on the interval [0, 1], i.e., with a uniform density over [0, 1]. We take the interval [0, 1] just for simplicity. Let Ω be an equally spaced hyperfinite approximation of the unit interval [0, 1]. That is, such that ν is an infinite natural number,t 0 = 0,t ν = 1, and we also asumme that for every n ≤ ν. It is clear that #Ω = ν + 1. Thus, we are assuming that . We define the distribution pr on Ω by setting for any n ≤ ν
Thus, for any internal subset A of Ω, we have the probability distribution
The probability measure Pr is called the counting measure on Ω. It is clear that 〈Ω, pr〉 is an internal probability space. We now show that Pr has a uniform density.
The partition that we need in order to apply Theorem X.4 is formed by cubes, which in this case are intervals, of length dt = 1/ν. Then
Since the uniform density for [0, 1] is f(x) = 1, for 0 ≤ x ≤ 1, by Theorem X.4, we get that Pr has this uniform density.
Thus, if c, d ∈ [0, 1] with c < d, we have
We also can see that, in a sense, Theorem X.4, 2, cannot be improved: it is not true that for any partition with infinitesimal intervals, any interval r in the partition, and x ∈ r, satifies
This formula is true, as it is asserted in the theorem, only for partitions into cubes of volume ≥ 1/μ, for μ ≤ ν, for a certain fixed ν. In order to find a counterexample, take a partition with subintervals of length 1/2ν. Then there are some subintervals which do not contain any point in Ω, and, thus .
The same construction can be carried out in the unit circle C. As in Example VIII.4, let
where and ν is an infinite natural number, be a hyperfinite approximation of C. Again, let pr be the counting measure on C. Then, as above, if A is an arc of C, we have that the length of A.
We discuss, now, Bertrand's mixture paradox (see page 61) from an infinitesimal standpoint. The uniform distribution of the ratio of water to wine between 1 and 2 is obtained by a near interval, , for [1,2]. We must assume that . Every point in T 1 is equally probable, that is .
On the other hand, the uniform distribution of the ratio of wine to water between 1/2 and 1, is obtained by an equally spaced near interval for [1/2,1]. If we take the reciprocals of the elements of T 1, we do get a near interval, , for [1/2, 1], but its points are not equally spaced. Thus, we do not get in this way a uniform distribution. Suppose that we assign the same probability to each point, that is
We have that
Then
Thus, the density is, in this case, , which is clearly not uniform.
We conclude this section with a corollary to the theorem on change of variables (Theorem IX.18). Notice, that if R is a region and T is continuously differentiate transformation defined on R, then T(R) is also a region.
Theorem X.7
Let be a continuous random vector with density over the region R, and let T be a continuously differentiable one-one transformation of R. Then the random vector is also continuous with density
for any .
Proof
Let be defined over 〈Ω, pr〉, and let R ν and be as in (2) of Theorem X.7, such that is composed of infinitesimal cubes of area 1/μ. Thus, we have, for any r ∈ and
Let be the partition of T(R) obtained by applying T to the elements of . Then is an infinitesimal partition of T(R), and, by Lemma IX.16
We have for any s in , since and are finite
Since is continuous and, by Theorem IX.18, integrable over T(R), by Theorem X.7 (1) it is a density of .
Recall that T is called an affine transformation of a region R if there exists an n × n matrix A and a vector such that , for . If , then T is called linear. The function T is one-one, if A is nonsingular, and then
for . We have the following corollary.
Corollary X.8
Suppose that is a continuous random vector with density and that T is a one-one affine transformation, as defined above. Then is continuous with density
for every ,where det A is the determinant of A.
Proof
The corollary follows from the preceding theorem and
Read full chapter
URL:
https://www.sciencedirect.com/science/article/pii/S0304020808724216
Lattice Vector Quantization for Wavelet-Based Image Coding
Mikhail Shnaider , Andrew P. Papliński , in Advances in Imaging and Electron Physics, 1999
B Distortion Measure and Quantization Regions
Let us consider an input stream of continuous random vectors x = (x 1, x 2, ..., xN ) with the probability density function p( x ). Let us also consider an output stream y = (y 1, y 2, ..., yN ) related to an input stream x by the conditional probability density function q(y|x ). Assuming that the stream x is the input to a quantizer, then the corresponding values from the output stream y are obtained by a reference to the quantizer codebook.
According to Shannon (1948), the amount of uncertainty R( x , y ) of a value of x when we receive its quantized counterpart y from a transmission channel is given by
(17)
where h( x ) = ∫p( x ) log p( x )dx is the differential entropy of a variable x The value of R( x , y ) is effectively the actual transmission rate. With this in mind, we can pose the problem of optimal quantization in terms of minimization of the amount of uncertainty, that is,
(18)
with QD being a set specified by the conditional probability q( y | x ) so that
(19)
where d(·,·) is adistance function (to be defined). The substitution of Eq. (17) in Eq. (18) leads to
(20)
As the entropy h( x ) is independent of the conditional probability q( y, x ) and the entropy h( x | y ) = h( x – x | y ) ≤ h( x − y ), we can further modify Eq. (20) in the following way:
(21)
In information theory the last equation is known as the Shannon lower bound(Berger, 1971; Sakrison, 1979; Gibson and Sayood, 1988).
Now let us return to the process of quantization. We collect the input vectors x j = (x 1, x 2..., x N) in blocks X = x 1, x 2..., x L) and, for each input block, the quantizer finds a corresponding output block Y = y 1, y 2..., y L consisting of vectors yj = (y 1, y 2..., y N). The average bit rate per vector for a vector quantizer can be calculated as
(22)
Assuming that the distribution of x is uniform within each quantization region V and that all regions are simple translations of each other, we have the following approximation for the average bit rate per vector (Gibson and Sayood, 1988):
(23)
If the input vectors x are independent we have that
Furthermore, it can be shown (Berger, 1971; Sakrison, 1979) that for sufficiently high dimensionality of vectors the following equality is satisfied:
(24)
Thus, we can conclude that the performance of an uniform vector quantizer asymptotically achieves the Shannon lower bound for large dimensions.
An average distortion of a quantizer can be defined in the following way:
(25)
where for every input x with the PDF p(x) the quantizer produces an output y . The mth-power distortion d(m) introduced by this operation is defined by
(26)
By varying m, different distortion measures can be obtained. Closely related to the distortion measure is an optimal shape of the quantization region. It is easy to see that for a uniform source the optimal shape of quantization region is a pyramid, if m is selected to be 1, as depicted inFig. 11. For mequal to 2, Eq. (26) gives rise to the well-known mean squared error (MSE) distortion measure, the optimal quantization region for a uniform source being a sphere (Conway and Sloane, 1982). Furthermore, the radius of the sphere is equal to where D and L are the target distortion and dimensionality of quantization space, respectively.
From the foregoing considerations we can conjecture that, assuming the MSE distortion measure, an optimal quantizer for an L-dimensional uniform source constitutes a set of uniformly distributed spheres with code vectors as centroids of the spheres.
Read full chapter
URL:
https://www.sciencedirect.com/science/article/pii/S1076567008701988
Reliability Theory
N. Unnikrishnan Nair , ... N. Balakrishnan , in Reliability Modelling and Analysis in Discrete Time, 2018
Blomqvist's β
The Blomqvist measure is also based on concordance, but with the difference that instead of two random vectors here we choose a random vector and a fixed point with as the median of and as the median of . Thus, the measure in the case of a continuous random vector becomes
In the discrete case, this modifies to
(1.67)
The medians are chosen to be the nearest integer values.
Read full chapter
URL:
https://www.sciencedirect.com/science/article/pii/B9780128019139000014
Preliminaries
Félix Belzunce , ... Julio Mulero , in An Introduction to Stochastic Orders, 2016
1.3.1 The standard construction
The problem of defining a suitable multivariate extension of the univariate quantile function has a long history in statistics and probability. Next, the definition of the multivariate quantile transform is recalled. This transformation is also known as the standard construction, and is introduced by Arjas and Lehtonen [28], O'Brien [29], Rosenblatt [30], or Rüschendorf [31 ]. Given a continuous random vector X = (X 1,…,X n ), its standard construction is recursively defined as
(1.11)
for all (p 1,p 2,…,p n ) ∈ (0,1) n , where denotes the quantile function of X 1 and
denotes the quantile functions of the univariate conditional random variables given by
It is worth mentioning that this well-known transform is widely used in simulation theory and it plays the role of the quantile in the multivariate case. Hence, given n independent uniformly distributed random variables U 1,…,U n on the interval (0,1), then
(1.12)
where
for all (p 1,…,p n ) ∈ (0,1) n . As we shall see later, another transform of interest is the multivariate distributional transform. Given a continuous random vector X = (X 1,…,X n ), its multivariate distributional transform is recursively defined as
(1.13)
for all (x 1,…,x n ) in the support of X, where denotes the distribution function of X 1 and
denote the distribution function of the conditional random variables given by
According to the previous notation, we find that
(1.14)
where
for all . Given two continuous random vectors X = (X 1,…,X n ) and Y = (Y 1,…,Y n ), we can consider the transformation
(1.15)
defined for all (x 1,…,x n ) in the support of X. Observe that
for all (x 1,…,x n ) in the support of X. Since the distribution and quantile functions are increasing, then Φ i (x 1,…,x i ) is also increasing in x i , for all i = 1,…,n. In fact, in cases of differentiability, the Jacobian matrix of Φ is always a lower triangular matrix with strictly positive diagonal elements.
The most relevant property of Φ is that, from (1.12) and (1.14), we have
and, consequently, the function Φ maps the random vector X onto Y.
In addition, Fernández-Ponce and Suárez-Llorens [32] prove in their Theorem 3.1 that if we take a function such that Y =st k(X) and k has a lower triangular Jacobian matrix with strictly positive diagonal elements, then k necessarily has the form of the function Φ given in (1.15).
Read full chapter
URL:
https://www.sciencedirect.com/science/article/pii/B9780128037683000016
Fundamental of predictive filtering
Lu Cao , ... Bing Xiao , in Predictive Filtering for Microsatellite Control System, 2021
2.2.1 Random vector
In Probability Theory, a random vector is defined as , where is a random variable, . Its joint cumulative distribution function is defined as the probability of an n-dimensional semi-infinite rectangle associated with the sample point . It is given by
(2.1)
If X is continuous random vector, then its joint probability density function is defined as
(2.2)
For a discrete vector X , its joint probability mass function is defined as
(2.3)
The joint probability density function and the joint probability mass function satisfy
(2.4)
(2.5)
Moreover, for all continuous random vectors, is such that
(2.6)
Let and be two random vectors having n and m components, respectively. The joint cumulative distribution function of X and Y at and is defined as
(2.7)
When X and Y are discrete, we define the joint probability mass function of X and Y as
(2.8)
and their marginal probability functions as
(2.9)
(2.10)
For continuous X and Y , we define their joint probability density function as
(2.11)
and their marginal probability functions as
(2.12)
(2.13)
Then, the random vectors X and Y are independent if and only if
(2.14)
Read full chapter
URL:
https://www.sciencedirect.com/science/article/pii/B9780128218655000152
The Analysis of Structural Equation Model with Ranking Data using Mx
Wai-Yin Poon , in Handbook of Latent Variable and Related Models, 2007
1 Introduction
Ranking data are obtained when subjects are asked to rank p objects from 1 (most preferred) to p (least preferred). A number of approaches have been developed to model and analyze ranking data. Among others, researches on using Thurstonian models (Thurstone, 1927) remain active for more than 70 years (see, e.g., Critchlow and Fligner, 1991; Böckenholt, 1992, 1996, 2001; Chan and Bentler, 1998; Mayedu-Olivares, 1999), confirming their usefulness in a wide range of disciplines. Thurstonian models postulate that the ranking of the p objects are determined by a p × 1 latent continuous random vector Y which is distributed as multivariate normal with mean μ and covariance matrix Σ. Different models are achieved by putting different restrictions on the elements in Σ. In particular, the use of structural equation models by imposing the structures on Σ substantially enriches the horizon of modeling ranking data and hence its practical applicability (see, e.g., Currim, 1982; Böckenholt, 1992; Elrod and Keane, 1995).
In effect, structural equation modeling remains an extremely active area of research. The interest derives from both the practical applicability of the approach in addressing and verifying substantive theories and the technical difficulties involving in modeling of various types of data together with the estimation of the resulted model. User-friendly structural equation modeling packages are widely available, such as PRELIS and LISREL (Jöreskog and Sörbom, 1996a, 1996b), EQS (Bentler and Wu, 1993), Mx (Neale et al., 1999) as well as Mplus (Muthén and Muthén, 1998). The capabilities of these packages continue to be extended and nowadays, they all provide options for analyzing ordinal categorical data in a convenient manner. Although ranking data and ordinal categorical data are different in nature and require different modeling and estimation techniques, the approaches adopted by popular packages for analyzing ordinal categorical data and the Thurstonian approach for analyzing ranking data both operate on the assumption that the observed variables are associated with some underlying continuous variables distributed as multivariate normal. With reference to this similarity, the current study establishes a relationship between the analysis of ordinal categorical data and ranking data with a view to making use of readily available structural equation modeling procedures for analyzing ordinal categorical data to analyze ranking data. Specifically, the implementation of the analysis of the Thurstonian models using options designated for analyzing ordinal categorical data in the widely available software program Mx (Neale et al., 1999) is addressed. Some initial effort along similar direction has been made by Mayedu-Olivares (1999). He has developed a procedure for analyzing ranking data using the package MECOSA (Arminger et al., 1996).
A description of the Thurstonian model of ranking data, the multivariate normal model for analyzing ordinal categorical data, and their similarity is given in Section 2. A summary of the Mx program together with its option for analyzing ordinal categorical data, and the procedure on using the option to implement the Thurstonian models are given in Section 3. It will be seen that the procedure is very flexible. Different identification and across parameters constraints together with various structures on the mean vector and on the covariance/correlation matrix can be incorporated in an efficient and easy manner. Some examples available in the literature are used for illustration. In effect, the procedure can also be applied to analyze models for partial ranking data (Böckenholt, 1992). Two applications are discussed in Section 4. The first application relates to an analysis of rankings of compound objects. The objective is to examine whether or not a mean score for a compound choice alternative consisting of two objects can be predicted by an additive combination of means scores obtained for each of the two objects separately. The second application relates to the analysis of a set of 8 soft drinks, and the rankings of the soft drinks are obtained via a balanced incomplete design. The paper is concluded with a discussion in Section 5.
Read full chapter
URL:
https://www.sciencedirect.com/science/article/pii/B9780444520449500124
Elicitation of multivariate prior distributions: A nonparametric Bayesian approach
Fernando A. Moala Anthony O'Hagan , in Journal of Statistical Planning and Inference, 2010
We assume that the expert's prior density is a relatively smooth function of an unknown continuous random vector . We will model a nonparametric prior for hierarchically in terms of a vector of hyperparameters. An appropriate choice to represent the analyst's prior beliefs about is a Gaussian process, which we denote by
(1)
where the mean and covariance functions
(2)
depend on the hyperparameters .
Read full article
URL:
https://www.sciencedirect.com/science/article/pii/S0378375810000157
petersonsomuckledge.blogspot.com
Source: https://www.sciencedirect.com/topics/mathematics/continuous-random-vector
0 Response to "Marginal Distributions of Continuous Random Variable"
Enviar um comentário