# Reasoning with fuzzy and uncertain evidence using epistemic random fuzzy sets: general framework and practical models111This paper was published in Fuzzy Sets and Systems, 453:1–36, 2023. This version corrects an error in Equation (23).
**Authors**: Thierry Denœux
> Université de technologie de Compiègne, CNRS UMR 7253 Heudiasyc, Compiègne, France Institut universitaire de France, Paris, France
## Abstract
We introduce a general theory of epistemic random fuzzy sets for reasoning with fuzzy or crisp evidence. This framework generalizes both the Dempster-Shafer theory of belief functions, and possibility theory. Independent epistemic random fuzzy sets are combined by the generalized product-intersection rule, which extends both Dempster’s rule for combining belief functions, and the product conjunctive combination of possibility distributions. We introduce Gaussian random fuzzy numbers and their multi-dimensional extensions, Gaussian random fuzzy vectors, as practical models for quantifying uncertainty about scalar or vector quantities. Closed-form expressions for the combination, projection and vacuous extension of Gaussian random fuzzy numbers and vectors are derived.
keywords: Belief functions, evidence theory, possibility theory, random sets, uncertainty. journal: Fuzzy Sets and Systems
## 1 Introduction
The Dempster-Shafer (DS) theory of belief functions [29] and possibility theory [38] were introduced independently in the late 1970’s as non-probabilistic frameworks for reasoning with uncertainty [11, 10]. The former approach is based on the idea of representing elementary pieces of evidence as completely monotone capacities, or belief functions, and combining them using an operator known as the product-intersection rule or Dempster’s rule. As probability measures are special belief functions, and Dempster’s rule extends Bayesian conditioning, DS can be seen as an extension of Bayesian probability theory, particularly suitable to reasoning with severe uncertainty. There is also a strong relation between DS theory and the theory of random sets [23]: specifically, any random set induces a belief function and, conversely, any belief function can be seen as being induced by some random set [26]. In DS theory, a random set underlying a belief function does not represent a random mechanism for generating sets of outcomes, but the imprecise meanings of a piece of evidence under different interpretations with known probabilities [30]. To avoid confusion, we use the term epistemic random set for random sets representing evidence in DS theory.
In contrast, possibility theory originates from the theory of fuzzy sets [36]. In this approach, a fuzzy statement about the variable of interest, seen as a flexible constraint on its precise but unknown value in some domain $Θ$ , induces a possibility measure and a dual necessity measure on $Θ$ . Interestingly, a necessity measure is a belief function, and the dual possibility measure is the corresponding plausibility function, but the converse is not true (a belief function is not, in general, a necessity measure). For this reason, possibility theory has sometimes been presented as “a special branch of evidence theory” (another name for DS theory) [21, page 187]. However, combining two necessity measures by Dempster’s rule yields a belief function that is no longer a necessity measure: this combination rule is, thus, not compatible with possibilistic reasoning. In contrast, possibility theory has its own conjunctive combination operators based on triangular norms (or t-norms) [16]. Possibility and DS theory are, thus, two distinct models of uncertain reasoning based on related knowledge representation languages but different information processing mechanisms.
In a companion paper [9], we have revisited Zadeh’s notion of “evidence of the second kind”, defined as a pair $(X,Π_(Y\mid X))$ in which $X$ is a discrete random variable on a set $Ω$ and $Π_(Y\mid X)$ a collection of conditional possibility distributions of a variable $Y$ given $X=x$ , for all $x∈Ω$ . If random variable $X$ is constant, we get a unique possibility distribution for variable $Y$ ; if the conditional possibility distributions $Π_(Y\mid X)$ take values in $\{0,1\}$ , then the pair $(X,Π_(Y\mid X))$ defines a random set equivalent to a DS mass function. The mappings associating, to each event, its expected necessity and its expected possibility are, respectively, belief and plausibility functions. In this framework, a possibility distribution thus represents certain but fuzzy evidence, while a DS mass function is a model of uncertain and crisp evidence. In general, a pair $(X,Π_(Y\mid X))$ defines an epistemic random fuzzy set, allowing us to describe evidence that is both uncertain and fuzzy. (The term “epistemic” emphasizes the distinction between this interpretation and that of random fuzzy sets as mechanisms for generating fuzzy data considered, for instance in [28, 17]). In [9], we have proposed a family of combination rules for epistemic random fuzzy sets in the finite setting, generalizing both Dempster’s rule and the conjunctive combination rules of possibility theory. One of these rules, based on the product t-norm, is associative and arguably well suited for combining independent evidence. Equipped with this combination rule (called here the generalized product-intersection rule), the theory of epistemic random fuzzy sets can be seen as an extension of both DS theory and possibility theory, making it possible to combine evidence of various types, including expert assessments (possibly expressed in natural language), sensor information, and statistical evidence about a model parameter.
In this paper, drawing from mathematical results presented by Couso and Sánchez in [2], we give a more general exposition of the theory of epistemic fuzzy sets, considering arbitrary probability and measurable spaces. We define combination, marginalization and vacuous extension operations of random fuzzy sets in this general setting, laying the foundations of a wide-ranging theory of uncertainty encompassing DS and possibility theories as special cases. Finally, for the important case where the frame of discernment is $ℝ^p$ , we propose Gaussian random fuzzy numbers and vectors as a practical model, generalizing both Gaussian random variables and vectors on the one hand, and Gaussian possibility distributions on the other hand.
The rest of this paper is organized as follows. Classical models (including random sets, fuzzy sets and possibility theory) are first recalled in Section 2. Epistemic random fuzzy sets are then introduced in a general setting in Section 3. Finally, Gaussian random fuzzy numbers and vectors are studied, respectively, in Sections 4 and 5, and Section 6 concludes the paper.
## 2 Classical models
In this section, we recall the main definitions and results pertaining to the two models of uncertainty generalized in this paper: random sets and belief functions on the one hand (Section 2.1), fuzzy sets and possibility theory on the other hand (Section 2.2).
### 2.1 Random sets and belief functions
Whereas belief functions in the finite setting can be introduced without any reference to random sets [29], the mathematical framework of random sets is useful to analyze belief functions in more general spaces, and to define the practical models needed, e.g., in statistical applications. Important references about the link between random sets and belief functions include [26] and [2].
Let $(Ω,σ_Ω,P)$ be a probability space, $(Θ,σ_Θ)$ a measurable space, and ${\overline{X}}$ a mapping from $Ω$ to $2^Θ$ . The upper and lower inverses of ${\overline{X}}$ are defined, respectively, as follows:
$$
\displaystyle{\overline{X}}^*(B) \displaystyle=B^*=\{ω∈Ω:{\overline{X}}(ω)∩ B≠∅\} \displaystyle{\overline{X}}_*(B) \displaystyle=B_*=\{ω∈Ω:∅≠{\overline{X}}(ω)
⊆ B\}
$$
for all $B⊆Θ$ . It is easy to check that
$$
B^*∩(B^c)_*=∅
$$
and
$$
B^*∪(B^c)_*=\{ω∈Ω:{\overline{X}}(ω)≠∅\}=
Θ^*,
$$
where $B^c$ denotes the complement of $B$ in $Θ$ .
The mapping ${\overline{X}}$ is said to be $σ_Ω-σ_Θ$ strongly measurable [26] if, for all $B∈σ_Θ$ , $B^*∈σ_Ω$ (or, equivalently, if for all $B∈σ_Θ$ , $B_*∈σ_Ω$ ). The tuple $(Ω,σ_Ω,P,Θ,σ_Θ,{\overline{X}})$ is called a random set. When there is no confusion about the domain and co-domain, we will call the $σ_Ω-σ_Θ$ strongly measurable mapping ${\overline{X}}$ itself a random set.
In the special case where $|{\overline{X}}(ω)|=1$ for all $ω∈Ω$ , we can define the mapping $X:Ω→Θ$ such that ${\overline{X}}(ω)=\{X(ω)\}$ for all $ω∈Ω$ . We then have $B^*=B_*=X^-1(B)$ for all $B⊆Θ$ , and $X$ is $σ_Ω-σ_Θ$ measurable. The notion of random set thus extends that of random variable.
#### Belief and plausibility functions
From now on, we will assume, for simplicity, that $P(Θ^*)=1$ . (If not verified, this property can be enforced by conditioning $P$ on $Θ^*$ ). Let $P^*$ and $P_*$ be the lower and upper probability measures associated with random set ${\overline{X}}$ , defined as the mappings from $σ_Θ$ to $[0,1]$ such that
$$
P_*(B)=P(B_*) \tag{2}
$$
and
$$
P^*(B)=P(B^*)=1-P_*(B^c), \tag{3}
$$
for all $B∈σ_Θ$ . Mapping $P_*$ is a completely monotone capacity, i.e., a belief function, and $P^*$ is the dual plausibility function [26, Proposition 1]. In the following, they will be denoted, respectively, as $Bel_\overline{X}$ and $Pl_\overline{X}$ . The corresponding contour function is defined as the mapping $pl_\overline{X}$ from $Θ$ to $[0,1]$ such that
$$
pl_\overline{X}(θ)=Pl_\overline{X}(\{θ\})
$$
for all $θ∈Θ$ . The subsets ${\overline{X}}(ω)⊆Θ$ , for all $ω∈Ω$ , are called the focal sets of $Bel_\overline{X}$ .
#### Interpretation
In DS theory, $Ω$ represents a set of interpretations of a piece of evidence about a variable $\boldsymbol{θ}$ taking values in set $Θ$ (called the frame of discernment). If interpretation $ω∈Ω$ holds, we know that $\boldsymbol{θ}∈{\overline{X}}(ω)$ , and nothing more. For any $A∈σ_Ω$ , $P(A)$ is the (subjective) probability that the true interpretation lies in $A$ . For any $B∈σ_Θ$ , the degree of belief $Bel_\overline{X}(B)$ is then a measure of support of the proposition “ $\boldsymbol{θ}∈ B$ ” given the evidence, while the degree of plausibility $Pl_\overline{X}(B)$ is a measure of lack of support for the proposition “ $\boldsymbol{θ}\not∈ B$ ”. Under this interpretation, the random set ${\overline{X}}$ represents a state of knowledge: it can be said to be epistemic.
#### Vacuous random set
A constant random set $(Ω,σ_Ω,P,Θ,σ_Θ,{\overline{X}})$ such that ${\overline{X}}(ω)=Θ$ for all $ω∈Ω$ is said to be vacuous. For such a random set, we have $Bel_\overline{X}(A)=0$ for all $A∈σ_Θ∖\{Θ\}$ and $Pl_\overline{X}(A)=1$ for all $A∈σ_Θ∖\{∅\}$ . A vacuous random set represents complete ignorance about $\boldsymbol{θ}$ .
#### Finite case
Assume that $Θ$ is finite, and $σ_Θ=2^Θ$ . The Möbius inverse of $Bel_\overline{X}$ is the mapping $m_\overline{X}$ from $2^Θ$ to [0,1] such that
$$
m_\overline{X}(B)=∑_C⊆ B(-1)^|B|-|C|Bel_\overline{X}(C),
$$
for all $B⊆Θ$ . It verifies $m(B)≥ 0$ for all $B⊆Ω$ , $∑_B⊆Ωm(B)=1$ and $m(∅)=0$ . The belief and plausibility can be computed from $m_\overline{X}$ , respectively, as
$$
Bel_\overline{X}(B)=∑_C⊆ Bm_\overline{X}(C) and
Pl_\overline{X}(B)=∑_C∩ B≠∅m_\overline{X}(C),
$$
for all $B⊆Θ$ .
#### Random closed intervals
Random closed intervals are particularly simple models allowing us to define belief functions on the real line [4, 34, 6]. Let $(Ω,σ_Ω,P)$ be a probability space and $X,Y$ two random variables $Ω→ℝ$ such that $P(\{ω∈Ω:X(ω)≤ Y(ω)\})=1$ . Then, the mapping ${\overline{X}}:Ω→ 2^ℝ$ defined by ${\overline{X}}(ω)=[X(ω),Y(ω)]$ is $σ_Ω-β_ℝ$ strongly measurable, where $β_ℝ$ is the Borel $σ$ -algebra on $ℝ$ (see a formal proof in [22]). This mapping defines a random closed interval. For a random closed interval ${\overline{X}}=[X,Y]$ , we have [4]
$$
Bel_\overline{X}([x,y])=P([X,Y]⊆[x,y])=P(X≥ x;Y≤ y) Pl_\overline{X}([x,y])=P([X,Y]∩[x,y]≠∅)=1-P(X>y)-P(Y<x),
$$
for all $(x,y)∈ℝ^2$ such that $x≤ y$ . In particular, by letting $x$ tend to $-∞$ in (4), we obtain the lower and upper cumulative distribution functions (cdf’s) of ${\overline{X}}$ as
$$
F_*(y)=Bel_\overline{X}((-∞,y])=P(Y≤ y)=F_Y(y) F^*(y)=Pl_\overline{X}((-∞,y])=P(X≤ y)=F_X(y).
$$
#### Lower and upper expectation
Let ${\overline{X}}$ be a random set from $(Ω,σ_Ω,P)$ to $(ℝ,β_ℝ)$ . Following Dempster [3], we can define its lower and upper expectations, respectively, as
$$
E_*({\overline{X}})=∫_-∞^+∞x dF^*(x)
$$
and
$$
E^*({\overline{X}})=∫_-∞^+∞x dF_*(x),
$$
where $F_*(x)=Bel_\overline{X}((-∞,x])$ and $F^*(x)=Pl_\overline{X}((-∞,x])$ are the lower and upper cdf’s of ${\overline{X}}$ . When ${\overline{X}}$ is a random closed interval $[X,Y]$ , it follows from (5) that $E_*({\overline{X}})=E(X)$ and $E^*({\overline{X}})=E(Y)$ .
#### Dempster’s rule
Consider two pieces of evidence represented by random sets
$$
(Ω_1,σ_1,P_1,Θ,σ_Θ,{\overline{X}}_1)
and (Ω_2,σ_2,P_2,Θ,σ_Θ,{\overline{
X}}_2),
$$
and the mapping ${\overline{X}}_∩$ from $Ω_1×Ω_2$ to $2^Θ$ defined by ${\overline{X}}_∩(ω_1,ω_2)={\overline{X}}_1(ω_1) ∩{\overline{X}}_2(ω_2)$ . If interpretations $ω_1∈Ω_1$ and $ω_2∈Ω_2$ both hold, we know that $\boldsymbol{θ}∈{\overline{X}}_∩(ω_1,ω_2)$ , provided that ${\overline{X}}_1(ω_1)∩{\overline{X}}_2(ω_2)≠∅$ . Assume that ${\overline{X}}_∩$ is $(σ_1⊗σ_2)-σ_Θ$ strongly measurable, where $σ_1⊗σ_2$ is the tensor product $σ$ -algebra over the Cartesian product $Ω_1×Ω_2$ . The two pieces of evidence are said to be independent if, for any $A∈σ_1⊗σ_2$ , the probability that $A$ contains the true interpretations of the two pieces of evidence is the conditional probability
$$
P_12(A)=(P_1× P_2)(A\midΘ^*)=\frac{(P_1× P_2)(A∩
Θ^*)}{(P_1× P_2)(Θ^*)}, \tag{6}
$$
where $P_1× P_2$ is the product measure satisfying $(P_1× P_2)(A_1× A_2)=P_1(A_1)P_2(A_2)$ for all $A_1∈σ_1$ , $A_2∈σ_2$ , and
$$
Θ^*=\{(ω_1,ω_2)∈Ω_1×Ω_2:{\overline{X
}}_∩(ω_1,ω_2)≠∅\}
$$
is the set of noncontradictory pairs of interpretations. The quantity
$$
κ=1-(P_1× P_2)(Θ^*)=(P_1× P_2)(\{(ω_1,
ω_2)∈Ω_1×Ω_2:{\overline{X}}_∩(ω_1,
ω_2)=∅\})
$$
is called the degree of conflict between the two pieces of evidence. The combined random set
$$
(Ω_1×Ω_2,σ_1⊗σ_2,P_12,Θ,σ_
Θ,{\overline{X}}_∩)
$$
is called the orthogonal sum of the two pieces of evidence, and is denoted by ${\overline{X}}_1⊕{\overline{X}}_2$ . This combination rule, first introduced by Dempster in [3], is called the product-intersection rule, or Dempster’s rule of combination.
We can remark that Dempster’s rule is usually viewed as an operation to combine belief functions, whereas it is defined here as an operation to combine random sets. This distinction is immaterial in the standard setting, as the orthogonal sum of two belief functions does not depend on their particular random set representations and can be defined without reference to the random set framework [32]. However, it becomes crucial when considering random fuzzy sets as a model for generating belief functions, as done in this paper. We will come back to this important point in Section 3.2.
Any vacuous random set is obviously a neutral element for Dempster’s rule. The following important proposition states that pieces of evidence can be combined by Dempster’s rule in any order.
**Proposition 1**
*Dempster’s rule is commutative and associative.*
* Proof*
See A. ∎
**Example 1**
*Let $X_1∼ N(μ_1,σ_1^2)$ and $X_2∼ N(μ_2,σ_2^2)$ be two independent normal random variables and consider the random intervals ${\overline{X}}_1=[X_1,+∞)$ and ${\overline{X}}_2=(-∞,X_2]$ . The degree of conflict between ${\overline{X}}_1$ and ${\overline{X}}_2$ is
$$
κ=P(X_1>X_2)=P(X_2-X_1<0)=Φ≤ft(\frac{μ_1-μ_2}{√{
σ_1^2+σ_2^2}}\right),
$$
where $Φ$ is the standard normal cdf. The orthogonal sum of ${\overline{X}}_1$ and ${\overline{X}}_2$ is the random closed interval $[X^\prime_1,X^\prime_2]$ , where $(X^\prime_1,X^\prime_2)$ is the two-dimensional random vector with distribution equal the conditional distribution of $(X_1,X_2)$ given $X_1≤ X_2$ . Its density is
$$
f_X^\prime_1,X^{\prime_2}(x_1,x_2)=\frac{σ_1^-1σ_2
^-1φ≤ft(\frac{x_1-μ_1}{σ_1}\right)φ≤ft(\frac{x_2-
μ_2}{σ_2}\right)I(x_1≤ x_2)}{Φ≤ft(\frac{μ_2-μ_1
}{√{σ_1^2+σ_2^2}}\right)},
$$
where $φ$ is the standard normal probability density function (pdf) and $I(·)$ is the indicator function.*
The following proposition states that the contour function of the orthogonal sum of two independent random sets ${\overline{X}}_1$ and ${\overline{X}}_2$ is proportional to the product of the contour functions of ${\overline{X}}_1$ and ${\overline{X}}_2$ .
**Proposition 2**
*Let ${\overline{X}}_1$ and ${\overline{X}}_2$ be two independent random sets on the same domain $Θ$ , with contour functions $pl_{\overline{X}_1}$ and $pl_{\overline{X}_2}$ . For any $θ∈Θ$ ,
$$
pl_{\overline{X}_1⊕{\overline{X}}_2}(θ)=\frac{pl_{\overline{X
}_1}(θ)pl_{\overline{X}_2}(θ)}{1-κ}, \tag{7}
$$
where $κ$ is the degree of conflict between ${\overline{X}}_1$ and ${\overline{X}}_2$ .*
* Proof*
We have
| | $\displaystyle pl_{\overline{X}_1⊕{\overline{X}}_2}(θ)$ | $\displaystyle=\frac{(P_1× P_2)(\{(ω_1,ω_2)∈Ω_1 ×Ω_2:θ∈{\overline{X}}_∩(ω_1,ω_2)\})}{1-κ}$ | |
| --- | --- | --- | --- |
∎
**Example 2**
*Let us consider again the two random intervals of Example 1. The contour functions of ${\overline{X}}_1$ and ${\overline{X}}_2$ are, respectively,
$$
pl_{\overline{X}_1}(x)=P(X_1≤ x)=Φ≤ft(\frac{x-μ_1}{σ_1
}\right)
$$
and
$$
pl_{\overline{X}_2}(x)=P(X_2≥ x)=1-Φ≤ft(\frac{x-μ_2}{σ_
{2}}\right).
$$
Now, the contour function of ${\overline{X}}_1⊕{\overline{X}}_2$ is
| | $\displaystyle pl_{\overline{X}_1⊕{\overline{X}}_2}(x)$ | $\displaystyle=P(X^\prime_1≤ x≤ X^\prime_2)$ | |
| --- | --- | --- | --- |*
#### Marginalization and vacuous extension
Let us now consider the case where we have two variables $\boldsymbol{θ}_1$ and $\boldsymbol{θ}_2$ with domains $Θ_1$ and $Θ_2$ . (The case of $n$ variables is not more difficult conceptually but it requires heavier notations). Let $σ_Θ_1$ and $σ_Θ_2$ be $σ$ -algebras defined, respectively, on $Θ_1$ and $Θ_2$ . Let $Θ_12=Θ_1×Θ_2$ and $σ_Θ_12=σ_Θ_1⊗σ_Θ_2$ . Let ${\overline{X}}_12$ be a random set from $(Ω,σ_Ω,P)$ to $(Θ_12,σ_Θ_12)$ , and ${\overline{X}}_1$ the mapping from $Ω$ to $2^Θ_1$ that maps each $ω∈Ω$ to the projection of ${\overline{X}}_12(ω)$ onto $Θ_1$ :
$$
{\overline{X}}_1(ω)={\overline{X}}_12(ω)↓Θ_1=\{
θ_1∈Θ_1:∃θ_2∈Θ_2,(θ_1,θ_2)
∈{\overline{X}}_12(ω)\}.
$$
It is easy to see that ${\overline{X}}_1$ is $σ_Ω-σ_Θ_1$ measurable: for any $B∈σ_Θ_1$ ,
| | $\displaystyle{\overline{X}}_1^*(B)$ | $\displaystyle=\{ω∈Ω:{\overline{X}}_1(ω)∩ B≠∅\}$ | |
| --- | --- | --- | --- |
As $B×Θ_2∈σ_Θ_12$ and ${\overline{X}}_12$ is $σ_Ω-σ_Θ_12$ strongly measurable, it thus follows that ${\overline{X}}_1^*(B)∈σ_Ω$ . The random set ${\overline{X}}_1$ will be called the marginal of ${\overline{X}}_12$ on $Θ_1$ .
Conversely, let ${\overline{X}}_1$ be a random set from $(Ω,σ_Ω)$ to $(Θ_1,σ_Θ_1)$ and let ${\overline{X}}_1↑ 2$ be the mapping from $Ω$ to $Θ_12$ defined by
$$
{\overline{X}}_1↑(1,2)(ω)={\overline{X}}_1(ω)×Θ
_2.
$$
For any $B∈σ_Θ_12$ ,
| | $\displaystyle{\overline{X}}_1↑(1,2)^*(B)$ | $\displaystyle=\{ω∈Ω:{\overline{X}}_1↑ 2(ω)∩ B≠∅\}$ | |
| --- | --- | --- | --- |
If for all $B∈σ_Θ_12$ , ${\overline{X}}_1^*(B↓Θ_1)∈σ_Ω$ , then ${\overline{X}}_1↑(1,2)$ is $σ_Ω-σ_Θ_12$ strongly measurable. It is said to be the vacuous extension of ${\overline{X}}_1$ in $Θ_1×Θ_2$ .
We say that a random set ${\overline{X}}_12$ from $(Ω,σ_Ω,P)$ to $(Θ_12,σ_Θ_12)$ with marginals ${\overline{X}}_1$ and ${\overline{X}}_2$ is noninteractive if it is equal to the orthogonal sum of its marginals, i.e.,
$$
{\overline{X}}_12={\overline{X}}_1↑(1,2)⊕{\overline{X}}_2
↑(1,2) denoted by {\overline{X}}_1⊕{\overline{X}
}_2.
$$
**Example 3**
*Let $(X_1,X_2)$ be a two dimensional random vector from $(Ω,σ_Ω,P)$ to $(ℝ^2,β_ℝ^2)$ and consider the mapping ${\overline{X}}_12:Ω→ 2^ℝ^2$ defined as
$$
{\overline{X}}_12(ω)=(-∞,X_1(ω)]×(-∞,X_2(ω
)].
$$
This mapping defines a random set [23, page 3]. Its marginals are the random closed intervals $(-∞,X_1]$ and $(-∞,X_2]$ . If $X_1$ and $X_2$ are independent, then ${\overline{X}}_12=(-∞,X_1]⊕(-∞,X_2]$ and ${\overline{X}}_12$ is noninteractive.*
### 2.2 Fuzzy sets and possibility theory
A fuzzy subset of a set $Θ$ is a pair ${\widetilde{F}}=(Θ,μ_{\widetilde{F}})$ , where $μ_{\widetilde{F}}$ is a mapping from $Θ$ to $[0,1]$ , called the membership function of ${\widetilde{F}}$ . Each number $μ_{\widetilde{F}}(θ)$ is interpreted as a degree of membership of element $θ$ to the fuzzy set ${\widetilde{F}}$ . In the following, to simplify the notation, we will identify fuzzy sets to their membership functions and write ${\widetilde{F}}(θ)$ for $μ_{\widetilde{F}}(θ)$ . The height of fuzzy set ${\widetilde{F}}$ is defined as
$$
0pt({\widetilde{F}})=\sup_θ∈Θ{\widetilde{F}}(θ).
$$
If $0pt({\widetilde{F}})=1$ , ${\widetilde{F}}$ is said to be normal. For any $α∈[0,1]$ , the (weak) $α$ -cut of ${\widetilde{F}}$ is the set
$$
{}^α{\widetilde{F}}=\{θ∈Θ:{\widetilde{F}}(θ)≥α\}.
$$
#### Possibility and necessity measures
Let $\boldsymbol{θ}$ be a variable taking values in $Θ$ . Assume that we receive a piece of evidence telling us that “ $\boldsymbol{θ}$ is ${\widetilde{F}}$ ”, where ${\widetilde{F}}$ is a normal fuzzy subset of $Θ$ . This evidence induces a possibility measure $Π_\widetilde{F}$ from $2^Θ$ to $[0,1]$ defined by
$$
Π_\widetilde{F}(B)=\sup_θ∈ B{\widetilde{F}}(θ), \tag{8}
$$
for all $B⊆Θ$ . The number $Π_\widetilde{F}(B)$ is interpreted as the degree of possibility that $\boldsymbol{θ}∈ B$ , given that $\boldsymbol{θ}$ is ${\widetilde{F}}$ [38]. The corresponding possibility distribution is the mapping from $Θ$ to $[0,1]$ defined by
$$
π_\widetilde{F}(θ)=Π_\widetilde{F}(\{θ\})={\widetilde{F}}(
θ),
$$
i.e., it is identical to the membership function ${\widetilde{F}}$ . The dual necessity measure is defined as
$$
N_\widetilde{F}(B)=1-Π_\widetilde{F}(B^c)=∈f_θ\not∈ B≤ft[
1-{\widetilde{F}}(θ)\right]. \tag{9}
$$
It can easily be shown that mapping $N_\widetilde{F}:2^Ω→[0,1]$ is completely monotone, i.e., it is a belief function, and $Π_\widetilde{F}$ is the dual plausibility function [15]. These belief and plausibility functions are formally induced by the random set $([0,1],β_[0,1],λ,Θ,2^Θ,{\overline{X}})$ , where $β_[0,1]$ is the Borel $σ$ -algebra on $[0,1]$ , $λ$ is the uniform probability measure, and ${\overline{X}}$ is the mapping $[0,1]→ 2^Θ$ defined by ${\overline{X}}(α)={}^α{\widetilde{F}}$ . However, as we will see in Section 3.2, it is important, when combining evidence, to distinguish between possibility distributions induced by fuzzy sets, and consonant belief functions induced by random sets.
#### Conjunctive combination of possibility distributions
Assume that we receive two independent pieces of information telling us that “ $\boldsymbol{θ}$ is ${\widetilde{F}}$ ” and “ $\boldsymbol{θ}$ is ${\widetilde{G}}$ ”, where ${\widetilde{F}}$ and ${\widetilde{G}}$ are two fuzzy subsets of $Θ$ . The conjunctive combination of these two pieces of evidence requires some notion of intersection between fuzzy sets. As reviewed in [13], the intersection operation can be extended to fuzzy sets using triangular norms (or t-norms for short). Given a t-norm $⊤$ , the $⊤$ -intersection of two fuzzy subsets ${\widetilde{F}}$ and ${\widetilde{G}}$ of the same domain $Θ$ can be defined as
$$
({\widetilde{F}}∩_⊤{\widetilde{G}})(θ)={\widetilde{F}}(θ)
⊤{\widetilde{G}}(θ)
$$
for all $θ∈Θ$ . The most common choices for $⊤$ are the minimum and product t-norms, as originally proposed by Zadeh [36]; the corresponding operations are called, respectively, the minimum and product intersections. However, the intersection of two normal fuzzy sets is generally not normal. To obtain a normal fuzzy set, as needed for the definitions of possibility and necessity measures in (8)-(9), we define the normalized $⊤$ -intersection as
$$
({\widetilde{F}}∩^*_⊤{\widetilde{G}})(θ)=\begin{cases}
\displaystyle\frac{{\widetilde{F}}(θ)⊤{\widetilde{G}}(θ)}{0pt({
\widetilde{F}}∩_⊤{\widetilde{G}})}&if 0pt({\widetilde{F}}∩_
⊤{\widetilde{G}})>0\\
0&otherwise.\end{cases}
$$
The fuzzy set ${\widetilde{F}}∩^*_⊤{\widetilde{G}}$ is normal provided that $0pt({\widetilde{F}}∩_⊤{\widetilde{G}})>0$ . In general, the normalized intersection $∩^*_⊤$ associated with a t-norm $⊤$ is not associative. A notable exception is the case where $⊤$ is the product t-norm: the normalized product intersection, denoted by $\varodot$ , is associative (see [16], and a simple proof in [9]). By abuse of notation, we can use the same symbol to denote the conjunctive combination of possibility measures and the normalized product intersection of fuzzy sets, and write
$$
Π_\widetilde{F}\varodotΠ_\widetilde{G}=Π_{\widetilde{F}\varodot{
\widetilde{G}}}.
$$
As noted by Dubois and Prade [16, page 352], product intersection has a reinforcement effect that is appropriate when the information sources are assumed to be independent. The choice of the normalized product intersection for combining possibility distributions makes possibility theory fit in the framework of valuation-based systems [33] and allows for possibilistic reasoning with a large number of variables. The normalized product intersection operator also has an interesting property with respect to Gaussian fuzzy numbers, as recalled in the next paragraph.
#### Gaussian fuzzy numbers
A fuzzy number (or fuzzy interval) can be defined as a normal and convex fuzzy subset of the real line. In particular, a Gaussian fuzzy number (GFN) is a normal fuzzy subset of $ℝ$ with membership function
$$
φ(x;m,h)=\exp≤ft(-\frac{h}{2}(x-m)^2\right),
$$
where $m∈ℝ$ is the mode and $h∈[0,+∞]$ is the precision. Such a fuzzy number will be denoted by $\textsf{GFN}(m,h)$ . If $h=0$ , $φ(x;m,h)=1$ for all $x∈ℝ$ : $\textsf{GFN}(m,0)$ is then maximally imprecise and identical to the whole real line, whatever the value of $m$ . If $h=+∞$ , $φ(x;m,h)=I(x=m)$ , where $I(·)$ is the indicator function; the fuzzy number $\textsf{GFN}(m,+∞)$ is then maximally precise and equivalent to the real number $m$ .
It can easily be shown that the family of GFN’s is closed under the normalized product intersection (see, e.g., [1]). More precisely, we have the following proposition, proved in [1].
**Proposition 3**
*For any $x∈ℝ$ ,
$$
φ(x;m_1,h_1)·φ(x;m_2,h_2)=\exp≤ft(-\frac{h_1h_2(
m_1-m_2)^2}{2(h_1+h_2)}\right)φ(x;m_12,h_12),
$$
with
$$
m_12=\frac{h_1m_1+h_2m_2}{h_1+h_2} and h_12=h_
{1}+h_2.
$$
Consequently,
$$
\textsf{GFN}(m_1,h_1)\varodot\textsf{GFN}(m_2,h_2)=\textsf{GFN}(m_12
,h_12),
$$
and
$$
0pt≤ft[\textsf{GFN}(m_1,h_1)·\textsf{GFN}(m_2,h_2)\right]=\exp
≤ft(-\frac{h_1h_2(m_1-m_2)^2}{2(h_1+h_2)}\right). \tag{10}
$$*
#### Marginalization and cylindrical extension
Let us now assume that we have two variables $\boldsymbol{θ}_1$ and $\boldsymbol{θ}_2$ jointly constrained by a possibility distribution $π_\widetilde{F}$ , where ${\widetilde{F}}$ is a fuzzy subset of $Θ_12=Θ_1×Θ_2$ . As a result of (8), variable $\boldsymbol{θ}_1$ alone is constrained by the possibility distribution
$$
π_1(θ_1)=Π(\{θ_1\}×Θ_2)=\sup_θ_{2∈
Θ_2}π_\widetilde{F}(θ_1,θ_2)=\sup_θ_{2∈
Θ_2}{\widetilde{F}}(θ_1,θ_2)=({\widetilde{F}}↓
Θ_1)(θ_1),
$$
where ${\widetilde{F}}↓Θ_1$ is the projection of ${\widetilde{F}}$ on $Θ_1$ . We say that $π_1$ is the marginal of $π_\widetilde{F}$ on $Θ_1$ . Conversely, given a possibility distribution $π_{\widetilde{F}_1}$ , where ${\widetilde{F}}_1$ is a fuzzy subset of $Θ_1$ , its cylindrical extension in $Θ_1×Θ_2$ is the possibility distribution $π_{\widetilde{F}_1×Θ_2}$ defined as
$$
π_{\widetilde{F}_1×Θ_2}(θ_1,θ_2)=π_{
\widetilde{F}_1}(θ_1)
$$
for all $(θ_1,θ_2)∈Θ_1×Θ_2$ . We say that the joint possibility distribution $π_{\widetilde{F}}$ on $Θ_12$ is noninteractive with respect to the product intersection if it is the product of its marginals:
$$
π_{\widetilde{F}}(θ_1,θ_2)=π_{\widetilde{F}↓
Θ_1}(θ_1)·π_{\widetilde{F}↓Θ_2}(θ_
2).
$$
**Example 4**
*Let $π_12$ be the possibility distribution on $ℝ^2$ defined by
| | $\displaystyleπ_12(x_1,x_2)$ | $\displaystyle=\exp≤ft(-\frac{h_1}{2}(x_1-m_1)^2-\frac{h_2}{2}(x_2 -m_2)^2\right)$ | |
| --- | --- | --- | --- |
Its marginals are
$$
π_1(x_1)=\max_θ_2π_12(x_1,x_2)=\exp≤ft(-\frac{h_1}{
2}(x_1-m_1)^2\right)
$$
and
$$
π_2(x_2)=\max_θ_1π_12(x_1,x_2)=\exp≤ft(-\frac{h_2}{
2}(x_2-m_2)^2\right).
$$
Consequenty, $π_12$ is noninteractive with respect to the product intersection.*
## 3 Epistemic random fuzzy sets
The proposed epistemic random fuzzy set model is introduced in this section. The main definitions are first given in Section 3.1, and the generalized product-intersection rule is introduced in Section 3.2. Marginalization and vacuous extension are then addressed in Section 3.3, and an application to statistical inference is briefly discussed in Section 3.4.
### 3.1 General definitions
As before, let $(Ω,σ_Ω,P)$ be a probability space and let $(Θ,σ_Θ)$ be a measurable space. Let ${\widetilde{X}}$ by a mapping from $Ω$ to the set $[0,1]^Θ$ of fuzzy subsets of $Θ$ . For any $α∈[0,1]$ , let ${}^α{\widetilde{X}}$ be the mapping from $Ω$ to $2^Θ$ defined as
$$
{}^α{\widetilde{X}}(ω)={}^α[{\widetilde{X}}(ω)],
$$
where ${}^α[{\widetilde{X}}(ω)]$ is the weak $α$ -cut of ${\widetilde{X}}(ω)$ . If for any $α∈[0,1]$ , ${}^α{\widetilde{X}}$ is $σ_Ω-σ_Θ$ strongly measurable, the tuple $(Ω,σ_Ω,P,Θ,σ_Θ,{\widetilde{X}})$ is said to be a random fuzzy set (also called a fuzzy random variable) [2]. It is clear that the class of random fuzzy sets includes that of random sets, just as the class of fuzzy sets includes that of classical (crisp) sets.
**Example 5**
*Let $M$ be a Gaussian random variable from $(Ω,σ_Ω,P)$ to $(ℝ,β_ℝ)$ , with mean $μ$ and standard deviation $σ$ , and let ${\widetilde{X}}$ be the mapping from $Ω$ to $[0,1]^ℝ$ that maps each $ω∈Ω$ to the triangular fuzzy number with mode $M(ω)$ and support $[M(ω)-a,M(ω)+a]$ :
$$
{\widetilde{X}}(ω)(x)=\begin{cases}\frac{a-|x-M(ω)|}{a}&if |x
-M(ω)|≤ a\\
0&otherwise. \end{cases}
$$
for some $a>0$ . For any $α∈[0,1]$ , the $α$ -cut of ${\widetilde{X}}(ω)$ is
$$
{}^α{\widetilde{X}}(ω)=≤ft[M(ω)-a(1-α),M(ω)+a(1-
α)\right].
$$
The random set ${}^α{\widetilde{X}}:ω→{}^α{\widetilde{X}}(ω)$ is $σ_Ω-β_ℝ$ strongly measurable (it is a random closed interval). Consequently, ${\widetilde{X}}$ is a random fuzzy set. In the following, such random fuzzy sets with domain $[0,1]^ℝ$ will be called random fuzzy numbers.*
#### Interpretation
Here, as in [9], we use random fuzzy sets as a model of unreliable and fuzzy evidence. In this model, we see $Ω$ as a set of interpretations of a piece of evidence about a variable $\boldsymbol{θ}$ taking values in $Θ$ . If interpretation $ω∈Ω$ holds, we know that “ $\boldsymbol{θ} is {\widetilde{X}}(ω)$ ”, i.e., $\boldsymbol{θ}$ is constrained by the possibility distribution $π_{\widetilde{X}(ω)}$ . We qualify such random fuzzy sets as epistemic, because they encode a state of knowledge about some variable $\boldsymbol{θ}$ . It should be noted that this semantics of random fuzzy sets is different from those reviewed in [2]. The conditional possibility interpretation developed in [2] is the closest to ours, since we also see the fuzzy sets ${\widetilde{X}}(ω)$ as defining conditional possibility measures. However, in [2], the authors use the random fuzzy set formalism to model a situation in which we have two random experiments, one of which is completely determined; the family of possibility distributions $\{π_{\widetilde{X}(ω)}:ω∈Ω\}$ then models our knowledge about the relationship between the outcomes $ω$ of the first experiment and the possible outcomes of the second one. This formalism allows the authors of [2] to compute lower and upper bounds on the probability of any event related to the second experiment. In contrast, our model does not rely on the notion of random experiment. In particular, we do not postulate the existence of an objective probability measure on $Θ$ , and the belief and plausibility measures introduced below are not interpreted as lower and upper bounds on “true” probabilities.
#### Belief and plausibility
We say that random fuzzy set ${\widetilde{X}}$ is normalized if it verifies the following conditions:
1. For all $ω∈Ω$ , ${\widetilde{X}}(ω)$ is either the empty set, or a normal fuzzy set, i.e., $0pt({\widetilde{X}}(ω))∈\{0,1\}$ .
1. $P(\{ω∈Ω:{\widetilde{X}}(ω)=∅\})=0$ .
These conditions will be assumed in the rest of this section. For any $ω∈Ω$ , let $Π_\widetilde{X}(·\midω)$ be the possibility measure on $Θ$ induced by ${\widetilde{X}}(ω)$ :
$$
Π_\widetilde{X}(B\midω)=\sup_θ∈ B{\widetilde{X}}(ω)(
θ), \tag{11}
$$
and let $N_\widetilde{X}(·\midω)$ be the dual necessity measure:
$$
N_\widetilde{X}(B\midω)=\begin{cases}1-Π_\widetilde{X}(B^c\mid
ω)&if {\widetilde{X}}(ω)≠∅\\
0&otherwise. \end{cases}
$$
Let $Bel_\widetilde{X}$ and $Pl_\widetilde{X}$ be the mappings from $σ_Θ$ to $[0,1]$ defined as
$$
Bel_\widetilde{X}(B)=∫_ΩN(B\midω)dP(ω) \tag{12}
$$
and
$$
Pl_\widetilde{X}(B)=∫_ΩΠ(B\midω)dP(ω). \tag{13}
$$
Function $Bel_\widetilde{X}$ is a belief function, and $Pl_\widetilde{X}$ is the dual plausibility function. As shown in [2, Lemma 6.2], they are induced by the random set $(Ω×[0,1],σ_Ω⊗β_[0,1],P⊗λ,Θ, σ_Θ,{\overline{X}})$ , where ${\overline{X}}:Ω×[0,1]→ 2^Θ$ is the multi-valued mapping defined as
$$
{\overline{X}}(ω,α)={}^α{\widetilde{X}}(ω). \tag{14}
$$
As a consequence, $Bel_\widetilde{X}(B)$ and $Pl_\widetilde{X}(B)$ can also be written as follows:
$$
Bel_\widetilde{X}(B)=∫_0^1Bel_{^α{\widetilde{X}}}(B)dα Pl_\widetilde{X}(B)=∫_0^1Pl_{^α{\widetilde{X}}}(B)dα.
$$
#### Lower and upper expectations of a random fuzzy number
Let ${\widetilde{X}}$ be a random fuzzy number (i.e., a random fuzzy set with domain $[0,1]^ℝ$ ), and let ${\overline{X}}$ be the corresponding random set defined by (14). We define the lower and upper expectations of ${\widetilde{X}}$ as the lower and upper expectations of ${\overline{X}}$ , i.e., $E_*({\widetilde{X}})=E_*({\overline{X}})$ and $E^*({\widetilde{X}})=E^*({\overline{X}})$ . It follows from (15) that
$$
E_*({\widetilde{X}})=∫_0^1E_*({}^α{
\widetilde{X}})dα and E^*({\widetilde{X}})=∫
_0^1E^*({}^α{\widetilde{X}})dα. \tag{16}
$$
**Example 6**
*Let us consider again the random fuzzy number of Example 5. Its lower and upper cdf’s are, respectively, the mappings $x→ Bel_\widetilde{X}((-∞,x])$ and $x→ Pl_\widetilde{X}((-∞,x])$ . Let us illustrate the calculation of the upper cdf first, using two methods.*
#### Method 1
From (11),
$$
Π≤ft((-∞,x]\midω\right)=\sup_x^\prime≤ x{\widetilde{X}}(
ω)(x^\prime)=\begin{cases}1&if M(ω)≤ x\\
\frac{x-M(ω)+a}{a}&if x<M(ω)≤ x+a\\
0&otherwise.\end{cases}
$$
Using (13), we get
| | $\displaystyle Pl_\widetilde{X}((-∞,x])$ | $\displaystyle=P(M≤ x)× 1+P(x<M≤ x+a)E≤ft[\frac{x-M+a}{a} \mid x<M≤ x+a\right]$ | |
| --- | --- | --- | --- |
Now, using a well-known result about the truncated normal distribution,
$$
E≤ft[M\mid x<M≤ x+a\right]=μ+σ\frac{φ≤ft(\frac{x-μ
}{σ}\right)-φ≤ft(\frac{x+a-μ}{σ}\right)}{Φ≤ft(\frac{x+a-
μ}{σ}\right)-Φ≤ft(\frac{x-μ}{σ}\right)}.
$$
After rearranging the terms, we finally obtain
$$
Pl_\widetilde{X}((-∞,x])=≤ft(\frac{x+a-μ}{a}\right)Φ≤ft(\frac{
x+a-μ}{σ}\right)-≤ft(\frac{x-μ}{a}\right)Φ≤ft(\frac{x-μ}{
σ}\right)+\\
\frac{σ}{a}≤ft[φ≤ft(\frac{x+a-μ}{σ}\right)-φ≤ft(\frac{
x-μ}{σ}\right)\right]. \tag{17}
$$
#### Method 2
Let us now use (15b). We have
| | $\displaystyle Pl_\widetilde{X}((-∞,x])$ | $\displaystyle=∫_0^1P(M-a(1-α)≤ x)dα$ | |
| --- | --- | --- | --- |
Using the formula
$$
∫Φ(u+vx)dx=\frac{1}{v}≤ft[(u+vx)Φ(u+vx)+φ(u+vx)\right]+C,
$$
we get the same result as (17). Using any of the two methods demonstrated above, we obtain the following expression for the lower cdf:
$$
Bel_\widetilde{X}((-∞,x])=≤ft(\frac{x-μ}{a}\right)Φ≤ft(\frac{x
-μ}{σ}\right)-≤ft(\frac{x-a-μ}{a}\right)Φ≤ft(\frac{x-a-μ}{
σ}\right)+\\
\frac{σ}{a}≤ft[φ≤ft(\frac{x-μ}{σ}\right)-φ≤ft(\frac{x-
a-μ}{σ}\right)\right]. \tag{18}
$$
It can easily be checked that, when $a=0$ ,
$$
Bel_\widetilde{X}((-∞,x])=Pl_\widetilde{X}((-∞,x])=Φ≤ft(
\frac{x-μ}{σ}\right).
$$
Examples of functions $Bel_\widetilde{X}((-∞,x])$ and $Pl_\widetilde{X}((-∞,x])$ for different values of $a$ are shown in Figure 1.
Figure 1: Lower and upper cdf’s for the random fuzzy numbers studied in Examples 5 and 6, with $μ=0$ , $σ=1$ , and $a=0.5$ (blue curves) or $a=1.5$ (red curves). The Gaussian cdf corresponding to $a=0$ is shown as a broken line.
Now, the lower and upper expectations of ${\widetilde{X}}$ can be computed from (16) as
$$
E_*({\widetilde{X}})=∫_0^1E_*({}^α{
\widetilde{X}})dα=∫_0^1[μ-a(1-α)]dα=μ-\frac{a}{2},
$$
and
$$
E^*({\widetilde{X}})=∫_0^1E^*({}^α{
\widetilde{X}})dα=∫_0^1[μ+a(1-α)]dα=μ+\frac{a}{2}.
$$
### 3.2 Generalized product-intersection rule
Dempster’s rule and the possibilistic product intersection rule recalled, respectively, in Sections 2.1 and 2.2 can be generalized to combine epistemic random fuzzy sets. Consider two epistemic random fuzzy sets $(Ω_1,σ_1,P_1,Θ,σ_Θ,{\widetilde{X}}_1)$ and $(Ω_2,σ_2,P_2,Θ,σ_Θ,{\widetilde{X}}_2)$ encoding independent pieces of evidence. The independence assumption means here that the relevant probability measure on the joint measurable space $(Ω_1×Ω_2,σ_1⊗σ_2$ ) is the product measure $P_1× P_2$ .
If interpretations $ω_1∈Ω_1$ and $ω_2∈Ω_2$ both hold, we know that “ $\boldsymbol{θ} is {\widetilde{X}}_1(ω_1)$ ” and “ $\boldsymbol{θ} is {\widetilde{X}}_2(ω_2)$ ”. It is then natural to combine the fuzzy sets ${\widetilde{X}}_1(ω_1)$ and ${\widetilde{X}}_2(ω_2)$ by an intersection operator. As discussed in Section 2.2, normalized product intersection is a good candidate as it suitable for combining fuzzy information from independent sources and it is associative. We will thus consider the mapping ${\widetilde{X}}_\varodot(ω_1,ω_2)={\widetilde{X}}_1(ω_ 1)\varodot{\widetilde{X}}_2(ω_2)$ , which we will assume to be $σ_1⊗σ_2$ - $σ_Θ$ strongly measurable.
As in the crisp case recalled in Section 2.1, if $0pt({\widetilde{X}}_1(ω_1){\widetilde{X}}_2(ω_2))=0$ , the two interpretations $ω_1$ and $ω_2$ are inconsistent and they must be discarded. If $0pt({\widetilde{X}}_1(ω_1){\widetilde{X}}_2(ω_2))=1$ , the two interpretations are fully consistent. If $0<0pt({\widetilde{X}}_1(ω_1){\widetilde{X}}_2(ω_2))<1$ , $ω_1$ and $ω_2$ are partially consistent. As proposed in [9], instead of simply discarding only fully inconsistent pairs $(ω_1,ω_2)$ , it makes sense to give all pairs $(ω_1,ω_2)$ a weight proportional to the degree of consistency between ${\widetilde{X}}_1(ω_1)$ and ${\widetilde{X}}_2(ω_2)$ . This can be achieved by conditioning $P_1× P_2$ on the fuzzy set ${\widetilde{Θ}}^*$ of consistent pairs of interpretations, with membership function
$$
{\widetilde{Θ}}^*(ω_1,ω_2)=0pt≤ft({\widetilde{X}}_1(
ω_1)·{\widetilde{X}}_2(ω_2)\right).
$$
Using Zadeh’s definition of a fuzzy event [37], we get the following expression for the conditional probability measure ${\widetilde{P}}_12=(P_1× P_2)(·\mid{\widetilde{Θ}}^*)$ , for any $B∈σ_1⊗σ_2$ :
$$
{\widetilde{P}}_12(B)=\frac{(P_1× P_2)(B∩{\widetilde{Θ}}^*
)}{(P_1× P_2)({\widetilde{Θ}}^*)}=\frac{∫_Ω_1∫
_Ω_2B(ω_1,ω_2)0pt≤ft({\widetilde{X}}_1(ω_1)
·{\widetilde{X}}_2(ω_2)\right)dP_2(ω_2)dP_1(ω_1
)}{∫_Ω_1∫_Ω_20pt≤ft({\widetilde{X}}_1(ω_1)
·{\widetilde{X}}_2(ω_2)\right)dP_2(ω_2)dP_1(ω_1
)},
$$
where $B(·,·)$ denotes the indicator function of $B$ . This conditioning operation, called soft normalization was first proposed in [35] in the finite case and with a different justification.
The combined random fuzzy set
$$
(Ω_1×Ω_2,σ_1⊗σ_2,{\widetilde{P}}_12,
Θ,σ_Θ,{\widetilde{X}}_\varodot)
$$
is called the orthogonal sum of the two pieces of evidence. This operation generalizes both Dempster’s rule and the normalized product of possibility distribution. We will refer to it as the generalized product-intersection rule, and it will be denoted by the same symbol $⊕$ as Dempster’s rule. It is clear that ${\widetilde{X}}⊕{\overline{X}}_0={\widetilde{X}}$ for any random fuzzy set ${\widetilde{X}}$ and any vacuous random set ${\overline{X}}_0$ on the same domain $Θ$ . The degree of conflict between two random fuzzy sets ${\widetilde{X}}_1$ and ${\widetilde{X}}_2$ is naturally defined as
$$
κ=1-(P_1× P_2)({\widetilde{Θ}}^*)=1-∫_Ω_1∫
_Ω_20pt≤ft({\widetilde{X}}_1(ω_1){\widetilde{X}}_2(
ω_2)\right)dP_2(ω_2)dP_1(ω_1). \tag{19}
$$
The associativity of $⊕$ was proved in [9] in the finite case; we give a similar proof in the general case.
**Proposition 4**
*The generalized product-intersection rule $⊕$ for random fuzzy sets is commutative and associative.*
* Proof*
See B ∎
The following proposition states that a counterpart of Proposition 2 is still valid when combining independent random fuzzy sets, i.e., the combined contour function is still proportional to the product of the contour functions.
**Proposition 5**
*Let ${\widetilde{X}}_1$ and ${\widetilde{X}}_2$ be two random fuzzy sets on the same domain $Θ$ , with contour functions $pl_{\widetilde{X}_1}$ and $pl_{\widetilde{X}_2}$ and with degree of conflict $κ$ defined by (19). The contour function $pl_{\widetilde{X}_1⊕{\widetilde{X}}_2}$ of ${\widetilde{X}}_1⊕{\widetilde{X}}_2$ verifies
$$
(pl_{\widetilde{X}_1⊕{\widetilde{X}}_2})(θ)=\frac{pl_{
\widetilde{X}_1}(θ)pl_{\widetilde{X}_2}(θ)}{1-κ}, \tag{20}
$$
for all $θ∈Θ$ .*
* Proof*
We have
| | $\displaystyle(pl_{\widetilde{X}_1⊕{\widetilde{X}}_2})(θ)$ | $\displaystyle=\frac{∫_Ω_1∫_Ω_20pt≤ft({\widetilde{X}} _1(ω_1)·{\widetilde{X}}_2(ω_2)\right){\widetilde{X}}_ \varodot(ω_1,ω_2)(θ)dP_2(ω_2)dP_1(ω_1)}{ 1-κ}$ | |
| --- | --- | --- | --- |
∎
As remarked in Section 2.2, a belief function induced by a random fuzzy set is also induced by a random (crisp) set. However, combining random fuzzy sets or random crisp sets does not result in the same belief function in general. In particular, it is well-known that Dempster’s rule does not preserve consonance. To combine two belief functions, we must, therefore, examine the evidence on which they are based, not only to determine whether the bodies of evidence are independent or not, but also to determine whether the evidence is fuzzy or crisp. This point is illustrated by the following example.
**Example 7**
*Consider the following two mappings from $ℝ$ to $[0,1]$ represented in Figure 2a:
$$
π_1(x)=\textsf{GFN}(0,0.3), π_2(x)=\textsf{GFN}≤ft(1,0.5\right).
$$
If these two mappings are possibility distributions encoding fully reliable but fuzzy evidence, they correspond to “constant random fuzzy sets”, i.e., mappings ${\widetilde{X}}_1(ω)=π_1$ and ${\widetilde{X}}_2(ω)=π_2$ with $P(\{ω\})=1$ . The combined random fuzzy set ${\widetilde{X}}_1⊕{\widetilde{X}}_2$ is then defined by $({\widetilde{X}}_1⊕{\widetilde{X}}_2)(ω)=π_1\varodotπ_2$ . From Proposition 3, the normalized product of two GFN’s is a GFN. Here, we get the combined possibility distribution plotted as a red broken curve in Figure 2a:
$$
(π_1\varodotπ_2)(x)=\textsf{GFN}(0.625,0.8).
$$
(a)
(b)
Figure 2: (a): Two Gaussian possibility distributions (black solid curves) with their normalized product intersection (red broken curve) and the contour function of the combined random set (blue solid curve). (b): Lower and upper cdf’s of the combined possibility distribution (red broken curves) and of the combined random set (blue solid curves). The corresponding lower and upper cumulative distribution functions (cdf’s) are, respectively
$$
Bel_{\widetilde{X}_1⊕{\widetilde{X}}_2}((-∞,x])=\begin{cases}0
&if x≤ 0.625\\
1-\exp≤ft(-0.4(x-0.625)^2\right)&if x>0.625\end{cases}
$$
and
$$
Pl_{\widetilde{X}_1⊕{\widetilde{X}}_2}((-∞,x])=\begin{cases}
\exp≤ft(-0.4(x-0.625)^2\right)&if x≤ 0.625\\
1&if x>0.625.\end{cases}
$$ These two functions are plotted as red broken curves in Figure 2b. Alternatively, as explained in Section 2.1, we may see $π_1$ and $π_2$ as encoding crisp but partially reliable evidence, in which case they define two independent consonant random intervals ${\overline{X}}_1(α_1)={}^α_1π_1$ and ${\overline{X}}(α_2)={}^α_2π_2$ , where $(α_1,α_2)$ has a uniform distribution on $[0,1]^2$ . These two random intervals can be combined numerically using Monte-Carlo simulation, as explained in [20]. The contour function and the lower and upper cdf’s are plotted as solid blue lines in Figures 2a and 2b, respectively. We notice that the contour functions are proportional, as a consequence of Proposition 2.*
### 3.3 Marginalization and vacuous extension
Let us now consider again the case where we have two variables $\boldsymbol{θ}_1$ and $\boldsymbol{θ}_2$ with respective domains $Θ_1$ and $Θ_2$ . Let ${\widetilde{X}}_12$ be a random fuzzy set from a probability space $(Ω,σ_Ω,P)$ to the measurable space $(Θ_12,σ_Θ_12)$ with $Θ_12=Θ_1×Θ_2$ and $σ_Θ_12=σ_Θ_1⊗σ_Θ_2$ , where $σ_Θ_1$ and $σ_Θ_2$ are $σ$ -algebras on $Θ_1$ and $Θ_2$ , respectively. Let ${\widetilde{X}}_1$ be the mapping from $Ω$ to $[0,1]^Θ_1$ defined by
$$
{\widetilde{X}}_1(ω)={\widetilde{X}}_12(ω)↓Θ_1,
$$
where, as before, $↓$ denotes fuzzy set projection. If, for all $α∈[0,1]$ , the mapping ${}^α{\widetilde{X}}_1$ is $σ_Ω-σ_Θ_1$ strongly measurable, then the random fuzzy set ${\widetilde{X}}_1$ is called the marginal of ${\widetilde{X}}_12$ on $Θ_1$ .
Conversely, given a random fuzzy set ${\widetilde{X}}_1$ from $(Ω,σ_Ω,P)$ to $(Θ_1,σ_Θ_1)$ , let ${\widetilde{X}}_1↑(1,2)$ be the mapping from $Ω$ to $[0,1]^Θ_12$ that maps each $ω∈Ω$ to the cylindrical extension of ${\widetilde{X}}_1(ω)$ in $Θ_12$
$$
{\widetilde{X}}_1↑(1,2)(ω)={\widetilde{X}}_1(ω)×
Θ_2,
$$
i.e., for all $(θ_1,θ_2)∈Θ_12$ ,
$$
{\widetilde{X}}_1↑(1,2)(ω)(θ_1,θ_2)={\widetilde{X}
}_1(ω)(θ_1).
$$
If the mapping ${\widetilde{X}}_1↑(1,2)$ is $σ_Ω-σ_Θ_12$ strongly measurable, then the random fuzzy set ${\widetilde{X}}_1↑(1,2)$ is called the vacuous extension of ${\widetilde{X}}_1$ in $Θ_12$ .
We say that a joint random fuzzy set is noninteractive if it is equal to the orthogonal sum of the vacuous extensions of its projections:
$$
{\widetilde{X}}_12={\widetilde{X}}_1↑(1,2)⊕{\widetilde{X}}_2
↑(1,2) denoted as {\widetilde{X}}_1⊕{\widetilde{
X}}_2.
$$
A particular kind of noninteractive random fuzzy sets will be studied in Section 5.3.
### 3.4 Application to statistical inference
Epistemic random fuzzy sets naturally arise in the context of statistical inference. As proposed by Shafer [29] and formally justified in [7] [8], the information conveyed by the likelihood function in statistical inference problems can be represented by a consonant belief function, whose contour function is equal to the relative likelihood function. For a statistical model $f({\boldsymbol{x}},θ)$ , where ${\boldsymbol{x}}∈{\cal X}$ is the observation and $θ∈Θ$ is the unknown parameter, the likelihood-based belief function $Bel(·,{\boldsymbol{x}})$ on $Θ$ after observing ${\boldsymbol{x}}$ is, thus, consonant and defined by the contour function
$$
pl(θ;{\boldsymbol{x}})=\frac{L(θ;{\boldsymbol{x}})}{\sup_θ^
\prime∈ΘL(θ^\prime;{\boldsymbol{x}})}, \tag{21}
$$
where $L(·,{\boldsymbol{x}}):θ→ f({\boldsymbol{x}};θ)$ is the likelihood function, and it is assumed that the denominator in (21) is finite. The corresponding plausibility function is, thus, defined by
$$
Pl(A;{\boldsymbol{x}})=\sup_θ∈ Apl(θ;{\boldsymbol{x}})
$$
for any $A⊂Θ$ , i.e., it is a possibility measure. However, as noticed by Shafer in [29] and [31], and also discussed in [7], this construction is not compatible with Dempster’s rule: if we consider two independent observations ${\boldsymbol{x}}$ and ${\boldsymbol{x}}^\prime$ , the belief function $Bel(·;{\boldsymbol{x}},{\boldsymbol{x}}^\prime)$ is not equal to the orthogonal sum $Bel(·;{\boldsymbol{x}})⊕ Bel(·;{\boldsymbol{x}}^\prime)$ , which is not consonant. As argued in [9], this problem disappears if we do not consider the likelihood-based belief function to be induced by a consonant random crisp set, but by a constant random fuzzy set ${\widetilde{θ}}_\boldsymbol{x}$ with membership function ${\widetilde{θ}}_\boldsymbol{x}(θ)=pl(θ;{\boldsymbol{x}})$ . We can interpret ${\widetilde{θ}}_\boldsymbol{x}$ as the fuzzy set of likely values of $θ$ after observing ${\boldsymbol{x}}$ . Combining the contour functions (21) by the normalized product intersection rule then yields the correct result, i.e., the constant random fuzzy set ${\widetilde{θ}}_{\boldsymbol{x},{\boldsymbol{x}}^\prime}$ with membership function ${\widetilde{θ}}_{\boldsymbol{x},{\boldsymbol{x}}^\prime}(θ)={ \widetilde{θ}}_\boldsymbol{x}(θ)\varodot{\widetilde{θ}}_{ \boldsymbol{x}^\prime}(θ)$ .
Now, consider a prediction problem, where we want to predict the value of a random variable $Y$ whose distribution also depends on $θ$ . We can always write $Y$ in the form $Y=φ(θ,U)$ , where $U$ is a pivotal random variable with known distribution [19, 20]. After observing the data ${\boldsymbol{x}}$ , our knowledge of $θ$ is represented by the fuzzy set ${\widetilde{θ}}_\boldsymbol{x}$ . Conditionally on $U=u$ , our knowledge of $Y$ is, thus, represented by the fuzzy set ${\widetilde{Y}}(u)=φ({\widetilde{θ}}_\boldsymbol{x},u)$ , with membership function
$$
{\widetilde{Y}}(u)(y)=\sup_θ:φ(θ,u)=y{\widetilde{θ}}_
\boldsymbol{x}(θ).
$$
The mapping ${\widetilde{Y}}:u→{\widetilde{Y}}(u)$ is, then, a random fuzzy set representing statistical evidence about $Y$ .
**Example 8**
*Let ${\boldsymbol{X}}=(X_1,…,X_n)$ be an independent and identically distributed (iid) Gaussian sample with parent distribution $N(θ,1)$ , and let $Y∼ N(θ,1)$ . After observing a realization ${\boldsymbol{x}}$ of ${\boldsymbol{X}}$ , the likelihood function is
$$
L(θ;{\boldsymbol{x}})=(2π)^-n/2\exp≤ft(-\frac{1}{2}∑_i=1^n(x
_i-θ)^2\right).
$$
Denoting by $\widehat{θ}$ the sample mean, the fuzzy set ${\widetilde{θ}}_\boldsymbol{x}$ of likely values of $θ$ after observing ${\boldsymbol{x}}$ is the relative likelihood
$$
{\widetilde{θ}}_\boldsymbol{x}(θ)=\frac{L(θ;{\boldsymbol{x}})
}{L(\widehat{θ};{\boldsymbol{x}})}=\exp≤ft(-\frac{n}{2}(θ-\widehat
{θ})^2\right).
$$
It is the Gaussian fuzzy number $\textsf{GFN}(\widehat{θ},n)$ with mode $\widehat{θ}$ and precision $n$ . Now, $Y$ can be written as $Y=θ+U$ , with $U∼ N(0,1)$ . Consequently, the conditional possibility distribution on $Y$ given $U=u$ is the Gaussian fuzzy number ${\widetilde{θ}}_\boldsymbol{x}+u=\textsf{GFN}(\widehat{θ}+u,n)$ , and our knowledge of $Y$ is described by the random fuzzy set $U→\textsf{GFN}(\widehat{θ}+U,n)$ , with $U∼ N(0,1)$ . This is a Gaussian fuzzy number with fixed precision $h=n$ and normal random mode $M=\widehat{θ}+U∼ N(\widehat{θ},1)$ . This important class of random fuzzy sets will be studied in the next section.*
## 4 Gaussian random fuzzy numbers
In this section, we introduce Gaussian random fuzzy numbers (GRFN’s) as a practical model for representing uncertainty on a real variable. As we will see, this model encompasses Gaussian random variables and Gaussian fuzzy numbers as special cases. A GRFN can be seen, equivalently, as a Gaussian random variable with fuzzy mean, or as a Gaussian fuzzy number with random mode. The definition and main properties will first be presented in Section 4.1. The expression of the orthogonal sum of two GRFN’s will then be derived in Section 4.2. Finally, arithmetic operations on GRFN’s will be addressed in Section 4.3.
### 4.1 Definition and main properties
**Definition 1**
*Let $(Ω,σ_Ω,P)$ be a probability space and let $M:Ω→ℝ$ be a Gaussian random variable with mean $μ$ and variance $σ^2$ . The random fuzzy set ${\widetilde{X}}:Ω→[0,1]^ℝ$ defined as
$$
{\widetilde{X}}(ω)=\textsf{GFN}(M(ω),h)
$$
is called a Gaussian random fuzzy number (GRFN) with mean $μ$ , variance $σ^2$ and precision $h$ , which we write ${\widetilde{X}}∼{\widetilde{N}}(μ,σ^2,h)$ .*
In the definition of a GRFN, $μ$ is a location parameter, while parameters $h$ and $σ^2$ correspond, respectively, to possibilistic and probabilistic uncertainty. If $h=0$ , imprecision is maximal whatever the values of $μ$ and $σ^2$ : the GRFN ${\widetilde{X}}$ then induces the vacuous belief function on $ℝ$ , in which case $Bel_\widetilde{X}(A)=0$ for all $A⊂ℝ$ , and $Pl_\widetilde{X}(A)=1$ for all $A⊆ℝ$ such that $A≠∅$ ; such a GRFN will be said to be vacuous and will be denoted by ${\widetilde{X}}∼{\widetilde{N}}(0,1,0)$ . If $h=+∞$ , each fuzzy number $\textsf{GFN}(M(ω),h)$ is reduced to a point: the GRFN ${\widetilde{X}}$ is then equivalent to a Gaussian random variable with mean $μ$ and variance $σ^2$ , which we can write: ${\widetilde{N}}(μ,σ^2,+∞)=N(μ,σ^2)$ . Another special case of interest is that where $σ^2=0$ , in which case $M$ is a constant random variable taking value $μ$ , and ${\widetilde{X}}$ is a possibilistic variable with possibility distribution $\textsf{GFN}(μ,h)$ .
The following proposition gives the expression of the contour functions $pl_\widetilde{X}(x)$ associated to ${\widetilde{X}}$ .
**Proposition 6**
*The contour function of GRFN ${\widetilde{X}}∼{\widetilde{N}}(μ,σ^2,h)$ is
$$
pl_\widetilde{X}(x)=\frac{1}{√{1+hσ^2}}\exp≤ft(-\frac{h(x-μ)^
{2}}{2(1+hσ^2)}\right). \tag{22}
$$*
* Proof*
See C. ∎
A shown by Proposition 6, the contour function $pl_\widetilde{X}$ is constant in two cases: if $h=0$ , ${\widetilde{X}}$ is vacuous, and $pl_\widetilde{X}(x)=1$ for all $x∈ℝ$ ; if $h=+∞$ , ${\widetilde{X}}$ is a random variable, and $pl_\widetilde{X}(x)=0$ for all $x∈ℝ$ . We also note that, if $σ^2=0$ , $pl_\widetilde{X}$ is equal to the possibility distribution $\textsf{GFN}(μ,h)$ . When $σ^2→+∞$ and $h>0$ , $pl_\widetilde{X}(x)→ 0$ for all $x$ . The next proposition gives the expressions of the belief and plausibility of any real interval.
**Proposition 7**
*For any real interval $[x,y]$ , the degrees of belief and plausibility of $[x,y]$ induced by the GRFN ${\widetilde{X}}∼{\widetilde{N}}(μ,σ^2,h)$ are, respectively,
$$
Bel_\widetilde{X}([x,y])=Φ≤ft(\frac{y-μ}{σ}\right)-Φ≤ft(
\frac{x-μ}{σ}\right)-\\
pl_\widetilde{X}(x)≤ft[Φ≤ft(\frac{(x+y)/2-μ+(y-x)hσ^2/2}{
σ√{hσ^2+1}}\right)-Φ≤ft(\frac{x-μ}{σ√{hσ^
{2}+1}}\right)\right]-\\
pl_\widetilde{X}(y)≤ft[Φ≤ft(\frac{y-μ}{σ√{hσ^2+1}}
\right)-Φ≤ft(\frac{(x+y)/2-μ-(y-x)hσ^2/2}{σ√{hσ^2
+1}}\right)\right], \tag{23}
$$
and
$$
Pl_\widetilde{X}([x,y])=Φ≤ft(\frac{y-μ}{σ}\right)-Φ≤ft(
\frac{x-μ}{σ}\right)+pl_\widetilde{X}(x)Φ≤ft(\frac{x-μ}{σ
√{hσ^2+1}}\right)+\\
pl_\widetilde{X}(y)≤ft[1-Φ≤ft(\frac{y-μ}{σ√{hσ^2+1}
}\right)\right]. \tag{24}
$$*
* Proof*
See D. ∎
**Corollary 1**
*The lower and upper cdf’s of the GRFN ${\widetilde{X}}∼{\widetilde{N}}(μ,σ^2,h)$ are, respectively
$$
Bel_\widetilde{X}((-∞,y])=Φ≤ft(\frac{y-μ}{σ}\right)-pl_
\widetilde{X}(y)Φ≤ft(\frac{y-μ}{σ√{hσ^2+1}}\right) \tag{25}
$$
and
$$
Pl_\widetilde{X}((-∞,y])=Φ≤ft(\frac{y-μ}{σ}\right)+pl_
\widetilde{X}(y)≤ft[1-Φ≤ft(\frac{y-μ}{σ√{hσ^2+1}}
\right)\right]. \tag{26}
$$*
* Proof*
Immediate from Proposition 7 by letting $x$ tend to $-∞$ in (23) and (24) ∎
We can easily check from (23) and (24) that $Bel_\widetilde{X}([x,y])$ and $Pl_\widetilde{X}([x,y])$ both tend to $Φ≤ft(\frac{y-μ}{σ}\right)-Φ≤ft(\frac{x-μ}{σ}\right)$ when $h→∞$ , which is consistent with the fact that a GRFN with infinite precision is equivalent to a Gaussian random variable. Finally, the following proposition gives the expressions of the lower and upper expectations of a GRFN.
**Proposition 8**
*Let ${\widetilde{X}}∼{\widetilde{N}}(μ,σ^2,h)$ be a GRFN with $h>0$ . Its lower and upper expectations are, respectively,
$$
E_*({\widetilde{X}})=μ-√{\frac{π}{2h}} and
E^*({\widetilde{X}})=μ+√{\frac{π}{2h}}. \tag{27}
$$*
* Proof*
See E. ∎
As expected, we can see from (27) that the lower and upper expectations boil down to the usual expectation $μ$ when $h=+∞$ .
### 4.2 Orthogonal sum of Gaussian random fuzzy numbers
In this section, we derive the expression of the orthogonal sum ${\widetilde{X}}_1⊕{\widetilde{X}}_2$ of two GRFN’s ${\widetilde{X}}_1$ and ${\widetilde{X}}_2$ . We start with the following lemma.
**Lemma 1**
*Let $M_1∼ N(μ_1,σ_1^2)$ and $M_2∼ N(μ_2,σ_2^2)$ be two independent Gaussian random variables, and let ${\widetilde{F}}$ be the fuzzy subset of $ℝ^2$ with membership function
$$
{\widetilde{F}}(m_1,m_2)=0pt≤ft(\textsf{GFN}(m_1,h_1)·\textsf{
GFN}(m_2,h_2)\right)=\exp≤ft(-\frac{h_1h_2(m_1-m_2)^2}{2(h_1
+h_2)}\right).
$$
The conditional probability distribution of $(M_1,M_2)$ given ${\widetilde{F}}$ is two-dimensional Gaussian with mean vector ${\widetilde{\boldsymbol{μ}}}=({\widetilde{μ}}_1,{\widetilde{μ}}_2)^T$ and covariance matrix
$$
{\widetilde{\boldsymbol{Σ}}}=\begin{pmatrix}{\widetilde{σ}}_1^2&
ρ{\widetilde{σ}}_1{\widetilde{σ}}_2\\
ρ{\widetilde{σ}}_1{\widetilde{σ}}_2&{\widetilde{σ}}_2^
{2}\end{pmatrix},
$$
with
$$
\displaystyle{\widetilde{μ}}_1 \displaystyle=\frac{μ_1(1+{\overline{h}}σ_2^2)+μ_2{\overline{
h}}σ_1^2}{1+{\overline{h}}(σ_1^2+σ_2^2)} \displaystyle{\widetilde{μ}}_2 \displaystyle=\frac{μ_2(1+{\overline{h}}σ_1^2)+μ_1{\overline{
h}}σ_2^2}{1+{\overline{h}}(σ_1^2+σ_2^2)} \displaystyle{\widetilde{σ}}_1^2 \displaystyle=\frac{σ_1^2(1+{\overline{h}}σ_2^2)}{1+{
\overline{h}}(σ_1^2+σ_2^2)} \displaystyle{\widetilde{σ}}_2^2 \displaystyle=\frac{σ_2^2(1+{\overline{h}}σ_1^2)}{1+{
\overline{h}}(σ_1^2+σ_2^2)} \displaystyleρ \displaystyle=\frac{{\overline{h}}σ_1σ_2}{√{(1+{\overline{h}
}σ_1^2)(1+{\overline{h}}σ_2^2)}}, {\overline{h}}=\frac{h_1h_2}{h_1+h_2}.
$$
Furthermore, the degree of conflict between two independent GRFN’s ${\widetilde{X}}_1∼{\widetilde{N}}(μ_1,σ_1^2,h_1)$ and ${\widetilde{X}}_2∼{\widetilde{N}}(μ_2,σ_2^2,h_2)$ is
$$
κ=1-\iint f(m_1,m_2){\widetilde{F}}(m_1,m_2)dm_1dm_2=\\
1-\frac{{\widetilde{σ}}_1{\widetilde{σ}}_2}{σ_1σ_2
}√{1-ρ^2}\exp≤ft\{-\frac{1}{2}≤ft[\frac{μ_1^2}{σ_1^
2}+\frac{μ_2^2}{σ_2^2}\right]+\frac{1}{2(1-ρ^2)}≤ft[
\frac{{\widetilde{μ}}_1^2}{{\widetilde{σ}}_1^2}+\frac{{
\widetilde{μ}}_2^2}{{\widetilde{σ}}_2^2}-2ρ\frac{{\widetilde
{μ}}_1{\widetilde{μ}}_2}{{\widetilde{σ}}_1{\widetilde{σ}}_
{2}}\right]\right\},
$$
where $f(m_1,m_2)$ is the pdf of random vector $(M_1,M_2)$ .*
* Proof*
See F. ∎
**Proposition 9**
*Let ${\widetilde{X}}_1∼{\widetilde{N}}(μ_1,σ_1^2,h_1)$ and ${\widetilde{X}}_2∼{\widetilde{N}}(μ_2,σ_2^2,h_2)$ be two independent GRFN’s, and assume that $h_1>0$ or $h_2>0$ . We have
$$
{\widetilde{X}}_1⊕{\widetilde{X}}_2∼{\widetilde{N}}({\widetilde{
μ}}_12,{\widetilde{σ}}_12^2,h_12),
$$
with
$$
h_12=h_1+h_2, \tag{29}
$$
$$
{\widetilde{μ}}_12=\frac{h_1{\widetilde{μ}}_1+h_2{\widetilde{μ}}
_2}{h_1+h_2}, \tag{30}
$$
and
$$
{\widetilde{σ}}_12^2=\frac{h^2_1{\widetilde{σ}}^2_1+h^2
_2{\widetilde{σ}}^2_2+2ρ h_1h_2{\widetilde{σ}}_1{
\widetilde{σ}}_2}{(h_1+h_2)^2}, \tag{31}
$$
where ${\widetilde{μ}}_1$ , ${\widetilde{μ}}_2$ , ${\widetilde{σ}}_1$ , ${\widetilde{σ}}_2$ and $ρ$ are given by (28) in Lemma 1.*
* Proof*
Let $M_1$ and $M_2$ be the Gaussian random variables from $(Ω_1,σ_1,P_1)$ and $(Ω_2,σ_2,P_2)$ to $(ℝ,β_ℝ)$ corresponding, respectively, to GRFN’s ${\widetilde{X}}_1∼{\widetilde{N}}(μ_1,σ_1^2,h_1)$ and ${\widetilde{X}}_2∼{\widetilde{N}}(μ_2,σ_2^2,h_2)$ . The orthogonal sum of ${\widetilde{X}}_1$ and ${\widetilde{X}}_2$ is the random fuzzy set $(Ω_1×Ω_2,σ_1⊗σ_2,{\widetilde{P}}_12, ℝ,β_ℝ,{\widetilde{X}}_\varodot)$ , where ${\widetilde{X}}_\varodot$ is the mapping
$$
{\widetilde{X}}_\varodot:(ω_1,ω_2)→\textsf{GFN}(M_1
2(ω_1,ω_2),h_1+h_2),
$$
with
$$
M_12(ω_1,ω_2)=\frac{h_1M_1(ω_1)+h_2M_2(ω_
2)}{h_1+h_2},
$$
and ${\widetilde{P}}_12$ is the probability measure on $Ω_1×Ω_2$ obtained by conditioning $P_1× P_2$ on the fuzzy set ${\widetilde{Θ}}^*(ω_1,ω_2)=0pt≤ft(\textsf{GFN}(M_1( ω_1),h_1),\textsf{GFN}(M_2(ω_2),h_2)\right)$ . From Lemma 1, the pushforward measure of ${\widetilde{P}}_12$ by the random vector $(M_1,M_2)$ is the two-dimensional Gaussian distribution with parameters $({\widetilde{μ}}_1,{\widetilde{μ}}_2,{\widetilde{σ}}_1,{ \widetilde{σ}}_2,ρ)$ . Consequently, $M_12$ is a Gaussian random variable with mean
$$
E(M_12)=\frac{h_1E(M_1)+h_2E(M_2)}{h_1+
h_2}=\frac{h_1{\widetilde{μ}}_1+h_2{\widetilde{μ}}_2}{h_1+h_2
},
$$
and variance
| | $\displaystyleVar(M_12)$ | $\displaystyle=\frac{h_1^2Var(M_1)+h_2^2Var(M_2)+2h_ 1h_2Cov(M_1,M_2)}{(h_1+h_2)^2}$ | |
| --- | --- | --- | --- |
∎
Let us now consider some special cases in which one of two GRFN’s is a Gaussian random variable. The next proposition states that the orthogonal sum of a Gaussian random variable and an arbitrary GRFN with finite precision is a Gaussian random variable.
**Proposition 10**
*Let $X_1∼ N(μ_1,σ_1^2)$ be a Gaussian random variable and ${\widetilde{X}}_2∼{\widetilde{N}}(μ_2,σ_2^2,h_2)$ a GRFN with finite precision $h_2<+∞$ . Their orthogonal sum is a Gaussian random variable $X_1⊕{\widetilde{X}}_2∼ N({\widetilde{μ}}_12,{\widetilde{σ }}_12^2)$ with
$$
{\widetilde{μ}}_12=\frac{μ_1(1+h_2σ_2^2)+μ_2h_2σ
_1^2}{1+h_2(σ_1^2+σ_2^2)}, \tag{32}
$$
$$
{\widetilde{σ}}_12^2=\frac{σ_1^2(1+h_2σ_2^2)}{1+h
_2(σ_1^2+σ_2^2)}, \tag{33}
$$
and the probability density of $X_1⊕{\widetilde{X}}_2$ is proportional to the product of the pdf of $X_1$ and the contour function of ${\widetilde{X}}_2$ .*
* Proof*
See G. ∎
The following corollary addresses the special case where ${\widetilde{X}}_2$ is a possibilistic GRFN.
**Corollary 2**
*Let $X_1∼ N(μ_1,σ_1^2)$ be a Gaussian random variable and ${\widetilde{X}}_2∼{\widetilde{N}}(μ_2,0,h_2)$ a possibilistic GRFN. Their orthogonal sum $X_1⊕{\widetilde{X}}_2$ is a Gaussian random variable and its distribution is the conditional distribution of $X_1$ given the fuzzy event $\textsf{GFN}(μ_2,h_2)$ .*
* Proof*
From Proposition 10, $X_1⊕{\widetilde{X}}_2∼{\widetilde{N}}({\widetilde{μ}}_12,{ \widetilde{σ}}_12^2)$ with
$$
{\widetilde{μ}}_12=\frac{μ_1+μ_2h_2σ_1^2}{1+h_2σ
_1^2} and {\widetilde{σ}}_12^2=\frac{σ_1^2
}{1+h_2σ_1^2}.
$$
Now, we know from Proposition 10 that the density of $X_1⊕{\widetilde{X}}_2$ is proportional to the product of the density of $X_1$ and the contour function of ${\widetilde{X}}_2$ , which is $φ(x;μ_2,h_2)$ . Consequently, we have
$$
f_X_{1⊕{\widetilde{X}}_2}(x)=\frac{\frac{1}{σ_1^2}\exp≤ft(
-\frac{1}{2}\frac{(x-μ_1)^2}{σ_1^2}\right)\exp≤ft(-\frac{h_2
(x-μ_2)^2}{2(1+h_2σ_2^2)}\right)}{∫\frac{1}{σ_1^2
}\exp≤ft(-\frac{1}{2}\frac{(x-μ_1)^2}{σ_1^2}\right)\exp≤ft
(-\frac{h_2(x-μ_2)^2}{2(1+h_2σ_2^2)}\right)dx},
$$
which is the conditional density $f_X_1(x|\textsf{GFN}(μ_2,h_2))$ . ∎
Finally, another special case of interest is when both GRFN’s are Gaussian random variables. This case is addressed by the following corollary.
**Corollary 3**
*Let $X_1∼ N(μ_1,σ_1^2)$ and $X_2∼ N(μ_2,σ_2^2)$ be two Gaussian random variables. We have $X_1⊕ X_2∼ N({\widetilde{μ}}_12,σ_12^2)$ with
$$
{\widetilde{μ}}_12=\frac{μ_1σ_2^2+μ_2σ_1^2}{
σ_1^2+σ_2^2} and {\widetilde{σ}}_12^2
=\frac{σ_1^2σ_2^2}{σ_1^2+σ_2^2}.
$$*
* Proof*
Immediate from Proposition 10 by letting $h_2$ tend to $+∞$ in (32) and (33). ∎
### 4.3 Arithmetic operations on GRFN’s
Arithmetic operations can be extended to fuzzy numbers using Zadeh’s extension principles [14, 12]. More precisely, let ${\widetilde{A}}$ and ${\widetilde{B}}$ be two fuzzy numbers, and let $*$ be a binary operation on reals. Then the fuzzy number ${\widetilde{C}}={\widetilde{A}}*{\widetilde{B}}$ is defined as
$$
{\widetilde{C}}(c)=\sup_c=a*b\min({\widetilde{A}}(a),{\widetilde{B}}(b)).
$$
The membership function ${\widetilde{C}}$ is equal to the possibility distribution on $c=a*b$ , if $a$ and $b$ are constrained, respectively, by possibility distributions ${\widetilde{A}}$ and ${\widetilde{B}}$ . Unary or $n$ -ary operations can be extended from real to fuzzy numbers in the same way. For a certain class of fuzzy number called LR-fuzzy numbers [14, page 54], closed-form expressions exist for the addition, subtraction and scalar multiplication of fuzzy numbers. In particular, Gaussian fuzzy numbers with positive precision are LR fuzzy numbers and they verify the following equalities [25]:
| | $\displaystyle\textsf{GFN}(m_1,h_1)+\textsf{GFN}(m_2,h_2)$ | $\displaystyle=\textsf{GFN}(m_1+m_2,(h_1^-1/2+h_2^-1/2)^-2)$ | |
| --- | --- | --- | --- |
As addition of fuzzy numbers is associative, we can express the linear combination of $n$ GFN’s as
$$
∑_i=1^nλ_i·\textsf{GFN}(m_i,h_i)=\textsf{GFN}≤ft(∑
_i=1^nλ_im_i,≤ft(∑_i=1^n|λ_i|h_i^-1/2\right
)^-2\right). \tag{34}
$$
Now, let us consider $n$ independent GRFN’s ${\widetilde{X}}_i$ from probability spaces $(Ω_i,σ_i,P_i)$ to $[0,1]^ℝ$ defined by
$$
{\widetilde{X}}_i(ω)=\textsf{GFN}(M_i(ω),h_i)
$$
for all $ω∈Ω_i$ , where $M_i$ is a Gaussian random variable with mean $μ_i$ and standard deviation $σ_i$ , and $h_i>0$ . Let
$$
{\widetilde{X}}=∑_i=1^nλ_i{\widetilde{X}}_i
$$
be the random fuzzy set from $(Ω_1×…×Ω_n,σ_1⊗…⊗σ_ n,P_1×…× P_n)$ to $[0,1]^ℝ$ defined by
$$
{\widetilde{X}}(ω_1,…,ω_n)=∑_i=1^nλ_i·
\textsf{GFN}(M_i(ω_i),h_i).
$$
If each GRFN ${\widetilde{X}}_i$ represents our knowledge about the value of some quantity $X_i$ , ${\widetilde{X}}$ represents our knowledge about $X=∑_i=1^nλ_iX_i$ . From (34), ${\widetilde{X}}∼{\widetilde{N}}(μ,σ,h)$ with
$$
μ=∑_i=1^nλ_iμ_i, σ^2=∑_i=1^nλ_i
^2σ_i^2, and h=≤ft(∑_i=1^n|λ_i|h_
i^-1/2\right)^-2.
$$
## 5 Gaussian random fuzzy vectors
In this section, we introduce Gaussian random fuzzy vectors (GRFV’s), an extension of the model presented in Section 4 allowing us to describe knowledge about multidimensional quantities. The main definitions and properties are first introduced in Section 5.1. The expression of the orthogonal sum of two GRFV’s is then given in Section 5.2, after which the marginalization and vacuous extension of GRFV’s are described in Section 5.3. Finally, our model is compared to Dempster’s normal belief function model in Section 5.4.
### 5.1 Definition and main properties
We consider a $p$ -dimensional variable $\boldsymbol{θ}$ taking values in $ℝ^p$ . Knowledge about $\boldsymbol{θ}$ may be encoded as a $p$ -dimensional Gaussian fuzzy vector, defined as follows.
**Definition 2**
*We define the $p$ -dimensional Gaussian fuzzy vector (GFV) with center ${\boldsymbol{m}}∈ℝ^p$ and $p× p$ symmetric and positive semidefinite precision matrix ${\boldsymbol{H}}$ as the normalized fuzzy subset of $ℝ^p$ with membership function
$$
φ({\boldsymbol{x}};{\boldsymbol{m}},{\boldsymbol{H}})=\exp≤ft(-\frac{1
}{2}({\boldsymbol{x}}-{\boldsymbol{m}})^T{\boldsymbol{H}}({\boldsymbol{x}}-{
\boldsymbol{m}})\right),
$$
denoted as $\textsf{GFV}({\boldsymbol{m}},{\boldsymbol{H}})$ .*
As shown in [27], the normalized product of two GFV’s is still a GFV. The following proposition generalizes Proposition 3.
**Proposition 11**
*Let $\textsf{GFV}({\boldsymbol{m}}_1,{\boldsymbol{H}}_1)$ and $\textsf{GFV}({\boldsymbol{m}}_2,{\boldsymbol{H}}_2)$ be two $p$ -dimensional GFV’s with positive definite precision matrices ${\boldsymbol{H}}_1$ and ${\boldsymbol{H}}_2$ . We have
$$
φ({\boldsymbol{x}};{\boldsymbol{m}}_1,{\boldsymbol{H}}_1)·
φ({\boldsymbol{x}};{\boldsymbol{m}}_2,{\boldsymbol{H}}_2)=φ({
\boldsymbol{x}};{\boldsymbol{m}}_12,{\boldsymbol{H}}_12)×\\
\exp≤ft(-\frac{1}{2}({\boldsymbol{m}}_1-{\boldsymbol{m}}_2)^T({
\boldsymbol{H}}_1^-1+{\boldsymbol{H}}_2^-1)^-1({\boldsymbol{m}}_1-
{\boldsymbol{m}}_2)\right),
$$
with
$$
{\boldsymbol{m}}_12=({\boldsymbol{H}}_1+{\boldsymbol{H}}_2)^-1({
\boldsymbol{H}}_1{\boldsymbol{m}}_1+{\boldsymbol{H}}_2{\boldsymbol{m}}_
2) and {\boldsymbol{H}}_12={\boldsymbol{H}}_1+{\boldsymbol
{H}}_2.
$$
Consequently, the following equation holds:
$$
\textsf{GFV}({\boldsymbol{m}}_1,{\boldsymbol{H}}_1)\varodot\textsf{GFV}({
\boldsymbol{m}}_2,{\boldsymbol{H}}_2)=\textsf{GFV}({\boldsymbol{m}}_12,{
\boldsymbol{H}}_12),
$$
and the height of the product intersection between $\textsf{GFV}({\boldsymbol{m}}_1,{\boldsymbol{H}}_1)$ and $\textsf{GFV}({\boldsymbol{m}}_1,{\boldsymbol{H}}_2)$ is
$$
\displaystyle 0pt≤ft(\textsf{GFV}({\boldsymbol{m}}_1,{\boldsymbol{H}}_1)
,\textsf{GFV}({\boldsymbol{m}}_1,{\boldsymbol{H}}_2)\right) \displaystyle=\max_\boldsymbol{x}φ({\boldsymbol{x}};{\boldsymbol{m}}_
1,{\boldsymbol{H}}_1)φ({\boldsymbol{x}};{\boldsymbol{m}}_2,{
\boldsymbol{H}}_2) \displaystyle=\exp≤ft(-\frac{1}{2}({\boldsymbol{m}}_1-{\boldsymbol{m}}_2
)^T({\boldsymbol{H}}_1^-1+{\boldsymbol{H}}_2^-1)^-1({\boldsymbol{m
}}_1-{\boldsymbol{m}}_2)\right).
$$*
Equipped with the notion of GFV, we can now introduce a model of random fuzzy set that can be seen as a GFV whose mode is a multidimensional Gaussian random variable. This model is defined formally as follows.
**Definition 3**
*Let $(Ω,σ_Ω,P)$ be a probability space, ${\boldsymbol{M}}:Ω→ℝ^p$ a $p$ -dimensional Gaussian random vector with mean $\boldsymbol{μ}$ and variance matrix $\boldsymbol{Σ}$ , and ${\boldsymbol{H}}$ a $p× p$ symmetric and positive semidefinite real matrix. The random fuzzy set ${\widetilde{X}}:Ω→[0,1]^ℝ^{p}$ defined as
$$
{\widetilde{X}}(ω)=\textsf{GFV}({\boldsymbol{M}}(ω),{\boldsymbol{H}})
$$
is called a Gaussian random fuzzy vector (GRFV), which we denote as ${\widetilde{X}}∼{\widetilde{N}}(\boldsymbol{μ},\boldsymbol{Σ},{ \boldsymbol{H}})$ .*
The following proposition generalizes Proposition 6.
**Proposition 12**
*The contour function of GRFV ${\widetilde{X}}∼{\widetilde{N}}(\boldsymbol{μ},\boldsymbol{Σ},{ \boldsymbol{H}})$ with positive definite precision matrix ${\boldsymbol{H}}$ is
$$
pl_\widetilde{X}({\boldsymbol{x}})=\frac{1}{|{\boldsymbol{I}}_p+
\boldsymbol{Σ}{\boldsymbol{H}}|^1/2}\exp≤ft(-\frac{1}{2}({\boldsymbol
{x}}-\boldsymbol{μ})^T({\boldsymbol{H}}^-1+\boldsymbol{Σ})^-1({
\boldsymbol{x}}-\boldsymbol{μ})\right),
$$
where ${\boldsymbol{I}}_p$ is the $p$ -dimensional identity matrix.*
* Proof*
See H. ∎
### 5.2 Orthogonal sum of Gaussian random fuzzy vectors
The practical interest of GRFV’s arises from the fact that they can be easily combined by the generalized product-intersection rule. The following lemma and proposition, which generalize, respectively, Lemma 1 and Proposition 9, give us the expression of the orthogonal sum of two GRFV’s.
**Lemma 2**
*Let ${\boldsymbol{M}}_1∼{\cal N}(\boldsymbol{μ}_1,\boldsymbol{Σ}_1)$ and ${\boldsymbol{M}}_2∼{\cal N}(\boldsymbol{μ}_2,\boldsymbol{Σ}_2)$ be two independent Gaussian $p$ -dimensional random vectors and let ${\boldsymbol{H}}_1$ and ${\boldsymbol{H}}_2$ be two symmetric and positive definite $p× p$ matrices. Let ${\widetilde{F}}$ be the fuzzy subset of $ℝ^2p$ with membership function
$$
{\widetilde{F}}({\boldsymbol{m}}_1,{\boldsymbol{m}}_2)=0pt≤ft(\textsf{
GFV}({\boldsymbol{m}}_1,{\boldsymbol{H}}_1)·\textsf{GFV}({\boldsymbol{
m}}_2,{\boldsymbol{H}}_2)\right),
$$
and let ${\boldsymbol{M}}$ be the $2p$ -dimensional vector $({\boldsymbol{M}}_1,{\boldsymbol{M}}_2)$ . The conditional probability distribution of ${\boldsymbol{M}}$ given ${\widetilde{F}}$ is $2p$ -dimensional Gaussian with mean vector ${\widetilde{\boldsymbol{μ}}}$ and covariance matrix ${\widetilde{\boldsymbol{Σ}}}$ defined as follows:
$$
{\widetilde{\boldsymbol{Σ}}}=\begin{pmatrix}\boldsymbol{Σ}_1^-1+
{\overline{{\boldsymbol{H}}}}&-{\overline{{\boldsymbol{H}}}}\\
-{\overline{{\boldsymbol{H}}}}&\boldsymbol{Σ}_2^-1+{\overline{{
\boldsymbol{H}}}}\end{pmatrix}^-1,
$$
$$
{\widetilde{\boldsymbol{μ}}}=\begin{pmatrix}{\overline{{\boldsymbol{H}}}}^-
1\boldsymbol{Σ}_1^-1+{\boldsymbol{I}}_p&-{\boldsymbol{I}}_p\\
-{\boldsymbol{I}}_p&{\overline{{\boldsymbol{H}}}}^-1\boldsymbol{Σ}_2
^-1+{\boldsymbol{I}}_p\end{pmatrix}^-1\begin{pmatrix}{\overline{{
\boldsymbol{H}}}}^-1\boldsymbol{Σ}_1^-1&\boldsymbol{0}\\
\boldsymbol{0}&{\overline{{\boldsymbol{H}}}}^-1\boldsymbol{Σ}_2^-1
\end{pmatrix}\begin{pmatrix}\boldsymbol{μ}_1\\
\boldsymbol{μ}_2\end{pmatrix}, \tag{36}
$$
with
$$
{\overline{{\boldsymbol{H}}}}=({\boldsymbol{H}}_1^-1+{\boldsymbol{H}}_2^
{-1})^-1.
$$
Furthermore, the degree of conflict between two GRFV’s ${\widetilde{X}}_1∼{\widetilde{N}}(\boldsymbol{μ}_1,\boldsymbol{Σ }_1,{\boldsymbol{H}}_1)$ and ${\widetilde{X}}_2∼{\widetilde{N}}(\boldsymbol{μ}_2,\boldsymbol{Σ }_2,{\boldsymbol{H}}_2)$ is
$$
κ=1-∫_ℝ^2pf({\boldsymbol{m}}_1,{\boldsymbol{m}}_2){
\widetilde{F}}({\boldsymbol{m}}_1,{\boldsymbol{m}}_2)d{\boldsymbol{m}}_1
d{\boldsymbol{m}}_2=\\
1-√{\frac{|{\widetilde{\boldsymbol{Σ}}}|}{|\boldsymbol{Σ}_1||
\boldsymbol{Σ}_2|}}\exp≤ft\{-\frac{1}{2}≤ft[\boldsymbol{μ}_1^T
\boldsymbol{Σ}_1^-1\boldsymbol{μ}_1+\boldsymbol{μ}_2^T
\boldsymbol{Σ}_2^-1\boldsymbol{μ}_2-{\widetilde{\boldsymbol{μ}}
}^T{\widetilde{\boldsymbol{Σ}}}^-1{\widetilde{\boldsymbol{μ}}}
\right]\right\}.
$$*
* Proof*
See I ∎
**Proposition 13**
*Let ${\widetilde{X}}_1∼{\widetilde{N}}(\boldsymbol{μ}_1,\boldsymbol{Σ }_1,{\boldsymbol{H}}_1)$ and ${\widetilde{X}}_2∼{\widetilde{N}}(\boldsymbol{μ}_2,\boldsymbol{Σ }_2,{\boldsymbol{H}}_2)$ be two independent GRFV’s. We have
$$
{\widetilde{X}}_1⊕{\widetilde{X}}_2∼{\widetilde{N}}({\widetilde{
\boldsymbol{μ}}}_12,{\widetilde{\boldsymbol{Σ}}}_12,{\boldsymbol{H}
}_12)
$$
with
$$
{\boldsymbol{H}}_12={\boldsymbol{H}}_1+{\boldsymbol{H}}_2,
$$
$$
{\widetilde{\boldsymbol{μ}}}_12={\boldsymbol{A}}{\widetilde{\boldsymbol{
μ}}},
$$
and
$$
{\widetilde{\boldsymbol{Σ}}}_12={\boldsymbol{A}}{\widetilde{\boldsymbol
{Σ}}}{\boldsymbol{A}}^T,
$$
where ${\boldsymbol{A}}$ is the constant $p× 2p$ matrix defined as
$$
{\boldsymbol{A}}={\boldsymbol{H}}_12^-1\begin{pmatrix}{\boldsymbol{H}}_1
&{\boldsymbol{H}}_2\end{pmatrix}.
$$*
* Proof*
Let ${\boldsymbol{M}}_1$ and ${\boldsymbol{M}}_2$ be the Gaussian random vector from $(Ω_1,σ_1,P_1)$ and $(Ω_2,σ_2,P_2)$ to $(ℝ^p,β_ℝ^p)$ corresponding, respectively, to GRFV’s ${\widetilde{X}}_1∼{\widetilde{N}}(\boldsymbol{μ}_1,\boldsymbol{Σ }_1,{\boldsymbol{H}}_1)$ and ${\widetilde{X}}_2∼{\widetilde{N}}(\boldsymbol{μ}_2,\boldsymbol{Σ }_2,{\boldsymbol{H}}_2)$ . The orthogonal sum of ${\widetilde{X}}_1$ and ${\widetilde{X}}_2$ is defined by the mapping
$$
{\widetilde{X}}_\varodot:(ω_1,ω_2)→\textsf{GFV}({
\boldsymbol{M}}_12(ω_1,ω_2),{\boldsymbol{H}}_1+{\boldsymbol{
H}}_2)
$$
with
$$
{\boldsymbol{M}}_12=({\boldsymbol{H}}_1+{\boldsymbol{H}}_2)^-1({
\boldsymbol{H}}_1{\boldsymbol{M}}_1+{\boldsymbol{H}}_2{\boldsymbol{M}}_
2)={\boldsymbol{A}}\begin{pmatrix}{\boldsymbol{M}}_1\\
{\boldsymbol{M}}_2\end{pmatrix},
$$
where ${\boldsymbol{A}}$ is the $p× 2p$ matrix
$$
{\boldsymbol{A}}=({\boldsymbol{H}}_1+{\boldsymbol{H}}_2)^-1\begin{
pmatrix}{\boldsymbol{H}}_1&{\boldsymbol{H}}_2\end{pmatrix},
$$
and the probability measure ${\widetilde{P}}_12$ on $Ω_1×Ω_2$ obtained by conditioning $P_1× P_2$ on the fuzzy set ${\widetilde{Θ}}^*(ω_1,ω_2)=0pt≤ft(\textsf{GFV}({ \boldsymbol{M}}_1(ω_1),{\boldsymbol{H}}_1),\textsf{GFV}({ \boldsymbol{M}}_2(ω_2),{\boldsymbol{H}}_2)\right)$ . From Lemma 2, the pushforward measure of ${\widetilde{P}}_12$ by the random vector $({\boldsymbol{M}}_1,{\boldsymbol{M}}_2)$ is the $p$ -dimensional Gaussian distribution with parameters $({\widetilde{\boldsymbol{μ}}},{\widetilde{\boldsymbol{Σ}}})$ . Consequently, ${\boldsymbol{M}}_12$ is a Gaussian random vector with mean
$$
E({\boldsymbol{M}}_12)={\boldsymbol{A}}{\widetilde{\boldsymbol{μ}}}
$$
and variance
$$
Var({\boldsymbol{M}}_12)={\boldsymbol{A}}{\widetilde{\boldsymbol{
Σ}}}{\boldsymbol{A}}^T.
$$ ∎
### 5.3 Marginalization and vacuous extension
In this section, we consider the marginalization and vacuous extension (defined in Section 3.3) of a GRFV. We assume that variable $\boldsymbol{θ}$ taking values in $ℝ^p$ is decomposed as $\boldsymbol{θ}=(\boldsymbol{θ}_1,\boldsymbol{θ}_2)$ with $\boldsymbol{θ}_1∈Θ_1=ℝ^p-k$ and $\boldsymbol{θ}_2∈Θ_2=ℝ^k$ for $0<k<p$ .
#### Marginalization
We start with the following lemma.
**Lemma 3**
*Let ${\widetilde{F}}=\textsf{GFV}({\boldsymbol{m}},{\boldsymbol{H}})$ be a $p$ -dimensional Gaussian fuzzy vector with mode ${\boldsymbol{m}}=({\boldsymbol{m}}_1,{\boldsymbol{m}}_2)$ , where ${\boldsymbol{m}}_1∈Θ_1=ℝ^p-k$ and ${\boldsymbol{m}}_2∈Θ_2=ℝ^k$ for $0<k<p$ , and precision matrix ${\boldsymbol{H}}$ with block decomposition
$$
{\boldsymbol{H}}=\begin{pmatrix}{\boldsymbol{H}}_11&{\boldsymbol{H}}_12\\
{\boldsymbol{H}}_21&{\boldsymbol{H}}_22\end{pmatrix}.
$$
Assume that ${\boldsymbol{H}}_22$ is nonsingular. The projection of ${\widetilde{F}}$ on $Θ_1$ , denoted as ${\widetilde{F}}↓Θ_1$ is the Gaussian fuzzy vector $\textsf{GFV}({\boldsymbol{m}}_1,{\boldsymbol{H}}^\prime_11)$ with
$$
{\boldsymbol{H}}^\prime_11={\boldsymbol{H}}_11-{\boldsymbol{H}}_12{
\boldsymbol{H}}_22^-1{\boldsymbol{H}}_21.
$$*
* Proof*
See J ∎
Let us now consider a $p$ -dimensional GRFV ${\widetilde{X}}∼{\widetilde{N}}(\boldsymbol{μ},\boldsymbol{Σ},{ \boldsymbol{H}})$ representing partial knowledge about $\boldsymbol{θ}=(\boldsymbol{θ}_1,\boldsymbol{θ}_2)$ . The marginal RFS for $\boldsymbol{θ}_1$ is given by the following proposition, which follows directly from Lemma 3.
**Proposition 14**
*Let ${\widetilde{X}}∼{\widetilde{N}}(\boldsymbol{μ},\boldsymbol{Σ},{ \boldsymbol{H}})$ by a $p$ -dimensional GRFV taking values in $2^Θ$ , with $Θ=Θ_1×Θ_2$ , where $Θ_1=ℝ^p-k$ and $Θ_2=ℝ^k$ for $0<k<p$ . Let $\boldsymbol{μ}=(\boldsymbol{μ}_1,\boldsymbol{μ}_2)$ with $\boldsymbol{μ}_1∈Θ_1$ and $\boldsymbol{μ}_2∈Θ_2$ , and consider the block decompositions
$$
\boldsymbol{Σ}=\begin{pmatrix}\boldsymbol{Σ}_11&\boldsymbol{Σ
}_12\\
\boldsymbol{Σ}_21&\boldsymbol{Σ}_22\end{pmatrix} and
{\boldsymbol{H}}=\begin{pmatrix}{\boldsymbol{H}}_11&{\boldsymbol{H}}_1
2\\
{\boldsymbol{H}}_21&{\boldsymbol{H}}_22\end{pmatrix}.
$$
Assume that ${\boldsymbol{H}}_22$ is nonsingular. The marginal of ${\widetilde{X}}$ on $Θ_1$ is the GRFV ${\widetilde{X}}_1∼{\widetilde{N}}(\boldsymbol{μ}_1,\boldsymbol{Σ }_11,{\boldsymbol{H}}^\prime_11)$ with
$$
{\boldsymbol{H}}^\prime_11={\boldsymbol{H}}_11-{\boldsymbol{H}}_12{
\boldsymbol{H}}_22^-1{\boldsymbol{H}}_21.
$$*
#### Vacuous extension
We now consider a Gaussian fuzzy vector $\textsf{GFV}({\boldsymbol{m}}_1,{\boldsymbol{H}}_11)$ in $Θ_1=ℝ^p-k$ for $0<k<p$ . Its cylindrical extension in $Θ=Θ_1×Θ_2$ , with $Θ_2=ℝ^k$ has the following membership function
$$
φ({\boldsymbol{x}})=\exp≤ft(-\frac{1}{2}({\boldsymbol{x}}_1-{
\boldsymbol{m}}_1)^T{\boldsymbol{H}}_11({\boldsymbol{x}}_1-{
\boldsymbol{m}}_1)\right),
$$
which can be written as
$$
φ({\boldsymbol{x}})=\exp≤ft(-\frac{1}{2}({\boldsymbol{x}}-{\boldsymbol
{m}})^T{\boldsymbol{H}}({\boldsymbol{x}}-{\boldsymbol{m}})\right),
$$
where ${\boldsymbol{m}}$ is the $p$ -dimensional vector
$$
{\boldsymbol{m}}=\begin{pmatrix}{\boldsymbol{m}}_1\\
\boldsymbol{0}\end{pmatrix}
$$
and ${\boldsymbol{H}}$ is the $p× p$ matrix
$$
{\boldsymbol{H}}=\begin{pmatrix}{\boldsymbol{H}}_11&\boldsymbol{0}\\
\boldsymbol{0}&\boldsymbol{0}\end{pmatrix}. \tag{37}
$$
Given a GRFV ${\widetilde{X}}_1∼{\widetilde{N}}(\boldsymbol{μ}_1,\boldsymbol{Σ }_11,{\boldsymbol{H}}_11)$ taking values in $2^Θ_1$ , it follows immediately that its vacuous extension in $Θ=Θ_1×Θ_2$ is the GRFV
$$
{\widetilde{X}}_1↑(1,2)∼{\widetilde{N}}(\boldsymbol{μ},
\boldsymbol{Σ},{\boldsymbol{H}})
$$
with
$$
\boldsymbol{μ}=\begin{pmatrix}\boldsymbol{μ}_1\\
\boldsymbol{0}\end{pmatrix}, \boldsymbol{Σ}=\begin{pmatrix}
\boldsymbol{Σ}_11&\boldsymbol{0}\\
\boldsymbol{0}&{\boldsymbol{I}}_k\end{pmatrix},
$$
where ${\boldsymbol{I}}_k$ is the $k× k$ identity matrix, and ${\boldsymbol{H}}$ given by (37).
#### Noninteractivity
In Section 3.3, we defined the notion of noninteractive random fuzzy vector. The following proposition gives a necessary and sufficient condition for a GRFV to be noninteractive.
**Proposition 15**
*A $p$ -dimensional GRFV ${\widetilde{X}}∼{\widetilde{N}}(\boldsymbol{μ},\boldsymbol{Σ},{ \boldsymbol{H}})$ is noninteractive iff matrices $\boldsymbol{Σ}$ and ${\boldsymbol{H}}$ are diagonal.*
* Proof*
Let ${\widetilde{X}}_1,…,{\widetilde{X}}_p$ be the marginals of ${\widetilde{X}}$ on each of the $p$ coordinates. Let $σ^2_1,…,σ^2_p$ and $h_1,…,h_p$ be the diagonal elements of, respectively, $\boldsymbol{Σ}$ and ${\boldsymbol{H}}$ . Let $Ω$ be the set of departure of ${\widetilde{X}}$ . Let ${\widetilde{X}}_i↑(1:p)$ denote the vacuous extension of ${\widetilde{X}}_i$ in $ℝ^p$ , defined by
$$
{\widetilde{X}}_i↑(1:p)(ω)({\boldsymbol{x}})=\exp≤ft(-\frac{h}
{2}(x_i-M_i(ω))^2\right)
$$
with $M_i∼ N(μ_i,σ^2_i)$ . The orthogonal sum
$$
{\widetilde{X}}^\prime={\widetilde{X}}_1↑(1:p)⊕…⊕{
\widetilde{X}}_p↑(1:p)
$$
is given by
$$
{\widetilde{X}}^\prime(ω)({\boldsymbol{x}})=∏_i=1^p\exp≤ft(-
\frac{h}{2}(x_i-M_i(ω))^2\right)=\exp≤ft(-\frac{1}{2}({
\boldsymbol{x}}-{\boldsymbol{M}}^\prime(ω))^T{\boldsymbol{H}}^\prime
({\boldsymbol{x}}-{\boldsymbol{M}}^\prime(ω))\right),
$$
where ${\boldsymbol{H}}^\prime$ is the diagonal matrix with diagonal elements $h_1,…,h_p$ , and ${\boldsymbol{M}}^\prime$ is a random vector with mean $\boldsymbol{μ}$ and diagonal covariance matrix $\boldsymbol{Σ}^\prime$ with diagonal elements $σ^2_1,…,σ^2_p$ . We have ${\widetilde{X}}={\widetilde{X}}^\prime$ iff ${\boldsymbol{H}}={\boldsymbol{H}}^\prime$ and $\boldsymbol{Σ}=\boldsymbol{Σ}^\prime$ , i.e., if both ${\boldsymbol{H}}$ and $\boldsymbol{Σ}$ are diagonal. ∎
### 5.4 Comparison with Dempster’s normal belief functions
In [5], Dempster introduced another class of continuous belief functions in $ℝ^p$ , called normal belief functions Ref. [5] was actually available as a working paper from the Statistical Department of Harvard University since 1990, but it only appeared as a book chapter in 2001.. It is interesting to compare Dempster’s model with ours, as both models generalize the multivariate Gaussian distribution. A normal belief function $Bel$ on $ℝ^p$ as defined in [5] is specified by the following components:
- An $n$ -dimensional subspace ${\cal S}$ of $ℝ^p$ ;
- A $q$ -dimensional partition $Π$ of ${\cal S}$ into parallel $n-q$ dimensional subspaces; (If $q=0$ , $Π=\{{\cal S}\}$ );
- A full-rank $q$ -dimensional Gaussian distribution $N(μ,Σ)$ on $Π$ if $q>0$ , or the discrete probability measure with mass function $m({\cal S})=1$ if $q=0$ .
Belief function $Bel$ is then induced by a random set from $Π$ , equipped with the normal distribution $N(μ,Σ)$ if $q>0$ or probability mass function $m$ if $q=0$ , to the corresponding family of parallel $n-q$ dimensional subspaces of ${\cal S}$ . The following special cases are of interest:
1. If $p=n=q$ , $Bel$ is a Gaussian probability distribution on $ℝ^p$ ;
1. If $p>n=q$ , $Bel$ is a Gaussian probability distribution limited to an $n$ -dimensional subspace of $ℝ^p$ ;
1. If $p=n$ and $q=0$ , $Bel$ is vacuous;
1. If $q=0$ while $p>n>0$ , $Bel$ is logical with ${\cal S}$ as its only focal set; it is then equivalent to specifying $p-n$ linear equations;
1. If $n=q=0$ , the true point in $ℝ^p$ is known with certainty.
Like GRFV’s, Dempster’s normal belief functions thus include the vacuous belief function, Gaussian probability distributions, as well as vacuous extensions of marginal Gaussian distributions. However, the two models are clearly distinct. Dempster’s model is based on the combination of Gaussian probability distributions and linear equations, and is specially useful in relation with linear statistical models such as the Kalman filter [5] or linear regression [24]. In contrast, in the GRFV model, focal sets are fuzzy subsets of $ℝ^n$ ( $n≤ p$ ) with Gaussian membership functions, or cylindrical extensions of such fuzzy subsets. This model allows us to represent not only probabilistic and logical evidence, but also fuzzy information. In particular, it includes Gaussian probability distribution and Gaussian possibility distributions as special cases. We could attempt to design an even more general model that would contain both Dempster’s normal belief functions and belief functions induced by GRFV’s as special cases. Such a model would allow us to reason with Gaussian probability and possibility distributions as well as with linear equations. The rigorous development of such a model is left for further research.
## 6 Conclusions
In this paper, continuing a study started in [9] with the finite case, we have introduced a theory of epistemic random fuzzy sets in a general setting. An epistemic random fuzzy set represents a piece of evidence, which may be crisp or fuzzy. This framework generalizes both epistemic random sets as considered in the Dempster-Shafer theory of belief functions, and possibility distributions considered in possibility theory. Independent epistemic random fuzzy sets are combined by the generalized product-intersection rule, which extends both Dempster’s rule for combining belief functions and the product intersection rule for combining possibility distributions.
In addition, we have also introduced Gaussian random fuzzy numbers (GRFN’s) and their multidimensional extensions, Gaussian random fuzzy vectors (GRFV’s) as practical models of random fuzzy subsets of, respectively, $ℝ$ and $ℝ^p$ with $p≥ 2$ . A GRFN is described by three parameters: its mode $m$ , its variance $σ^2$ and its precision $h$ . In this setting, a Gaussian random variable can be seen as an infinitely precise GRFN ( $h=+∞$ ), while a Gaussian possibility distribution is a constant GRFN ( $σ^2=0$ ). A maximally imprecise GRFN such that $h=0$ is said to be vacuous: it represents complete ignorance. In GRFV’s, the mode becomes a $p$ -dimensional vector, while the variance and precision become positive semi-definite $p× p$ square matrices. The practical convenience of GRFN’s and GRFV’s arises from the fact that they can easily be combined by the generalized product-intersection rule. Also, formulas for the projection and marginal extension fo GRFV’s have been derived.
This work opens up several perspectives. Using random fuzzy sets and, in particular, GRFN’s to represent expert knowledge about numerical quantities will require the development of adequate elicitation procedures. We also consider using this framework in machine learning, to quantify prediction uncertainty in regression problems. Finally, the extension of the model introduced in this paper to take into account linear equations, as well as the development of computational procedures for reasoning with GRFV’s over many variables are promising avenues for further research.
## References
## References
- [1] P. A. Bromiley. Products and convolutions of Gaussian probability density functions. Technical Report 2003-003, TINA, 2014. https://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.583.3007&rep=rep1&type=pdf.
- [2] I. Couso and L. Sánchez. Upper and lower probabilities induced by a fuzzy random variable. Fuzzy Sets and Systems, 165(1):1–23, 2011.
- [3] A. P. Dempster. Upper and lower probabilities induced by a multivalued mapping. Annals of Mathematical Statistics, 38:325–339, 1967.
- [4] A. P. Dempster. Upper and lower probabilities generated by a random closed interval. Annals of Mathematical Statistics, 39(3):957–966, 1968.
- [5] A. P. Dempster. Normal belief functions and the Kalman filter. In A. K. M. E. Saleh, editor, Data Analysis from Statistical Foundations, pages 65–68. Nova Science Publishers, 2001.
- [6] T. Denœux. Extending stochastic ordering to belief functions on the real line. Information Sciences, 179:1362–1376, 2009.
- [7] T. Denœux. Likelihood-based belief function: justification and some extensions to low-quality data. International Journal of Approximate Reasoning, 55(7):1535–1547, 2014.
- [8] T. Denœux. Rejoinder on “likelihood-based belief function: Justification and some extensions to low-quality data”. International Journal of Approximate Reasoning, 55(7):1614–1617, 2014.
- [9] T. Denœux. Belief functions induced by random fuzzy sets: A general framework for representing uncertain and fuzzy evidence. Fuzzy Sets and Systems, 424:63–91, 2021.
- [10] T. Denœux, D. Dubois, and H. Prade. Representations of uncertainty in artificial intelligence: Beyond probability and possibility. In P. Marquis, O. Papini, and H. Prade, editors, A Guided Tour of Artificial Intelligence Research, volume 1, chapter 4, pages 119–150. Springer Verlag, 2020.
- [11] T. Denœux, D. Dubois, and H. Prade. Representations of uncertainty in artificial intelligence: Probability and possibility. In P. Marquis, O. Papini, and H. Prade, editors, A Guided Tour of Artificial Intelligence Research, volume 1, chapter 3, pages 69–117. Springer Verlag, 2020.
- [12] D. Dubois, E. Kerre, R. Mesiar, and H. Prade. Fuzzy interval analysis. In D. Dubois and H. Prade, editors, Fundamentals of Fuzzy sets, pages 483–581. Kluwer Academic Publishers, Boston, 2000.
- [13] D. Dubois, H. T. Nguyen, and H. Prade. Possibility theory, probability and fuzzy sets: Misunderstandings, bridges and gaps. In D. Dubois and H. Prade, editors, Fundamentals of Fuzzy sets, pages 343–438. Kluwer Academic Publishers, Boston, 2000.
- [14] D. Dubois and H. Prade. Fuzzy Sets and Systems: Theory and Applications. Academic Press, New York, 1980.
- [15] D. Dubois and H. Prade. Possibility theory: qualitative and quantitative aspects. In D. M. Gabbay and P. Smets, editors, Handbook of Defeasible reasoning and uncertainty management systems, volume 1, pages 169–226. Kluwer Academic Publishers, Dordrecht, 1998.
- [16] D. Dubois, H. Prade, and R. Yager. Merging fuzzy information. In J. C. Bezdek, D. Dubois, and H. Prade, editors, Fuzzy sets in approximate reasoning and information systems, pages 335–401. Kluwer Academic Publishers, Boston, 1999.
- [17] M. A. Gil, M. López-Díaz, and D. A. Ralescu. Overview on the development of fuzzy random variables. Fuzzy Sets and Systems, 157(19):2546–2557, 2006.
- [18] P. R. Halmos. Measure theory. Springer Science, New York, 1950.
- [19] O. Kanjanatarakul, S. Sriboonchitta, and T. Denœux. Forecasting using belief functions: an application to marketing econometrics. International Journal of Approximate Reasoning, 55(5):1113–1128, 2014.
- [20] O. Kanjanatarakul, S. Sriboonchitta, and T. Denœux. Prediction of future observations using belief functions: A likelihood-based approach. International Journal of Approximate Reasoning, 72:71–94, 2016.
- [21] G. J. Klir and B. Yuan. Fuzzy sets and fuzzy logic. Theory and applications. Prentice-Hall, Upper Saddle River, NJ., 1995.
- [22] E. Miranda, I. Couso, and P. Gil. Random intervals as a model for imprecise information. Fuzzy Sets and Systems, 154(3):386–412, 2005.
- [23] I. Molchanov. Theory of Random Sets. Springer, New York, 2005.
- [24] P.-A. Monney. A Mathematical Theory of Arguments for Statistical Evidence. Contributions to Statistics. Physica-Verlag, Heidelberg, 2003.
- [25] S. Nahmias. Fuzzy variables. Fuzzy Sets and Systems, 1(2):97–110, 1978.
- [26] H. T. Nguyen. On random sets and belief functions. Journal of Mathematical Analysis and Applications, 65:531–542, 1978.
- [27] K. B. Petersen and M. S. Pedersen. The matrix cookbook, nov 2012. http://www2.compute.dtu.dk/pubdb/pubs/3274-full.html.
- [28] M. L. Puri and D. A. Ralescu. Fuzzy random variables. Journal of Mathematical Analysis and Applications, 114(2):409–422, 1986.
- [29] G. Shafer. A mathematical theory of evidence. Princeton University Press, Princeton, N.J., 1976.
- [30] G. Shafer. Constructive probability. Synthese, 48(1):1–60, 1981.
- [31] G. Shafer. Belief functions and parametric models (with discussion). J. Roy. Statist. Soc. Ser. B, 44:322–352, 1982.
- [32] G. Shafer. Dempster’s rule of combination. International Journal of Approximate Reasoning, 79:26–40, 2016.
- [33] P. P. Shenoy. Using possibility theory in expert systems. Fuzzy Sets and Systems, 52(2):129–142, 1992.
- [34] P. Smets. Belief functions on real numbers. International Journal of Approximate Reasoning, 40(3):181–223, 2005.
- [35] R. R. Yager. On the normalization of fuzzy belief structure. International Journal of Approximate Reasoning, 14:127–153, 1996.
- [36] L. A. Zadeh. Fuzzy sets. Inform. Control, 8:338–353, 1965.
- [37] L. A. Zadeh. Probability measures of fuzzy events. J. Math. Analysis and Appl., 10:421–427, 1968.
- [38] L. A. Zadeh. Fuzzy sets as a basis for a theory of possibility. Fuzzy Sets and Systems, 1:3–28, 1978.
## Appendix A Proof of Proposition 1
Commutativity is obvious. To prove associativity, let us consider three random sets $(Ω_i,σ_i,P_i,Θ,σ_Θ,{\overline{X}}_i)$ , $i=1,2,3$ . Consider the combined random set
$$
(Ω_1×Ω_2×Ω_3,σ_1⊗σ_2⊗
σ_3,P_123,Θ,σ_Θ,{\overline{X}}_1∩ 2∩ 3), \tag{38}
$$
where
$$
{\overline{X}}_1∩ 2∩ 3(ω_1,ω_2,ω_3)={\overline{X}
}_1(ω_1)∩{\overline{X}}_2(ω_2)∩{\overline{X}}_3(
ω_3),
$$
$$
P_123=(P_1× P_2× P_3)(·\midΘ^*_123),
$$
and
$$
Θ^*_123=\{(ω_1,ω_2,ω_3)∈Ω_1×Ω
_2×Ω_2:{\overline{X}}_1∩ 2∩ 3(ω_1,ω_2,
ω_3)≠∅\}.
$$
We will show that we get the same result by combining ${\overline{X}}_1$ with ${\overline{X}}_2$ first, and then combining the result with ${\overline{X}}_3$ . Combining the first two random sets, we get
$$
(Ω_1×Ω_2,σ_1⊗σ_2,P_12,Θ,σ_
Θ,{\overline{X}}_1∩ 2),
$$
with ${\overline{X}}_1∩ 2(ω_1,ω_2)={\overline{X}}_1(ω_1) ∩{\overline{X}}_2(ω_2)$ , $P_12=(P_1× P_2)(·\midΘ^*_12)$ and
$$
Θ^*_12=\{(ω_1,ω_2)∈Ω_1×Ω_2:{
\overline{X}}_1∩ 2(ω_1,ω_2)≠∅\}.
$$
Combining it with ${\overline{X}}_3$ we get
$$
(Ω_1×Ω_2×Ω_3,σ_1⊗σ_2⊗
σ_3,P_(12)3,Θ,σ_Θ,{\overline{X}}_1∩ 2∩ 3), \tag{12}
$$
with $P_(12)3=(P_12× P_3)(·\midΘ^*_123)$ . Comparing (38) and (39), we see that we only need to show that $P_(12)3=P_123$ . For any event $C⊆Θ^*_123$ and any $ω_3∈Ω_3$ , let $C_ω_3=\{(ω_1,ω_2)∈Ω_1×Ω_2:(ω_ {1},ω_2,ω_3)∈ C\}$ . By definition of the product measure $P_12× P_3$ (see [18, page 144]), we have
$$
P_(12)3(C)=\frac{(P_12× P_3)(C)}{(P_12× P_3)(Θ^*_1
23)}=\frac{1}{(P_12× P_3)(Θ^*_123)}∫ P_12(C_ω_{3
})dP_3(ω_3) \tag{12}
$$
Now, as $C⊆Θ^*_123$ , for any $(ω_1,ω_2)∈ C_ω_3$ , ${\overline{X}}_1(ω_1)∩{\overline{X}}_2(ω_2)≠∅$ . Consequently, $C_ω_3⊆Θ^*_12$ , so
$$
P_12(C_ω_3)=\frac{(P_1× P_2)(C_ω_3)}{(P_1×
P
_2)(Θ^*_12)}. \tag{41}
$$
From (40) and (41), we get
$$
\displaystyle P_(12)3(C) \displaystyle=\frac{1}{(P_12× P_3)(Θ^*_123)(P_1× P_2
)(Θ^*_12)}∫(P_1× P_2)(C_ω_3)dP_3(ω_3) \displaystyle=\frac{(P_1× P_2× P_3)(C)}{(P_12× P_3)(
Θ^*_123)(P_1× P_2)(Θ^*_12)}. \tag{12}
$$
Now,
$$
P_123(C)=\frac{(P_1× P_2× P_3)(C)}{(P_1× P_2× P
_3)(Θ^*_123)}. \tag{43}
$$
As $P_(12)3(Θ^*_123)=P_123(Θ^*_123)=1$ , the denominators in (42b) and (43) are equal, and $P_(12)3=P_123$ .
## Appendix B Proof of Proposition 4
Commutativity is obvious. To prove associativity, consider three random fuzzy sets
$$
(Ω_i,σ_i,P_i,Θ,σ_Θ,{\widetilde{X}}_i),
i
=1,2,3.
$$
Let ${\widetilde{Θ}}^*_12$ be the fuzzy subset of $Ω_1×Ω_2$ with membership function
$$
{\widetilde{Θ}}_12^*(ω_1,ω_2)=0pt≤ft({\widetilde{X}}_
{1}(ω_1){\widetilde{X}}_2(ω_2)\right),
$$
and let ${\widetilde{Θ}}^*_(12)3$ and ${\widetilde{Θ}}^*_123$ be the fuzzy subsets of $Ω_1×Ω_2×Ω_3$ defined, respectively, as
$$
{\widetilde{Θ}}_(12)3^*(ω_1,ω_2,ω_3)=0pt≤ft(
≤ft[{\widetilde{X}}_1(ω_1)\varodot{\widetilde{X}}_2(ω_1)
\right]{\widetilde{X}}_3(ω_3)\right) \tag{12}
$$
and
$$
{\widetilde{Θ}}_123^*(ω_1,ω_2,ω_3)=0pt≤ft({
\widetilde{X}}_1(ω_1){\widetilde{X}}_2(ω_2){\widetilde{X}}_
3(ω_3)\right).
$$
Let ${\widetilde{P}}_12=(P_1× P_2)(·\mid{\widetilde{Θ}}_12^*)$ , ${\widetilde{P}}_(12)3=({\widetilde{P}}_12× P_3)(·\mid{ \widetilde{Θ}}_(12)3^*)$ , and ${\widetilde{P}}_123=(P_1× P_2× P_3)(·\mid{\widetilde{ Θ}}_123^*)$ . We only need to show that ${\widetilde{P}}_(12)3={\widetilde{P}}_123$ . For any $B∈σ_1⊗σ_2⊗σ_3$ , we have
$$
\displaystyle{\widetilde{P}}_(12)3(B) \displaystyle∝∫_Ω_{1×Ω_2}∫_Ω_3B(ω
_1,ω_2,ω_3)0pt≤ft(≤ft[{\widetilde{X}}_1(ω_1)
\varodot{\widetilde{X}}_2(ω_1)\right]{\widetilde{X}}_3(ω_3)
\right)dP_3(ω_3)d{\widetilde{P}}_12(ω_1,ω_2) \displaystyle∝∫_Ω_1∫_Ω_2∫_Ω_3B(
ω_1,ω_2,ω_3)0pt≤ft(≤ft[{\widetilde{X}}_1(ω_1
)\varodot{\widetilde{X}}_2(ω_1)\right]{\widetilde{X}}_3(ω_3)
\right)× \displaystyle\hskip 142.26378pt0pt≤ft({\widetilde{X}}_1(ω_1){
\widetilde{X}}_2(ω_2)\right)dP_3(ω_3)dP_2(ω_2)dP_1
(ω_1). \tag{12}
$$
Now,
| | $\displaystyle 0pt≤ft(≤ft[{\widetilde{X}}_1(ω_1)\varodot{ \widetilde{X}}_2(ω_1)\right]{\widetilde{X}}_3(ω_3)\right)$ | $\displaystyle=0pt≤ft(\frac{{\widetilde{X}}_1(ω_1){\widetilde{X}}_2 (ω_1)}{0pt({\widetilde{X}}_1(ω_1){\widetilde{X}}_2(ω_ 2))}{\widetilde{X}}_3(ω_3)\right)$ | |
| --- | --- | --- | --- |
Hence,
$$
{\widetilde{P}}_(12)3(B)∝∫_Ω_1∫_Ω_2∫_
Ω_3B(ω_1,ω_2,ω_3)0pt≤ft({\widetilde{X}}_1(
ω_1){\widetilde{X}}_2(ω_2){\widetilde{X}}_3(ω_3)
\right)dP_3(ω_3)dP_2(ω_2)dP_1(ω_1), \tag{12}
$$
which proves that ${\widetilde{P}}_(12)3={\widetilde{P}}_123$ , and the associativity of $⊕$ .
## Appendix C Proof of Proposition 6
We have
$$
\displaystyle pl_\widetilde{X}(x)= \displaystyleE_M[φ(x;M,h)] \displaystyle= \displaystyle∫_-∞^+∞φ(x;m,h)φ(m;μ,σ)dm \displaystyle= \displaystyle\frac{1}{σ√{2π}}∫_-∞^+∞\exp≤ft(-
\frac{h}{2}(x-m)^2\right)\exp≤ft(-\frac{(m-μ)^2}{2σ^2}\right)dm. \tag{44}
$$
From Proposition 3, the integrand can be written as
$$
\exp≤ft(-\frac{(m-μ_0)^2}{2σ_0^2}\right)\exp≤ft(-\frac{h(x-
μ)^2}{2(1+hσ^2)}\right),
$$
with
$$
μ_0=\frac{xh+μ/σ^2}{h+1/σ^2}=\frac{xhσ^2+μ}{h
σ^2+1}
$$
and
$$
σ_0=√{\frac{1}{h+1/σ^2}}=\frac{σ}{√{1+hσ^2}}.
$$
Consequently,
$$
\displaystyle pl_\widetilde{X}(x)= \displaystyle\frac{1}{σ√{2π}}\exp≤ft(-\frac{h(x-μ)^2}{2(1+h
σ^2)}\right)\underbrace{∫_-∞^+∞\exp≤ft(-\frac{(m-μ
_0)^2}{2σ_0^2}\right)dm}_σ_{0√{2π}} \displaystyle=\frac{1}{√{1+hσ^2}}\exp≤ft(-\frac{h(x-μ)^2}{2(1
+hσ^2)}\right). \tag{47}
$$
## Appendix D Proof of Proposition 7
If $h=0$ , we have, trivially, $Bel_\widetilde{X}([x,y])=0$ and $Pl_\widetilde{X}([x,y])=1$ for all $x≤ y$ . Let us assume that $h>0$ . We have
$$
Pl_\widetilde{X}([x,y])=ℙ(M≤ x)E[φ(x;M,h)\mid M
≤ x]+\\
ℙ(x<M≤ y)× 1+ℙ(M>y)E[φ(y;M,h)\mid M>y], \tag{49}
$$
which can be written as
$$
Pl_\widetilde{X}([x,y])=Φ≤ft(\frac{x-μ}{σ}\right)E[
φ(x;M,h)\mid M≤ x]+\\
Φ≤ft(\frac{y-μ}{σ}\right)-Φ≤ft(\frac{x-μ}{σ}\right)+\\
≤ft[1-Φ≤ft(\frac{y-μ}{σ}\right)\right]E[φ(y;M,h)
\mid M>y]. \tag{50}
$$
Conditionally on $M≤ x$ , $M$ has a truncated normal distribution on $(-∞,x]$ with pdf
$$
f(m)=\frac{1}{σ√{2π}}\frac{\exp≤ft(\frac{-(m-μ)^2}{2σ^2
}\right)}{Φ≤ft(\frac{x-μ}{σ}\right)}1_(-∞,x](m).
$$
Consequently,
$$
E[φ(x;M,h)\mid M≤ x]=\frac{1}{σ√{2π}}\frac{1}{
Φ≤ft(\frac{x-μ}{σ}\right)}\underbrace{∫_-∞^x\exp≤ft(
-\frac{h}{2}(x-m)^2\right)\exp≤ft(-\frac{(m-μ)^2}{2σ^2}\right)
dm}_I. \tag{51}
$$
From Proposition 3, integral $I$ in (51) can be written as
$$
I=σ_0√{2π}Φ≤ft(\frac{x-μ_0}{σ_0}\right)\exp≤ft(
-\frac{(x-μ)^2}{2(h^-1+σ^2)}\right),
$$
with
$$
μ_0=\frac{xhσ^2+μ}{hσ^2+1} and σ_0=
\frac{σ}{√{hσ^2+1}}.
$$
Consequently,
$$
E[φ(x;M,h)\mid M≤ x]=\frac{1}{Φ≤ft(\frac{x-μ}{σ}
\right)}pl_\widetilde{X}(x)Φ≤ft(\frac{x-μ}{σ√{hσ^2+1}
}\right).
$$
Using similar calculations, we find
$$
E[φ(y;M,h)\mid M>y]=\frac{1}{1-Φ≤ft(\frac{y-μ}{σ}
\right)}pl_\widetilde{X}(y)≤ft[1-Φ≤ft(\frac{y-μ}{σ√{h
σ^2+1}}\right)\right],
$$
which concludes the proof of (24).
Now, let us consider (23). We have
$$
Bel_\widetilde{X}([x,y])=1-Pl_\widetilde{X}((-∞,x]∪[y,+∞)),
$$
and
$$
Pl_\widetilde{X}((-∞,x]∪[y,+∞))=ℙ(M≤ x)× 1+\\
ℙ(x<M≤(x+y)/2)E[φ(x;M,h)\mid x<M≤(x+y)/2]+\\
ℙ((x+y)/2<M≤ y)E[φ(y;M,h)\mid(x+y)/2<M≤ y]+
ℙ(M>y)× 1, \tag{52}
$$
which can be written as
$$
Pl_\widetilde{X}((-∞,x]∪[y,+∞))=Φ≤ft(\frac{x-μ}{σ}
\right)+\\
≤ft[Φ≤ft(\frac{(x+y)/2-μ}{σ}\right)-Φ≤ft(\frac{x-μ}{
σ}\right)\right]E[φ(x;M,h)\mid x<M≤(x+y)/2]+\\
≤ft[Φ≤ft(\frac{y-μ}{σ}\right)-Φ≤ft(\frac{(x+y)/2-μ}{
σ}\right)\right]E[φ(y;M,h)\mid(x+y)/2<M≤ y]+\\
1-Φ≤ft(\frac{y-μ}{σ}\right). \tag{53}
$$
Conditionally on $x<M≤(x+y)/2$ , $M$ has a truncated normal distribution on $(x,(x+y)/2]$ with pdf
$$
f(m)=\frac{1}{σ√{2π}}\frac{\exp≤ft(\frac{-(m-μ)^2}{2σ^2
}\right)}{Φ≤ft(\frac{(x+y)/2-μ}{σ}\right)-Φ≤ft(\frac{x-μ}{
σ}\right)}1_(x,(x+y)/2](m).
$$
Consequently,
$$
E[φ(x;M,h)\mid x<M≤(x+y)/2]=\frac{1}{σ√{2π}}\frac
{1}{Φ≤ft(\frac{(x+y)/2-μ}{σ}\right)-Φ≤ft(\frac{x-μ}{σ}
\right)}×\\
\underbrace{∫_x^(x+y)/2\exp≤ft(-\frac{h}{2}(x-m)^2\right)\exp≤ft(
-\frac{(m-μ)^2}{2σ^2}\right)dm}_I^\prime. \tag{54}
$$
The integral in (54) is, with the same notations as before,
$$
I^\prime=σ_0√{2π}≤ft[Φ≤ft(\frac{(x+y)/2-μ_0}{σ_
{0}}\right)-Φ≤ft(\frac{x-μ_0}{σ_0}\right)\right]\exp≤ft(-
\frac{(x-μ)^2}{2(h^-1+σ^2)}\right).
$$
Consequently,
$$
E[φ(x;M,h)\mid x<M≤(x+y)/2]=\\
\frac{1}{Φ≤ft(\frac{(x+y)/2-μ}{σ}\right)-Φ≤ft(\frac{x-μ}{
σ}\right)}pl_\widetilde{X}(x)≤ft[Φ≤ft(\frac{(x+y)/2-μ+hσ^
{2}(y-x)/2}{σ√{hσ^2+1}}\right)\right.\\
≤ft.-Φ≤ft(\frac{x-μ}{σ√{hσ^2+1}}\right)\right]. \tag{55}
$$
Similarly, we find
$$
E[φ(y;M,h)\mid(x+y)/2<M≤ y]=\\
\frac{1}{Φ≤ft(\frac{y-μ}{σ}\right)-Φ≤ft(\frac{(x+y)/2-μ}{
σ}\right)}pl_\widetilde{X}(y)≤ft[Φ≤ft(\frac{y-μ}{σ√{h
σ^2+1}}\right)-\right.\\
≤ft.Φ≤ft(\frac{(x+y)/2-μ-(y-x)hσ^2/2}{σ√{hσ^2+
1}}\right)\right]. \tag{56}
$$
The expressions of $Pl_\widetilde{X}((-∞,x]∪[y,+∞))$ and $Bel_\widetilde{X}([x,y])$ follow.
## Appendix E Proof of Proposition 8
Let ${\widetilde{X}}(ω)=\textsf{GFN}(M(ω),h)$ be the image of $ω∈Ω$ by ${\widetilde{X}}$ , with $M∼ N(μ,σ^2)$ . For any $α∈(0,1]$ , its alpha-cut is the random interval
$$
{}^α{\widetilde{X}}(ω)=≤ft[M(ω)-√{\frac{-2\lnα}{h}
},M(ω)+√{\frac{-2\lnα}{h}}\right].
$$
Consequently, from (16), the lower and upper expectation of ${\widetilde{X}}$ are
$$
E_*({\widetilde{X}})=μ-∫_0^1√{\frac{-2\lnα}{h}}dα,
$$
and
$$
E^*({\widetilde{X}})=μ+∫_0^1√{\frac{-2\lnα}{h}}dα.
$$
By the change of variable $β=√{-2(\lnα)/h}$ , we get
$$
∫_0^1√{\frac{-2\lnα}{h}}dα=h∫_0^+∞β^2
\exp≤ft(-\frac{hβ^2}{2}\right)dβ.
$$
Now, the second-order moment of the normal distribution $N(0,1/h)$ is
$$
√{\frac{h}{2π}}∫_-∞^+∞β^2\exp≤ft(-\frac{hβ^
{2}}{2}\right)dβ=\frac{1}{h},
$$
from which we get
$$
h∫_0^+∞β^2\exp≤ft(-\frac{hβ^2}{2}\right)dβ=h
·\frac{1}{h}√{\frac{π}{2h}}=√{\frac{π}{2h}}.
$$
## Appendix F Proof of Lemma 1
The conditional density of $(M_1,M_2)$ is
$$
f(m_1,m_2\mid{\widetilde{F}})=\frac{f(m_1,m_2){\widetilde{F}}(m_1,m_
{2})}{\iint f(m_1,m_2){\widetilde{F}}(m_1,m_2)dm_1dm_2}. \tag{57}
$$
The numerator on the right-hand side of (57) is
$$
\frac{1}{2πσ_1σ_2}\exp≤ft\{-\frac{1}{2}≤ft[≤ft(\frac{m_
1-μ_1}{σ_1}\right)^2+≤ft(\frac{m_2-μ_2}{σ_2}
\right)^2\right]\right\}\exp≤ft\{-\frac{{\overline{h}}(m_1-m_2)^2}{2
}\right\}\\
=\frac{1}{2πσ_1σ_2}\exp≤ft\{-\frac{1}{2}≤ft[m_1^2≤ft
(\frac{1}{σ_1^2}+{\overline{h}}\right)-\frac{2m_1μ_1}{σ_1
^2}+\frac{μ_1^2}{σ_1^2}+\right.\right.\\
≤ft.≤ft.m_2^2≤ft(\frac{1}{σ_2^2}+{\overline{h}}\right)-
\frac{2m_2μ_2}{σ_2^2}+\frac{μ_2^2}{σ_2^2}-2{
\overline{h}}m_1m_2\right]\right\}. \tag{58}
$$
Now, the two-dimensional Gaussian density with parameters $({\widetilde{μ}}_1,{\widetilde{μ}}_2,{\widetilde{σ}}_1,{ \widetilde{σ}}_2,ρ)$ equals
$$
\frac{1}{2π{\widetilde{σ}}_1{\widetilde{σ}}_2√{1-ρ^2}
}\exp≤ft\{-\frac{1}{2(1-ρ)^2}≤ft[≤ft(\frac{m_1-{\widetilde{μ}}_
{1}}{{\widetilde{σ}}_1}\right)^2-\right.\right.\\
≤ft.≤ft.2ρ≤ft(\frac{m_1-{\widetilde{μ}}_1}{{\widetilde{σ}}
_1}\right)≤ft(\frac{m_2-{\widetilde{μ}}_2}{{\widetilde{σ}}_2}
\right)+≤ft(\frac{m_2-{\widetilde{μ}}_2}{{\widetilde{σ}}_2}
\right)^2\right]\right\}. \tag{59}
$$
Equating the second and first-order terms inside the exponentials in (58) and (59) gives us
$$
\displaystyle{\widetilde{σ}}_1 \displaystyle=\frac{1}{1-ρ^2}≤ft(\frac{1}{σ_1^2}+{\overline{h}
}\right)^-1 \displaystyle{\widetilde{σ}}_2 \displaystyle=\frac{1}{1-ρ^2}≤ft(\frac{1}{σ_2^2}+{\overline{h}
}\right)^-1 \displaystyleρ \displaystyle=\frac{{\overline{h}}σ_1σ_2}{√{(1+{\overline{h}
}σ_1^2)(1+{\overline{h}}σ_2^2)}} \displaystyle{\widetilde{μ}}_1 \displaystyle=\frac{μ_1{\widetilde{σ}}_1^2}{σ_1^2}+ρ
μ_2\frac{{\widetilde{σ}}_1{\widetilde{σ}}_2}{σ_2^2} \displaystyle{\widetilde{μ}}_2 \displaystyle=\frac{μ_2{\widetilde{σ}}_2^2}{σ_2^2}+ρ
μ_1\frac{{\widetilde{σ}}_1{\widetilde{σ}}_2}{σ_1^2}.
$$
Replacing $ρ$ by its expression (60c) in (60a) and (60b) yields (28c) and (28d). Replacing $ρ$ , ${\widetilde{σ}}_1$ and ${\widetilde{σ}}_2$ by their expressions in (60d) and (60e) gives (28a) and (28b).
Finally, the degree of conflict between GRFN’s ${\widetilde{X}}_1∼{\widetilde{N}}(μ_1,σ_1^2,h_1)$ and ${\widetilde{X}}_2∼{\widetilde{N}}(μ_2,σ_2^2,h_2)$ is
$$
κ=1-(P_1× P_2)({\widetilde{Θ}}^*),
$$
with
$$
(P_1× P_2)({\widetilde{Θ}}^*)=\iint f(m_1,m_2){\widetilde{
F}}(m_1,m_2)dm_1dm_2.
$$
Taking the ratio of (58) to (59), we get
$$
\iint f(m_1,m_2){\widetilde{F}}(m_1,m_2)dm_1dm_2=\\
\frac{{\widetilde{σ}}_1{\widetilde{σ}}_2}{σ_1σ_2}
√{1-ρ^2}\exp≤ft\{-\frac{1}{2}≤ft[\frac{μ_1^2}{σ_1^2
}+\frac{μ_2^2}{σ_2^2}\right]+\frac{1}{2(1-ρ^2)}≤ft[
\frac{{\widetilde{μ}}_1^2}{{\widetilde{σ}}_1^2}+\frac{{
\widetilde{μ}}_2^2}{{\widetilde{σ}}_2^2}-2ρ\frac{{\widetilde
{μ}}_1{\widetilde{μ}}_2}{{\widetilde{σ}}_1{\widetilde{σ}}_
{2}}\right]\right\}.
$$
## Appendix G Proof of Proposition 10
From (29), $h_12=+∞$ and the combined GRFN ${\widetilde{N}}({\widetilde{μ}}_12,{\widetilde{σ}}_12^2,h_12)$ is probabilistic. From (30) and (31),
$$
{\widetilde{μ}}_12=\lim_h_{1→+∞}\frac{{\widetilde{μ}}_
1+\frac{h_2}{h_1}{\widetilde{μ}}_2}{1+\frac{h_2}{h_1}}={
\widetilde{μ}}_1,
$$
and
$$
{\widetilde{σ}}_12^2=\lim_h_{1→+∞}\frac{{\widetilde{
σ}}^2_1+\frac{h^2_2}{h^2_1}{\widetilde{σ}}^2_2+2ρ
\frac{h_2}{h_1}{\widetilde{σ}}_1{\widetilde{σ}}_2}{(1+\frac{
h_2}{h_1})^2}={\widetilde{σ}}^2_1.
$$
From (28f),
$$
{\overline{h}}=\lim_h_{1→+∞}\frac{h_2}{1+\frac{h_2}{h_1
}}=h_2.
$$
From (28a) and (28c),
$$
{\widetilde{μ}}_1=\frac{μ_1(1+h_2σ_2^2)+μ_2h_2σ_
{1}^2}{1+h_2(σ_1^2+σ_2^2)},
$$
and
$$
{\widetilde{σ}}_1^2=\frac{σ_1^2(1+h_2σ_2^2)}{1+h_
{2}(σ_1^2+σ_2^2)}.
$$
Now, using Proposition 3, the product of the probability density of $X_1$ and the contour function of ${\widetilde{X}}_2$ can be written as
| | $\displaystyle f_X_1(x)pl_{\widetilde{X}_2}(x)$ | $\displaystyle∝\exp≤ft(-\frac{1}{2}\frac{(x-μ_1)^2}{σ_1^2 }\right)\exp≤ft(-\frac{h_2(x-μ_2)^2}{2(1+h_2σ_2^2)}\right)$ | |
| --- | --- | --- | --- |
with
$$
\frac{1}{σ_12^2}=\frac{1}{σ_1^2}+\frac{h_2}{1+h_2σ_
{2}^2}=\frac{1+h_2(σ_1^2+σ_2^2)}{σ_1^2(1+h_2
σ_2^2)}
$$
and
$$
μ_12=\frac{\frac{1}{σ_1^2}μ_1+\frac{h_2}{1+h_2σ_2^
{2}}μ_2}{\frac{1}{σ_1^2}+\frac{h_2}{1+h_2σ_2^2}}=
\frac{μ_1(1+h_2σ_2^2)+μ_2h_2σ_1^2}{1+h_2(
σ_1^2+σ_2^2)}.
$$
We can check that $μ_12={\widetilde{μ}}_1$ and $σ_12^2={\widetilde{σ}}_1^2$ .
## Appendix H Proof of Proposition 12
We have
$$
\displaystyle pl_\widetilde{X}({\boldsymbol{x}}) \displaystyle=E_\boldsymbol{M}[φ({\boldsymbol{x}};{
\boldsymbol{M}},{\boldsymbol{H}})] \displaystyle=∫_ℝ^pφ({\boldsymbol{x}};{\boldsymbol{m}},{
\boldsymbol{H}})φ({\boldsymbol{m}};\boldsymbol{μ},Σ)d{\boldsymbol{m}} \displaystyle=\frac{1}{(2π)^p/2|\boldsymbol{Σ}|^1/2}∫_R
^p\exp≤ft(-\frac{1}{2}({\boldsymbol{x}}-{\boldsymbol{m}})^T{
\boldsymbol{H}}({\boldsymbol{x}}-{\boldsymbol{m}})\right)× \displaystyle\hskip 113.81102pt\exp≤ft(-\frac{1}{2}({\boldsymbol{m}}-
\boldsymbol{μ})\boldsymbol{Σ}^-1({\boldsymbol{m}}-\boldsymbol{μ})
\right)d{\boldsymbol{m}}. \tag{61}
$$
From Proposition 3, the integrand can be written as
$$
\exp≤ft(-\frac{1}{2}({\boldsymbol{m}}-\boldsymbol{μ}_0)^T\boldsymbol{
Σ}_0^-1({\boldsymbol{m}}-\boldsymbol{μ}_0)\right)\exp≤ft(-\frac
{1}{2}({\boldsymbol{x}}-\boldsymbol{μ})^T({\boldsymbol{H}}^-1+
\boldsymbol{Σ})^-1({\boldsymbol{x}}-\boldsymbol{μ})\right),
$$
with
$$
\boldsymbol{μ}_0=({\boldsymbol{H}}+\boldsymbol{Σ}^-1)^-1({
\boldsymbol{H}}{\boldsymbol{x}}+\boldsymbol{Σ}^-1\boldsymbol{μ})
$$
and
$$
\boldsymbol{Σ}_0=({\boldsymbol{H}}+\boldsymbol{Σ}^-1)^-1.
$$
Consequently,
$$
\displaystyle pl_\widetilde{X}({\boldsymbol{x}}) \displaystyle=\frac{1}{(2π)^p/2|\boldsymbol{Σ}|^1/2}\exp≤ft(-
\frac{1}{2}({\boldsymbol{x}}-\boldsymbol{μ})^T({\boldsymbol{H}}^-1+
\boldsymbol{Σ})^-1({\boldsymbol{x}}-\boldsymbol{μ})\right)× \displaystyle\hskip 113.81102pt\underbrace{∫_ℝ^p\exp≤ft(-
\frac{1}{2}({\boldsymbol{m}}-\boldsymbol{μ}_0)^T\boldsymbol{Σ}_0^
{-1}({\boldsymbol{m}}-\boldsymbol{μ}0)\right)d{\boldsymbol{m}}}_(2π)^p/2
|\boldsymbol{Σ_0|^1/2} \displaystyle=≤ft(\frac{|\boldsymbol{Σ}_0|}{|\boldsymbol{Σ}|}
\right)^1/2\exp≤ft(-\frac{1}{2}({\boldsymbol{x}}-\boldsymbol{μ})^T({
\boldsymbol{H}}^-1+\boldsymbol{Σ})^-1({\boldsymbol{x}}-\boldsymbol{
μ})\right) \displaystyle=\frac{1}{|I_p+\boldsymbol{Σ}H|^1/2}\exp≤ft(-\frac{1}{
2}({\boldsymbol{x}}-\boldsymbol{μ})^T({\boldsymbol{H}}^-1+\boldsymbol{
Σ})^-1({\boldsymbol{x}}-\boldsymbol{μ})\right). \tag{65}
$$
## Appendix I Proof of Lemma 2
The conditional density of ${\boldsymbol{M}}=({\boldsymbol{M}}_1,{\boldsymbol{M}}_2)$ is
$$
f({\boldsymbol{m}}_1,{\boldsymbol{m}}_2\mid{\widetilde{F}})=\frac{f({
\boldsymbol{m}}_1,{\boldsymbol{m}}_2){\widetilde{F}}({\boldsymbol{m}}_1,
{\boldsymbol{m}}_2)}{∫_ℝ^2pf({\boldsymbol{m}}_1,{
\boldsymbol{m}}_2){\widetilde{F}}({\boldsymbol{m}}_1,{\boldsymbol{m}}_2)
d{\boldsymbol{m}}_1d{\boldsymbol{m}}_2}. \tag{69}
$$
The numerator on the right-hand side of (69) is
$$
f({\boldsymbol{m}}_1,{\boldsymbol{m}}_2){\widetilde{F}}({\boldsymbol{m}}_
1,{\boldsymbol{m}}_2)=φ({\boldsymbol{m}}_1;\boldsymbol{μ}_1,
\boldsymbol{Σ}_1)φ({\boldsymbol{m}}_2;\boldsymbol{μ}_2,
\boldsymbol{Σ}_2)×\\
\exp≤ft\{-\frac{1}{2}({\boldsymbol{m}}_1-{\boldsymbol{m}}_2)^T{
\overline{{\boldsymbol{H}}}}({\boldsymbol{m}}_1-{\boldsymbol{m}}_2)\right\}, \tag{70}
$$
which can be written as
$$
f({\boldsymbol{m}}_1,{\boldsymbol{m}}_2){\widetilde{F}}({\boldsymbol{m}}_
1,{\boldsymbol{m}}_2)=\frac{1}{(2π)^p|\boldsymbol{Σ}_1
\boldsymbol{Σ}_2|^1/2}\exp≤ft(-\frac{Z}{2}\right)
$$
with
$$
Z={\boldsymbol{m}}_1^T(\boldsymbol{Σ}_1^-1+{\overline{{
\boldsymbol{H}}}}){\boldsymbol{m}}_1+{\boldsymbol{m}}_2^T(\boldsymbol{
Σ}_2^-1+{\overline{{\boldsymbol{H}}}}){\boldsymbol{m}}_2-2{
\boldsymbol{m}}_1^T{\overline{{\boldsymbol{H}}}}{\boldsymbol{m}}_2-2{
\boldsymbol{m}}_1^T\boldsymbol{Σ}_1^-1\boldsymbol{μ}_1-\\
2{\boldsymbol{m}}_2^T\boldsymbol{Σ}_2^-1\boldsymbol{μ}_2+
\boldsymbol{μ}_1^T\boldsymbol{Σ}_1^-1\boldsymbol{μ}_1+
\boldsymbol{μ}_2^T\boldsymbol{Σ}_2^-1\boldsymbol{μ}_2. \tag{71}
$$
Now, the $2p$ -dimensional Gaussian density with mean ${\widetilde{\boldsymbol{μ}}}$ and covariance matrix ${\widetilde{\boldsymbol{Σ}}}$ equals
$$
φ({\boldsymbol{m}};{\widetilde{\boldsymbol{μ}}},{\widetilde{\boldsymbol{
Σ}}})=\frac{1}{(2π)^p|{\widetilde{\boldsymbol{Σ}}}|^1/2}\exp
≤ft\{-\frac{1}{2}({\boldsymbol{m}}-\boldsymbol{μ})^T{\widetilde{
\boldsymbol{Σ}}}^-1({\boldsymbol{m}}-\boldsymbol{μ})\right\}. \tag{72}
$$
Decomposing vector ${\widetilde{\boldsymbol{μ}}}$ as ${\widetilde{\boldsymbol{μ}}}=({\widetilde{\boldsymbol{μ}}}_1,{\widetilde {\boldsymbol{μ}}}_2)$ , with ${\widetilde{\boldsymbol{μ}}}_1,{\widetilde{\boldsymbol{μ}}}_2∈ ℝ^p$ , and ${\widetilde{\boldsymbol{Σ}}}^-1$ as
$$
{\widetilde{\boldsymbol{Σ}}}^-1=\begin{pmatrix}{\boldsymbol{A}}&{
\boldsymbol{B}}\\
{\boldsymbol{B}}&{\boldsymbol{C}}\end{pmatrix},
$$
where ${\boldsymbol{A}}$ , ${\boldsymbol{B}}$ and ${\boldsymbol{C}}$ are $p× p$ matrices, we can rewrite (72) as
$$
φ({\boldsymbol{m}};{\widetilde{\boldsymbol{μ}}},{\widetilde{\boldsymbol{
Σ}}})=\frac{1}{(2π)^p|{\widetilde{\boldsymbol{Σ}}}|^1/2}\exp
≤ft\{-\frac{1}{2}Z^\prime\right\}
$$
with
$$
Z^\prime={\boldsymbol{m}}_1^T{\boldsymbol{A}}{\boldsymbol{m}}_1-2{
\boldsymbol{m}}_1^T{\boldsymbol{A}}{\widetilde{μ}}_1+{\widetilde{μ}}
_1^T{\boldsymbol{A}}{\widetilde{μ}}_1+{\boldsymbol{m}}_2^T{
\boldsymbol{C}}{\boldsymbol{m}}_2-2{\boldsymbol{m}}_2^T{\boldsymbol{C}}{
\widetilde{μ}}_2+{\widetilde{μ}}_2^T{\boldsymbol{C}}{\widetilde{μ}
}_u+\\
2{\boldsymbol{m}}_2^T{\boldsymbol{B}}{\boldsymbol{m}}_1-2{\boldsymbol{m}
}_2^T{\boldsymbol{B}}\boldsymbol{μ}_1-2{\boldsymbol{m}}_1^T{
\boldsymbol{B}}\boldsymbol{μ}_2+2\boldsymbol{μ}_2^T{\boldsymbol{B}}
\boldsymbol{μ}_1. \tag{73}
$$
Equating the second-order terms in (71) and (73), we get
$$
{\boldsymbol{A}}=\boldsymbol{Σ}_1^-1+{\overline{{\boldsymbol{H}}}},
{\boldsymbol{C}}=\boldsymbol{Σ}_2^-1+{\overline{{\boldsymbol{H}}
}}, {\boldsymbol{B}}=-{\overline{{\boldsymbol{H}}}}.
$$
Equating the first-order terms, we get
$$
\displaystyle\boldsymbol{Σ}_1^-1\boldsymbol{μ}_1 \displaystyle={\boldsymbol{A}}{\widetilde{\boldsymbol{μ}}}_1+{\boldsymbol{
B}}{\widetilde{\boldsymbol{μ}}}_2=(\boldsymbol{Σ}_1^-1+{\overline
{{\boldsymbol{H}}}}){\widetilde{\boldsymbol{μ}}}_1-{\overline{{\boldsymbol
{H}}}}{\widetilde{\boldsymbol{μ}}}_2, \displaystyle\boldsymbol{Σ}_2^-1\boldsymbol{μ}_2 \displaystyle={\boldsymbol{B}}{\widetilde{\boldsymbol{μ}}}_1+{\boldsymbol{
C}}{\widetilde{\boldsymbol{μ}}}_2=-{\overline{{\boldsymbol{H}}}}{
\widetilde{\boldsymbol{μ}}}_1+(\boldsymbol{Σ}_2^-1+{\overline{{
\boldsymbol{H}}}}){\widetilde{\boldsymbol{μ}}}_2.
$$
Multiplying both sides of (74a) and (74b) by ${\overline{{\boldsymbol{H}}}}^-1$ , we get
$$
\displaystyle({\overline{{\boldsymbol{H}}}}^-1\boldsymbol{Σ}_1^-1+{
\boldsymbol{I}}_p){\widetilde{\boldsymbol{μ}}}_1-{\widetilde{\boldsymbol
{μ}}}_2 \displaystyle={\overline{{\boldsymbol{H}}}}^-1\boldsymbol{Σ}_1^-1
\boldsymbol{μ}_1 \displaystyle-{\widetilde{\boldsymbol{μ}}}_1+({\overline{{\boldsymbol{H}}}
}^-1\boldsymbol{Σ}_2^-1+{\boldsymbol{I}}_p){\widetilde{
\boldsymbol{μ}}}_2 \displaystyle={\overline{{\boldsymbol{H}}}}^-1\boldsymbol{Σ}_2^-1
\boldsymbol{μ}_2, \tag{75}
$$
which can be written in matrix form
$$
\begin{pmatrix}{\overline{{\boldsymbol{H}}}}^-1\boldsymbol{Σ}_1^-1+
{\boldsymbol{I}}_p&-{\boldsymbol{I}}_p\\
-{\boldsymbol{I}}_p&{\overline{{\boldsymbol{H}}}}^-1\boldsymbol{Σ}_2
^-1+{\boldsymbol{I}}_p\end{pmatrix}\begin{pmatrix}{\widetilde{\boldsymbol
{μ}}}_1\\
{\widetilde{\boldsymbol{μ}}}_2\end{pmatrix}=\begin{pmatrix}{\overline{{
\boldsymbol{H}}}}^-1\boldsymbol{Σ}_1^-1&\boldsymbol{0}\\
\boldsymbol{0}&{\overline{{\boldsymbol{H}}}}^-1\boldsymbol{Σ}_2^-1
\end{pmatrix}\begin{pmatrix}\boldsymbol{μ}_1\\
\boldsymbol{μ}_2\end{pmatrix},
$$
from which we obtain (36).
Finally, the degree of conflict between GRFV’s ${\widetilde{X}}_1∼{\widetilde{N}}(\boldsymbol{μ}_1,\boldsymbol{Σ }_1,{\boldsymbol{H}}_1)$ and ${\widetilde{X}}_2∼{\widetilde{N}}(\boldsymbol{μ}_2,\boldsymbol{Σ }_2,{\boldsymbol{H}}_2)$ is
$$
κ=1-(P_1× P_2)({\widetilde{Θ}}^*)=1-∫_ℝ^2p
f({\boldsymbol{m}}_1,{\boldsymbol{m}}_2){\widetilde{F}}({\boldsymbol{m}}_
{1},{\boldsymbol{m}}_2)d{\boldsymbol{m}}_1d{\boldsymbol{m}}_2.
$$
Taking the ratio of (70) to (72), we get
$$
∫_ℝ^2pf({\boldsymbol{m}}_1,{\boldsymbol{m}}_2){\widetilde{
F}}({\boldsymbol{m}}_1,{\boldsymbol{m}}_2)d{\boldsymbol{m}}_1d{
\boldsymbol{m}}_2=\\
√{\frac{|{\widetilde{\boldsymbol{Σ}}}|}{|\boldsymbol{Σ}_1||
\boldsymbol{Σ}_2|}}\exp≤ft\{-\frac{1}{2}≤ft[\boldsymbol{μ}_1^T
\boldsymbol{Σ}_1^-1\boldsymbol{μ}_1+\boldsymbol{μ}_2^T
\boldsymbol{Σ}_2^-1\boldsymbol{μ}_2-{\widetilde{\boldsymbol{μ}}
}^T{\widetilde{\boldsymbol{Σ}}}^-1{\widetilde{\boldsymbol{μ}}}
\right]\right\}.
$$
## Appendix J Proof of Lemma 3
The membership function of the projection of fuzzy vector $\textsf{GFV}({\boldsymbol{m}},{\boldsymbol{H}})$ on $Θ_1$ is
$$
φ({\boldsymbol{x}}_1)=\max_{\boldsymbol{x}_2}\exp≤ft(-\frac{1}{2
}({\boldsymbol{x}}-{\boldsymbol{m}})^T{\boldsymbol{H}}({\boldsymbol{x}}-{
\boldsymbol{m}})\right)=\exp≤ft(-\frac{1}{2}\min_{\boldsymbol{x}_2}Z
\right), \tag{77}
$$
with $Z=({\boldsymbol{x}}-{\boldsymbol{m}})^T{\boldsymbol{H}}({\boldsymbol{x}}-{ \boldsymbol{m}})$ . Now,
$$
\displaystyle Z \displaystyle=({\boldsymbol{x}}_1-{\boldsymbol{m}}_1,{\boldsymbol{x}}_2-
{\boldsymbol{m}}_2)\begin{pmatrix}{\boldsymbol{H}}_11&{\boldsymbol{H}}_12
\\
{\boldsymbol{H}}_21&{\boldsymbol{H}}_22\end{pmatrix}\begin{pmatrix}{
\boldsymbol{x}}_1-{\boldsymbol{m}}_1\\
{\boldsymbol{x}}_2-{\boldsymbol{m}}_2\end{pmatrix} \displaystyle=({\boldsymbol{x}}_1-{\boldsymbol{m}}_1)^T{\boldsymbol{H}}_
{11}({\boldsymbol{x}}_1-{\boldsymbol{m}}_1)+({\boldsymbol{x}}_2-{
\boldsymbol{m}}_2)^T{\boldsymbol{H}}_21({\boldsymbol{x}}_1-{
\boldsymbol{m}}_1)+ \displaystyle\hskip 56.9055pt({\boldsymbol{x}}_1-{\boldsymbol{m}}_1)^T{
\boldsymbol{H}}_12({\boldsymbol{x}}_2-{\boldsymbol{m}}_2)+({\boldsymbol{
x}}_2-{\boldsymbol{m}}_2)^T{\boldsymbol{H}}_22({\boldsymbol{x}}_2-{
\boldsymbol{m}}_2).
$$
Using ${\boldsymbol{H}}_21={\boldsymbol{H}}_12^T$ , the gradient of $Z$ with respect to ${\boldsymbol{x}}_2$ can be written as
$$
\frac{∂ Z}{∂{\boldsymbol{x}}_2}=2{\boldsymbol{H}}_21({
\boldsymbol{x}}_1-{\boldsymbol{m}}_1)+2{\boldsymbol{H}}_22({\boldsymbol{
x}}_2-{\boldsymbol{m}}_2).
$$
Setting $\frac{∂ Z}{∂{\boldsymbol{x}}_2}=0$ , and assuming ${\boldsymbol{H}}_22$ to be nonsingular, we get
$$
({\boldsymbol{x}}_2-{\boldsymbol{m}}_2)=-{\boldsymbol{H}}_22^-1{
\boldsymbol{H}}_21({\boldsymbol{x}}_1-{\boldsymbol{m}}_1). \tag{79}
$$
Replacing $({\boldsymbol{x}}_2-{\boldsymbol{m}}_2)$ by its expression (79) in (78) and using (77), we finally get
$$
φ({\boldsymbol{x}}_1)=\exp≤ft(-\frac{1}{2}({\boldsymbol{x}}_1-{
\boldsymbol{m}}_1)^T{\boldsymbol{H}}_11^\prime({\boldsymbol{x}}_1-{
\boldsymbol{m}}_1)\right),
$$
with
$$
{\boldsymbol{H}}^\prime_11={\boldsymbol{H}}_11-{\boldsymbol{H}}_12{
\boldsymbol{H}}_22^-1{\boldsymbol{H}}_21.
$$