Evidence-based subjective logic

Evidence-based subjective logic (EBSL) is a variant [1] of subjective logic in which the transitivity of opinions (discounting) is handled by applying weights to the evidence underlying the opinions. Subjective Logic is based on Dempster–Shafer belief theory. The discounting rule in EBSL makes it possible to handle arbitrary trust networks.

Relation between evidence and opinions

Consider a proposition P. Let p be the amount of evidence supporting P, and n the amount of evidence supporting ¬P. We write the evidence as a vector (p, n). Let c be a positive constant representing a "unit" of evidence. An opinion (b, d, u) is formed on the basis of the evidence (p, n). There is a one-to-one mapping between the opinion and the evidence,


(b,d,u)=\frac{(p,n,c)}{p+n+c}
\quad\quad
(p,n)=c\frac{(b,d)}{u}.
\quad\quad (1)

In the original literature on subjective logic the constant was set to c = 2. The mapping (1) is the unique solution of the following set of constraints,[1]

Alternatively, (1) can be derived from an analysis of a posteriori probability distributions[2] (beta distributions).
There are three "corner points" in opinion space: the full Belief B = (1,0,0), the full Disbelief D = (0,1,0), and full Uncertainty U = (0,0,1). Opinions on the line between B and D (including B and D) are called "dogmatic opinions". They have zero uncertainty, which is achievable only with an infinite amount of evidence. Dogmatic opinions are often excluded from the algebra.

Consensus / fusion

The consensus operation combines two opinions about the same predicate into one opinion by piling up the evidence. Let x = (xb,xd,xu) and y = (yb,yd,yu) be the opinions that are to be fused, and z = x y the result. We denote their evidence vectors as (px, nx), (py, ny) and (pz, nz) respectively. In evidence space the consensus is straightforwardly defined as


(p_z,n_z) = (p_x,n_x)+(p_y,n_y).

In opinion space this yields


x\oplus y = \frac{(p_x+p_y,n_x+n_y,c)}{p_x+p_y+n_x+n_y+c}

which using (1) can be rewritten as


x\oplus y = \frac{(x_{\rm u} y_{\rm b}+y_{\rm u} x_{\rm b},x_{\rm u} y_{\rm d}+y_{\rm u} x_{\rm d},x_{\rm u} y_{\rm u})}
{x_{\rm u}+y_{\rm u}-x_{\rm u} y_{\rm u}}.
\quad\quad (2)

The consensus rule can only be applied if the evidence underlying x and y is independent, otherwise double counting of evidence occurs.

Discounting

Old discounting rule ()

The traditional discounting operation in Subjective Logic is denoted as and defined as


x\otimes y \stackrel{\rm def}{=} 
(x_{\rm b}y_{\rm b}, x_{\rm b}y_{\rm d},1-x_{\rm b}y_{\rm b} - x_{\rm b}y_{\rm d}).

This operation suffers from a number of serious problems, e.g.

These problems make it very difficult to handle trust networks in subjective logic.

The product of a scalar and an opinion

Let x = (xb, xd, xu) be an opinion based on evidence (p, n). Let λ 0 be a scalar. The product λ x is defined[1] as (λ p, λ n) in evidence space, which corresponds to


\lambda\cdot x=\frac{(\lambda x_{\rm b},\lambda x_{\rm d},x_{\rm u})}
{\lambda(x_{\rm b}+x_{\rm d})+x_{\rm u}}
\quad\quad (3)

in opinion space.

The discounting rule () in EBSL

Let x and y be opinions. Let g be a mapping from opinion space to [0,1] satisfying g(B) = 1 and g(D) = 0.

In EBSL the discounting of y through x is denoted as x y and defined as[1]


\quad
x \boxtimes y \stackrel{\rm def}{=} g(x)\cdot y, 
\quad \quad (4)

with the "dot" product as specified in (3).

The function g can be chosen at will, depending on the context. The rule has a very simple interpretation in evidence space: Due to the disbelief and uncertainty present in x, only a fraction g(x) of the evidence in y survives.

The operation avoids all the inconsistencies of the operation. The following properties hold,

There is no associativity, i.e. x_1\boxtimes(x_2\boxtimes y)\neq(x_1\boxtimes x_2)\boxtimes y, in contrast to the operation. This is not a problem, since the flow if information in a trust network has a well defined direction.

Also, we have

 
(x_1\oplus x_2)\boxtimes y \neq (x_1\boxtimes y)\oplus (x_2\boxtimes y).

Computation of opinions in arbitrary trust networks

EBSL makes it possible to compute trust values even when the graph connecting the users in the trust network is complicated. This makes EBSL interesting e.g. for Reputation systems

Let Aij be the opinion that user i has about the trustworthiness of user j, based on direct evidence, e.g. direct interactions between i and j. We set Aii = U. Let every user publish these direct opinions in a reliable way; the matrix A is public and its integrity is guaranteed.
Based on all the available trust information, direct as well as indirect, what should a user conclude about the trustworthiness of all the other users? In general this is a nontrivial problem because of the complicated connection graphs, in which loops may occur. The problem is to find a "reputation" matrix R that consistently combines the direct and indirect evidence. In EBSL the following "self-consistent" (self-containing) equation must be satisfied[1] by R,

 \mbox{For }i\neq j:\quad
R_{ij}=A_{ij}\oplus \sum_{k:\,k\neq i}R_{ik}\boxtimes A_{kj}
\quad\quad (5)

\mbox{Diagonal: }R_{ii}=B.

Here the "Σ" stands for operations. The diagonal is set to full belief since everybody trusts himself implicitly, independent of other users' opinions.
User i forms an opinion about j by combining his direct opinion Aij with other users' opinions Akj. The indirect evidence is weighted with a scalar that depends on the reputation of the intermediary: g(Rik).

Equation (5) can be written compactly in matrix form,


R = B{\mathbf 1}\oplus (R\boxtimes A).\quad\quad (6)

Here {\mathbf 1} is the identity matrix, and summation should be executed as . Eq.(6) is a fixed point equation similar to the case of Markov chains. It can be solved recursively. Let X be a square matrix with the same dimensions as A. Define a function f as

f(X)=B{\mathbf 1}\oplus (X\boxtimes A).

Pick a starting matrix X_0 and repeatedly apply f until the result does not change any more, i.e. a fixed point R=f(R) is reached. Independent of the choice of X_0, after one iteration the diagonal is B1 and stays B1 in all further iterations.
If Eq.(6) were an ordinary matrix equation R={\mathbf 1}+RA for real-valued (or complex) Aij, it would have solution R=({\mathbf 1}-A)^{-1}={\mathbf 1}+\sum_{k\geq 1}A^k. However, the opinion algebra does not allow for such a simple expression. Instead we have

         f2(X0) = B1 ((B1 (X0 A)) A)

         f3(X0) = B1 ((B1 ( B1 (X0 A) A)) A)

         ...

which in general cannot be simplified.

References

  1. 1 2 3 4 5 B. Skoric, S.J.A. de Hoogh, N. Zannone. Flow-based reputation with uncertainty: Evidence-Based Subjective Logic. http://link.springer.com/article/10.1007/s10207-015-0298-5 . http://arxiv.org/abs/1402.3319
  2. A. Jøsang. A Logic for Uncertain Probabilities. International Journal of Uncertainty, Fuzziness and Knowledge-Based Systems. 9(3), pp. 279–311, June 2001. PDF
This article is issued from Wikipedia - version of the Monday, August 31, 2015. The text is available under the Creative Commons Attribution/Share Alike but additional terms may apply for the media files.