Supriya Ghosh (Editor)

Evidence based subjective logic

Updated on
Edit
Like
Comment
Share on FacebookTweet on TwitterShare on LinkedInShare on Reddit

Evidence-based subjective logic (EBSL) is a variant of subjective logic in which the transitivity of opinions (discounting) is handled by applying weights to the evidence underlying the opinions. Subjective Logic is based on Dempster–Shafer belief theory. The discounting rule in EBSL makes it possible to handle arbitrary trust networks.

Contents

Relation between evidence and opinions

Consider a proposition P. Let p be the amount of evidence supporting P, and n the amount of evidence supporting ¬P. We write the evidence as a vector (p, n). Let c be a positive constant representing a "unit" of evidence. An opinion (b, d, u) is formed on the basis of the evidence (p, n). There is a one-to-one mapping between the opinion and the evidence,

( b , d , u ) = ( p , n , c ) p + n + c ( p , n ) = c ( b , d ) u . ( 1 )

In the original literature on subjective logic the constant was set to c = 2. The mapping (1) is the unique solution of the following set of constraints,

  • b/d = p/n.
  • b + d + u = 1.
  • p + n = 0 implies u = 1.
  • p implies u 0 .
  • Alternatively, (1) can be derived from an analysis of a posteriori probability distributions (beta distributions).
    There are three "corner points" in opinion space: the full Belief B = (1,0,0), the full Disbelief D = (0,1,0), and full Uncertainty U = (0,0,1). Opinions on the line between B and D (including B and D) are called "dogmatic opinions". They have zero uncertainty, which is achievable only with an infinite amount of evidence. Dogmatic opinions are often excluded from the algebra.

    Consensus / fusion

    The consensus operation combines two opinions about the same predicate into one opinion by piling up the evidence. Let x = (xb,xd,xu) and y = (yb,yd,yu) be the opinions that are to be fused, and z = xy the result. We denote their evidence vectors as (px, nx), (py, ny) and (pz, nz) respectively. In evidence space the consensus is straightforwardly defined as

    ( p z , n z ) = ( p x , n x ) + ( p y , n y ) .

    In opinion space this yields

    x y = ( p x + p y , n x + n y , c ) p x + p y + n x + n y + c

    which using (1) can be rewritten as

    x y = ( x u y b + y u x b , x u y d + y u x d , x u y u ) x u + y u x u y u . ( 2 )

    The consensus rule can only be applied if the evidence underlying x and y is independent, otherwise double counting of evidence occurs.

    Old discounting rule (⊗)

    The traditional discounting operation in Subjective Logic is denoted as ⊗ and defined as

    x y = d e f ( x b y b , x b y d , 1 x b y b x b y d ) .

    This operation suffers from a number of serious problems, e.g.

  • There is no natural interpretation of the evidence underlying xy when (1) is applied.
  • Consider the following. Alice has trust x in Bob. Bob gathers two independent evidence vectors, (p1, n1) and (p2, n2), about some proposition P.
    Scenario I: Bob forms two indendent opinions, y1 and y2, based on the evidence. He sends y1 and y2 to Alice. Alice forms opinion ( x y 1 ) ( x y 2 ) about P.
    Scenario II: Bob combines his evidence and forms opinion y 1 y 2 . He then sends y 1 y 2 to Alice. Alice forms opinion x ( y 1 y 2 ) about P.
    It is obvious that these scenarios should yield the same result for Alice. Yet the traditional discounting rule gives: x ( y 1 y 2 ) ( x y 1 ) ( x y 2 ) .
  • These problems make it very difficult to handle trust networks in subjective logic.

    The product of a scalar and an opinion

    Let x = (xb, xd, xu) be an opinion based on evidence (p, n). Let λ ≥ 0 be a scalar. The product λ ⋅ x is defined as (λ p, λ n) in evidence space, which corresponds to

    λ x = ( λ x b , λ x d , x u ) λ ( x b + x d ) + x u ( 3 )

    in opinion space.

    The discounting rule (☒) in EBSL

    Let x and y be opinions. Let g be a mapping from opinion space to [0,1] satisfying g(B) = 1 and g(D) = 0.

    In EBSL the discounting of y through x is denoted as xy and defined as

    x y = d e f g ( x ) y , ( 4 )

    with the "dot" product as specified in (3).

    The function g can be chosen at will, depending on the context. The ☒ rule has a very simple interpretation in evidence space: Due to the disbelief and uncertainty present in x, only a fraction g(x) of the evidence in y survives.

    The ☒ operation avoids all the inconsistencies of the ⊗ operation. The following properties hold,

  • x ( y 1 y 2 ) = ( x y 1 ) ( x y 2 )
  • x 1 ( x 2 y ) = x 2 ( x 1 y ) .
  • There is no associativity, i.e. x 1 ( x 2 y ) ( x 1 x 2 ) y , in contrast to the ⊗ operation. This is not a problem, since the flow if information in a trust network has a well defined direction.

    Also, we have

    ( x 1 x 2 ) y ( x 1 y ) ( x 2 y ) .

    Computation of opinions in arbitrary trust networks

    EBSL makes it possible to compute trust values even when the graph connecting the users in the trust network is complicated. This makes EBSL interesting e.g. for Reputation systems

    Let Aij be the opinion that user i has about the trustworthiness of user j, based on direct evidence, e.g. direct interactions between i and j. We set Aii = U. Let every user publish these direct opinions in a reliable way; the matrix A is public and its integrity is guaranteed.
    Based on all the available trust information, direct as well as indirect, what should a user conclude about the trustworthiness of all the other users? In general this is a nontrivial problem because of the complicated connection graphs, in which loops may occur. The problem is to find a "reputation" matrix R that consistently combines the direct and indirect evidence. In EBSL the following "self-consistent" (self-containing) equation must be satisfied by R,

    For  i j : R i j = A i j k : k i R i k A k j ( 5 ) Diagonal:  R i i = B .

    Here the "Σ" stands for ⊕ operations. The diagonal is set to full belief since everybody trusts himself implicitly, independent of other users' opinions.
    User i forms an opinion about j by combining his direct opinion Aij with other users' opinions Akj. The indirect evidence is weighted with a scalar that depends on the reputation of the intermediary: g(Rik).

    Equation (5) can be written compactly in matrix form,

    R = B 1 ( R A ) . ( 6 )

    Here 1 is the identity matrix, and summation should be executed as ⊕. Eq.(6) is a fixed point equation similar to the case of Markov chains. It can be solved recursively. Let X be a square matrix with the same dimensions as A. Define a function f as

    f ( X ) = B 1 ( X A ) .

    Pick a starting matrix X 0 and repeatedly apply f until the result does not change any more, i.e. a fixed point R = f ( R ) is reached. Independent of the choice of X 0 , after one iteration the diagonal is B1 and stays B1 in all further iterations.
    If Eq.(6) were an ordinary matrix equation R = 1 + R A for real-valued (or complex) Aij, it would have solution R = ( 1 A ) 1 = 1 + k 1 A k . However, the opinion algebra does not allow for such a simple expression. Instead we have

             f2(X0) = B1 ⊕ ((B1 ⊕ (X0A)) ☒ A)

             f3(X0) = B1 ⊕ ((B1 ⊕ ( B1 ⊕ (X0A) ☒ A)) ☒ A)

             ...

    which in general cannot be simplified.

    References

    Evidence-based subjective logic Wikipedia


    Similar Topics