Relevance Feedback for Association Rules by Leveraging Concepts from - - PowerPoint PPT Presentation

relevance feedback for association rules by leveraging
SMART_READER_LITE
LIVE PREVIEW

Relevance Feedback for Association Rules by Leveraging Concepts from - - PowerPoint PPT Presentation

Relevance Feedback for Association Rules by Leveraging Concepts from Information Retrieval Georg Ru, russ@iws.cs.uni-magdeburg.de Mirko Bttcher, mail@mirkoboettcher.de Prof. Dr. Rudolf Kruse, kruse@iws.cs.uni-magdeburg.de December 12th,


slide-1
SLIDE 1

Relevance Feedback for Association Rules by Leveraging Concepts from Information Retrieval

Georg Ruß, russ@iws.cs.uni-magdeburg.de Mirko Böttcher, mail@mirkoboettcher.de

  • Prof. Dr. Rudolf Kruse, kruse@iws.cs.uni-magdeburg.de

December 12th, 2007

slide-2
SLIDE 2

Outline

Association Rules Concepts from Information Retrieval Rule Similarity Relevance Scoring Conclusion

slide-3
SLIDE 3

Motivation

◮ Large amounts of transactional data ◮ Association rule mining yields rules as a condensed

representation

◮ Form: IF item1, item2, . . . , itemn THEN itemm ◮ Problem: still too many rules to analyze ◮ Topic: find interesting association rules

slide-4
SLIDE 4

Association Rules – Formalization

◮ Set D of transactions T ∈ D. ◮ Transaction T is a subset of a set of items L. ◮ A subset X ⊆ L is called itemset. ◮ A transaction T supports an itemset X if X ⊆ T . ◮ An association rule r is an expression X → Y where X and Y

are itemsets, |Y| > 0 and X ∩ Y = ∅.

◮ X: body, Y: head ◮ Rule reliability: confidence conf(r) = P(Y | X) ◮ Statistical significance: support supp(r) = P(XY) ◮ Time series: confidence and support of one rule over time

slide-5
SLIDE 5

Linking to Information Retrieval

◮ Interestingness of rules is subjective. ◮ Finding interesting rules requires user input. ◮ Manual specification of user’s knowledge

◮ key aspects are often forgotten ◮ requires expert user ◮ knowledge changes ◮ hard to specify at beginning of analysis

slide-6
SLIDE 6

Information Retrieval – Relevance Feedback

◮ Automatic acquisition of user’s knowledge through actions

◮ user rates what he sees ◮ easy (binary) decision: interesting / not interesting ◮ system collects user’s choices and updates results

◮ Relevance Feedback known from Information Retrieval

◮ association rules are presented (possibly pre-ordered) ◮ user can examine and rate them ◮ an internal ranking is adapted ◮ best results are presented ◮ cycle starts over

slide-7
SLIDE 7

Rule Representation – Informal

◮ Use existing algorithms for relevance feedback from IR ◮ Represent rules as vectors

  • r = (

body

  • r1, . . . , rb,

head

  • rb+1, . . . , rb+h,
  • symbolic

rb+h+1, . . . , rb+h+t

  • timeseries

) (1)

◮ item weights: TF-IDF approach ◮ high weight: term frequent in rule (TF), but less frequent in

rule set (IDF)

◮ filters commonly used terms, captures perceived relevance

slide-8
SLIDE 8

Rule Representation – Maths

◮ term frequency

tf (x, r) =

  • 1

if x ∈ r,

  • therwise.

(2)

◮ inverse document frequency

idf (x, R) = 1 − ln |r : r ∈ R ∧ x ∈ r| ln |R| (3)

◮ A rule’s feature vector is filled as follows:

ri = tf (xi, r) · idf (xi, R), i = 1, . . . , b (4)

rb+j = tf (xj, r) · idf (xj, R), j = 1, . . . , h (5)

◮ rtimeseries . . . respective time-variant properties of rule

slide-9
SLIDE 9

Interestingness of Rules

◮ Idea: compare the same features of different rules ◮ Interestingness based on (dis-)similarity ◮ Six combinations deemed interesting:

similar dissimilar head body time series symbolic head

  • ω4

ω5

  • body

ω1

  • ω6
  • time series
  • ω2

symbolic

  • ω3
  • Table: Interestingness Matrix
slide-10
SLIDE 10

Pairwise Similarity

◮ Similarity between rules as measure of interestingness ◮ Similarity can easily be computed by similarity measures for

vectors

◮ Cosine similarity:

sim( r, s) = n

i=1 risi

  • r2

i

  • s2

i

(6)

◮ Dissimilarity:

dissim( r, s) = 1 − sim( r, s) (7)

slide-11
SLIDE 11

Similarity Aggregation – first step

◮ Similarity between rule and rule set:

simrs( r, R) = Ω({sim( r, s1), . . . , sim( r, sm)}) (8)

◮ Dissimilarity analogously to Equation 7:

dissimrs( r, R) = 1 − simrs( r, R) (9)

slide-12
SLIDE 12

Similarity Aggregation – second step

◮ Use OWA operator to aggregate single similarities:

◮ weighting vector W = (w1, w2, . . . , wn)T with wj ∈ [0, 1] and

n

j=1 wj = 1

Ω({s1, s2, . . . , sn}) =

n

  • j=1

wjbj , (10) with bj being the j-th largest of the si.

slide-13
SLIDE 13

Relevance Scoring

◮ Score calculation for each association rule ◮ User selects rule r as interesting

◮ determine interesting combinations: ◮ rules with similar head, but different body ◮ rules with similar body, but dissimilar head ◮ six combinations (see Table 1)

◮ Calculate weighted sum of the score part in those six

combinations

◮ Yields a relevance score for each association rule ◮ Sort rules by score – interesting ones assumed to have high

score

slide-14
SLIDE 14

Conclusion

◮ Similarity-based interestingness of association rules ◮ Incorporation of relevance feedback to find interesting rules ◮ User-specific, automatic adaptation ◮ Simple relevance scoring to assess interestingness