Beyond Keyword Filtering for Message and Conversation Detection - PowerPoint PPT Presentation

Loading...

PPT – Beyond Keyword Filtering for Message and Conversation Detection PowerPoint presentation | free to download - id: 686348-OGFjY



Loading


The Adobe Flash plugin is needed to view this content

Get the plugin now

View by Category
About This Presentation
Title:

Beyond Keyword Filtering for Message and Conversation Detection

Description:

Beyond Keyword Filtering for Message and Conversation Detection David Skillicorn School of Computing, Queen s University Math and CS, Royal Military College – PowerPoint PPT presentation

Number of Views:2
Avg rating:3.0/5.0
Date added: 9 January 2020
Slides: 30
Provided by: DavidSki4
Category:

less

Write a Comment
User Comments (0)
Transcript and Presenter's Notes

Title: Beyond Keyword Filtering for Message and Conversation Detection


1
Beyond Keyword Filtering for Message and
Conversation Detection
  • David Skillicorn
  • School of Computing, Queens University
  • Math and CS, Royal Military College
  • skill_at_cs.queensu.ca

2
  • The problem
  • Pick out the most interesting intercepted
    messages when conventional markers
    (sender/receivers etc.) are missing.
  • The solution
  • Look for correlated use of words that are used
    with the wrong frequency, caused by
    substitution to evade keyword filtering.
  • The technique
  • Use singular value decomposition and independent
    component analysis applied to noun frequency
    profiles suspicious related messages appear as
    outliers.
  • Messages with ordinary word frequencies and lone
    eccentrics do not show up. So it can be applied
    to large sets of messages to select the
    interesting few.

3
  • THE PROBLEM

4
  • Many governments collect and analyze message
    traffic (e.g.
  • Echelon) email, file traffic/web, cellphone
    traffic, radio.
  • There are 3 levels of analysis
  • 1. Match the content of individual messages
    against a watch list of words that suggest the
    message is suspicious.
  • German Federal Intelligence Service nuclear
    proliferation (2000 terms), arms trade (1000),
    terrorism (500), drugs (400), as of 2000
    (certainly changed now).
  • Countermeasures use a speech code (hard in
    realtime) or use locutions (the package is
    ready).
  • Main benefit Changes behavior of those who DONT
    want their messages intercepted.

5
  • 2. Look for sets of messages that are
    connected, that form a conversation, based on
    some of their properties sender/receiver
    identities, time of transmission, specialized
    word use, etc..
  • (Social Network Analysis)
  • Countermeasures conceal the connections between
    the messages by making sure they share no obvious
    attributes
  • use temporary email addresses, stolen cell
    phones
  • decouple by using intermediaries
  • smear time factors e.g. by using web sites
  • In general, hide in the background noise .

6
  • 3. Look for sets of messages that are
    connected in more subtle ways because of
    correlation among their properties.
  • Workable countermeasures are hard to find
    because
  • conversations are about something, so that
    correlation in their
  • content arises naturally
  • sensitivity to watch list surveillance
    alters the way words are used
  • We hypothesize that related messages among a
    threat group in the context of watch list
    surveillance will be characterized by correlated
    word use but that the words will be used with
    the wrong frequencies.
  • Common words will be used as if they were
    uncommon uncommon words will be used as if they
    were common.

7
  • THE DATA

8
  • The frequency of words in English (and many other
    languages) is Zipf frequent words are very
    frequent, and frequency drops off very quickly.
  • We restrict our attention to nouns.
  • In English
  • Most common noun time
  • 3262nd most common noun quantum
  • We assume that messages are reduced to a
    frequency histogram of their nouns (this can be
    done reliably with a tagger).

9
  • A message-frequency matrix has a row
    corresponding to each message, and a column
    corresponding to each noun. The ij th entry is
    the frequency of noun j in message i .
  • The matrix is very sparse.
  • We generate artificial datasets using a Poisson
    distribution with mean f 1/j1 , where f models
    the base frequency.
  • We add 10 extra rows representing the correlated
    threat messages, using a block of 6 columns,
    uniformly randomly 0s and 1s, added at columns
    301306.

10
  • A message-rank matrix has a row corresponding to
    each message, and a column corresponding to the
    rank, in English, of the j th most frequent noun
    in the message.
  • Message-rank matrices have many fewer columns,
    which makes them easier and faster to work with
    (e.g. Enron email dataset 200,000 words but
    average number of nouns per message lt200).
  • Message-frequency matrices have been extensively
    studied in IR, but message-rank matrices not at
    all.
  • Message-rank messages are insensitive to
    countermeasures such as using words with almost
    the right frequency.

11
messages
nouns
12
messages
rank of jth noun in message
13
  • THE TECHNIQUES

14
  • Matrix decompositions.
  • The basic idea
  • Treat the dataset as a matrix, A, with n rows
    and m columns
  • Factor A into the product of two matrices, C
    and F
  • A C F
  • where C is n x r, F is r x m and r is smaller
    than m.
  • Think of F as a set of underlying real
    somethings and C as a way of mixing these
    somethings together to get the observed attribute
    values. Choosing r smaller than m forces the
    decomposition to somehow represent the data more
    compactly.

  • F
  • A C

15
  • Two matrix decompositions are useful
  • Singular value decomposition (SVD) the rows of
    F are orthogonal axes such that the maximum
    possible variation in the data lies along the
    first axis the maximum of what remains along the
    second, and so on. The rows of C are coordinates
    in this space.
  • Independent component analysis (ICA) the rows
    of F are statistically independent factors. The
    rows of C describe how to mix these factors to
    produce the original data.
  • Strictly speaking, the row of C are not
    coordinates, but we can plot them to get some
    idea of structure.

16
The messages with correlated unusual word usage
are marked with red circles
First 3 dimensions SVD
17
First 3 dimensions ICA
18
  • (Fortunately) both unusual word use and
    correlated word use are necessary to make such
    messages detectable.

Correlation with proper word frequencies (SVD)
So ordinary conversations dont show up as false
positives!!
19
Correlation with proper word frequencies (ICA)
20
Uncorrelated with unusual word frequencies (SVD)
Conversations about unusual things dont show up
as false positives either!!
21
Uncorrelated with unusual word frequencies (ICA)
22
  • This trick permits a new level of sophistication
    in connecting related messages into conversations
    when the usual indicators are not available.
  • It does exactly the right thing ignoring
    conversations about ordinary topics, and
    conversations about unusual topics, but homing in
    on conversations about unusual topics using
    inappropriate words.
  • Because the dataset is sparse, SVD takes time
    linear in the number of messages. The complexity
    of ICA is less clear but there are direct
    hardware implementations.

23
  • Message-rank matrices are useful because they
    defend against the countermeasure of rules like
    use the word 5 ranks below the one you want to
    use.
  • Such rules are easy to apply with access to the
    internet, for example the site www.fabrica.it/word
    count/main.php.
  • However, this isnt so easy in real-time
    communication.

24
SVD of message-rank matrix has a fan
shape. Points are labelled with the length of
each message
25
Same plot with messages labelled by the average
rank of the nouns they contain.
Length of message and average rank are correlated
partly because of opportunity, but its not
clear that this the whole story.
26
  • Replacing words with those, say, five positions
    down the list does not show up in the SVD of a
    message-frequency matrix

27
  • But its very clear in the SVD of a message-rank
    matrix

28
  • We have been applying these techniques to the
    Enron email dataset, which is a good surrogate
    for intercepted communications
  • about 500,000 emails
  • about 1500 people
  • partially known command and control
    structure
  • Early results from several groups were presented
    at the Workshop on Link Analysis,
    Counterterrorism and Security
  • www.cs.queensu.ca/home/skill/siamworkshop.html
  • also
  • New York Times Week in Review this weekend

29
  • ?
About PowerShow.com