Privacy Risk Models for Designing PrivacySensitive Ubiquitous Computing Systems - PowerPoint PPT Presentation

About This Presentation
Title:

Privacy Risk Models for Designing PrivacySensitive Ubiquitous Computing Systems

Description:

Quotes from nurses using locator badges 'I think this is disrespectful, demeaning and degrading' ... Locator badges, Video cameras. Client-based approach ... – PowerPoint PPT presentation

Number of Views:108
Avg rating:3.0/5.0
Slides: 24
Provided by: jason88
Learn more at: http://www.cs.cmu.edu
Category:

less

Transcript and Presenter's Notes

Title: Privacy Risk Models for Designing PrivacySensitive Ubiquitous Computing Systems


1
Privacy Risk Models forDesigning
Privacy-Sensitive Ubiquitous Computing Systems
2
MotivationUbiquitous Computing is Coming
  • Advances in wireless networking, sensors, devices
  • Greater awareness of and interaction with
    physical world

But what about my privacy?
E911
Find Friends
3
MotivationBut Hard to Design Privacy-Sensitive
Ubicomp Apps
  • Discussions on privacy generate lots of heat but
    not light
  • Big brother, overprotective parents,
    telemarketers, genetics
  • Many conflicting values
  • Often end up talking over each other
  • Hard to have reasoned debates and create designs
    that address the issues
  • Need a design method that helps design teams
  • Identify
  • Prioritize
  • Manage privacy risks for specific applications
  • Propose Privacy Risk Models for doing this

4
Privacy Risk Model AnalogySecurity Threat Model
  • The first rule of security analysis is this
    understand your threat model. Experience teaches
    that if you dont have a clear threat model
    a clear idea of what you are trying to prevent
    and what technical capabilities your adversaries
    have then you wont be able to think
    analytically about how to proceed. The threat
    model is the starting point of any security
    analysis.
  • - Ed Felten

5
Privacy Risk ModelTwo Parts Risk Analysis and
Risk Management
  • Privacy Risk Analysis
  • Common questions to help design teams identify
    potential risks
  • Like a task analysis
  • Privacy Risk Management
  • Helps teams prioritize and manage risks
  • Like severity rankings in heuristic evaluation
  • Will present a specific privacy risk model for
    ubicomp
  • Draws on previous work, plus surveys and
    interviews
  • Provide reasonable level of protection for
    foreseeable risks

6
Outline
  • Motivation
  • Privacy Risk Analysis
  • Privacy Risk Management
  • Case Study Location-enhanced Instant Messenger

7
Privacy Risk AnalysisCommon Questions to Help
Design Teams Identify Risks
  • Social and Organizational Context
  • Who are the users?
  • What kinds of personal info are shared?
  • Relationships between sharers and observers?
  • Value proposition for sharing?

8
Social and Organizational ContextWho are the
users? Who shares info? Who sees it?
  • Different communities have different needs and
    norms
  • An app appropriate for families might not be for
    work settings
  • Affects conditions and types of info willing to
    be shared
  • Location information with spouse vs co-workers
  • Real-time monitoring of ones health
  • Start with most likely users
  • Ex. Find Friends
  • Likely sharers are people using mobile phone
  • Likely observers are friends, family, co-workers

Find Friends
9
Social and Organizational ContextWhat kinds of
personal info are shared?
  • Different kinds of info have different risks and
    norms
  • Current location vs home phone vs hobbies
  • Some information already known between people
  • Ex. Dont need to protect identity with your
    friends and family
  • Different ways of protecting different kinds of
    info
  • Ex. Can revoke access to location, cannot for
    birthday or name

10
Social and Organizational ContextRelationships
between sharers and observers?
  • Kinds of risks and concerns
  • Ex. Risks w/ friends are unwanted intrusions,
    embarrassment
  • Ex. Risks w/ paid services are spam, 2nd use,
    hackers
  • Incentives for protecting personal information
  • Ex. Most friends dont have reason to
    intentionally cause harm
  • Ex. Neither do paid services, but want to make
    more money
  • Mechanisms for recourse
  • Ex. Kindly ask friends and family to stop being
    nosy
  • Ex. Recourse for paid services include formally
    complaining, switching services, suing

11
Social and Organizational ContextValue
proposition for sharing personal information?
  • What incentive do users have for sharing?
  • Quotes from nurses using locator badges
  • I think this is disrespectful, demeaning and
    degrading
  • At first, we hated it for various reasons, but
    mostly we felt we couldnt take a bathroom break
    without someone knowing where we werebut now
    requests for medications go right to the nurse
    and bedpans etc go to the techs first... I just
    love the locator system.
  • When those who share personal info do not benefit
    in proportion to perceived risks, then the tech
    is likely to fail

12
Privacy Risk AnalysisCommon Questions to Help
Design Teams Identify Risks
  • Social and Organizational Context
  • Who are the users?
  • What kinds of personal info are shared?
  • Relationships between sharers and observers?
  • Value proposition for sharing?
  • Technology
  • How is personal info collected?
  • Push or pull?
  • One-time or continuous?
  • Granularity of info?

13
TechnologyHow is personal info collected?
  • Different technologies have different tradeoffs
    for privacy
  • Network-based approach
  • Info captured and processed by external computers
    that users have no practical control over
  • Ex. Locator badges, Video cameras
  • Client-based approach
  • Info captured and processed on end-users device
  • Ex. GPS, beacons
  • Stronger privacy guarantees, all info starts with
    you first

14
TechnologyPush or pull?
  • Push is when user sends info first
  • Ex. you send your location info on E911 call
  • Few people seem to have problems with push
  • Pull is when another person requests info first
  • Ex. a friend requests your current location
  • Design space much harder here
  • need to make people aware of requests
  • want to provide understandable level of control
  • dont want to overwhelm

E911
Find Friends
15
TechnologyOne-time or continuous disclosures?
  • One-time disclosure
  • Ex. observer gets snapshot
  • Fewer privacy concerns
  • Continuous disclosure
  • Ex. observer repeatedly gets info
  • Greater privacy concerns
  • Its stalking, man.

Find Friends
Active Campus
16
TechnologyGranularity of info shared?
  • Different granularities have different utility
    and risks
  • Spatial granularity
  • Ex. City? Neighborhood? Street? Room?
  • Temporal granularity
  • Ex. at Boston last month vs at Boston August 2
    2004
  • Identification granularity
  • Ex. a person vs a woman vs alice_at_blah.com
  • Keep and use coarsest granularity needed
  • Least specific data, fewer inferences, fewer risks

17
Outline
  • Motivation
  • Privacy Risk Analysis
  • Privacy Risk Management
  • Case Study Location-enhanced Instant Messenger

18
Privacy Risk ManagementHelps teams prioritize
and manage risks
  • First step is to prioritize risks by estimating
  • Likelihood that unwanted disclosure occurs
  • Damage that will happen on such a disclosure
  • Cost of adequate privacy protection
  • Focus on high likelihood, high damage, low cost
    risks first
  • Like heuristic eval, fix high severity and/or low
    cost
  • Difficult to get exact numbers, more important is
    the process

19
Privacy Risk ManagementHelps teams prioritize
and manage risks
  • Next step is to help manage those risks
  • How does the disclosure happen?
  • Accident? Bad user interface? Poor conceptual
    model?
  • Malicious? Inside job? Scammers?
  • What kinds of choice, control, and awareness are
    there?
  • Opt-in? Opt-out?
  • What mechanisms? Ex. Buddy list, Invisible mode
  • What are the default settings?
  • Better to prevent or to detect abuses?
  • Bob has asked for your location five times in
    the past hour

20
Case StudyLocation-enhanced Instant Messenger
  • New features
  • Request a friends current location
  • Automatically show your location
  • Invisible mode, reject requests
  • Default location is unknown
  • Who are the users?
  • Typical IM users
  • Relationships?
  • Friends, family, classmates,
  • One-time or continuous?
  • One-time w/ notifications

21
Case StudyLocation-enhanced Instant Messenger
  • Identifying potential privacy risks
  • Over-monitoring by friends and family
  • Over-monitoring at work place
  • Being found by malicious person (ex. stalker,
    mugger)
  • Assessing the first risk, over-monitoring by
    family
  • Likelihood depends on family, conservatively
    assign high
  • Damage might be embarrassing but not
    life-threatening, assign medium
  • Managing the first risk
  • Buddy list, Notifications for awareness,
    invisible mode, unknown if location not
    disclosed
  • All easy to implement, cost is low

22
Discussion
  • Privacy risk models are only a starting point
  • Like task analysis, should try to verify
    assumptions and answers
  • Can be combined with field studies, interviews,
    low-fi prototypes

23
Summary
  • Privacy risk models for helping design teams
    identify, prioritize, and manage risks
  • Privacy risk analysis for identifying risks
  • Series of common questions, like a task analysis
  • Privacy risk management for prioritizing
    managing risks
  • Like severity ratings in heuristic evaluation
  • Described our first iteration of privacy risk
    model
  • Help us evolve and advance it!
Write a Comment
User Comments (0)
About PowerShow.com