400 likes | 411 Views
Explore the intersection of usability and security in technology, including human memory, password systems, biometrics, mental models, research challenges, and the importance of user knowledge. Learn about the impact of user behavior on system security and discover strategies for improving usability and enhancing security measures.
E N D
Usability and Security –Why we need to look at the big picture M. Angela Sasse Professor of Human-Centred Technology Department of Computer Science University College London, UK a.sasse@cs.ucl.ac.uk www.ucl.cs.ac.uk/staff/A.Sasse
Acknowledgements • My Doctoral Students • Anne Adams - Dirk Weirich • Sacha Brostoff - Ivan Flechais • Marek Rejmann-Greene (BT Exact) • BIOVISION (EU Roadmap Project) • UK e-Science Programme • German Federal Office for Information Security
Overview • Yes, security needs usability • Beyond UI design: changing user behaviour • Organisational factors • Designing and maintaining security culture • Taking responsibility • Changing the development process • Involving all stakeholders • Changing the design/development process
Human Memory • Limited capacity • Decays over time (items cannot be recalled at all or not 100% correct) • Frequent recall improves memorability • Unaided recall is harder than recognition • Non-meaningful items much harder to recall than meaningful ones • Similar items are easily confused • Items linger - cannot “forget on demand”
Password Systems • Require unaided recall • Entry must be 100% correct • Not meaningful • Many similar items compete • Frequently change • Proliferation passwords and PINs (banking, phones, websites)
Consequences of unusable password systems • Cost of secure re-setting (help desks) is high • security undermined by • cheap reset techniques (reminders) • user workarounds • Organisations where everybody has password problems are vulnerable to social engineering attacks
Research challenges for authentication • Different mechanisms for frequent and infrequently used passwords • Make mechanisms more forgiving • Replace “all or nothing” with more forgiving mechanism • Provide feedback and instructions • Cued rather than unaided recall • Keep password changes to a minimum • while making it sufficiently secure (including spouse-proof), AND • providing universal access
Is biometrics the answer? • Potential to reduce mental workload • But • No single biometric provides universal access • False rejection rates still high • Current equipment has raft of usability issues
Mental Models/Metaphors • “Why Johnny Can’t Encrypt” [Whitten & Tygar, Procs USENIX 1999] • UI problems • User Tasks not represented • Misleading labels • Lack of feedback • Problem lies deeper: • “key” cues the wrong mental model, usage of “public” “private” does not match everyday use of language
Research Challenges: simple concepts or metaphors • Move from “shorthand” metaphors for security community to metaphors that work for wider user base • Adopt/adapt conceptual design approach to make security concepts & tools more accessible • Identify suitable metaphors • Engineer system and discourse to communicate these
User Knowledge • “Users are not the Enemy” [Adams & Sasse, 1999]: employees’ and managers’ knowledge about security is sketchy • Replicated by Fitzgibbons et al. [Univ. of Colorado, 2003] • “Namedropping” of security concepts and unwarranted assumptions among software developers [Flechais & Sasse, 2003] • User education about security needed
User knowledge of security “You know, if you think about, who’s actuallygoing to go through all that struggle to hack a departmental computer scienceaccount of some academic at xxxx college. It’s not like NASA or anything,nothing of interest. “What would make it more likely?”Answer: “Maybe if I was more famous, or…” [laughs] Weirich & Sasse, Procs NSPW 2003
“Adams & Sasse propose that educating users in security is a solution for the problem of chosing weak passwords. They claim that if users receive specific security training and understand security models, they will select secure passwords and refrain from insecure behaviour. In our study, however, we discovered that the level of security training did not prevent users from choosing trivial passwords and refrain from engaging in insecure behaviour.” [Dhamija & Perrig, 2001]
Education vs. Training • Education not one-way information transfer • Aim of Training: change behaviour • Form good habits and change bad ones • Checked to establish correctness, and provide feedback • Repeat and reinforce sufficiently often to form habit • Checked again after certain time to ensure desired behaviours have been established, and have desired effect
Changing User Behaviour • More than increasing user knowledge: • Changing user behaviour beyond their interaction with the system • Motivation • Persuasion • Social norms
Perceptions of, and attitudes to security • Weirich & Sasse, 2001 • “How would you describe a person who cares about security?”
“People who would want to be more secure. I don’tknow. That’s really a question for psychologists. What sort of people keeptheir desks tidy. What sort of people comb their hair in the morning.” “People therefore who areobedient. People who follow the crowd.”
“So, you could probably be changing yourpassword every week, for no obvious reason apart from your paranoia, whereas I am not terribly paranoid about this sort of thing.”
Research Challenges:Perceptions & Attitudes • Ways of persuading and motivating users to be secure • Appeal to self-interest: • Link security to goals that matter to people • Economic impact • Part of professional and ethical conduct • Make threats believable, appear real • Changing the image of security • Social marketing • Role models, “it’s cool to be secure” • Persuasive technology (based on Fogg 2003) • Can we make security fun?
Task Factors • For most people, most of the time, security is enabling task to one or more production tasks • Enabling tasks perceived as “hurdles” if relevance to production task not clear • Human nature to take short-cuts, especially when workload is felt to be high
“We in engineering like to leave things fairly unprotected so we can go and access other people’s directories, so if he people I’m working with are changing files, I can work with their latest revisions.” From Fitzgibbons et al., Univ. of Colorado, 2003
Example: Passfaces • Good recall rates even after long periods of non-use (90+ % after 3 months • But: in field trial, Passfaces users only had 30% login frequency of password users Brostoff & Sasse, Procs HCI 2000
Design for production tasks • If competing with production tasks, it will be eliminated/circumvented whenever • Security behaviour most fit production tasks • No competing demands for user resources (physical/mental workload) • Cost of keeping out legitimate users: viable contingencies • Performance criteria: Speed, errors
Research Challenges:Task design • Develop security mechanisms that can be configured to match requirements of production tasks • Support individuals and organisations to identify and deal with competing goals • Resolve locally vs. globally optimal solutions
Physical Context • Outdoors usage is different • Lighting, pollution, temperature, noise, etc. • Limit performance especially of novel complex mechanisms (such as biometrics) • Mobile and handheld systems • Physical ease of use • Nomadic use of devices and networks • creates new threats
Research Challenges:Physical Context • Develop usable & secure mechanisms for interaction with pervasive/ubiquitous systems • Specific? • General? • Multiple? • Consider implications for • physical & mental workload • economic viability
Organisational Context • Security culture • “Do as I say, not as I do” • Being able to violate “petty” security regulations is badge of seniority • Link security into business goals • Design of specific (goal- and risk-based) security policies that are enforced, and are seen to be enforced
Research Challenges:Organisations & Security • Integrate security into organisation & business model • Socio-technical design approach • Adapt safety-critical design approaches [e.g. Reason 1990] • Apply risk analysis and economic principles to decision-making about security
Cultural Context: Trust • Social norms, key: trust • People want to trust, and be trusted, but this can create risks [e.g. Mitnick, 2002] • But • low-trust systems are expensive to run [Handy, 1985] • do not allow building of social capital • May be counterproductive since people are only defense against novel attacks.
Research Challenges:Social Norms • Identify social norms that may interfere with desired security behaviour • Trust is a key norm • Create clear conceptual basis for role of trust in security systems
Conclusions • Usable and effective security needs a systemic approach • Security technology, and improving user interfaces to that technology, is by itself not the answer.
Any Questions? contact: a.sasse@cs.ucl.ac.uk www.ucl.cs.ac.uk/staff/A.Sasse